code stringlengths 22 1.05M | apis listlengths 1 3.31k | extract_api stringlengths 75 3.25M |
|---|---|---|
import urllib as U
__all__ = ('sequence', )
def _seq_from_xml(xml):
start = xml.find(">", xml.find("<DNA")) + 1
end = xml.rfind("</DNA>")
return xml[start:end].replace(' ', '').replace('\n', '').strip()
def sequence(db, chrom, start, end):
"""
return the sequence for a region using the UCSC DAS
... | [
"doctest.testmod",
"urllib.urlopen"
] | [((764, 781), 'doctest.testmod', 'doctest.testmod', ([], {}), '()\n', (779, 781), False, 'import doctest\n'), ((639, 675), 'urllib.urlopen', 'U.urlopen', (['(url % (chrom, start, end))'], {}), '(url % (chrom, start, end))\n', (648, 675), True, 'import urllib as U\n')] |
# emacs: -*- mode: python; py-indent-offset: 4; indent-tabs-mode: nil -*-
# vi: set ft=python sts=4 ts=4 sw=4 et:
import os
from tempfile import mkdtemp
from shutil import rmtree
from nipype.testing import (assert_equal, assert_true, assert_false,
assert_raises, skipif, parametric)
import ... | [
"nipype.testing.assert_true",
"nipype.interfaces.matlab.MatlabCommand",
"nipype.interfaces.base.CommandLine",
"nipype.testing.assert_equal",
"nipype.testing.assert_false",
"os.path.join",
"os.getcwd",
"os.chdir",
"nipype.testing.assert_raises",
"tempfile.mkdtemp",
"shutil.rmtree",
"nipype.inte... | [((896, 913), 'nipype.testing.skipif', 'skipif', (['no_matlab'], {}), '(no_matlab)\n', (902, 913), False, 'from nipype.testing import assert_equal, assert_true, assert_false, assert_raises, skipif, parametric\n'), ((1772, 1789), 'nipype.testing.skipif', 'skipif', (['no_matlab'], {}), '(no_matlab)\n', (1778, 1789), Fals... |
"""A filter block.
"""
import control
import numpy as np
import scipy
from .base import Block
class Filter(Block):
"""A Filter block class
This is simply a single-input-single-output LTI system defined by a
single TransferFunction object.
Parameters
----------
tf : control.TransferFunction
... | [
"numpy.dot",
"numpy.zeros_like",
"scipy.signal.cont2discrete"
] | [((2083, 2112), 'numpy.dot', 'np.dot', (['num_d', 'input_register'], {}), '(num_d, input_register)\n', (2089, 2112), True, 'import numpy as np\n'), ((2130, 2168), 'numpy.dot', 'np.dot', (['den_d[1:]', 'output_register[1:]'], {}), '(den_d[1:], output_register[1:])\n', (2136, 2168), True, 'import numpy as np\n'), ((5163,... |
import pandas as pd
import numpy as np
import xml.etree.ElementTree as ElementTree
from traffic_analysis.d00_utils.bbox_helpers import bboxcv2_to_bboxcvlib
from traffic_analysis.d05_evaluation.parse_annotation import parse_annotation
from traffic_analysis.d05_evaluation.compute_mean_average_precision import get_avg_p... | [
"traffic_analysis.d00_utils.bbox_helpers.bboxcv2_to_bboxcvlib",
"numpy.mean",
"traffic_analysis.d05_evaluation.parse_annotation.parse_annotation",
"xml.etree.ElementTree.parse",
"pandas.merge",
"pandas.DataFrame.from_dict",
"traffic_analysis.d05_evaluation.compute_mean_average_precision.get_avg_precisio... | [((781, 797), 'pandas.DataFrame', 'pd.DataFrame', (['{}'], {}), '({})\n', (793, 797), True, 'import pandas as pd\n'), ((831, 847), 'pandas.DataFrame', 'pd.DataFrame', (['{}'], {}), '({})\n', (843, 847), True, 'import pandas as pd\n'), ((2556, 2594), 'pandas.concat', 'pd.concat', (['frame_level_map_dfs'], {'axis': '(0)'... |
from load_to_s3 import export_to_s3_as_dataframe, export_to_s3_as_csv
from transform_form_specifications import get_custom_form_schema_df
from transform_form_results import get_form_results_df
from utils.clients import Clients
from utils.constants import PuenteTables
def run_transform_jobs(event, context):
"""
... | [
"load_to_s3.export_to_s3_as_csv",
"transform_form_specifications.get_custom_form_schema_df",
"load_to_s3.export_to_s3_as_dataframe",
"transform_form_results.get_form_results_df"
] | [((615, 652), 'transform_form_results.get_form_results_df', 'get_form_results_df', ([], {'raw_results': '(True)'}), '(raw_results=True)\n', (634, 652), False, 'from transform_form_results import get_form_results_df\n'), ((1093, 1120), 'transform_form_specifications.get_custom_form_schema_df', 'get_custom_form_schema_df... |
# Generated by Django 2.1.2 on 2019-06-21 15:16
from django.db import migrations, models
class Migration(migrations.Migration):
initial = True
dependencies = [
]
operations = [
migrations.CreateModel(
name='ContactQuery',
fields=[
('id', models.AutoF... | [
"django.db.models.EmailField",
"django.db.models.TextField",
"django.db.models.AutoField",
"django.db.models.CharField"
] | [((308, 401), 'django.db.models.AutoField', 'models.AutoField', ([], {'auto_created': '(True)', 'primary_key': '(True)', 'serialize': '(False)', 'verbose_name': '"""ID"""'}), "(auto_created=True, primary_key=True, serialize=False,\n verbose_name='ID')\n", (324, 401), False, 'from django.db import migrations, models\... |
"""
Script to post grades back to EdX
"""
import requests
import json
import os
import time
from oauthlib.oauth1.rfc5849 import signature, parameters
from lxml import etree
from hashlib import sha1
import argparse
import base64
class GradePostException(Exception):
def __init__(self, response=None):
self.r... | [
"oauthlib.oauth1.rfc5849.signature.sign_hmac_sha1",
"json.loads",
"requests.post",
"oauthlib.oauth1.rfc5849.signature.collect_parameters",
"argparse.ArgumentParser",
"oauthlib.oauth1.rfc5849.parameters.prepare_headers",
"oauthlib.oauth1.rfc5849.signature.normalize_base_string_uri",
"hashlib.sha1",
"... | [((1786, 1792), 'hashlib.sha1', 'sha1', ([], {}), '()\n', (1790, 1792), False, 'from hashlib import sha1\n'), ((2374, 2434), 'oauthlib.oauth1.rfc5849.signature.sign_hmac_sha1', 'signature.sign_hmac_sha1', (['base_string', 'consumer_secret', 'None'], {}), '(base_string, consumer_secret, None)\n', (2398, 2434), False, 'f... |
import copy
import torch
import torch.nn as nn
from pytorch_metric_learning.losses import NTXentLoss
from transformers import BertForMaskedLM, BertForPreTraining, BertTokenizer
def mask_tokens(inputs, tokenizer, not_mask_pos=None):
""" Prepare masked tokens of inputs and labels for masked language modeling (80% MA... | [
"torch.bernoulli",
"transformers.BertForMaskedLM.from_pretrained",
"torch.full",
"pytorch_metric_learning.losses.NTXentLoss",
"transformers.BertTokenizer.from_pretrained",
"torch.max",
"torch.tensor",
"torch.sum",
"copy.deepcopy",
"torch.nn.BCEWithLogitsLoss",
"torch.cat"
] | [((911, 932), 'copy.deepcopy', 'copy.deepcopy', (['inputs'], {}), '(inputs)\n', (924, 932), False, 'import copy\n'), ((1098, 1128), 'torch.full', 'torch.full', (['labels.shape', '(0.15)'], {}), '(labels.shape, 0.15)\n', (1108, 1128), False, 'import torch\n'), ((1292, 1343), 'torch.tensor', 'torch.tensor', (['special_to... |
import pandas as pd
import os
if __name__ == "__main__":
data_dir = os.environ.get("DATA_DIRECTORY")
data_file = os.environ.get("DATA_FILE")
print(data_dir)
data_df = pd.read_pickle(data_dir + data_file)
test_df = pd.DataFrame()
# Creating two non-overlapping datasets for training and validati... | [
"pandas.read_pickle",
"os.environ.get",
"pandas.DataFrame",
"pandas.concat"
] | [((73, 105), 'os.environ.get', 'os.environ.get', (['"""DATA_DIRECTORY"""'], {}), "('DATA_DIRECTORY')\n", (87, 105), False, 'import os\n'), ((122, 149), 'os.environ.get', 'os.environ.get', (['"""DATA_FILE"""'], {}), "('DATA_FILE')\n", (136, 149), False, 'import os\n'), ((184, 220), 'pandas.read_pickle', 'pd.read_pickle'... |
## ACL Import Module
# ACL CSV Import
# Version 5
# 2015-10-30
# we only need the datetime class & the static function strptime from datetime module
from datetime import datetime
import re
import sys
import os
import logging
# best postgresql module so far, install it "yum install python-psycopg2"
import psycopg2
im... | [
"logging.getLogger",
"os.path.exists",
"os.listdir",
"os.makedirs",
"os.rename",
"tempfile.mkstemp",
"csv.writer",
"os.path.join",
"os.path.isfile",
"os.path.isdir",
"os.path.basename",
"os.unlink",
"csv.reader",
"re.search"
] | [((390, 414), 'logging.getLogger', 'logging.getLogger', (['"""acl"""'], {}), "('acl')\n", (407, 414), False, 'import logging\n'), ((6289, 6313), 'os.path.isfile', 'os.path.isfile', (['filename'], {}), '(filename)\n', (6303, 6313), False, 'import os\n'), ((8608, 8694), 're.search', 're.search', (['"""^(.*?)homeofficerol... |
import os, queue
from tablet import Tablet
f = open(os.path.join(os.path.dirname(__file__), '../input/18/part1.txt'), 'r')
def main():
instructionStrings = []
line = f.readline()
while line:
instructionStrings.append(line.rstrip())
line = f.readline()
q0 = queue.Queue()
q1 = queue... | [
"os.path.dirname",
"tablet.Tablet",
"queue.Queue"
] | [((292, 305), 'queue.Queue', 'queue.Queue', ([], {}), '()\n', (303, 305), False, 'import os, queue\n'), ((315, 328), 'queue.Queue', 'queue.Queue', ([], {}), '()\n', (326, 328), False, 'import os, queue\n'), ((339, 376), 'tablet.Tablet', 'Tablet', (['instructionStrings', '(0)', 'q0', 'q1'], {}), '(instructionStrings, 0,... |
import datetime
import os
from zipfile import ZipFile
import requests
from icecream import ic
from wget import download
ic.enable()
date_from = (datetime.date.today() - datetime.timedelta(days=5)).isoformat()
date_to = (datetime.date.today()).isoformat()
ic(f'Date Range: {date_from} to {date_to}')
def getbhav_dat... | [
"icecream.ic.enable",
"os.path.exists",
"icecream.ic",
"wget.download",
"requests.get",
"datetime.timedelta",
"os.chdir",
"os.mkdir",
"datetime.date.today",
"datetime.date.fromisoformat"
] | [((122, 133), 'icecream.ic.enable', 'ic.enable', ([], {}), '()\n', (131, 133), False, 'from icecream import ic\n'), ((259, 302), 'icecream.ic', 'ic', (['f"""Date Range: {date_from} to {date_to}"""'], {}), "(f'Date Range: {date_from} to {date_to}')\n", (261, 302), False, 'from icecream import ic\n'), ((365, 403), 'datet... |
# coding=utf-8
# !/usr/bin/python3
# Name: ruuvitag calc - calculations
# Copyright: (c) 2019 TK
# Licence: MIT
# -------------------------------------------------------------------------------
import logging
logger = logging.getLogger('ruuvitag')
import math
# ----------------------------------------... | [
"logging.getLogger",
"math.exp"
] | [((234, 263), 'logging.getLogger', 'logging.getLogger', (['"""ruuvitag"""'], {}), "('ruuvitag')\n", (251, 263), False, 'import logging\n'), ((1403, 1446), 'math.exp', 'math.exp', (['(17.67 * l_temp / (243.5 + l_temp))'], {}), '(17.67 * l_temp / (243.5 + l_temp))\n', (1411, 1446), False, 'import math\n')] |
# -*- coding: utf-8 -*-
'''
Copyright (c) 2021, Trustworthy AI, Inc. All rights reserved.
Redistribution and use in source and binary forms, with or without modification,
are permitted provided that the following conditions are met:
1. Redistributions of source code must retain the above copyright notice, this
list ... | [
"geometry_msgs.msg.Vector3",
"geometry_msgs.msg.Twist",
"math.radians",
"geometry_msgs.msg.Transform",
"numpy.array",
"geometry_msgs.msg.Point",
"geometry_msgs.msg.Quaternion",
"tf.transformations.quaternion_from_euler",
"tf.transformations.euler_matrix",
"geometry_msgs.msg.Accel",
"geometry_msg... | [((2477, 2545), 'numpy.array', 'numpy.array', (['[carla_location.x, -carla_location.y, carla_location.z]'], {}), '([carla_location.x, -carla_location.y, carla_location.z])\n', (2488, 2545), False, 'import numpy\n'), ((2962, 2971), 'geometry_msgs.msg.Vector3', 'Vector3', ([], {}), '()\n', (2969, 2971), False, 'from geom... |
import numpy as np
class KNearestNeighbors:
def __init__(self, distances, labels, k=10):
self.distances = distances
self.labels = labels
self.k = k
def _kNN(self, instance, train, k):
nearest = np.argpartition(self.distances[instance][train], k)
nearest_labels = self.... | [
"numpy.argmax",
"numpy.zeros",
"numpy.unique",
"numpy.argpartition"
] | [((238, 289), 'numpy.argpartition', 'np.argpartition', (['self.distances[instance][train]', 'k'], {}), '(self.distances[instance][train], k)\n', (253, 289), True, 'import numpy as np\n'), ((372, 417), 'numpy.unique', 'np.unique', (['nearest_labels'], {'return_counts': '(True)'}), '(nearest_labels, return_counts=True)\n... |
# coding: utf-8
# In[1]:
from path import Path
from matplotlib import pyplot as plt
import numpy as np
import skimage.io as io
import os
from PIL import Image
import cv2
import random
import shutil
def crop_by_sequence(image_path,img_class_path,crop_size_w,crop_size_h,prefix,save_dir ,same_scale = False):
... | [
"cv2.imwrite",
"shutil.move",
"path.Path",
"skimage.io.imread",
"numpy.random.randint",
"numpy.zeros",
"os.mkdir"
] | [((703, 728), 'skimage.io.imread', 'io.imread', (['img_class_path'], {}), '(img_class_path)\n', (712, 728), True, 'import skimage.io as io\n'), ((3340, 3365), 'skimage.io.imread', 'io.imread', (['img_class_path'], {}), '(img_class_path)\n', (3349, 3365), True, 'import skimage.io as io\n'), ((3576, 3624), 'numpy.random.... |
import cProfile
import functools
import pstats
import resource
import signal
import time
from typing import Any, Callable
# This list is passed to @profile() which aggregates the cumulative runtime
# into these functions, which represent the primary stages of task solutioning.
PROFILE_BREAKOUT_STD: list[str] = [
"... | [
"signal.signal",
"resource.getrusage",
"functools.wraps",
"pstats.Stats",
"signal.alarm",
"cProfile.Profile",
"time.time"
] | [((809, 830), 'functools.wraps', 'functools.wraps', (['func'], {}), '(func)\n', (824, 830), False, 'import functools\n'), ((2152, 2192), 'resource.getrusage', 'resource.getrusage', (['resource.RUSAGE_SELF'], {}), '(resource.RUSAGE_SELF)\n', (2170, 2192), False, 'import resource\n'), ((2643, 2664), 'functools.wraps', 'f... |
# -*- coding: utf-8 -*-
#Requires Win32 Python Extensions
import os
import servicemanager
import shutil
import subprocess
import sys
import win32api
import win32event
import win32service
import win32serviceutil
import tempfile
class VulnService(win32serviceutil.ServiceFramework):
_svc_na... | [
"subprocess.check_output",
"servicemanager.LogInfoMsg",
"win32serviceutil.HandleCommandLine",
"servicemanager.LogMsg",
"win32event.WaitForSingleObject",
"win32serviceutil.ServiceFramework.__init__",
"shutil.copyfile",
"os.path.dirname",
"tempfile.gettempdir",
"win32api.SetConsoleCtrlHandler",
"o... | [((4090, 4139), 'win32api.SetConsoleCtrlHandler', 'win32api.SetConsoleCtrlHandler', (['ctrlHandler', '(True)'], {}), '(ctrlHandler, True)\n', (4120, 4139), False, 'import win32api\n'), ((4149, 4196), 'win32serviceutil.HandleCommandLine', 'win32serviceutil.HandleCommandLine', (['VulnService'], {}), '(VulnService)\n', (4... |
import copy
import logging
import torch
import numpy as np
from torch.utils.data import DataLoader
from torchvision import datasets, transforms
log = logging.getLogger(__name__)
def balanced_batches(dataset, batch_size):
unlabled_idx = dataset.unlabeled_idx
labeled_idx = list(filter(lambda _: _ not in unlab... | [
"logging.getLogger",
"copy.deepcopy",
"numpy.random.choice",
"torch.LongTensor",
"torch.stack",
"numpy.array_split",
"numpy.array",
"torch.utils.data.DataLoader",
"torchvision.transforms.ToTensor"
] | [((152, 179), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (169, 179), False, 'import logging\n'), ((411, 432), 'numpy.array', 'np.array', (['labeled_idx'], {}), '(labeled_idx)\n', (419, 432), True, 'import numpy as np\n'), ((590, 629), 'numpy.array_split', 'np.array_split', (['unlabled... |
# emacs: -*- mode: python; py-indent-offset: 4; indent-tabs-mode: nil -*-
# vi: set ft=python sts=4 ts=4 sw=4 et:
### ### ### ### ### ### ### ### ### ### ### ### ### ### ### ### ### ### ### ##
#
# See COPYING file distributed along with the PyMVPA package for the
# copyright and license terms.
#
### ### ### ### ###... | [
"numpy.abs",
"mvpa2.clfs.gnb.GNB",
"numpy.exp",
"numpy.sum",
"mvpa2.generators.splitters.Splitter"
] | [((712, 717), 'mvpa2.clfs.gnb.GNB', 'GNB', ([], {}), '()\n', (715, 717), False, 'from mvpa2.clfs.gnb import GNB\n'), ((735, 761), 'mvpa2.clfs.gnb.GNB', 'GNB', ([], {'common_variance': '(False)'}), '(common_variance=False)\n', (738, 761), False, 'from mvpa2.clfs.gnb import GNB\n'), ((778, 797), 'mvpa2.clfs.gnb.GNB', 'GN... |
# MIT License
#
# Copyright (c) 2020 <NAME>
#
# Permission is hereby granted, free of charge, to any person obtaining a copy
# of this software and associated documentation files (the "Software"), to deal
# in the Software without restriction, including without limitation the rights
# to use, copy, modify, merge, publi... | [
"io_mesh_amf.export_amf.ExportAMF",
"mathutils.Matrix.Identity",
"tempfile.gettempdir",
"unittest.mock.Mock"
] | [((2393, 2399), 'unittest.mock.Mock', 'Mock', ([], {}), '()\n', (2397, 2399), False, 'from unittest.mock import Mock\n'), ((2893, 2899), 'unittest.mock.Mock', 'Mock', ([], {}), '()\n', (2897, 2899), False, 'from unittest.mock import Mock\n'), ((3042, 3060), 'mathutils.Matrix.Identity', 'Matrix.Identity', (['(4)'], {}),... |
"""
amplitude.py
measure the maximum peak-to-peak amplitude
"""
import obspy
import types
import numpy as np
import pandas as pd
import madpy.noise as n
from typing import Tuple
import madpy.checks as ch
import madpy.config as config
import matplotlib.pyplot as plt
import madpy.plotting.amp as plot
def measure_ampl... | [
"numpy.abs",
"madpy.noise.arrival_time_utc",
"madpy.checks.check_amplitude",
"madpy.noise.rms_noise",
"numpy.divide",
"numpy.where",
"madpy.checks.check_window",
"numpy.diff",
"numpy.subtract",
"madpy.plotting.amp.amplitude_plot",
"numpy.array",
"numpy.isnan",
"pandas.DataFrame",
"numpy.na... | [((1808, 1829), 'madpy.checks.check_waveform', 'ch.check_waveform', (['tr'], {}), '(tr)\n', (1825, 1829), True, 'import madpy.checks as ch\n'), ((2460, 2474), 'numpy.diff', 'np.diff', (['peaks'], {}), '(peaks)\n', (2467, 2474), True, 'import numpy as np\n'), ((2538, 2561), 'madpy.checks.check_amplitude', 'ch.check_ampl... |
# External Import
from django.contrib.auth import get_user_model
from django import forms
from django.contrib.auth.forms import UserCreationForm, AuthenticationForm
from django.core.exceptions import ValidationError
from django.contrib import messages
from django.urls import reverse
from django.contrib.sites.shortcuts ... | [
"django.contrib.auth.get_user_model",
"django.urls.reverse",
"django.contrib.messages.error",
"django.forms.PasswordInput",
"django.core.exceptions.ValidationError",
"django.forms.ValidationError",
"django.forms.EmailInput",
"django.forms.TextInput",
"django.forms.EmailField"
] | [((404, 420), 'django.contrib.auth.get_user_model', 'get_user_model', ([], {}), '()\n', (418, 420), False, 'from django.contrib.auth import get_user_model\n'), ((2419, 2450), 'django.forms.EmailField', 'forms.EmailField', ([], {'required': '(True)'}), '(required=True)\n', (2435, 2450), False, 'from django import forms\... |
import torch
import torch.nn as nn
import torch.nn.functional as F
from src.utils import get_seed
class DuelingQNetwork(nn.Module):
def __init__(self, state_size, action_size):
super().__init__()
self.seed = torch.manual_seed(get_seed())
self.V_fc1 = nn.Linear(state_size, 64)
self... | [
"src.utils.get_seed",
"torch.nn.Linear"
] | [((282, 307), 'torch.nn.Linear', 'nn.Linear', (['state_size', '(64)'], {}), '(state_size, 64)\n', (291, 307), True, 'import torch.nn as nn\n'), ((329, 346), 'torch.nn.Linear', 'nn.Linear', (['(64)', '(64)'], {}), '(64, 64)\n', (338, 346), True, 'import torch.nn as nn\n'), ((368, 384), 'torch.nn.Linear', 'nn.Linear', ([... |
"""Class for generation using the GAN."""
from pathlib import Path
import tensorflow as tf
from tensorflow.keras import Model
from tqdm import tqdm
from ..utils import Config
class GANEvaluator:
"""Class to generate images using a GAN.
Attributes:
generator: The generator model to be evaluated
... | [
"tensorflow.random.normal",
"tensorflow.image.convert_image_dtype",
"tqdm.tqdm",
"tensorflow.map_fn",
"tensorflow.convert_to_tensor"
] | [((808, 867), 'tensorflow.random.normal', 'tf.random.normal', (['[digits.shape[0], self.config.noise_dims]'], {}), '([digits.shape[0], self.config.noise_dims])\n', (824, 867), True, 'import tensorflow as tf\n'), ((884, 912), 'tensorflow.convert_to_tensor', 'tf.convert_to_tensor', (['digits'], {}), '(digits)\n', (904, 9... |
# Copyright (C) 2019 Intel Corporation
#
# SPDX-License-Identifier: MIT
import argparse
from . import project as project_module
def build_parser(parser=argparse.ArgumentParser()):
project_module.build_create_parser(parser) \
.set_defaults(command=project_module.create_command)
return parser
def m... | [
"argparse.ArgumentParser"
] | [((157, 182), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {}), '()\n', (180, 182), False, 'import argparse\n')] |
#REQUESTS -> FAZER REQUISIÇÕES HTTP (URLLIB3)
#HTTP = Hypertext Transfer Protocol
#BEAUTIFULSOUP -> HTML ou XML -> Organiza-lo de maneira a ler no python
#PANDAS -> pegar esses objetos e tranformar em EXCEL
#USAR REQUESTS E ENTENDER COMO FUNCIONA.
import requests as req
url = 'https://g1.globo.com'
r = req.get(url)... | [
"bs4.BeautifulSoup",
"requests.get"
] | [((308, 320), 'requests.get', 'req.get', (['url'], {}), '(url)\n', (315, 320), True, 'import requests as req\n'), ((386, 421), 'bs4.BeautifulSoup', 'BeautifulSoup', (['conteudo_cru', '"""lxml"""'], {}), "(conteudo_cru, 'lxml')\n", (399, 421), False, 'from bs4 import BeautifulSoup\n'), ((788, 800), 'requests.get', 'req.... |
#!/usr/bin/env python
# -*- coding:Utf-8 -*-
import sys
import os
from collections import OrderedDict
from time import time
if sys.version_info >= (3, 0):
import configparser
else:
import ConfigParser as configparser
from subprocess import Popen
from tekamenu.const import *
def help():
print("usage : ")
... | [
"collections.OrderedDict",
"subprocess.Popen",
"os.access",
"os.path.join",
"os.path.split",
"ConfigParser.ConfigParser",
"os.path.isfile",
"sys.exit",
"time.time"
] | [((441, 451), 'sys.exit', 'sys.exit', ([], {}), '()\n', (449, 451), False, 'import sys\n'), ((1069, 1082), 'collections.OrderedDict', 'OrderedDict', ([], {}), '()\n', (1080, 1082), False, 'from collections import OrderedDict\n'), ((2885, 2907), 'os.path.split', 'os.path.split', (['program'], {}), '(program)\n', (2898, ... |
import ConfigParser
import logging
import os
import sys
import yaml
log = logging.getLogger()
def load_yaml(filename):
with open(filename, 'r') as stream:
return yaml.load(stream)
def set_env_aws_creds(account='default'):
"""
Parse ~/.aws/credentials for credentials
Allow OS environment var... | [
"logging.getLogger",
"os.path.exists",
"os.getenv",
"yaml.load",
"ConfigParser.ConfigParser",
"sys.exit",
"os.path.expanduser"
] | [((75, 94), 'logging.getLogger', 'logging.getLogger', ([], {}), '()\n', (92, 94), False, 'import logging\n'), ((569, 595), 'os.getenv', 'os.getenv', (['aws_key_env_var'], {}), '(aws_key_env_var)\n', (578, 595), False, 'import os\n'), ((617, 646), 'os.getenv', 'os.getenv', (['aws_secret_env_var'], {}), '(aws_secret_env_... |
from setuptools import setup
import setuptools
readme = ''
with open('README.md') as f:
readme = f.read()
setup(
name="discsocket",
author="<NAME>",
url="https://github.com/murillotadeo/discsocket",
project_urls={
"Issue tracker": "https://github.com/murillotadeo/discsocket/issues",
... | [
"setuptools.find_packages"
] | [((448, 479), 'setuptools.find_packages', 'setuptools.find_packages', (['"""src"""'], {}), "('src')\n", (472, 479), False, 'import setuptools\n')] |
# ==BEGIN LICENSE==
#
# MIT License
#
# Copyright (c) 2018 SRI Lab, ETH Zurich
#
# Permission is hereby granted, free of charge, to any person obtaining a copy
# of this software and associated documentation files (the "Software"), to deal
# in the Software without restriction, including without limitation the right... | [
"dpfinder.utils.utils.arr_to_str",
"os.path.dirname",
"importlib.import_module",
"re.compile"
] | [((1516, 1584), 're.compile', 're.compile', (['"""^// ==BEGIN LICENSE==[\\\\s\\\\S]+// ==END LICENSE==[\\\\s]+"""'], {}), "('^// ==BEGIN LICENSE==[\\\\s\\\\S]+// ==END LICENSE==[\\\\s]+')\n", (1526, 1584), False, 'import re\n'), ((3484, 3537), 'importlib.import_module', 'importlib.import_module', (["('.algs.' + name)",... |
####################################################
####################################################
# functions and classes used in conjunction with
# pipeline_metaomics.py
####################################################
####################################################
# import libraries
import sys
impo... | [
"numpy.mean",
"CGATPipelines.Pipeline.run",
"sqlite3.connect",
"CGATPipelines.Pipeline.snip",
"itertools.product",
"CGATPipelines.Pipeline.getTempFilename",
"os.unlink",
"os.path.basename",
"pandas.DataFrame",
"CGAT.IOTools.openFile",
"rpy2.robjects.r"
] | [((2859, 2881), 'sqlite3.connect', 'sqlite3.connect', (['rnadb'], {}), '(rnadb)\n', (2874, 2881), False, 'import sqlite3\n'), ((2926, 2948), 'sqlite3.connect', 'sqlite3.connect', (['dnadb'], {}), '(dnadb)\n', (2941, 2948), False, 'import sqlite3\n'), ((4401, 4420), 'sqlite3.connect', 'sqlite3.connect', (['db'], {}), '(... |
# ------------------------------------------------------------------------------
# Test Formatting weeks
# ------------------------------------------------------------------------------
import sys
import datetime as dt
from django.test import TestCase, override_settings
import ls.joyous.utils.weeks
from ls.joyous.utils... | [
"ls.joyous.utils.weeks._iso_year_start",
"ls.joyous.utils.weeks._iso_to_gregorian",
"ls.joyous.utils.weeks._ssweek_to_gregorian",
"ls.joyous.utils.weeks._ssweek_info",
"ls.joyous.utils.weeks._iso_num_weeks",
"ls.joyous.utils.weeks._ssweek_year_start",
"django.utils.translation.override",
"django.test.... | [((5694, 5711), 'django.utils.translation.override', 'override', (['"""en-gb"""'], {}), "('en-gb')\n", (5702, 5711), False, 'from django.utils.translation import override\n'), ((6620, 6637), 'django.utils.translation.override', 'override', (['"""en-au"""'], {}), "('en-au')\n", (6628, 6637), False, 'from django.utils.tr... |
#!/usr/bin/env python3
#
# MIT License
#
# Copyright (c) 2020-2021 EntySec
#
# Permission is hereby granted, free of charge, to any person obtaining a copy
# of this software and associated documentation files (the "Software"), to deal
# in the Software without restriction, including without limitation the rights
# to... | [
"core.cli.badges.badges",
"core.base.exceptions.exceptions",
"core.base.jobs.jobs"
] | [((1356, 1364), 'core.cli.badges.badges', 'badges', ([], {}), '()\n', (1362, 1364), False, 'from core.cli.badges import badges\n'), ((1385, 1391), 'core.base.jobs.jobs', 'jobs', ([], {}), '()\n', (1389, 1391), False, 'from core.base.jobs import jobs\n'), ((1418, 1430), 'core.base.exceptions.exceptions', 'exceptions', (... |
from collections import namedtuple
Genotype = namedtuple('Genotype', 'backbone rpn')
OP_NAMES = [
'sep_conv_3x3',
'sep_conv_3x3_dil3',
'sep_conv_5x5_dil6',
'skip_connect',
'def_conv_3x3',
]
AGG_NAMES = [
'psum',
'cat'
]
HEAD_OP_NAMES = [
'conv1x1',
'conv3x3',
'sep_conv_3x3',
... | [
"collections.namedtuple"
] | [((47, 85), 'collections.namedtuple', 'namedtuple', (['"""Genotype"""', '"""backbone rpn"""'], {}), "('Genotype', 'backbone rpn')\n", (57, 85), False, 'from collections import namedtuple\n')] |
# Copyright (c) 2014 Red Hat, Inc.
#
# Licensed under the Apache License, Version 2.0 (the "License"); you may
# not use this file except in compliance with the License. You may obtain
# a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writ... | [
"rally.task.atomic.action_timer",
"rally.task.atomic.ActionTimer"
] | [((781, 822), 'rally.task.atomic.action_timer', 'atomic.action_timer', (['"""zaqar.create_queue"""'], {}), "('zaqar.create_queue')\n", (800, 822), False, 'from rally.task import atomic\n'), ((1177, 1218), 'rally.task.atomic.action_timer', 'atomic.action_timer', (['"""zaqar.delete_queue"""'], {}), "('zaqar.delete_queue'... |
# -*- coding: utf-8 -*-
'''
Script Name: ping_Utility.py
Path: \IPS_DecisionFabric\Exception Handling\
Description: This script is considered as a module for the Application level Exception handling in DF framework.
Author: <NAME>
Version: 1.0
Revision History:
---------------------------------------------... | [
"logging.basicConfig",
"pandas.Series",
"sys.path.insert",
"logging.debug",
"mongo_operations.Mongo",
"time.sleep",
"datetime.datetime.now",
"os.getpid",
"sys.exit",
"pandas.DataFrame",
"logging.info",
"ping_services.check_for_status"
] | [((1119, 1178), 'sys.path.insert', 'sys.path.insert', (['(0)', '"""/IPS_DecisionFabric/Control Framework"""'], {}), "(0, '/IPS_DecisionFabric/Control Framework')\n", (1134, 1178), False, 'import sys, os\n'), ((1331, 1384), 'ping_services.check_for_status', 'ping_services.check_for_status', (['config_file', 'csv_file'],... |
from setuptools import setup
setup(
name='NotebookScripter',
version='6.0.0',
packages=('NotebookScripter',),
url='https://github.com/breathe/NotebookScripter',
license='MIT',
author='<NAME>',
author_email='<EMAIL>',
install_requires=(
"ipython",
"nbformat"
),
te... | [
"setuptools.setup"
] | [((30, 839), 'setuptools.setup', 'setup', ([], {'name': '"""NotebookScripter"""', 'version': '"""6.0.0"""', 'packages': "('NotebookScripter',)", 'url': '"""https://github.com/breathe/NotebookScripter"""', 'license': '"""MIT"""', 'author': '"""<NAME>"""', 'author_email': '"""<EMAIL>"""', 'install_requires': "('ipython',... |
# Third-party
import astropy.units as u
import numpy as np
import pymc3 as pm
from pymc3.distributions import generate_samples
import aesara_theano_fallback.tensor as tt
import exoplanet.units as xu
__all__ = ['UniformLog', 'FixedCompanionMass']
class UniformLog(pm.Continuous):
def __init__(self, a, b, **kwargs... | [
"numpy.sqrt",
"numpy.log",
"aesara_theano_fallback.tensor.as_tensor_variable",
"numpy.zeros",
"numpy.random.uniform",
"pymc3.distributions.generate_samples",
"astropy.units.quantity_input"
] | [((1908, 1973), 'astropy.units.quantity_input', 'u.quantity_input', ([], {'sigma_K0': '(u.km / u.s)', 'P0': 'u.day', 'max_K': '(u.km / u.s)'}), '(sigma_K0=u.km / u.s, P0=u.day, max_K=u.km / u.s)\n', (1924, 1973), True, 'import astropy.units as u\n'), ((945, 973), 'numpy.random.uniform', 'np.random.uniform', ([], {'size... |
import json
import os
import unittest
import price_checker
from product import Product
def load_json(filename):
with open(filename) as json_file:
imported_file = json.load(json_file)
return imported_file
cwd = os.path.dirname(os.path.realpath(__file__))
products_data = load_json(r"" + cwd + "/t... | [
"price_checker.compare_prices",
"price_checker.open_policy",
"product.Product",
"unittest.main",
"os.path.realpath",
"json.load"
] | [((251, 277), 'os.path.realpath', 'os.path.realpath', (['__file__'], {}), '(__file__)\n', (267, 277), False, 'import os\n'), ((1689, 1704), 'unittest.main', 'unittest.main', ([], {}), '()\n', (1702, 1704), False, 'import unittest\n'), ((177, 197), 'json.load', 'json.load', (['json_file'], {}), '(json_file)\n', (186, 19... |
from collections import (
OrderedDict,
)
from unittest.mock import Mock
import pandas as pd
from datetime import (
datetime,
)
from fireant import *
from fireant.slicer.references import ReferenceType
from fireant.slicer.totals import get_totals_marker_for_dtype
from fireant.utils import (
format_dimensio... | [
"pandas.DataFrame.from_records",
"collections.OrderedDict",
"unittest.mock.Mock",
"fireant.utils.format_dimension_key",
"fireant.utils.format_metric_key",
"fireant.slicer.totals.get_totals_marker_for_dtype",
"pypika.functions.Count",
"pypika.functions.Sum",
"pandas.Index",
"pypika.Table",
"firea... | [((682, 720), 'pypika.Table', 'Table', (['"""politician"""'], {'schema': '"""politics"""'}), "('politician', schema='politics')\n", (687, 720), False, 'from pypika import JoinType, Table, functions as fn\n'), ((736, 769), 'pypika.Table', 'Table', (['"""voter"""'], {'schema': '"""politics"""'}), "('voter', schema='polit... |
import requests
from datetime import timedelta
from app.spotify_api import get_auth_header
def get_albums(data):
result = []
for item in data:
skip = False
header = get_auth_header()
songs_r = requests.get(f'https://api.spotify.com/v1/albums/{item["id"]}/tracks', headers=header)
... | [
"app.spotify_api.get_auth_header",
"datetime.timedelta",
"requests.get"
] | [((192, 209), 'app.spotify_api.get_auth_header', 'get_auth_header', ([], {}), '()\n', (207, 209), False, 'from app.spotify_api import get_auth_header\n'), ((228, 318), 'requests.get', 'requests.get', (['f"""https://api.spotify.com/v1/albums/{item[\'id\']}/tracks"""'], {'headers': 'header'}), '(f"https://api.spotify.com... |
import cv2
import numpy as np
from PyQt5.QtGui import QIntValidator
from PyQt5.QtWidgets import QDialog
from PyQt5.uic import loadUi
from utils import processing_utils as utils
IMAGE_DESCRIPT_DIALOG_UI = 'coreUI/image_description_dialog.ui'
class ImageDescriptionDialog(QDialog):
"""Image Description Dialog Windo... | [
"PyQt5.QtGui.QIntValidator",
"PyQt5.uic.loadUi",
"utils.processing_utils.display_img",
"cv2.filter2D",
"numpy.sum"
] | [((420, 458), 'PyQt5.uic.loadUi', 'loadUi', (['IMAGE_DESCRIPT_DIALOG_UI', 'self'], {}), '(IMAGE_DESCRIPT_DIALOG_UI, self)\n', (426, 458), False, 'from PyQt5.uic import loadUi\n'), ((1016, 1061), 'utils.processing_utils.display_img', 'utils.display_img', (['image', 'self.imageViewLabel'], {}), '(image, self.imageViewLab... |
import tensorflow as tf
from tensorflow import keras
import numpy as np
imdb = keras.datasets.imdb
(train_data, train_labels), (test_data, test_labels) = imdb.load_data(num_words=10000)
# A dictionary mapping words to an integer index
word_index = imdb.get_word_index()
# The first indices are reserved
word_index =... | [
"tensorflow.keras.preprocessing.sequence.pad_sequences",
"tensorflow.keras.Sequential",
"tensorflow.Session",
"tensorflow.keras.layers.Embedding",
"tensorflow.global_variables_initializer",
"tensorflow.keras.layers.Dense",
"tensorflow.keras.layers.GlobalAveragePooling1D",
"tensorflow.keras.backend.set... | [((1083, 1197), 'tensorflow.keras.preprocessing.sequence.pad_sequences', 'keras.preprocessing.sequence.pad_sequences', (['train_data'], {'value': "word_index['<PAD>']", 'padding': '"""post"""', 'maxlen': '(256)'}), "(train_data, value=word_index[\n '<PAD>'], padding='post', maxlen=256)\n", (1125, 1197), False, 'from... |
import vcr
from fast_arrow import Client
def gen_vcr():
return vcr.VCR(
cassette_library_dir='tests/fixtures_vcr',
record_mode='none',
match_on=['method', 'scheme', 'host', 'port', 'path', 'query'],
)
def gen_client():
auth_data = gen_auth_data()
client = Client(auth_data)
... | [
"fast_arrow.Client",
"vcr.VCR"
] | [((70, 208), 'vcr.VCR', 'vcr.VCR', ([], {'cassette_library_dir': '"""tests/fixtures_vcr"""', 'record_mode': '"""none"""', 'match_on': "['method', 'scheme', 'host', 'port', 'path', 'query']"}), "(cassette_library_dir='tests/fixtures_vcr', record_mode='none',\n match_on=['method', 'scheme', 'host', 'port', 'path', 'qu... |
import datetime
class Person(object):
def __init__(self, name):
self.name = name
try:
lastBlank = name.rindex(' ')
self.lastName = name[lastBlank+1:]
except:
self.lastName = name
self.birthday = None
def getName(self):
return sel... | [
"datetime.date.today",
"datetime.date"
] | [((1905, 1930), 'datetime.date', 'datetime.date', (['(1961)', '(8)', '(4)'], {}), '(1961, 8, 4)\n', (1918, 1930), False, 'import datetime\n'), ((1946, 1972), 'datetime.date', 'datetime.date', (['(1958)', '(8)', '(16)'], {}), '(1958, 8, 16)\n', (1959, 1972), False, 'import datetime\n'), ((557, 578), 'datetime.date.today... |
#
# Copyright 2022 DMetaSoul
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, ... | [
"argparse.ArgumentParser",
"pyspark.mllib.evaluation.RankingMetrics",
"subprocess.run",
"yaml.load",
"metaspore.SwingEstimator",
"sys.path.append",
"pyspark.sql.functions.collect_set"
] | [((760, 791), 'sys.path.append', 'sys.path.append', (['"""../../../../"""'], {}), "('../../../../')\n", (775, 791), False, 'import sys\n'), ((3667, 3723), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {'description': '"""Tuner information"""'}), "(description='Tuner information')\n", (3690, 3723), False, '... |
import shared_module
from shared_module import module_function as my_function, ModuleClass
class NewParent(object):
def do_useful_stuff(self):
i = shared_module.MODULE_CONTANT
my_function()
ModuleClass() | [
"shared_module.ModuleClass",
"shared_module.module_function"
] | [((198, 211), 'shared_module.module_function', 'my_function', ([], {}), '()\n', (209, 211), True, 'from shared_module import module_function as my_function, ModuleClass\n'), ((220, 233), 'shared_module.ModuleClass', 'ModuleClass', ([], {}), '()\n', (231, 233), False, 'from shared_module import module_function as my_fun... |
# -*- coding: utf-8 -*-
import os
os.environ['DJANGO_SETTINGS_MODULE']='settings'
from logistic import logisticdb
import webapp2 as webapp
from google.appengine.ext.webapp.util import run_wsgi_app
from google.appengine.ext.webapp import template
import numpy as np
import cgi
import cgitb
cgitb.enable()
def lesl... | [
"cgi.FieldStorage",
"os.path.dirname",
"numpy.dot",
"numpy.zeros",
"google.appengine.ext.webapp.util.run_wsgi_app",
"webapp2.WSGIApplication",
"cgitb.enable",
"google.appengine.ext.webapp.template.render"
] | [((291, 305), 'cgitb.enable', 'cgitb.enable', ([], {}), '()\n', (303, 305), False, 'import cgitb\n'), ((4709, 4772), 'webapp2.WSGIApplication', 'webapp.WSGIApplication', (["[('/.*', leslieOutputPage)]"], {'debug': '(True)'}), "([('/.*', leslieOutputPage)], debug=True)\n", (4731, 4772), True, 'import webapp2 as webapp\n... |
from cement import Controller, ex
from ..utils import databaseUtils, controllerUtils
class CollectionController(Controller):
class Meta:
label = 'collection controls'
@ex(
help='set currently active collection by id or name',
arguments=[
(
['-i', '--id'],
... | [
"cement.ex"
] | [((187, 337), 'cement.ex', 'ex', ([], {'help': '"""set currently active collection by id or name"""', 'arguments': "[(['-i', '--id'], {'help': 'collection id', 'action': 'store', 'dest': 'id'})]"}), "(help='set currently active collection by id or name', arguments=[(['-i',\n '--id'], {'help': 'collection id', 'actio... |
import copy
import typing
import splendor_sim.interfaces.coin.i_coin_reserve as i_coin_reserve
import splendor_sim.interfaces.coin.i_coin_type as i_coin_type
import splendor_sim.interfaces.coin.i_coin_type_manager as i_coin_type_manager
class CoinReserve(i_coin_reserve.ICoinReserve):
def __init__(
self,
... | [
"copy.copy"
] | [((1260, 1290), 'copy.copy', 'copy.copy', (['self._current_coins'], {}), '(self._current_coins)\n', (1269, 1290), False, 'import copy\n'), ((1383, 1413), 'copy.copy', 'copy.copy', (['self._max_coin_size'], {}), '(self._max_coin_size)\n', (1392, 1413), False, 'import copy\n')] |
import pytest
from dagger.serializer.as_pickle import AsPickle
from dagger.serializer.errors import DeserializationError, SerializationError
from dagger.serializer.protocol import Serializer
def test__conforms_to_protocol():
assert isinstance(AsPickle(), Serializer)
def test_extension():
assert AsPickle().... | [
"dagger.serializer.as_pickle.AsPickle",
"pytest.raises"
] | [((426, 436), 'dagger.serializer.as_pickle.AsPickle', 'AsPickle', ([], {}), '()\n', (434, 436), False, 'from dagger.serializer.as_pickle import AsPickle\n'), ((1105, 1115), 'dagger.serializer.as_pickle.AsPickle', 'AsPickle', ([], {}), '()\n', (1113, 1115), False, 'from dagger.serializer.as_pickle import AsPickle\n'), (... |
from torch.utils.data import Dataset
from PIL import Image
from pathlib import Path
import pandas
import torch
# from utils import data_utils
class InferenceDataset(Dataset):
def __init__(self, root, opts, split, transform=None, preprocess=None):
self.root = root
attributes_path = Path(root) / "list_attr_celeba... | [
"torch.as_tensor",
"PIL.Image.open",
"pandas.read_csv",
"pathlib.Path"
] | [((335, 400), 'pandas.read_csv', 'pandas.read_csv', (['attributes_path'], {'delim_whitespace': '(True)', 'header': '(1)'}), '(attributes_path, delim_whitespace=True, header=1)\n', (350, 400), False, 'import pandas\n'), ((588, 673), 'pandas.read_csv', 'pandas.read_csv', (['attributes_path'], {'delim_whitespace': '(True)... |
#!/usr/bin/python2.4
#
# Copyright 2009 Google Inc. All Rights Reserved.
"""Defines the generic robot classes.
This module provides the Robot class and RobotListener interface,
as well as some helper functions for web requests and responses.
"""
__author__ = '<EMAIL> (<NAME>)'
import events
import model
import ops
... | [
"util.CollapseJavaCollections",
"simplejson.dumps",
"util.ToUpperCamelCase",
"ops.CreateContext",
"util.Serialize",
"simplejson.loads",
"model.CreateEvent"
] | [((459, 486), 'simplejson.loads', 'simplejson.loads', (['json_body'], {}), '(json_body)\n', (475, 486), False, 'import simplejson\n'), ((554, 588), 'util.CollapseJavaCollections', 'util.CollapseJavaCollections', (['json'], {}), '(json)\n', (582, 588), False, 'import util\n'), ((601, 624), 'ops.CreateContext', 'ops.Crea... |
import pathlib
from setuptools import setup, find_packages
# The directory containing this file
HERE = pathlib.Path(__file__).parent
# The text of the README file
README = (HERE / "README.md").read_text()
setup(
description='Data extraction and processing for genre prediction using ML',
long_description=REA... | [
"setuptools.find_packages",
"pathlib.Path"
] | [((105, 127), 'pathlib.Path', 'pathlib.Path', (['__file__'], {}), '(__file__)\n', (117, 127), False, 'import pathlib\n'), ((872, 904), 'setuptools.find_packages', 'find_packages', ([], {'exclude': "('test',)"}), "(exclude=('test',))\n", (885, 904), False, 'from setuptools import setup, find_packages\n')] |
# Copyright (c) 2015-2019 The Switch Authors. All rights reserved.
# Licensed under the Apache License, Version 2.0, which is in the LICENSE file.
"""
Defines generation projects build-outs.
INPUT FILE FORMAT
Import data describing project builds. The following files are
expected in the input directory.
g... | [
"switch_model.financials.capital_recovery_factor",
"os.path.join",
"switch_model.tools.graph.graph",
"switch_model.utilities.scaling.get_assign_default_value_rule"
] | [((29938, 30029), 'switch_model.tools.graph.graph', 'graph', (['"""generation_capacity_per_period"""'], {'title': '"""Online Generation Capacity Per Period"""'}), "('generation_capacity_per_period', title=\n 'Online Generation Capacity Per Period')\n", (29943, 30029), False, 'from switch_model.tools.graph import gra... |
# https://www.codewars.com/kata/directions-reduction/train/python
# My solution
import re
def dirReduc(arr):
card = {"NORTH": "N", "SOUTH": "S", "EAST": "E", "WEST": "W"}
arr = "".join(map(lambda elem: card[elem], arr))
while "NS" in arr or "SN" in arr or "EW" in arr or "WE" in arr:
arr = r... | [
"re.sub"
] | [((319, 349), 're.sub', 're.sub', (['"""NS|SN|EW|WE"""', '""""""', 'arr'], {}), "('NS|SN|EW|WE', '', arr)\n", (325, 349), False, 'import re\n')] |
# Copyright 2020 Red Hat Inc.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing,... | [
"logging.getLogger",
"mrack.errors.ValidationError",
"datetime.datetime.now",
"mrack.errors.ProvisioningError",
"asyncio.gather"
] | [((784, 811), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (801, 811), False, 'import logging\n'), ((2694, 2708), 'datetime.datetime.now', 'datetime.now', ([], {}), '()\n', (2706, 2708), False, 'from datetime import datetime\n'), ((3370, 3384), 'datetime.datetime.now', 'datetime.now', (... |
"""
Capstone Project. Code to run on a LAPTOP (NOT the robot).
Displays the Graphical User Interface (GUI) and communicates with the robot.
Authors: Your professors (for the framework)
and <NAME>.
Winter term, 2018-2019.
"""
import mqtt_remote_method_calls as com
import tkinter
from tkinter import ttk
i... | [
"tkinter.ttk.Button",
"tkinter.ttk.Entry",
"random.randrange",
"tkinter.ttk.Frame",
"tkinter.ttk.Label",
"time.sleep",
"tkinter.Tk",
"tkinter.ttk.Progressbar",
"mqtt_remote_method_calls.MqttClient"
] | [((822, 846), 'mqtt_remote_method_calls.MqttClient', 'com.MqttClient', (['delegate'], {}), '(delegate)\n', (836, 846), True, 'import mqtt_remote_method_calls as com\n'), ((879, 892), 'time.sleep', 'time.sleep', (['(1)'], {}), '(1)\n', (889, 892), False, 'import time\n'), ((1104, 1116), 'tkinter.Tk', 'tkinter.Tk', ([], ... |
from aiogram.dispatcher import FSMContext
from aiogram.dispatcher.filters import Command
from aiogram.types import Message, CallbackQuery
from sqlalchemy.exc import NoResultFound
from loader import dp, logger_guru
from utils.database_manage.sql.sql_commands import DB_USERS
from utils.keyboards.admins_tools_kb import t... | [
"loader.logger_guru.exception",
"aiogram.dispatcher.filters.Command",
"loader.dp.callback_query_handler",
"utils.database_manage.sql.sql_commands.DB_USERS.update_personal_pass",
"utils.database_manage.sql.sql_commands.DB_USERS.check_personal_pass",
"loader.dp.message_handler",
"utils.misc.notify_users.s... | [((855, 958), 'loader.dp.callback_query_handler', 'dp.callback_query_handler', ([], {'text': "{'reset_user_codeword', 'make_newsletter'}", 'state': '"""admin_in_action"""'}), "(text={'reset_user_codeword', 'make_newsletter'},\n state='admin_in_action')\n", (880, 958), False, 'from loader import dp, logger_guru\n'), ... |
import math
#matplotlib keyboard
noteKeys = "<KEY>'"
class keyboard:
def __init__(self,keystring="<KEY>"):
self.keys = keystring
self._keepinscope = None
self.keysDown = set()
def plot(self,ival=50):
import matplotlib.pyplot as plt
import matplotlib.animation as animati... | [
"filters.biquadPeak",
"matplotlib.animation.FuncAnimation",
"math.log",
"math.exp",
"matplotlib.pyplot.subplots"
] | [((1047, 1068), 'math.exp', 'math.exp', (['(2 * math.pi)'], {}), '(2 * math.pi)\n', (1055, 1068), False, 'import math\n'), ((2865, 2876), 'math.log', 'math.log', (['(2)'], {}), '(2)\n', (2873, 2876), False, 'import math\n'), ((341, 355), 'matplotlib.pyplot.subplots', 'plt.subplots', ([], {}), '()\n', (353, 355), True, ... |
# -*- coding: utf-8 -*-
# Form implementation generated from reading ui file 'Calculation.ui'
#
# Created by: PyQt5 UI code generator 5.9.2
#
# WARNING! All changes made in this file will be lost!
from Calculation.CalculatingLast5Days import CalculatingLast5Days
from PyQt5 import QtCore, QtGui, QtWidgets
... | [
"PyQt5.QtWidgets.QWidget",
"PyQt5.QtWidgets.QMainWindow",
"PyQt5.QtGui.QFont",
"PyQt5.QtWidgets.QSizePolicy",
"PyQt5.QtCore.QMetaObject.connectSlotsByName",
"Calculation.CalculatingLast5Days.CalculatingLast5Days",
"PyQt5.QtWidgets.QHBoxLayout",
"PyQt5.QtCore.QRect",
"PyQt5.QtWidgets.QLabel",
"PyQt... | [((8245, 8277), 'PyQt5.QtWidgets.QApplication', 'QtWidgets.QApplication', (['sys.argv'], {}), '(sys.argv)\n', (8267, 8277), False, 'from PyQt5 import QtCore, QtGui, QtWidgets\n'), ((8296, 8319), 'PyQt5.QtWidgets.QMainWindow', 'QtWidgets.QMainWindow', ([], {}), '()\n', (8317, 8319), False, 'from PyQt5 import QtCore, QtG... |
import os
import sys
import inspect
from ruamel import yaml
def get_script_dir(follow_symlinks=True):
if getattr(sys, 'frozen', False): # py2exe, PyInstaller, cx_Freeze
path = os.path.abspath(sys.executable)
else:
path = inspect.getabsfile(get_script_dir)
if follow_symlinks:
... | [
"os.path.exists",
"os.listdir",
"os.makedirs",
"os.path.realpath",
"os.path.dirname",
"inspect.getabsfile",
"os.path.abspath"
] | [((370, 391), 'os.path.dirname', 'os.path.dirname', (['path'], {}), '(path)\n', (385, 391), False, 'import os\n'), ((453, 471), 'os.path.dirname', 'os.path.dirname', (['d'], {}), '(d)\n', (468, 471), False, 'import os\n'), ((696, 717), 'os.listdir', 'os.listdir', (['directory'], {}), '(directory)\n', (706, 717), False,... |
import random
from typing import Any
import factory
import pytz
from django.utils.text import slugify
from ap.apps.events.models import EVENT_TYPE_CHOICES, Event, Organization, Route
from ap.apps.users.constants import COUNTRY_CHOICES
from ap.apps.users.models import User
from ap.apps.events.factory_data import RANDO... | [
"django.utils.text.slugify",
"ap.apps.events.models.Organization.objects.all",
"random.choice",
"factory.Faker",
"ap.apps.users.models.User.objects.all"
] | [((552, 590), 'factory.Faker', 'factory.Faker', (['"""text"""'], {'max_nb_chars': '(60)'}), "('text', max_nb_chars=60)\n", (565, 590), False, 'import factory\n'), ((601, 621), 'factory.Faker', 'factory.Faker', (['"""url"""'], {}), "('url')\n", (614, 621), False, 'import factory\n'), ((1003, 1045), 'factory.Faker', 'fac... |
#!/usr/bin/env python
from __future__ import division
import math
import random
import time
import genpy
import rospy
import roslib
from qt_gui.plugin import Plugin
from python_qt_binding.QtCore import Slot, QSignalMapper, QTimer
from python_qt_binding.QtWidgets import QHeaderView, QTableWidgetItem
from rqt_command_... | [
"python_qt_binding.QtCore.Slot",
"rqt_command_publisher.command_publisher_widget.CommandPublisherWidget",
"python_qt_binding.QtCore.QTimer",
"roslib.message.get_message_class",
"rospy.loginfo",
"python_qt_binding.QtCore.QSignalMapper"
] | [((1935, 1964), 'python_qt_binding.QtCore.Slot', 'Slot', (['str', 'str', 'float', 'object'], {}), '(str, str, float, object)\n', (1939, 1964), False, 'from python_qt_binding.QtCore import Slot, QSignalMapper, QTimer\n'), ((3702, 3711), 'python_qt_binding.QtCore.Slot', 'Slot', (['int'], {}), '(int)\n', (3706, 3711), Fal... |
import time
import os
import arcade
import argparse
import gym
from gym import spaces
import swarm_env
import numpy as np
import random
import sys
sys.path.insert(0, '..')
from objects import SwarmSimulator
# Running experiment 22 in standalone file.
def experiment_runner(SWARM_SIZE = 15, ARENA_WIDTH = 600, ARENA_HEI... | [
"random.uniform",
"sys.path.insert",
"random.randint",
"numpy.argmax",
"numpy.max",
"numpy.exp",
"numpy.zeros",
"os.path.isdir",
"os.mkdir",
"arcade.run",
"objects.SwarmSimulator",
"time.time",
"gym.make"
] | [((147, 171), 'sys.path.insert', 'sys.path.insert', (['(0)', '""".."""'], {}), "(0, '..')\n", (162, 171), False, 'import sys\n'), ((352, 363), 'time.time', 'time.time', ([], {}), '()\n', (361, 363), False, 'import time\n'), ((1595, 1638), 'gym.make', 'gym.make', (['"""humanswarm-v0"""'], {'maze_size': 'GRID_X'}), "('hu... |
from helpers import *
from discord.ext import commands
from discord.ext.commands import Cog, Bot, command, Context
import database as db
import discord
import asyncio
def setup(bot):
bot.add_cog(Commands(bot))
class Commands(Cog):
def __init__(self, bot: Bot):
self.bot = bot
# List spread acros... | [
"database.subscribe",
"database.is_subscribed",
"database.get_subscriptions",
"database.get_all_tv_shows",
"asyncio.wait",
"database.unsubscribe",
"database.search_tv_show",
"discord.ext.commands.command"
] | [((2097, 2173), 'discord.ext.commands.command', 'command', ([], {'aliases': "['shows']", 'name': '"""tvshows"""', 'help': '"""Lists available TV Shows."""'}), "(aliases=['shows'], name='tvshows', help='Lists available TV Shows.')\n", (2104, 2173), False, 'from discord.ext.commands import Cog, Bot, command, Context\n'),... |
#!/usr/bin/python
# -*- coding: UTF-8 -*-
import numpy as np
class DMatrix:
def __init__(self, data_arr, missing={np.nan, 0}):
"""
:param data_arr: 样本特征 (不含标签)
:param missing: 缺失值的集合, 若特征值在此集合中, 则认为其为缺失值
"""
# N 样本总个数( 包含缺出现缺失值的样本 )
# m 特征的总数
self.N, sel... | [
"numpy.shape"
] | [((326, 344), 'numpy.shape', 'np.shape', (['data_arr'], {}), '(data_arr)\n', (334, 344), True, 'import numpy as np\n')] |
from bs4 import BeautifulSoup
import requests
import pandas as pd
url = 'https://mexico.as.com/resultados/futbol/mexico_clausura/clasificacion/'
page = requests.get(url)
soup = BeautifulSoup(page.content, 'html.parser')
#Equipos
eq = soup.find_all('span', class_='nombre-equipo')
equipos = list()
count = 0
for i in e... | [
"bs4.BeautifulSoup",
"requests.get"
] | [((153, 170), 'requests.get', 'requests.get', (['url'], {}), '(url)\n', (165, 170), False, 'import requests\n'), ((178, 220), 'bs4.BeautifulSoup', 'BeautifulSoup', (['page.content', '"""html.parser"""'], {}), "(page.content, 'html.parser')\n", (191, 220), False, 'from bs4 import BeautifulSoup\n')] |
import inspect
from flask import render_template, request, jsonify
from ..configurables import app
from .. import metrics
from ..metrics import metric_classes
def get_metrics(add_class=False):
records = []
for name, metric in metric_classes.iteritems():
if metric.show_in_ui:
new_record = {... | [
"flask.render_template",
"flask.jsonify"
] | [((1298, 1322), 'flask.jsonify', 'jsonify', ([], {'metrics': 'records'}), '(metrics=records)\n', (1305, 1322), False, 'from flask import render_template, request, jsonify\n'), ((1977, 2071), 'flask.render_template', 'render_template', (['"""forms/metric_configuration.html"""'], {'form': 'metric_form', 'action': 'reques... |
#!/usr/bin/python
# -*- coding: utf-8 -*-
from setuptools import setup
setup(
name='pyramid_zipkin-example',
version='0.1',
author='OpenZipkin',
author_email='<EMAIL>',
license='Apache 2.0',
url='https://github.com/openzipkin/pyramid_zipkin-example',
description='See how much time python se... | [
"setuptools.setup"
] | [((72, 402), 'setuptools.setup', 'setup', ([], {'name': '"""pyramid_zipkin-example"""', 'version': '"""0.1"""', 'author': '"""OpenZipkin"""', 'author_email': '"""<EMAIL>"""', 'license': '"""Apache 2.0"""', 'url': '"""https://github.com/openzipkin/pyramid_zipkin-example"""', 'description': '"""See how much time python s... |
import os
from glob import glob
import tensorflow as tf
from tensorflow.keras.applications.mobilenet_v2 import MobileNetV2, preprocess_input
from tensorflow.keras.preprocessing.image import img_to_array, load_img
from tensorflow.keras.models import Model, load_model
from tensorflow.keras.layers import Conv2D,... | [
"tensorflow.keras.applications.mobilenet_v2.MobileNetV2",
"os.path.dirname",
"glob.glob",
"tensorflow.keras.applications.mobilenet_v2.preprocess_input",
"numpy.expand_dims",
"tensorflow.keras.models.Model",
"cv2.resize",
"cv2.imread",
"numpy.save"
] | [((816, 865), 'tensorflow.keras.applications.mobilenet_v2.MobileNetV2', 'MobileNetV2', ([], {'weights': '"""imagenet"""', 'include_top': '(True)'}), "(weights='imagenet', include_top=True)\n", (827, 865), False, 'from tensorflow.keras.applications.mobilenet_v2 import MobileNetV2, preprocess_input\n'), ((906, 974), 'ten... |
import argparse
import matplotlib.pyplot as plt
import meshcut
import numpy as np
import pandas
import seaborn as sns
import pandas as pd
import sys, os
import math
#from scipy.stats import norm
SAVE_PATH = os.path.join(os.path.expanduser("~"),'PycharmProjects/Gibson_Exercise/examples/plot_result/')
WAY_PATH = os.path... | [
"matplotlib.pyplot.grid",
"argparse.ArgumentParser",
"matplotlib.pyplot.ylabel",
"meshcut.cross_section",
"matplotlib.pyplot.xlabel",
"numpy.absolute",
"matplotlib.pyplot.plot",
"matplotlib.pyplot.style.use",
"os.path.join",
"numpy.array",
"matplotlib.pyplot.figure",
"numpy.zeros",
"numpy.mi... | [((221, 244), 'os.path.expanduser', 'os.path.expanduser', (['"""~"""'], {}), "('~')\n", (239, 244), False, 'import sys, os\n'), ((326, 349), 'os.path.expanduser', 'os.path.expanduser', (['"""~"""'], {}), "('~')\n", (344, 349), False, 'import sys, os\n'), ((705, 720), 'numpy.array', 'np.array', (['verts'], {}), '(verts)... |
"""A module for the uFJC single-chain model in the isometric ensemble.
This module consist of the class ``uFJCIsometric`` which contains
methods for computing single-chain quantities
in the isometric (constant end-to-end vector) thermodynamic ensemble.
Example:
Import and instantiate the class:
... | [
"numpy.log",
"numpy.linalg.norm"
] | [((11454, 11494), 'numpy.linalg.norm', 'la.norm', (['(config[j, :] - config[j - 1, :])'], {}), '(config[j, :] - config[j - 1, :])\n', (11461, 11494), True, 'import numpy.linalg as la\n'), ((10199, 10234), 'numpy.log', 'np.log', (['(1 + eta * coth / self.kappa)'], {}), '(1 + eta * coth / self.kappa)\n', (10205, 10234), ... |
import inspect
import re
def debug_str(arg):
'''Return string of arg varible name and str(arg) value.'''
frame = inspect.currentframe().f_back
s = inspect.getframeinfo(frame).code_context[0]
r = re.search(r"\((.*)\)", s).group(1)
return str("{} = {}".format(r, arg))
def debug_print(arg):
'''... | [
"re.search",
"inspect.currentframe",
"inspect.getframeinfo"
] | [((123, 145), 'inspect.currentframe', 'inspect.currentframe', ([], {}), '()\n', (143, 145), False, 'import inspect\n'), ((378, 400), 'inspect.currentframe', 'inspect.currentframe', ([], {}), '()\n', (398, 400), False, 'import inspect\n'), ((161, 188), 'inspect.getframeinfo', 'inspect.getframeinfo', (['frame'], {}), '(f... |
"""
Author: <NAME>
Date: 10 April 2021
"""
import logging
import os
from typing import List
from dataclasses import dataclass, asdict
import click
from click import Context
from sla_cli.src.cli.context import COMMAND_CONTEXT_SETTINGS
from sla_cli.src.cli.utils import kwargs_to_dataclass, default_from_conte... | [
"logging.getLogger",
"click.argument",
"sla_cli.src.download.isic.IsicMetadataDownloader",
"click.option",
"sla_cli.src.cli.utils.default_from_context",
"sla_cli.src.db.accessors.AccessorFactory.create_datasets",
"sla_cli.src.cli.utils.kwargs_to_dataclass",
"click.command",
"sla_cli.src.download.Dow... | [((724, 751), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (741, 751), False, 'import logging\n'), ((931, 1021), 'click.command', 'click.command', ([], {'short_help': '"""Downloads available datasets."""'}), "(**COMMAND_CONTEXT_SETTINGS, short_help=\n 'Downloads available datasets.')... |
import json
print("Generating stamp files")
stamps = [
"armor_plate",
"armor_trim",
"arrow_head",
"arrow_shaft",
"axe_head",
"binding",
"boots_core",
"bow_limb",
"bow_string",
"broad_axe_head",
"chest_core",
"cross_guard",
"emerald",
"excavator_head",
"fletching",
"hammer_head",
"han... | [
"json.dump"
] | [((929, 963), 'json.dump', 'json.dump', (['data', 'outfile'], {'indent': '(2)'}), '(data, outfile, indent=2)\n', (938, 963), False, 'import json\n')] |
import torch
from uninas.modules.mixed.mixedop import AbstractDependentMixedOp
from uninas.methods.strategies.manager import StrategyManager
from uninas.register import Register
@Register.network_mixed_op()
class SplitWeightsMixedOp(AbstractDependentMixedOp):
"""
all op choices on one path in parallel,
th... | [
"uninas.register.Register.network_mixed_op",
"uninas.methods.strategies.manager.StrategyManager"
] | [((181, 208), 'uninas.register.Register.network_mixed_op', 'Register.network_mixed_op', ([], {}), '()\n', (206, 208), False, 'from uninas.register import Register\n'), ((1253, 1270), 'uninas.methods.strategies.manager.StrategyManager', 'StrategyManager', ([], {}), '()\n', (1268, 1270), False, 'from uninas.methods.strat... |
"""
Plot various visualizations
"""
import sys
import json
from collections import Counter
import numpy as np
import scipy.stats as scstats
import matplotlib
matplotlib.use("agg")
import matplotlib.pyplot as plt
import matplotlib.cm as cm
import matplotlib.patches as mp
from zipteedo.util import GzipFileType, load... | [
"zipteedo.stats.make_bias_table",
"matplotlib.pyplot.ylabel",
"numpy.polyfit",
"zipteedo.util.first",
"numpy.array",
"zipteedo.stats.get_correlations",
"matplotlib.pyplot.errorbar",
"sys.exit",
"argparse.ArgumentParser",
"matplotlib.pyplot.xlabel",
"matplotlib.pyplot.plot",
"zipteedo.viz.violi... | [((160, 181), 'matplotlib.use', 'matplotlib.use', (['"""agg"""'], {}), "('agg')\n", (174, 181), False, 'import matplotlib\n'), ((1343, 1365), 'zipteedo.stats.get_correlations', 'get_correlations', (['data'], {}), '(data)\n', (1359, 1365), False, 'from zipteedo.stats import get_correlations, get_data_efficiencies, make_... |
import plotly.plotly as py
import plotly.graph_objs as go
import plotly.offline as offline
import sys, os
from .CreateTableFromDatabase import getRankingsFromDatabase
# Using the plotly API creates a few plots
def createAndUploadPlots(table, plotName):
# Read plotly username and API key from file (to avoid acci... | [
"plotly.offline.plot",
"plotly.plotly.plot",
"os.path.dirname",
"plotly.graph_objs.Layout",
"plotly.graph_objs.Figure"
] | [((1285, 1334), 'plotly.graph_objs.Layout', 'go.Layout', ([], {'title': "('%s' % plotName)", 'barmode': '"""group"""'}), "(title='%s' % plotName, barmode='group')\n", (1294, 1334), True, 'import plotly.graph_objs as go\n'), ((1371, 1406), 'plotly.graph_objs.Figure', 'go.Figure', ([], {'data': 'data', 'layout': 'layout'... |
# -*- coding: utf-8 -*-
from __future__ import unicode_literals
from django.db import models, migrations
from django.conf import settings
class Migration(migrations.Migration):
dependencies = [
migrations.swappable_dependency(settings.AUTH_USER_MODEL),
('proposals', '0003_auto_20150113_1401'),
... | [
"django.db.models.ForeignKey",
"django.db.models.BooleanField",
"django.db.models.AutoField",
"django.db.models.DateTimeField",
"django.db.migrations.swappable_dependency",
"django.db.models.CharField"
] | [((210, 267), 'django.db.migrations.swappable_dependency', 'migrations.swappable_dependency', (['settings.AUTH_USER_MODEL'], {}), '(settings.AUTH_USER_MODEL)\n', (241, 267), False, 'from django.db import models, migrations\n'), ((523, 616), 'django.db.models.AutoField', 'models.AutoField', ([], {'verbose_name': '"""ID"... |
#!/usr/bin/env python3
import statistics
import os
import glob
from tkinter import filedialog
from tkinter import * # noqa
import pandas as pd
from eventcodes import eventcodes_dictionary
from natsort import natsorted, ns
import matplotlib.pyplot as plt
import numpy as np
import datetime
__all__ = ["loop_over_days",... | [
"statistics.mean",
"tkinter.filedialog.askdirectory",
"pandas.read_csv",
"datetime.datetime.strptime",
"numpy.delete",
"matplotlib.pyplot.xlabel",
"os.path.join",
"pandas.DataFrame.from_dict",
"matplotlib.pyplot.figure",
"glob.glob",
"natsort.natsorted",
"pandas.to_numeric",
"pandas.DataFram... | [((4274, 4307), 'pandas.DataFrame', 'pd.DataFrame', ([], {'columns': 'column_list'}), '(columns=column_list)\n', (4286, 4307), True, 'import pandas as pd\n'), ((5395, 5428), 'pandas.DataFrame', 'pd.DataFrame', ([], {'columns': 'column_list'}), '(columns=column_list)\n', (5407, 5428), True, 'import pandas as pd\n'), ((7... |
#!/usr/bin/env python3
# -*- coding: UTF-8 -*-
"""
boot script, do initial stuff here, similar to the setup() function on Arduino
"""
import esp
import gc
import network
import time
# custom packages
from be_helpers.led_helper import Led
# set clock speed to 240MHz instead of default 160MHz
# machine.freq(24000000... | [
"be_helpers.led_helper.Led",
"time.sleep",
"network.WLAN",
"gc.collect",
"esp.osdebug"
] | [((354, 371), 'esp.osdebug', 'esp.osdebug', (['None'], {}), '(None)\n', (365, 371), False, 'import esp\n'), ((379, 384), 'be_helpers.led_helper.Led', 'Led', ([], {}), '()\n', (382, 384), False, 'from be_helpers.led_helper import Led\n'), ((443, 471), 'network.WLAN', 'network.WLAN', (['network.STA_IF'], {}), '(network.S... |
#!/usr/bin/env python
# -*- coding: utf8 -*-
import os
from functools import reduce
from Log import log
__title__ = 'Structure runner'
__all__ = ['Structure']
__author__ = '<NAME> <<EMAIL>>'
class Structure(object):
@staticmethod
def get_directory_structure(confdir: str) -> dict:
"""
Create... | [
"Log.log.error",
"xml.etree.ElementTree.parse",
"configparser.ConfigParser",
"functools.reduce",
"os.path.join",
"yaml.load",
"json.load",
"os.walk"
] | [((565, 581), 'os.walk', 'os.walk', (['confdir'], {}), '(confdir)\n', (572, 581), False, 'import os\n'), ((998, 1014), 'os.walk', 'os.walk', (['confdir'], {}), '(confdir)\n', (1005, 1014), False, 'import os\n'), ((3732, 3746), 'configparser.ConfigParser', 'ConfigParser', ([], {}), '()\n', (3744, 3746), False, 'from con... |
import ray
from ray.experimental.workflow import storage
from ray.experimental.workflow import workflow_storage
def some_func(x):
return x + 1
def some_func2(x):
return x - 1
def test_raw_storage():
ray.init()
workflow_id = test_workflow_storage.__name__
raw_storage = storage.get_global_storag... | [
"ray.shutdown",
"ray.experimental.workflow.workflow_storage.WorkflowStorage",
"ray.experimental.workflow.storage.get_global_storage",
"ray.put",
"ray.init",
"ray.experimental.workflow.workflow_storage.StepInspectResult"
] | [((217, 227), 'ray.init', 'ray.init', ([], {}), '()\n', (225, 227), False, 'import ray\n'), ((295, 323), 'ray.experimental.workflow.storage.get_global_storage', 'storage.get_global_storage', ([], {}), '()\n', (321, 323), False, 'from ray.experimental.workflow import storage\n'), ((517, 541), 'ray.put', 'ray.put', (['ob... |
from django.db import models
from django.utils import timezone
class BaseCoin(models.Model):
date = models.DateField(default=timezone.now)
exchange = models.CharField(max_length=100, blank=True, default='')
price = models.DecimalField(max_digits=12, decimal_places=4, default=0)
high = models.DecimalFi... | [
"django.db.models.DecimalField",
"django.db.models.DateField",
"django.db.models.CharField"
] | [((106, 144), 'django.db.models.DateField', 'models.DateField', ([], {'default': 'timezone.now'}), '(default=timezone.now)\n', (122, 144), False, 'from django.db import models\n'), ((160, 216), 'django.db.models.CharField', 'models.CharField', ([], {'max_length': '(100)', 'blank': '(True)', 'default': '""""""'}), "(max... |
from __future__ import print_function
from ghidra.program.model.address import Address
import platform
import ghidra
if (platform.system() != "Java"):
from ghidra_builtins import *
monitor = ghidra.util.task.TaskMonitor.DUMMY
def getFuncInstructions(function):
instructions = []
instruction = getFirstIns... | [
"ghidra.app.cmd.disassemble.DisassembleCommand",
"ghidra.program.util.ProgramSelection",
"ghidra.util.datastruct.ListAccumulator",
"ghidra.program.model.data.PointerDataType",
"ghidra.util.search.memory.SearchInfo",
"platform.system",
"ghidra.app.plugin.core.searchmem.SearchData.createSearchData"
] | [((123, 140), 'platform.system', 'platform.system', ([], {}), '()\n', (138, 140), False, 'import platform\n'), ((1047, 1135), 'ghidra.app.plugin.core.searchmem.SearchData.createSearchData', 'ghidra.app.plugin.core.searchmem.SearchData.createSearchData', (['"""search"""', 'values', 'mask'], {}), "('search',\n values,... |
import unittest
import numpy as np
import sklearn_supp.random_coordinates as random_coordinates
class TestRandomCoordinateForestClassifier(unittest.TestCase):
"""These are just some simple sanity checks to make sure we don't get
exceptions.
"""
def test_simple(self):
X = [[0], [1]]
... | [
"numpy.array",
"numpy.all",
"sklearn_supp.random_coordinates.RandomCoordinateForestClassifier"
] | [((352, 420), 'sklearn_supp.random_coordinates.RandomCoordinateForestClassifier', 'random_coordinates.RandomCoordinateForestClassifier', ([], {'n_estimators': '(50)'}), '(n_estimators=50)\n', (403, 420), True, 'import sklearn_supp.random_coordinates as random_coordinates\n'), ((537, 556), 'numpy.all', 'np.all', (['(y_p... |
# Author : <NAME>
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# d... | [
"os.listdir",
"utils.fetch_action_ids",
"pandas.read_csv",
"os.path.join",
"numpy.array",
"numpy.sum",
"warnings.simplefilter",
"utils.fetch_parsed_recipe"
] | [((824, 886), 'warnings.simplefilter', 'warnings.simplefilter', ([], {'action': '"""ignore"""', 'category': 'FutureWarning'}), "(action='ignore', category=FutureWarning)\n", (845, 886), False, 'import warnings\n'), ((974, 992), 'os.listdir', 'os.listdir', (['folder'], {}), '(folder)\n', (984, 992), False, 'import os\n'... |
from flask import Flask, current_app, request
from flask_babelex import get_locale
from flask_sqlalchemy import SQLAlchemy
from flask_transalchemy.model import TranslationMixin
class TransAlchemy(object):
"""Flask-TransAlchemy extension class.
:param app: Flask application instance
:param db: Flask-SQLA... | [
"flask_babelex.get_locale",
"flask.current_app.extensions.get"
] | [((2719, 2762), 'flask.current_app.extensions.get', 'current_app.extensions.get', (['"""babel_alchemy"""'], {}), "('babel_alchemy')\n", (2745, 2762), False, 'from flask import Flask, current_app, request\n'), ((3159, 3202), 'flask.current_app.extensions.get', 'current_app.extensions.get', (['"""babel_alchemy"""'], {}),... |
# Copyright 2016, The NIG Authors. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License"); you may not
# use this file except in compliance with the License. You may obtain a copy of
# the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable la... | [
"timeit.default_timer",
"re.escape"
] | [((1001, 1016), 'timeit.default_timer', 'default_timer', ([], {}), '()\n', (1014, 1016), False, 'from timeit import default_timer\n'), ((1101, 1116), 'timeit.default_timer', 'default_timer', ([], {}), '()\n', (1114, 1116), False, 'from timeit import default_timer\n'), ((1336, 1351), 're.escape', 're.escape', (['char'],... |
from __future__ import print_function, absolute_import
from random import random
from time import sleep
import threading
import pygame
import sys
try:
from olympe.messages.ardrone3.GPSSettingsState import GPSFixStateChanged
from olympe.messages.ardrone3.Piloting import TakeOff, Landing
from olympe.messag... | [
"olympe.Drone",
"pygame.init",
"olympe.messages.ardrone3.Piloting.Landing",
"olympe.messages.ardrone3.PilotingState.FlyingStateChanged",
"pygame.joystick.Joystick",
"time.sleep",
"olympe.messages.ardrone3.Piloting.TakeOff",
"sys.exit",
"pygame.event.pump",
"olympe.messages.ardrone3.GPSSettingsStat... | [((4034, 4087), 'olympe.Drone', 'olympe.Drone', (['JoystickTeleop.SIMULATED_IP'], {'loglevel': '(0)'}), '(JoystickTeleop.SIMULATED_IP, loglevel=0)\n', (4046, 4087), False, 'import olympe\n'), ((4239, 4250), 'sys.exit', 'sys.exit', (['(0)'], {}), '(0)\n', (4247, 4250), False, 'import sys\n'), ((499, 510), 'sys.exit', 's... |
import argparse
import csv
import glob
import json
import os
import re
import time
from math import ceil
from shutil import copyfile, rmtree
from urllib.parse import urlparse
import requests
from pyproc import Lpse, __version__
from pyproc.helpers import DetilDownloader
from urllib3 import disable_warnings
from urll... | [
"csv.DictWriter",
"pyproc.helpers.DetilDownloader",
"pathlib.Path.home",
"time.sleep",
"argparse.ArgumentParser",
"pyproc.Lpse",
"csv.reader",
"glob.glob",
"os.rename",
"csv.writer",
"os.path.isfile",
"urllib3.disable_warnings",
"os.path.dirname",
"re.findall",
"requests.session",
"mat... | [((2671, 2709), 'os.makedirs', 'os.makedirs', (['detail_dir'], {'exist_ok': '(True)'}), '(detail_dir, exist_ok=True)\n', (2682, 2709), False, 'import os\n'), ((3397, 3436), 'os.path.join', 'os.path.join', (['folder_name', '"""detil"""', '"""*"""'], {}), "(folder_name, 'detil', '*')\n", (3409, 3436), False, 'import os\n... |
import tkinter as tk
from tkinter import ttk, messagebox, font, StringVar
from tkcalendar import DateEntry
from tkcrud.controller.client_controller import ClientController,\
saving_updating, get_clients, window_popup
class FormClientRegister(tk.Toplevel, ClientController):
def __init__(self, master, tree):
... | [
"tkinter.LabelFrame",
"tkinter.Entry",
"tkinter.Toplevel.__init__",
"tkcalendar.DateEntry",
"tkinter.ttk.Scrollbar",
"tkcrud.controller.client_controller.get_clients",
"tkinter.font.Font",
"tkinter.StringVar",
"tkinter.Label",
"tkinter.ttk.Combobox",
"tkcrud.controller.client_controller.window_p... | [((327, 361), 'tkinter.Toplevel.__init__', 'tk.Toplevel.__init__', (['self', 'master'], {}), '(self, master)\n', (347, 361), True, 'import tkinter as tk\n'), ((444, 487), 'tkinter.LabelFrame', 'tk.LabelFrame', (['self'], {'text': '"""Register client"""'}), "(self, text='Register client')\n", (457, 487), True, 'import t... |
import sys, os, re, json, pprint, traceback
from pathlib import Path
from aiohttp import web
routes = web.RouteTableDef()
async def app():
from web_chains_202105 import directories
app = web.Application(middlewares=[exception_middleware])
app.add_routes(routes)
app.router.add_static("/js/", path="js",... | [
"web_chains_202105.chart.get_map",
"pathlib.Path",
"web_chains_202105.table_page.table_page",
"web_chains_202105.directories.load",
"web_chains_202105.index_page.index_page",
"aiohttp.web.Application",
"sys.exc_info",
"aiohttp.web.RouteTableDef",
"web_chains_202105.chain_page.chain_page",
"os.getp... | [((103, 122), 'aiohttp.web.RouteTableDef', 'web.RouteTableDef', ([], {}), '()\n', (120, 122), False, 'from aiohttp import web\n'), ((197, 248), 'aiohttp.web.Application', 'web.Application', ([], {'middlewares': '[exception_middleware]'}), '(middlewares=[exception_middleware])\n', (212, 248), False, 'from aiohttp import... |
# Copyright 2019 Zuru Tech HK Limited. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable la... | [
"tensorflow.train.Checkpoint",
"tensorflow.metrics.Mean",
"tensorflow.math.log",
"tensorflow.metrics.Accuracy",
"tensorflow.keras.layers.Dense",
"tensorflow.math.exp",
"pathlib.Path",
"tensorflow.image.grayscale_to_rgb",
"tensorflow.math.reduce_mean",
"tensorflow_hub.KerasLayer",
"tensorflow.ker... | [((11863, 11889), 'tensorflow.math.exp', 'tf.math.exp', (['kl_divergence'], {}), '(kl_divergence)\n', (11874, 11889), True, 'import tensorflow as tf\n'), ((12157, 12220), 'tensorflow.keras.losses.SparseCategoricalCrossentropy', 'tf.keras.losses.SparseCategoricalCrossentropy', ([], {'from_logits': '(True)'}), '(from_log... |
# -*- coding: utf-8 -*-
"""Database module, including the SQLAlchemy database object and DB-related utilities."""
import json
from sqlalchemy.inspection import inspect
from sqlalchemy.orm import relationship
from .extensions import db
# Alias common SQLAlchemy names
Column = db.Column
relationship = relationship
c... | [
"sqlalchemy.inspection.inspect"
] | [((1383, 1396), 'sqlalchemy.inspection.inspect', 'inspect', (['self'], {}), '(self)\n', (1390, 1396), False, 'from sqlalchemy.inspection import inspect\n')] |
from owslib.wms import WebMapService
from owslib import crs
from PIL import Image, ImageEnhance, ImageFilter
import cv2
import numpy as np
from pyspark import SparkContext
from pyproj import Proj
c = crs.Crs('EPSG:3857')
wms = WebMapService('http://www.ign.es/wms-inspire/pnoa-ma', version='1.3.0')
box = 1000 # m?
x=... | [
"cv2.imwrite",
"cv2.drawContours",
"owslib.wms.WebMapService",
"cv2.inRange",
"owslib.crs.Crs",
"cv2.arcLength",
"cv2.bitwise_and",
"cv2.contourArea",
"numpy.array",
"cv2.approxPolyDP",
"cv2.cvtColor",
"pyproj.Proj",
"cv2.findContours",
"cv2.Canny",
"cv2.imread",
"cv2.boundingRect"
] | [((201, 221), 'owslib.crs.Crs', 'crs.Crs', (['"""EPSG:3857"""'], {}), "('EPSG:3857')\n", (208, 221), False, 'from owslib import crs\n'), ((228, 299), 'owslib.wms.WebMapService', 'WebMapService', (['"""http://www.ign.es/wms-inspire/pnoa-ma"""'], {'version': '"""1.3.0"""'}), "('http://www.ign.es/wms-inspire/pnoa-ma', ver... |
"""Set of basic widgets for BioImageIT
Classes
-------
BiWebBrowser
"""
from qtpy.QtWebEngineWidgets import QWebEngineView
from qtpy.QtWidgets import (QWidget, QPushButton,
QHBoxLayout, QVBoxLayout)
class BiWebBrowser(QWidget):
def __init__(self, parent: QWidget):
super(Bi... | [
"qtpy.QtWidgets.QVBoxLayout",
"qtpy.QtWidgets.QWidget",
"qtpy.QtWebEngineWidgets.QWebEngineView",
"qtpy.QtWidgets.QPushButton",
"qtpy.QtWidgets.QHBoxLayout"
] | [((377, 390), 'qtpy.QtWidgets.QVBoxLayout', 'QVBoxLayout', ([], {}), '()\n', (388, 390), False, 'from qtpy.QtWidgets import QWidget, QPushButton, QHBoxLayout, QVBoxLayout\n'), ((581, 601), 'qtpy.QtWebEngineWidgets.QWebEngineView', 'QWebEngineView', (['self'], {}), '(self)\n', (595, 601), False, 'from qtpy.QtWebEngineWi... |
import launch
import launch_ros
def generate_launch_description():
server_node_container = launch_ros.actions.ComposableNodeContainer(
node_name='server_node_container',
node_namespace='',
package='rclcpp_components',
node_executable='component_container',
... | [
"launch.LaunchDescription",
"launch_ros.descriptions.ComposableNode"
] | [((649, 698), 'launch.LaunchDescription', 'launch.LaunchDescription', (['[server_node_container]'], {}), '([server_node_container])\n', (673, 698), False, 'import launch\n'), ((370, 498), 'launch_ros.descriptions.ComposableNode', 'launch_ros.descriptions.ComposableNode', ([], {'package': '"""hello_world_cpp"""', 'node_... |