code
stringlengths
22
1.05M
apis
listlengths
1
3.31k
extract_api
stringlengths
75
3.25M
import urllib as U __all__ = ('sequence', ) def _seq_from_xml(xml): start = xml.find(">", xml.find("<DNA")) + 1 end = xml.rfind("</DNA>") return xml[start:end].replace(' ', '').replace('\n', '').strip() def sequence(db, chrom, start, end): """ return the sequence for a region using the UCSC DAS ...
[ "doctest.testmod", "urllib.urlopen" ]
[((764, 781), 'doctest.testmod', 'doctest.testmod', ([], {}), '()\n', (779, 781), False, 'import doctest\n'), ((639, 675), 'urllib.urlopen', 'U.urlopen', (['(url % (chrom, start, end))'], {}), '(url % (chrom, start, end))\n', (648, 675), True, 'import urllib as U\n')]
# emacs: -*- mode: python; py-indent-offset: 4; indent-tabs-mode: nil -*- # vi: set ft=python sts=4 ts=4 sw=4 et: import os from tempfile import mkdtemp from shutil import rmtree from nipype.testing import (assert_equal, assert_true, assert_false, assert_raises, skipif, parametric) import ...
[ "nipype.testing.assert_true", "nipype.interfaces.matlab.MatlabCommand", "nipype.interfaces.base.CommandLine", "nipype.testing.assert_equal", "nipype.testing.assert_false", "os.path.join", "os.getcwd", "os.chdir", "nipype.testing.assert_raises", "tempfile.mkdtemp", "shutil.rmtree", "nipype.inte...
[((896, 913), 'nipype.testing.skipif', 'skipif', (['no_matlab'], {}), '(no_matlab)\n', (902, 913), False, 'from nipype.testing import assert_equal, assert_true, assert_false, assert_raises, skipif, parametric\n'), ((1772, 1789), 'nipype.testing.skipif', 'skipif', (['no_matlab'], {}), '(no_matlab)\n', (1778, 1789), Fals...
"""A filter block. """ import control import numpy as np import scipy from .base import Block class Filter(Block): """A Filter block class This is simply a single-input-single-output LTI system defined by a single TransferFunction object. Parameters ---------- tf : control.TransferFunction ...
[ "numpy.dot", "numpy.zeros_like", "scipy.signal.cont2discrete" ]
[((2083, 2112), 'numpy.dot', 'np.dot', (['num_d', 'input_register'], {}), '(num_d, input_register)\n', (2089, 2112), True, 'import numpy as np\n'), ((2130, 2168), 'numpy.dot', 'np.dot', (['den_d[1:]', 'output_register[1:]'], {}), '(den_d[1:], output_register[1:])\n', (2136, 2168), True, 'import numpy as np\n'), ((5163,...
import pandas as pd import numpy as np import xml.etree.ElementTree as ElementTree from traffic_analysis.d00_utils.bbox_helpers import bboxcv2_to_bboxcvlib from traffic_analysis.d05_evaluation.parse_annotation import parse_annotation from traffic_analysis.d05_evaluation.compute_mean_average_precision import get_avg_p...
[ "traffic_analysis.d00_utils.bbox_helpers.bboxcv2_to_bboxcvlib", "numpy.mean", "traffic_analysis.d05_evaluation.parse_annotation.parse_annotation", "xml.etree.ElementTree.parse", "pandas.merge", "pandas.DataFrame.from_dict", "traffic_analysis.d05_evaluation.compute_mean_average_precision.get_avg_precisio...
[((781, 797), 'pandas.DataFrame', 'pd.DataFrame', (['{}'], {}), '({})\n', (793, 797), True, 'import pandas as pd\n'), ((831, 847), 'pandas.DataFrame', 'pd.DataFrame', (['{}'], {}), '({})\n', (843, 847), True, 'import pandas as pd\n'), ((2556, 2594), 'pandas.concat', 'pd.concat', (['frame_level_map_dfs'], {'axis': '(0)'...
from load_to_s3 import export_to_s3_as_dataframe, export_to_s3_as_csv from transform_form_specifications import get_custom_form_schema_df from transform_form_results import get_form_results_df from utils.clients import Clients from utils.constants import PuenteTables def run_transform_jobs(event, context): """ ...
[ "load_to_s3.export_to_s3_as_csv", "transform_form_specifications.get_custom_form_schema_df", "load_to_s3.export_to_s3_as_dataframe", "transform_form_results.get_form_results_df" ]
[((615, 652), 'transform_form_results.get_form_results_df', 'get_form_results_df', ([], {'raw_results': '(True)'}), '(raw_results=True)\n', (634, 652), False, 'from transform_form_results import get_form_results_df\n'), ((1093, 1120), 'transform_form_specifications.get_custom_form_schema_df', 'get_custom_form_schema_df...
# Generated by Django 2.1.2 on 2019-06-21 15:16 from django.db import migrations, models class Migration(migrations.Migration): initial = True dependencies = [ ] operations = [ migrations.CreateModel( name='ContactQuery', fields=[ ('id', models.AutoF...
[ "django.db.models.EmailField", "django.db.models.TextField", "django.db.models.AutoField", "django.db.models.CharField" ]
[((308, 401), 'django.db.models.AutoField', 'models.AutoField', ([], {'auto_created': '(True)', 'primary_key': '(True)', 'serialize': '(False)', 'verbose_name': '"""ID"""'}), "(auto_created=True, primary_key=True, serialize=False,\n verbose_name='ID')\n", (324, 401), False, 'from django.db import migrations, models\...
""" Script to post grades back to EdX """ import requests import json import os import time from oauthlib.oauth1.rfc5849 import signature, parameters from lxml import etree from hashlib import sha1 import argparse import base64 class GradePostException(Exception): def __init__(self, response=None): self.r...
[ "oauthlib.oauth1.rfc5849.signature.sign_hmac_sha1", "json.loads", "requests.post", "oauthlib.oauth1.rfc5849.signature.collect_parameters", "argparse.ArgumentParser", "oauthlib.oauth1.rfc5849.parameters.prepare_headers", "oauthlib.oauth1.rfc5849.signature.normalize_base_string_uri", "hashlib.sha1", "...
[((1786, 1792), 'hashlib.sha1', 'sha1', ([], {}), '()\n', (1790, 1792), False, 'from hashlib import sha1\n'), ((2374, 2434), 'oauthlib.oauth1.rfc5849.signature.sign_hmac_sha1', 'signature.sign_hmac_sha1', (['base_string', 'consumer_secret', 'None'], {}), '(base_string, consumer_secret, None)\n', (2398, 2434), False, 'f...
import copy import torch import torch.nn as nn from pytorch_metric_learning.losses import NTXentLoss from transformers import BertForMaskedLM, BertForPreTraining, BertTokenizer def mask_tokens(inputs, tokenizer, not_mask_pos=None): """ Prepare masked tokens of inputs and labels for masked language modeling (80% MA...
[ "torch.bernoulli", "transformers.BertForMaskedLM.from_pretrained", "torch.full", "pytorch_metric_learning.losses.NTXentLoss", "transformers.BertTokenizer.from_pretrained", "torch.max", "torch.tensor", "torch.sum", "copy.deepcopy", "torch.nn.BCEWithLogitsLoss", "torch.cat" ]
[((911, 932), 'copy.deepcopy', 'copy.deepcopy', (['inputs'], {}), '(inputs)\n', (924, 932), False, 'import copy\n'), ((1098, 1128), 'torch.full', 'torch.full', (['labels.shape', '(0.15)'], {}), '(labels.shape, 0.15)\n', (1108, 1128), False, 'import torch\n'), ((1292, 1343), 'torch.tensor', 'torch.tensor', (['special_to...
import pandas as pd import os if __name__ == "__main__": data_dir = os.environ.get("DATA_DIRECTORY") data_file = os.environ.get("DATA_FILE") print(data_dir) data_df = pd.read_pickle(data_dir + data_file) test_df = pd.DataFrame() # Creating two non-overlapping datasets for training and validati...
[ "pandas.read_pickle", "os.environ.get", "pandas.DataFrame", "pandas.concat" ]
[((73, 105), 'os.environ.get', 'os.environ.get', (['"""DATA_DIRECTORY"""'], {}), "('DATA_DIRECTORY')\n", (87, 105), False, 'import os\n'), ((122, 149), 'os.environ.get', 'os.environ.get', (['"""DATA_FILE"""'], {}), "('DATA_FILE')\n", (136, 149), False, 'import os\n'), ((184, 220), 'pandas.read_pickle', 'pd.read_pickle'...
## ACL Import Module # ACL CSV Import # Version 5 # 2015-10-30 # we only need the datetime class & the static function strptime from datetime module from datetime import datetime import re import sys import os import logging # best postgresql module so far, install it "yum install python-psycopg2" import psycopg2 im...
[ "logging.getLogger", "os.path.exists", "os.listdir", "os.makedirs", "os.rename", "tempfile.mkstemp", "csv.writer", "os.path.join", "os.path.isfile", "os.path.isdir", "os.path.basename", "os.unlink", "csv.reader", "re.search" ]
[((390, 414), 'logging.getLogger', 'logging.getLogger', (['"""acl"""'], {}), "('acl')\n", (407, 414), False, 'import logging\n'), ((6289, 6313), 'os.path.isfile', 'os.path.isfile', (['filename'], {}), '(filename)\n', (6303, 6313), False, 'import os\n'), ((8608, 8694), 're.search', 're.search', (['"""^(.*?)homeofficerol...
import os, queue from tablet import Tablet f = open(os.path.join(os.path.dirname(__file__), '../input/18/part1.txt'), 'r') def main(): instructionStrings = [] line = f.readline() while line: instructionStrings.append(line.rstrip()) line = f.readline() q0 = queue.Queue() q1 = queue...
[ "os.path.dirname", "tablet.Tablet", "queue.Queue" ]
[((292, 305), 'queue.Queue', 'queue.Queue', ([], {}), '()\n', (303, 305), False, 'import os, queue\n'), ((315, 328), 'queue.Queue', 'queue.Queue', ([], {}), '()\n', (326, 328), False, 'import os, queue\n'), ((339, 376), 'tablet.Tablet', 'Tablet', (['instructionStrings', '(0)', 'q0', 'q1'], {}), '(instructionStrings, 0,...
import datetime import os from zipfile import ZipFile import requests from icecream import ic from wget import download ic.enable() date_from = (datetime.date.today() - datetime.timedelta(days=5)).isoformat() date_to = (datetime.date.today()).isoformat() ic(f'Date Range: {date_from} to {date_to}') def getbhav_dat...
[ "icecream.ic.enable", "os.path.exists", "icecream.ic", "wget.download", "requests.get", "datetime.timedelta", "os.chdir", "os.mkdir", "datetime.date.today", "datetime.date.fromisoformat" ]
[((122, 133), 'icecream.ic.enable', 'ic.enable', ([], {}), '()\n', (131, 133), False, 'from icecream import ic\n'), ((259, 302), 'icecream.ic', 'ic', (['f"""Date Range: {date_from} to {date_to}"""'], {}), "(f'Date Range: {date_from} to {date_to}')\n", (261, 302), False, 'from icecream import ic\n'), ((365, 403), 'datet...
# coding=utf-8 # !/usr/bin/python3 # Name: ruuvitag calc - calculations # Copyright: (c) 2019 TK # Licence: MIT # ------------------------------------------------------------------------------- import logging logger = logging.getLogger('ruuvitag') import math # ----------------------------------------...
[ "logging.getLogger", "math.exp" ]
[((234, 263), 'logging.getLogger', 'logging.getLogger', (['"""ruuvitag"""'], {}), "('ruuvitag')\n", (251, 263), False, 'import logging\n'), ((1403, 1446), 'math.exp', 'math.exp', (['(17.67 * l_temp / (243.5 + l_temp))'], {}), '(17.67 * l_temp / (243.5 + l_temp))\n', (1411, 1446), False, 'import math\n')]
# -*- coding: utf-8 -*- ''' Copyright (c) 2021, Trustworthy AI, Inc. All rights reserved. Redistribution and use in source and binary forms, with or without modification, are permitted provided that the following conditions are met: 1. Redistributions of source code must retain the above copyright notice, this list ...
[ "geometry_msgs.msg.Vector3", "geometry_msgs.msg.Twist", "math.radians", "geometry_msgs.msg.Transform", "numpy.array", "geometry_msgs.msg.Point", "geometry_msgs.msg.Quaternion", "tf.transformations.quaternion_from_euler", "tf.transformations.euler_matrix", "geometry_msgs.msg.Accel", "geometry_msg...
[((2477, 2545), 'numpy.array', 'numpy.array', (['[carla_location.x, -carla_location.y, carla_location.z]'], {}), '([carla_location.x, -carla_location.y, carla_location.z])\n', (2488, 2545), False, 'import numpy\n'), ((2962, 2971), 'geometry_msgs.msg.Vector3', 'Vector3', ([], {}), '()\n', (2969, 2971), False, 'from geom...
import numpy as np class KNearestNeighbors: def __init__(self, distances, labels, k=10): self.distances = distances self.labels = labels self.k = k def _kNN(self, instance, train, k): nearest = np.argpartition(self.distances[instance][train], k) nearest_labels = self....
[ "numpy.argmax", "numpy.zeros", "numpy.unique", "numpy.argpartition" ]
[((238, 289), 'numpy.argpartition', 'np.argpartition', (['self.distances[instance][train]', 'k'], {}), '(self.distances[instance][train], k)\n', (253, 289), True, 'import numpy as np\n'), ((372, 417), 'numpy.unique', 'np.unique', (['nearest_labels'], {'return_counts': '(True)'}), '(nearest_labels, return_counts=True)\n...
# coding: utf-8 # In[1]: from path import Path from matplotlib import pyplot as plt import numpy as np import skimage.io as io import os from PIL import Image import cv2 import random import shutil def crop_by_sequence(image_path,img_class_path,crop_size_w,crop_size_h,prefix,save_dir ,same_scale = False): ...
[ "cv2.imwrite", "shutil.move", "path.Path", "skimage.io.imread", "numpy.random.randint", "numpy.zeros", "os.mkdir" ]
[((703, 728), 'skimage.io.imread', 'io.imread', (['img_class_path'], {}), '(img_class_path)\n', (712, 728), True, 'import skimage.io as io\n'), ((3340, 3365), 'skimage.io.imread', 'io.imread', (['img_class_path'], {}), '(img_class_path)\n', (3349, 3365), True, 'import skimage.io as io\n'), ((3576, 3624), 'numpy.random....
import cProfile import functools import pstats import resource import signal import time from typing import Any, Callable # This list is passed to @profile() which aggregates the cumulative runtime # into these functions, which represent the primary stages of task solutioning. PROFILE_BREAKOUT_STD: list[str] = [ "...
[ "signal.signal", "resource.getrusage", "functools.wraps", "pstats.Stats", "signal.alarm", "cProfile.Profile", "time.time" ]
[((809, 830), 'functools.wraps', 'functools.wraps', (['func'], {}), '(func)\n', (824, 830), False, 'import functools\n'), ((2152, 2192), 'resource.getrusage', 'resource.getrusage', (['resource.RUSAGE_SELF'], {}), '(resource.RUSAGE_SELF)\n', (2170, 2192), False, 'import resource\n'), ((2643, 2664), 'functools.wraps', 'f...
# -*- coding: utf-8 -*- #Requires Win32 Python Extensions import os import servicemanager import shutil import subprocess import sys import win32api import win32event import win32service import win32serviceutil import tempfile class VulnService(win32serviceutil.ServiceFramework): _svc_na...
[ "subprocess.check_output", "servicemanager.LogInfoMsg", "win32serviceutil.HandleCommandLine", "servicemanager.LogMsg", "win32event.WaitForSingleObject", "win32serviceutil.ServiceFramework.__init__", "shutil.copyfile", "os.path.dirname", "tempfile.gettempdir", "win32api.SetConsoleCtrlHandler", "o...
[((4090, 4139), 'win32api.SetConsoleCtrlHandler', 'win32api.SetConsoleCtrlHandler', (['ctrlHandler', '(True)'], {}), '(ctrlHandler, True)\n', (4120, 4139), False, 'import win32api\n'), ((4149, 4196), 'win32serviceutil.HandleCommandLine', 'win32serviceutil.HandleCommandLine', (['VulnService'], {}), '(VulnService)\n', (4...
import copy import logging import torch import numpy as np from torch.utils.data import DataLoader from torchvision import datasets, transforms log = logging.getLogger(__name__) def balanced_batches(dataset, batch_size): unlabled_idx = dataset.unlabeled_idx labeled_idx = list(filter(lambda _: _ not in unlab...
[ "logging.getLogger", "copy.deepcopy", "numpy.random.choice", "torch.LongTensor", "torch.stack", "numpy.array_split", "numpy.array", "torch.utils.data.DataLoader", "torchvision.transforms.ToTensor" ]
[((152, 179), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (169, 179), False, 'import logging\n'), ((411, 432), 'numpy.array', 'np.array', (['labeled_idx'], {}), '(labeled_idx)\n', (419, 432), True, 'import numpy as np\n'), ((590, 629), 'numpy.array_split', 'np.array_split', (['unlabled...
# emacs: -*- mode: python; py-indent-offset: 4; indent-tabs-mode: nil -*- # vi: set ft=python sts=4 ts=4 sw=4 et: ### ### ### ### ### ### ### ### ### ### ### ### ### ### ### ### ### ### ### ## # # See COPYING file distributed along with the PyMVPA package for the # copyright and license terms. # ### ### ### ### ###...
[ "numpy.abs", "mvpa2.clfs.gnb.GNB", "numpy.exp", "numpy.sum", "mvpa2.generators.splitters.Splitter" ]
[((712, 717), 'mvpa2.clfs.gnb.GNB', 'GNB', ([], {}), '()\n', (715, 717), False, 'from mvpa2.clfs.gnb import GNB\n'), ((735, 761), 'mvpa2.clfs.gnb.GNB', 'GNB', ([], {'common_variance': '(False)'}), '(common_variance=False)\n', (738, 761), False, 'from mvpa2.clfs.gnb import GNB\n'), ((778, 797), 'mvpa2.clfs.gnb.GNB', 'GN...
# MIT License # # Copyright (c) 2020 <NAME> # # Permission is hereby granted, free of charge, to any person obtaining a copy # of this software and associated documentation files (the "Software"), to deal # in the Software without restriction, including without limitation the rights # to use, copy, modify, merge, publi...
[ "io_mesh_amf.export_amf.ExportAMF", "mathutils.Matrix.Identity", "tempfile.gettempdir", "unittest.mock.Mock" ]
[((2393, 2399), 'unittest.mock.Mock', 'Mock', ([], {}), '()\n', (2397, 2399), False, 'from unittest.mock import Mock\n'), ((2893, 2899), 'unittest.mock.Mock', 'Mock', ([], {}), '()\n', (2897, 2899), False, 'from unittest.mock import Mock\n'), ((3042, 3060), 'mathutils.Matrix.Identity', 'Matrix.Identity', (['(4)'], {}),...
""" amplitude.py measure the maximum peak-to-peak amplitude """ import obspy import types import numpy as np import pandas as pd import madpy.noise as n from typing import Tuple import madpy.checks as ch import madpy.config as config import matplotlib.pyplot as plt import madpy.plotting.amp as plot def measure_ampl...
[ "numpy.abs", "madpy.noise.arrival_time_utc", "madpy.checks.check_amplitude", "madpy.noise.rms_noise", "numpy.divide", "numpy.where", "madpy.checks.check_window", "numpy.diff", "numpy.subtract", "madpy.plotting.amp.amplitude_plot", "numpy.array", "numpy.isnan", "pandas.DataFrame", "numpy.na...
[((1808, 1829), 'madpy.checks.check_waveform', 'ch.check_waveform', (['tr'], {}), '(tr)\n', (1825, 1829), True, 'import madpy.checks as ch\n'), ((2460, 2474), 'numpy.diff', 'np.diff', (['peaks'], {}), '(peaks)\n', (2467, 2474), True, 'import numpy as np\n'), ((2538, 2561), 'madpy.checks.check_amplitude', 'ch.check_ampl...
# External Import from django.contrib.auth import get_user_model from django import forms from django.contrib.auth.forms import UserCreationForm, AuthenticationForm from django.core.exceptions import ValidationError from django.contrib import messages from django.urls import reverse from django.contrib.sites.shortcuts ...
[ "django.contrib.auth.get_user_model", "django.urls.reverse", "django.contrib.messages.error", "django.forms.PasswordInput", "django.core.exceptions.ValidationError", "django.forms.ValidationError", "django.forms.EmailInput", "django.forms.TextInput", "django.forms.EmailField" ]
[((404, 420), 'django.contrib.auth.get_user_model', 'get_user_model', ([], {}), '()\n', (418, 420), False, 'from django.contrib.auth import get_user_model\n'), ((2419, 2450), 'django.forms.EmailField', 'forms.EmailField', ([], {'required': '(True)'}), '(required=True)\n', (2435, 2450), False, 'from django import forms\...
import torch import torch.nn as nn import torch.nn.functional as F from src.utils import get_seed class DuelingQNetwork(nn.Module): def __init__(self, state_size, action_size): super().__init__() self.seed = torch.manual_seed(get_seed()) self.V_fc1 = nn.Linear(state_size, 64) self...
[ "src.utils.get_seed", "torch.nn.Linear" ]
[((282, 307), 'torch.nn.Linear', 'nn.Linear', (['state_size', '(64)'], {}), '(state_size, 64)\n', (291, 307), True, 'import torch.nn as nn\n'), ((329, 346), 'torch.nn.Linear', 'nn.Linear', (['(64)', '(64)'], {}), '(64, 64)\n', (338, 346), True, 'import torch.nn as nn\n'), ((368, 384), 'torch.nn.Linear', 'nn.Linear', ([...
"""Class for generation using the GAN.""" from pathlib import Path import tensorflow as tf from tensorflow.keras import Model from tqdm import tqdm from ..utils import Config class GANEvaluator: """Class to generate images using a GAN. Attributes: generator: The generator model to be evaluated ...
[ "tensorflow.random.normal", "tensorflow.image.convert_image_dtype", "tqdm.tqdm", "tensorflow.map_fn", "tensorflow.convert_to_tensor" ]
[((808, 867), 'tensorflow.random.normal', 'tf.random.normal', (['[digits.shape[0], self.config.noise_dims]'], {}), '([digits.shape[0], self.config.noise_dims])\n', (824, 867), True, 'import tensorflow as tf\n'), ((884, 912), 'tensorflow.convert_to_tensor', 'tf.convert_to_tensor', (['digits'], {}), '(digits)\n', (904, 9...
# Copyright (C) 2019 Intel Corporation # # SPDX-License-Identifier: MIT import argparse from . import project as project_module def build_parser(parser=argparse.ArgumentParser()): project_module.build_create_parser(parser) \ .set_defaults(command=project_module.create_command) return parser def m...
[ "argparse.ArgumentParser" ]
[((157, 182), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {}), '()\n', (180, 182), False, 'import argparse\n')]
#REQUESTS -> FAZER REQUISIÇÕES HTTP (URLLIB3) #HTTP = Hypertext Transfer Protocol #BEAUTIFULSOUP -> HTML ou XML -> Organiza-lo de maneira a ler no python #PANDAS -> pegar esses objetos e tranformar em EXCEL #USAR REQUESTS E ENTENDER COMO FUNCIONA. import requests as req url = 'https://g1.globo.com' r = req.get(url)...
[ "bs4.BeautifulSoup", "requests.get" ]
[((308, 320), 'requests.get', 'req.get', (['url'], {}), '(url)\n', (315, 320), True, 'import requests as req\n'), ((386, 421), 'bs4.BeautifulSoup', 'BeautifulSoup', (['conteudo_cru', '"""lxml"""'], {}), "(conteudo_cru, 'lxml')\n", (399, 421), False, 'from bs4 import BeautifulSoup\n'), ((788, 800), 'requests.get', 'req....
#!/usr/bin/env python # -*- coding:Utf-8 -*- import sys import os from collections import OrderedDict from time import time if sys.version_info >= (3, 0): import configparser else: import ConfigParser as configparser from subprocess import Popen from tekamenu.const import * def help(): print("usage : ") ...
[ "collections.OrderedDict", "subprocess.Popen", "os.access", "os.path.join", "os.path.split", "ConfigParser.ConfigParser", "os.path.isfile", "sys.exit", "time.time" ]
[((441, 451), 'sys.exit', 'sys.exit', ([], {}), '()\n', (449, 451), False, 'import sys\n'), ((1069, 1082), 'collections.OrderedDict', 'OrderedDict', ([], {}), '()\n', (1080, 1082), False, 'from collections import OrderedDict\n'), ((2885, 2907), 'os.path.split', 'os.path.split', (['program'], {}), '(program)\n', (2898, ...
import ConfigParser import logging import os import sys import yaml log = logging.getLogger() def load_yaml(filename): with open(filename, 'r') as stream: return yaml.load(stream) def set_env_aws_creds(account='default'): """ Parse ~/.aws/credentials for credentials Allow OS environment var...
[ "logging.getLogger", "os.path.exists", "os.getenv", "yaml.load", "ConfigParser.ConfigParser", "sys.exit", "os.path.expanduser" ]
[((75, 94), 'logging.getLogger', 'logging.getLogger', ([], {}), '()\n', (92, 94), False, 'import logging\n'), ((569, 595), 'os.getenv', 'os.getenv', (['aws_key_env_var'], {}), '(aws_key_env_var)\n', (578, 595), False, 'import os\n'), ((617, 646), 'os.getenv', 'os.getenv', (['aws_secret_env_var'], {}), '(aws_secret_env_...
from setuptools import setup import setuptools readme = '' with open('README.md') as f: readme = f.read() setup( name="discsocket", author="<NAME>", url="https://github.com/murillotadeo/discsocket", project_urls={ "Issue tracker": "https://github.com/murillotadeo/discsocket/issues", ...
[ "setuptools.find_packages" ]
[((448, 479), 'setuptools.find_packages', 'setuptools.find_packages', (['"""src"""'], {}), "('src')\n", (472, 479), False, 'import setuptools\n')]
# ==BEGIN LICENSE== # # MIT License # # Copyright (c) 2018 SRI Lab, ETH Zurich # # Permission is hereby granted, free of charge, to any person obtaining a copy # of this software and associated documentation files (the "Software"), to deal # in the Software without restriction, including without limitation the right...
[ "dpfinder.utils.utils.arr_to_str", "os.path.dirname", "importlib.import_module", "re.compile" ]
[((1516, 1584), 're.compile', 're.compile', (['"""^// ==BEGIN LICENSE==[\\\\s\\\\S]+// ==END LICENSE==[\\\\s]+"""'], {}), "('^// ==BEGIN LICENSE==[\\\\s\\\\S]+// ==END LICENSE==[\\\\s]+')\n", (1526, 1584), False, 'import re\n'), ((3484, 3537), 'importlib.import_module', 'importlib.import_module', (["('.algs.' + name)",...
#################################################### #################################################### # functions and classes used in conjunction with # pipeline_metaomics.py #################################################### #################################################### # import libraries import sys impo...
[ "numpy.mean", "CGATPipelines.Pipeline.run", "sqlite3.connect", "CGATPipelines.Pipeline.snip", "itertools.product", "CGATPipelines.Pipeline.getTempFilename", "os.unlink", "os.path.basename", "pandas.DataFrame", "CGAT.IOTools.openFile", "rpy2.robjects.r" ]
[((2859, 2881), 'sqlite3.connect', 'sqlite3.connect', (['rnadb'], {}), '(rnadb)\n', (2874, 2881), False, 'import sqlite3\n'), ((2926, 2948), 'sqlite3.connect', 'sqlite3.connect', (['dnadb'], {}), '(dnadb)\n', (2941, 2948), False, 'import sqlite3\n'), ((4401, 4420), 'sqlite3.connect', 'sqlite3.connect', (['db'], {}), '(...
# ------------------------------------------------------------------------------ # Test Formatting weeks # ------------------------------------------------------------------------------ import sys import datetime as dt from django.test import TestCase, override_settings import ls.joyous.utils.weeks from ls.joyous.utils...
[ "ls.joyous.utils.weeks._iso_year_start", "ls.joyous.utils.weeks._iso_to_gregorian", "ls.joyous.utils.weeks._ssweek_to_gregorian", "ls.joyous.utils.weeks._ssweek_info", "ls.joyous.utils.weeks._iso_num_weeks", "ls.joyous.utils.weeks._ssweek_year_start", "django.utils.translation.override", "django.test....
[((5694, 5711), 'django.utils.translation.override', 'override', (['"""en-gb"""'], {}), "('en-gb')\n", (5702, 5711), False, 'from django.utils.translation import override\n'), ((6620, 6637), 'django.utils.translation.override', 'override', (['"""en-au"""'], {}), "('en-au')\n", (6628, 6637), False, 'from django.utils.tr...
#!/usr/bin/env python3 # # MIT License # # Copyright (c) 2020-2021 EntySec # # Permission is hereby granted, free of charge, to any person obtaining a copy # of this software and associated documentation files (the "Software"), to deal # in the Software without restriction, including without limitation the rights # to...
[ "core.cli.badges.badges", "core.base.exceptions.exceptions", "core.base.jobs.jobs" ]
[((1356, 1364), 'core.cli.badges.badges', 'badges', ([], {}), '()\n', (1362, 1364), False, 'from core.cli.badges import badges\n'), ((1385, 1391), 'core.base.jobs.jobs', 'jobs', ([], {}), '()\n', (1389, 1391), False, 'from core.base.jobs import jobs\n'), ((1418, 1430), 'core.base.exceptions.exceptions', 'exceptions', (...
from collections import namedtuple Genotype = namedtuple('Genotype', 'backbone rpn') OP_NAMES = [ 'sep_conv_3x3', 'sep_conv_3x3_dil3', 'sep_conv_5x5_dil6', 'skip_connect', 'def_conv_3x3', ] AGG_NAMES = [ 'psum', 'cat' ] HEAD_OP_NAMES = [ 'conv1x1', 'conv3x3', 'sep_conv_3x3', ...
[ "collections.namedtuple" ]
[((47, 85), 'collections.namedtuple', 'namedtuple', (['"""Genotype"""', '"""backbone rpn"""'], {}), "('Genotype', 'backbone rpn')\n", (57, 85), False, 'from collections import namedtuple\n')]
# Copyright (c) 2014 Red Hat, Inc. # # Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writ...
[ "rally.task.atomic.action_timer", "rally.task.atomic.ActionTimer" ]
[((781, 822), 'rally.task.atomic.action_timer', 'atomic.action_timer', (['"""zaqar.create_queue"""'], {}), "('zaqar.create_queue')\n", (800, 822), False, 'from rally.task import atomic\n'), ((1177, 1218), 'rally.task.atomic.action_timer', 'atomic.action_timer', (['"""zaqar.delete_queue"""'], {}), "('zaqar.delete_queue'...
# -*- coding: utf-8 -*- ''' Script Name: ping_Utility.py Path: \IPS_DecisionFabric\Exception Handling\ Description: This script is considered as a module for the Application level Exception handling in DF framework. Author: <NAME> Version: 1.0 Revision History: ---------------------------------------------...
[ "logging.basicConfig", "pandas.Series", "sys.path.insert", "logging.debug", "mongo_operations.Mongo", "time.sleep", "datetime.datetime.now", "os.getpid", "sys.exit", "pandas.DataFrame", "logging.info", "ping_services.check_for_status" ]
[((1119, 1178), 'sys.path.insert', 'sys.path.insert', (['(0)', '"""/IPS_DecisionFabric/Control Framework"""'], {}), "(0, '/IPS_DecisionFabric/Control Framework')\n", (1134, 1178), False, 'import sys, os\n'), ((1331, 1384), 'ping_services.check_for_status', 'ping_services.check_for_status', (['config_file', 'csv_file'],...
from setuptools import setup setup( name='NotebookScripter', version='6.0.0', packages=('NotebookScripter',), url='https://github.com/breathe/NotebookScripter', license='MIT', author='<NAME>', author_email='<EMAIL>', install_requires=( "ipython", "nbformat" ), te...
[ "setuptools.setup" ]
[((30, 839), 'setuptools.setup', 'setup', ([], {'name': '"""NotebookScripter"""', 'version': '"""6.0.0"""', 'packages': "('NotebookScripter',)", 'url': '"""https://github.com/breathe/NotebookScripter"""', 'license': '"""MIT"""', 'author': '"""<NAME>"""', 'author_email': '"""<EMAIL>"""', 'install_requires': "('ipython',...
# Third-party import astropy.units as u import numpy as np import pymc3 as pm from pymc3.distributions import generate_samples import aesara_theano_fallback.tensor as tt import exoplanet.units as xu __all__ = ['UniformLog', 'FixedCompanionMass'] class UniformLog(pm.Continuous): def __init__(self, a, b, **kwargs...
[ "numpy.sqrt", "numpy.log", "aesara_theano_fallback.tensor.as_tensor_variable", "numpy.zeros", "numpy.random.uniform", "pymc3.distributions.generate_samples", "astropy.units.quantity_input" ]
[((1908, 1973), 'astropy.units.quantity_input', 'u.quantity_input', ([], {'sigma_K0': '(u.km / u.s)', 'P0': 'u.day', 'max_K': '(u.km / u.s)'}), '(sigma_K0=u.km / u.s, P0=u.day, max_K=u.km / u.s)\n', (1924, 1973), True, 'import astropy.units as u\n'), ((945, 973), 'numpy.random.uniform', 'np.random.uniform', ([], {'size...
import json import os import unittest import price_checker from product import Product def load_json(filename): with open(filename) as json_file: imported_file = json.load(json_file) return imported_file cwd = os.path.dirname(os.path.realpath(__file__)) products_data = load_json(r"" + cwd + "/t...
[ "price_checker.compare_prices", "price_checker.open_policy", "product.Product", "unittest.main", "os.path.realpath", "json.load" ]
[((251, 277), 'os.path.realpath', 'os.path.realpath', (['__file__'], {}), '(__file__)\n', (267, 277), False, 'import os\n'), ((1689, 1704), 'unittest.main', 'unittest.main', ([], {}), '()\n', (1702, 1704), False, 'import unittest\n'), ((177, 197), 'json.load', 'json.load', (['json_file'], {}), '(json_file)\n', (186, 19...
from collections import ( OrderedDict, ) from unittest.mock import Mock import pandas as pd from datetime import ( datetime, ) from fireant import * from fireant.slicer.references import ReferenceType from fireant.slicer.totals import get_totals_marker_for_dtype from fireant.utils import ( format_dimensio...
[ "pandas.DataFrame.from_records", "collections.OrderedDict", "unittest.mock.Mock", "fireant.utils.format_dimension_key", "fireant.utils.format_metric_key", "fireant.slicer.totals.get_totals_marker_for_dtype", "pypika.functions.Count", "pypika.functions.Sum", "pandas.Index", "pypika.Table", "firea...
[((682, 720), 'pypika.Table', 'Table', (['"""politician"""'], {'schema': '"""politics"""'}), "('politician', schema='politics')\n", (687, 720), False, 'from pypika import JoinType, Table, functions as fn\n'), ((736, 769), 'pypika.Table', 'Table', (['"""voter"""'], {'schema': '"""politics"""'}), "('voter', schema='polit...
import requests from datetime import timedelta from app.spotify_api import get_auth_header def get_albums(data): result = [] for item in data: skip = False header = get_auth_header() songs_r = requests.get(f'https://api.spotify.com/v1/albums/{item["id"]}/tracks', headers=header) ...
[ "app.spotify_api.get_auth_header", "datetime.timedelta", "requests.get" ]
[((192, 209), 'app.spotify_api.get_auth_header', 'get_auth_header', ([], {}), '()\n', (207, 209), False, 'from app.spotify_api import get_auth_header\n'), ((228, 318), 'requests.get', 'requests.get', (['f"""https://api.spotify.com/v1/albums/{item[\'id\']}/tracks"""'], {'headers': 'header'}), '(f"https://api.spotify.com...
import cv2 import numpy as np from PyQt5.QtGui import QIntValidator from PyQt5.QtWidgets import QDialog from PyQt5.uic import loadUi from utils import processing_utils as utils IMAGE_DESCRIPT_DIALOG_UI = 'coreUI/image_description_dialog.ui' class ImageDescriptionDialog(QDialog): """Image Description Dialog Windo...
[ "PyQt5.QtGui.QIntValidator", "PyQt5.uic.loadUi", "utils.processing_utils.display_img", "cv2.filter2D", "numpy.sum" ]
[((420, 458), 'PyQt5.uic.loadUi', 'loadUi', (['IMAGE_DESCRIPT_DIALOG_UI', 'self'], {}), '(IMAGE_DESCRIPT_DIALOG_UI, self)\n', (426, 458), False, 'from PyQt5.uic import loadUi\n'), ((1016, 1061), 'utils.processing_utils.display_img', 'utils.display_img', (['image', 'self.imageViewLabel'], {}), '(image, self.imageViewLab...
import tensorflow as tf from tensorflow import keras import numpy as np imdb = keras.datasets.imdb (train_data, train_labels), (test_data, test_labels) = imdb.load_data(num_words=10000) # A dictionary mapping words to an integer index word_index = imdb.get_word_index() # The first indices are reserved word_index =...
[ "tensorflow.keras.preprocessing.sequence.pad_sequences", "tensorflow.keras.Sequential", "tensorflow.Session", "tensorflow.keras.layers.Embedding", "tensorflow.global_variables_initializer", "tensorflow.keras.layers.Dense", "tensorflow.keras.layers.GlobalAveragePooling1D", "tensorflow.keras.backend.set...
[((1083, 1197), 'tensorflow.keras.preprocessing.sequence.pad_sequences', 'keras.preprocessing.sequence.pad_sequences', (['train_data'], {'value': "word_index['<PAD>']", 'padding': '"""post"""', 'maxlen': '(256)'}), "(train_data, value=word_index[\n '<PAD>'], padding='post', maxlen=256)\n", (1125, 1197), False, 'from...
import vcr from fast_arrow import Client def gen_vcr(): return vcr.VCR( cassette_library_dir='tests/fixtures_vcr', record_mode='none', match_on=['method', 'scheme', 'host', 'port', 'path', 'query'], ) def gen_client(): auth_data = gen_auth_data() client = Client(auth_data) ...
[ "fast_arrow.Client", "vcr.VCR" ]
[((70, 208), 'vcr.VCR', 'vcr.VCR', ([], {'cassette_library_dir': '"""tests/fixtures_vcr"""', 'record_mode': '"""none"""', 'match_on': "['method', 'scheme', 'host', 'port', 'path', 'query']"}), "(cassette_library_dir='tests/fixtures_vcr', record_mode='none',\n match_on=['method', 'scheme', 'host', 'port', 'path', 'qu...
import datetime class Person(object): def __init__(self, name): self.name = name try: lastBlank = name.rindex(' ') self.lastName = name[lastBlank+1:] except: self.lastName = name self.birthday = None def getName(self): return sel...
[ "datetime.date.today", "datetime.date" ]
[((1905, 1930), 'datetime.date', 'datetime.date', (['(1961)', '(8)', '(4)'], {}), '(1961, 8, 4)\n', (1918, 1930), False, 'import datetime\n'), ((1946, 1972), 'datetime.date', 'datetime.date', (['(1958)', '(8)', '(16)'], {}), '(1958, 8, 16)\n', (1959, 1972), False, 'import datetime\n'), ((557, 578), 'datetime.date.today...
# # Copyright 2022 DMetaSoul # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, ...
[ "argparse.ArgumentParser", "pyspark.mllib.evaluation.RankingMetrics", "subprocess.run", "yaml.load", "metaspore.SwingEstimator", "sys.path.append", "pyspark.sql.functions.collect_set" ]
[((760, 791), 'sys.path.append', 'sys.path.append', (['"""../../../../"""'], {}), "('../../../../')\n", (775, 791), False, 'import sys\n'), ((3667, 3723), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {'description': '"""Tuner information"""'}), "(description='Tuner information')\n", (3690, 3723), False, '...
import shared_module from shared_module import module_function as my_function, ModuleClass class NewParent(object): def do_useful_stuff(self): i = shared_module.MODULE_CONTANT my_function() ModuleClass()
[ "shared_module.ModuleClass", "shared_module.module_function" ]
[((198, 211), 'shared_module.module_function', 'my_function', ([], {}), '()\n', (209, 211), True, 'from shared_module import module_function as my_function, ModuleClass\n'), ((220, 233), 'shared_module.ModuleClass', 'ModuleClass', ([], {}), '()\n', (231, 233), False, 'from shared_module import module_function as my_fun...
# -*- coding: utf-8 -*- import os os.environ['DJANGO_SETTINGS_MODULE']='settings' from logistic import logisticdb import webapp2 as webapp from google.appengine.ext.webapp.util import run_wsgi_app from google.appengine.ext.webapp import template import numpy as np import cgi import cgitb cgitb.enable() def lesl...
[ "cgi.FieldStorage", "os.path.dirname", "numpy.dot", "numpy.zeros", "google.appengine.ext.webapp.util.run_wsgi_app", "webapp2.WSGIApplication", "cgitb.enable", "google.appengine.ext.webapp.template.render" ]
[((291, 305), 'cgitb.enable', 'cgitb.enable', ([], {}), '()\n', (303, 305), False, 'import cgitb\n'), ((4709, 4772), 'webapp2.WSGIApplication', 'webapp.WSGIApplication', (["[('/.*', leslieOutputPage)]"], {'debug': '(True)'}), "([('/.*', leslieOutputPage)], debug=True)\n", (4731, 4772), True, 'import webapp2 as webapp\n...
from cement import Controller, ex from ..utils import databaseUtils, controllerUtils class CollectionController(Controller): class Meta: label = 'collection controls' @ex( help='set currently active collection by id or name', arguments=[ ( ['-i', '--id'], ...
[ "cement.ex" ]
[((187, 337), 'cement.ex', 'ex', ([], {'help': '"""set currently active collection by id or name"""', 'arguments': "[(['-i', '--id'], {'help': 'collection id', 'action': 'store', 'dest': 'id'})]"}), "(help='set currently active collection by id or name', arguments=[(['-i',\n '--id'], {'help': 'collection id', 'actio...
import copy import typing import splendor_sim.interfaces.coin.i_coin_reserve as i_coin_reserve import splendor_sim.interfaces.coin.i_coin_type as i_coin_type import splendor_sim.interfaces.coin.i_coin_type_manager as i_coin_type_manager class CoinReserve(i_coin_reserve.ICoinReserve): def __init__( self, ...
[ "copy.copy" ]
[((1260, 1290), 'copy.copy', 'copy.copy', (['self._current_coins'], {}), '(self._current_coins)\n', (1269, 1290), False, 'import copy\n'), ((1383, 1413), 'copy.copy', 'copy.copy', (['self._max_coin_size'], {}), '(self._max_coin_size)\n', (1392, 1413), False, 'import copy\n')]
import pytest from dagger.serializer.as_pickle import AsPickle from dagger.serializer.errors import DeserializationError, SerializationError from dagger.serializer.protocol import Serializer def test__conforms_to_protocol(): assert isinstance(AsPickle(), Serializer) def test_extension(): assert AsPickle()....
[ "dagger.serializer.as_pickle.AsPickle", "pytest.raises" ]
[((426, 436), 'dagger.serializer.as_pickle.AsPickle', 'AsPickle', ([], {}), '()\n', (434, 436), False, 'from dagger.serializer.as_pickle import AsPickle\n'), ((1105, 1115), 'dagger.serializer.as_pickle.AsPickle', 'AsPickle', ([], {}), '()\n', (1113, 1115), False, 'from dagger.serializer.as_pickle import AsPickle\n'), (...
from torch.utils.data import Dataset from PIL import Image from pathlib import Path import pandas import torch # from utils import data_utils class InferenceDataset(Dataset): def __init__(self, root, opts, split, transform=None, preprocess=None): self.root = root attributes_path = Path(root) / "list_attr_celeba...
[ "torch.as_tensor", "PIL.Image.open", "pandas.read_csv", "pathlib.Path" ]
[((335, 400), 'pandas.read_csv', 'pandas.read_csv', (['attributes_path'], {'delim_whitespace': '(True)', 'header': '(1)'}), '(attributes_path, delim_whitespace=True, header=1)\n', (350, 400), False, 'import pandas\n'), ((588, 673), 'pandas.read_csv', 'pandas.read_csv', (['attributes_path'], {'delim_whitespace': '(True)...
#!/usr/bin/python2.4 # # Copyright 2009 Google Inc. All Rights Reserved. """Defines the generic robot classes. This module provides the Robot class and RobotListener interface, as well as some helper functions for web requests and responses. """ __author__ = '<EMAIL> (<NAME>)' import events import model import ops ...
[ "util.CollapseJavaCollections", "simplejson.dumps", "util.ToUpperCamelCase", "ops.CreateContext", "util.Serialize", "simplejson.loads", "model.CreateEvent" ]
[((459, 486), 'simplejson.loads', 'simplejson.loads', (['json_body'], {}), '(json_body)\n', (475, 486), False, 'import simplejson\n'), ((554, 588), 'util.CollapseJavaCollections', 'util.CollapseJavaCollections', (['json'], {}), '(json)\n', (582, 588), False, 'import util\n'), ((601, 624), 'ops.CreateContext', 'ops.Crea...
import pathlib from setuptools import setup, find_packages # The directory containing this file HERE = pathlib.Path(__file__).parent # The text of the README file README = (HERE / "README.md").read_text() setup( description='Data extraction and processing for genre prediction using ML', long_description=REA...
[ "setuptools.find_packages", "pathlib.Path" ]
[((105, 127), 'pathlib.Path', 'pathlib.Path', (['__file__'], {}), '(__file__)\n', (117, 127), False, 'import pathlib\n'), ((872, 904), 'setuptools.find_packages', 'find_packages', ([], {'exclude': "('test',)"}), "(exclude=('test',))\n", (885, 904), False, 'from setuptools import setup, find_packages\n')]
# Copyright (c) 2015-2019 The Switch Authors. All rights reserved. # Licensed under the Apache License, Version 2.0, which is in the LICENSE file. """ Defines generation projects build-outs. INPUT FILE FORMAT Import data describing project builds. The following files are expected in the input directory. g...
[ "switch_model.financials.capital_recovery_factor", "os.path.join", "switch_model.tools.graph.graph", "switch_model.utilities.scaling.get_assign_default_value_rule" ]
[((29938, 30029), 'switch_model.tools.graph.graph', 'graph', (['"""generation_capacity_per_period"""'], {'title': '"""Online Generation Capacity Per Period"""'}), "('generation_capacity_per_period', title=\n 'Online Generation Capacity Per Period')\n", (29943, 30029), False, 'from switch_model.tools.graph import gra...
# https://www.codewars.com/kata/directions-reduction/train/python # My solution import re def dirReduc(arr): card = {"NORTH": "N", "SOUTH": "S", "EAST": "E", "WEST": "W"} arr = "".join(map(lambda elem: card[elem], arr)) while "NS" in arr or "SN" in arr or "EW" in arr or "WE" in arr: arr = r...
[ "re.sub" ]
[((319, 349), 're.sub', 're.sub', (['"""NS|SN|EW|WE"""', '""""""', 'arr'], {}), "('NS|SN|EW|WE', '', arr)\n", (325, 349), False, 'import re\n')]
# Copyright 2020 Red Hat Inc. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing,...
[ "logging.getLogger", "mrack.errors.ValidationError", "datetime.datetime.now", "mrack.errors.ProvisioningError", "asyncio.gather" ]
[((784, 811), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (801, 811), False, 'import logging\n'), ((2694, 2708), 'datetime.datetime.now', 'datetime.now', ([], {}), '()\n', (2706, 2708), False, 'from datetime import datetime\n'), ((3370, 3384), 'datetime.datetime.now', 'datetime.now', (...
""" Capstone Project. Code to run on a LAPTOP (NOT the robot). Displays the Graphical User Interface (GUI) and communicates with the robot. Authors: Your professors (for the framework) and <NAME>. Winter term, 2018-2019. """ import mqtt_remote_method_calls as com import tkinter from tkinter import ttk i...
[ "tkinter.ttk.Button", "tkinter.ttk.Entry", "random.randrange", "tkinter.ttk.Frame", "tkinter.ttk.Label", "time.sleep", "tkinter.Tk", "tkinter.ttk.Progressbar", "mqtt_remote_method_calls.MqttClient" ]
[((822, 846), 'mqtt_remote_method_calls.MqttClient', 'com.MqttClient', (['delegate'], {}), '(delegate)\n', (836, 846), True, 'import mqtt_remote_method_calls as com\n'), ((879, 892), 'time.sleep', 'time.sleep', (['(1)'], {}), '(1)\n', (889, 892), False, 'import time\n'), ((1104, 1116), 'tkinter.Tk', 'tkinter.Tk', ([], ...
from aiogram.dispatcher import FSMContext from aiogram.dispatcher.filters import Command from aiogram.types import Message, CallbackQuery from sqlalchemy.exc import NoResultFound from loader import dp, logger_guru from utils.database_manage.sql.sql_commands import DB_USERS from utils.keyboards.admins_tools_kb import t...
[ "loader.logger_guru.exception", "aiogram.dispatcher.filters.Command", "loader.dp.callback_query_handler", "utils.database_manage.sql.sql_commands.DB_USERS.update_personal_pass", "utils.database_manage.sql.sql_commands.DB_USERS.check_personal_pass", "loader.dp.message_handler", "utils.misc.notify_users.s...
[((855, 958), 'loader.dp.callback_query_handler', 'dp.callback_query_handler', ([], {'text': "{'reset_user_codeword', 'make_newsletter'}", 'state': '"""admin_in_action"""'}), "(text={'reset_user_codeword', 'make_newsletter'},\n state='admin_in_action')\n", (880, 958), False, 'from loader import dp, logger_guru\n'), ...
import math #matplotlib keyboard noteKeys = "<KEY>'" class keyboard: def __init__(self,keystring="<KEY>"): self.keys = keystring self._keepinscope = None self.keysDown = set() def plot(self,ival=50): import matplotlib.pyplot as plt import matplotlib.animation as animati...
[ "filters.biquadPeak", "matplotlib.animation.FuncAnimation", "math.log", "math.exp", "matplotlib.pyplot.subplots" ]
[((1047, 1068), 'math.exp', 'math.exp', (['(2 * math.pi)'], {}), '(2 * math.pi)\n', (1055, 1068), False, 'import math\n'), ((2865, 2876), 'math.log', 'math.log', (['(2)'], {}), '(2)\n', (2873, 2876), False, 'import math\n'), ((341, 355), 'matplotlib.pyplot.subplots', 'plt.subplots', ([], {}), '()\n', (353, 355), True, ...
# -*- coding: utf-8 -*- # Form implementation generated from reading ui file 'Calculation.ui' # # Created by: PyQt5 UI code generator 5.9.2 # # WARNING! All changes made in this file will be lost! from Calculation.CalculatingLast5Days import CalculatingLast5Days from PyQt5 import QtCore, QtGui, QtWidgets ...
[ "PyQt5.QtWidgets.QWidget", "PyQt5.QtWidgets.QMainWindow", "PyQt5.QtGui.QFont", "PyQt5.QtWidgets.QSizePolicy", "PyQt5.QtCore.QMetaObject.connectSlotsByName", "Calculation.CalculatingLast5Days.CalculatingLast5Days", "PyQt5.QtWidgets.QHBoxLayout", "PyQt5.QtCore.QRect", "PyQt5.QtWidgets.QLabel", "PyQt...
[((8245, 8277), 'PyQt5.QtWidgets.QApplication', 'QtWidgets.QApplication', (['sys.argv'], {}), '(sys.argv)\n', (8267, 8277), False, 'from PyQt5 import QtCore, QtGui, QtWidgets\n'), ((8296, 8319), 'PyQt5.QtWidgets.QMainWindow', 'QtWidgets.QMainWindow', ([], {}), '()\n', (8317, 8319), False, 'from PyQt5 import QtCore, QtG...
import os import sys import inspect from ruamel import yaml def get_script_dir(follow_symlinks=True): if getattr(sys, 'frozen', False): # py2exe, PyInstaller, cx_Freeze path = os.path.abspath(sys.executable) else: path = inspect.getabsfile(get_script_dir) if follow_symlinks: ...
[ "os.path.exists", "os.listdir", "os.makedirs", "os.path.realpath", "os.path.dirname", "inspect.getabsfile", "os.path.abspath" ]
[((370, 391), 'os.path.dirname', 'os.path.dirname', (['path'], {}), '(path)\n', (385, 391), False, 'import os\n'), ((453, 471), 'os.path.dirname', 'os.path.dirname', (['d'], {}), '(d)\n', (468, 471), False, 'import os\n'), ((696, 717), 'os.listdir', 'os.listdir', (['directory'], {}), '(directory)\n', (706, 717), False,...
import random from typing import Any import factory import pytz from django.utils.text import slugify from ap.apps.events.models import EVENT_TYPE_CHOICES, Event, Organization, Route from ap.apps.users.constants import COUNTRY_CHOICES from ap.apps.users.models import User from ap.apps.events.factory_data import RANDO...
[ "django.utils.text.slugify", "ap.apps.events.models.Organization.objects.all", "random.choice", "factory.Faker", "ap.apps.users.models.User.objects.all" ]
[((552, 590), 'factory.Faker', 'factory.Faker', (['"""text"""'], {'max_nb_chars': '(60)'}), "('text', max_nb_chars=60)\n", (565, 590), False, 'import factory\n'), ((601, 621), 'factory.Faker', 'factory.Faker', (['"""url"""'], {}), "('url')\n", (614, 621), False, 'import factory\n'), ((1003, 1045), 'factory.Faker', 'fac...
#!/usr/bin/env python from __future__ import division import math import random import time import genpy import rospy import roslib from qt_gui.plugin import Plugin from python_qt_binding.QtCore import Slot, QSignalMapper, QTimer from python_qt_binding.QtWidgets import QHeaderView, QTableWidgetItem from rqt_command_...
[ "python_qt_binding.QtCore.Slot", "rqt_command_publisher.command_publisher_widget.CommandPublisherWidget", "python_qt_binding.QtCore.QTimer", "roslib.message.get_message_class", "rospy.loginfo", "python_qt_binding.QtCore.QSignalMapper" ]
[((1935, 1964), 'python_qt_binding.QtCore.Slot', 'Slot', (['str', 'str', 'float', 'object'], {}), '(str, str, float, object)\n', (1939, 1964), False, 'from python_qt_binding.QtCore import Slot, QSignalMapper, QTimer\n'), ((3702, 3711), 'python_qt_binding.QtCore.Slot', 'Slot', (['int'], {}), '(int)\n', (3706, 3711), Fal...
import time import os import arcade import argparse import gym from gym import spaces import swarm_env import numpy as np import random import sys sys.path.insert(0, '..') from objects import SwarmSimulator # Running experiment 22 in standalone file. def experiment_runner(SWARM_SIZE = 15, ARENA_WIDTH = 600, ARENA_HEI...
[ "random.uniform", "sys.path.insert", "random.randint", "numpy.argmax", "numpy.max", "numpy.exp", "numpy.zeros", "os.path.isdir", "os.mkdir", "arcade.run", "objects.SwarmSimulator", "time.time", "gym.make" ]
[((147, 171), 'sys.path.insert', 'sys.path.insert', (['(0)', '""".."""'], {}), "(0, '..')\n", (162, 171), False, 'import sys\n'), ((352, 363), 'time.time', 'time.time', ([], {}), '()\n', (361, 363), False, 'import time\n'), ((1595, 1638), 'gym.make', 'gym.make', (['"""humanswarm-v0"""'], {'maze_size': 'GRID_X'}), "('hu...
from helpers import * from discord.ext import commands from discord.ext.commands import Cog, Bot, command, Context import database as db import discord import asyncio def setup(bot): bot.add_cog(Commands(bot)) class Commands(Cog): def __init__(self, bot: Bot): self.bot = bot # List spread acros...
[ "database.subscribe", "database.is_subscribed", "database.get_subscriptions", "database.get_all_tv_shows", "asyncio.wait", "database.unsubscribe", "database.search_tv_show", "discord.ext.commands.command" ]
[((2097, 2173), 'discord.ext.commands.command', 'command', ([], {'aliases': "['shows']", 'name': '"""tvshows"""', 'help': '"""Lists available TV Shows."""'}), "(aliases=['shows'], name='tvshows', help='Lists available TV Shows.')\n", (2104, 2173), False, 'from discord.ext.commands import Cog, Bot, command, Context\n'),...
#!/usr/bin/python # -*- coding: UTF-8 -*- import numpy as np class DMatrix: def __init__(self, data_arr, missing={np.nan, 0}): """ :param data_arr: 样本特征 (不含标签) :param missing: 缺失值的集合, 若特征值在此集合中, 则认为其为缺失值 """ # N 样本总个数( 包含缺出现缺失值的样本 ) # m 特征的总数 self.N, sel...
[ "numpy.shape" ]
[((326, 344), 'numpy.shape', 'np.shape', (['data_arr'], {}), '(data_arr)\n', (334, 344), True, 'import numpy as np\n')]
from bs4 import BeautifulSoup import requests import pandas as pd url = 'https://mexico.as.com/resultados/futbol/mexico_clausura/clasificacion/' page = requests.get(url) soup = BeautifulSoup(page.content, 'html.parser') #Equipos eq = soup.find_all('span', class_='nombre-equipo') equipos = list() count = 0 for i in e...
[ "bs4.BeautifulSoup", "requests.get" ]
[((153, 170), 'requests.get', 'requests.get', (['url'], {}), '(url)\n', (165, 170), False, 'import requests\n'), ((178, 220), 'bs4.BeautifulSoup', 'BeautifulSoup', (['page.content', '"""html.parser"""'], {}), "(page.content, 'html.parser')\n", (191, 220), False, 'from bs4 import BeautifulSoup\n')]
import inspect from flask import render_template, request, jsonify from ..configurables import app from .. import metrics from ..metrics import metric_classes def get_metrics(add_class=False): records = [] for name, metric in metric_classes.iteritems(): if metric.show_in_ui: new_record = {...
[ "flask.render_template", "flask.jsonify" ]
[((1298, 1322), 'flask.jsonify', 'jsonify', ([], {'metrics': 'records'}), '(metrics=records)\n', (1305, 1322), False, 'from flask import render_template, request, jsonify\n'), ((1977, 2071), 'flask.render_template', 'render_template', (['"""forms/metric_configuration.html"""'], {'form': 'metric_form', 'action': 'reques...
#!/usr/bin/python # -*- coding: utf-8 -*- from setuptools import setup setup( name='pyramid_zipkin-example', version='0.1', author='OpenZipkin', author_email='<EMAIL>', license='Apache 2.0', url='https://github.com/openzipkin/pyramid_zipkin-example', description='See how much time python se...
[ "setuptools.setup" ]
[((72, 402), 'setuptools.setup', 'setup', ([], {'name': '"""pyramid_zipkin-example"""', 'version': '"""0.1"""', 'author': '"""OpenZipkin"""', 'author_email': '"""<EMAIL>"""', 'license': '"""Apache 2.0"""', 'url': '"""https://github.com/openzipkin/pyramid_zipkin-example"""', 'description': '"""See how much time python s...
import os from glob import glob import tensorflow as tf from tensorflow.keras.applications.mobilenet_v2 import MobileNetV2, preprocess_input from tensorflow.keras.preprocessing.image import img_to_array, load_img from tensorflow.keras.models import Model, load_model from tensorflow.keras.layers import Conv2D,...
[ "tensorflow.keras.applications.mobilenet_v2.MobileNetV2", "os.path.dirname", "glob.glob", "tensorflow.keras.applications.mobilenet_v2.preprocess_input", "numpy.expand_dims", "tensorflow.keras.models.Model", "cv2.resize", "cv2.imread", "numpy.save" ]
[((816, 865), 'tensorflow.keras.applications.mobilenet_v2.MobileNetV2', 'MobileNetV2', ([], {'weights': '"""imagenet"""', 'include_top': '(True)'}), "(weights='imagenet', include_top=True)\n", (827, 865), False, 'from tensorflow.keras.applications.mobilenet_v2 import MobileNetV2, preprocess_input\n'), ((906, 974), 'ten...
import argparse import matplotlib.pyplot as plt import meshcut import numpy as np import pandas import seaborn as sns import pandas as pd import sys, os import math #from scipy.stats import norm SAVE_PATH = os.path.join(os.path.expanduser("~"),'PycharmProjects/Gibson_Exercise/examples/plot_result/') WAY_PATH = os.path...
[ "matplotlib.pyplot.grid", "argparse.ArgumentParser", "matplotlib.pyplot.ylabel", "meshcut.cross_section", "matplotlib.pyplot.xlabel", "numpy.absolute", "matplotlib.pyplot.plot", "matplotlib.pyplot.style.use", "os.path.join", "numpy.array", "matplotlib.pyplot.figure", "numpy.zeros", "numpy.mi...
[((221, 244), 'os.path.expanduser', 'os.path.expanduser', (['"""~"""'], {}), "('~')\n", (239, 244), False, 'import sys, os\n'), ((326, 349), 'os.path.expanduser', 'os.path.expanduser', (['"""~"""'], {}), "('~')\n", (344, 349), False, 'import sys, os\n'), ((705, 720), 'numpy.array', 'np.array', (['verts'], {}), '(verts)...
"""A module for the uFJC single-chain model in the isometric ensemble. This module consist of the class ``uFJCIsometric`` which contains methods for computing single-chain quantities in the isometric (constant end-to-end vector) thermodynamic ensemble. Example: Import and instantiate the class: ...
[ "numpy.log", "numpy.linalg.norm" ]
[((11454, 11494), 'numpy.linalg.norm', 'la.norm', (['(config[j, :] - config[j - 1, :])'], {}), '(config[j, :] - config[j - 1, :])\n', (11461, 11494), True, 'import numpy.linalg as la\n'), ((10199, 10234), 'numpy.log', 'np.log', (['(1 + eta * coth / self.kappa)'], {}), '(1 + eta * coth / self.kappa)\n', (10205, 10234), ...
import inspect import re def debug_str(arg): '''Return string of arg varible name and str(arg) value.''' frame = inspect.currentframe().f_back s = inspect.getframeinfo(frame).code_context[0] r = re.search(r"\((.*)\)", s).group(1) return str("{} = {}".format(r, arg)) def debug_print(arg): '''...
[ "re.search", "inspect.currentframe", "inspect.getframeinfo" ]
[((123, 145), 'inspect.currentframe', 'inspect.currentframe', ([], {}), '()\n', (143, 145), False, 'import inspect\n'), ((378, 400), 'inspect.currentframe', 'inspect.currentframe', ([], {}), '()\n', (398, 400), False, 'import inspect\n'), ((161, 188), 'inspect.getframeinfo', 'inspect.getframeinfo', (['frame'], {}), '(f...
""" Author: <NAME> Date: 10 April 2021 """ import logging import os from typing import List from dataclasses import dataclass, asdict import click from click import Context from sla_cli.src.cli.context import COMMAND_CONTEXT_SETTINGS from sla_cli.src.cli.utils import kwargs_to_dataclass, default_from_conte...
[ "logging.getLogger", "click.argument", "sla_cli.src.download.isic.IsicMetadataDownloader", "click.option", "sla_cli.src.cli.utils.default_from_context", "sla_cli.src.db.accessors.AccessorFactory.create_datasets", "sla_cli.src.cli.utils.kwargs_to_dataclass", "click.command", "sla_cli.src.download.Dow...
[((724, 751), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (741, 751), False, 'import logging\n'), ((931, 1021), 'click.command', 'click.command', ([], {'short_help': '"""Downloads available datasets."""'}), "(**COMMAND_CONTEXT_SETTINGS, short_help=\n 'Downloads available datasets.')...
import json print("Generating stamp files") stamps = [ "armor_plate", "armor_trim", "arrow_head", "arrow_shaft", "axe_head", "binding", "boots_core", "bow_limb", "bow_string", "broad_axe_head", "chest_core", "cross_guard", "emerald", "excavator_head", "fletching", "hammer_head", "han...
[ "json.dump" ]
[((929, 963), 'json.dump', 'json.dump', (['data', 'outfile'], {'indent': '(2)'}), '(data, outfile, indent=2)\n', (938, 963), False, 'import json\n')]
import torch from uninas.modules.mixed.mixedop import AbstractDependentMixedOp from uninas.methods.strategies.manager import StrategyManager from uninas.register import Register @Register.network_mixed_op() class SplitWeightsMixedOp(AbstractDependentMixedOp): """ all op choices on one path in parallel, th...
[ "uninas.register.Register.network_mixed_op", "uninas.methods.strategies.manager.StrategyManager" ]
[((181, 208), 'uninas.register.Register.network_mixed_op', 'Register.network_mixed_op', ([], {}), '()\n', (206, 208), False, 'from uninas.register import Register\n'), ((1253, 1270), 'uninas.methods.strategies.manager.StrategyManager', 'StrategyManager', ([], {}), '()\n', (1268, 1270), False, 'from uninas.methods.strat...
""" Plot various visualizations """ import sys import json from collections import Counter import numpy as np import scipy.stats as scstats import matplotlib matplotlib.use("agg") import matplotlib.pyplot as plt import matplotlib.cm as cm import matplotlib.patches as mp from zipteedo.util import GzipFileType, load...
[ "zipteedo.stats.make_bias_table", "matplotlib.pyplot.ylabel", "numpy.polyfit", "zipteedo.util.first", "numpy.array", "zipteedo.stats.get_correlations", "matplotlib.pyplot.errorbar", "sys.exit", "argparse.ArgumentParser", "matplotlib.pyplot.xlabel", "matplotlib.pyplot.plot", "zipteedo.viz.violi...
[((160, 181), 'matplotlib.use', 'matplotlib.use', (['"""agg"""'], {}), "('agg')\n", (174, 181), False, 'import matplotlib\n'), ((1343, 1365), 'zipteedo.stats.get_correlations', 'get_correlations', (['data'], {}), '(data)\n', (1359, 1365), False, 'from zipteedo.stats import get_correlations, get_data_efficiencies, make_...
import plotly.plotly as py import plotly.graph_objs as go import plotly.offline as offline import sys, os from .CreateTableFromDatabase import getRankingsFromDatabase # Using the plotly API creates a few plots def createAndUploadPlots(table, plotName): # Read plotly username and API key from file (to avoid acci...
[ "plotly.offline.plot", "plotly.plotly.plot", "os.path.dirname", "plotly.graph_objs.Layout", "plotly.graph_objs.Figure" ]
[((1285, 1334), 'plotly.graph_objs.Layout', 'go.Layout', ([], {'title': "('%s' % plotName)", 'barmode': '"""group"""'}), "(title='%s' % plotName, barmode='group')\n", (1294, 1334), True, 'import plotly.graph_objs as go\n'), ((1371, 1406), 'plotly.graph_objs.Figure', 'go.Figure', ([], {'data': 'data', 'layout': 'layout'...
# -*- coding: utf-8 -*- from __future__ import unicode_literals from django.db import models, migrations from django.conf import settings class Migration(migrations.Migration): dependencies = [ migrations.swappable_dependency(settings.AUTH_USER_MODEL), ('proposals', '0003_auto_20150113_1401'), ...
[ "django.db.models.ForeignKey", "django.db.models.BooleanField", "django.db.models.AutoField", "django.db.models.DateTimeField", "django.db.migrations.swappable_dependency", "django.db.models.CharField" ]
[((210, 267), 'django.db.migrations.swappable_dependency', 'migrations.swappable_dependency', (['settings.AUTH_USER_MODEL'], {}), '(settings.AUTH_USER_MODEL)\n', (241, 267), False, 'from django.db import models, migrations\n'), ((523, 616), 'django.db.models.AutoField', 'models.AutoField', ([], {'verbose_name': '"""ID"...
#!/usr/bin/env python3 import statistics import os import glob from tkinter import filedialog from tkinter import * # noqa import pandas as pd from eventcodes import eventcodes_dictionary from natsort import natsorted, ns import matplotlib.pyplot as plt import numpy as np import datetime __all__ = ["loop_over_days",...
[ "statistics.mean", "tkinter.filedialog.askdirectory", "pandas.read_csv", "datetime.datetime.strptime", "numpy.delete", "matplotlib.pyplot.xlabel", "os.path.join", "pandas.DataFrame.from_dict", "matplotlib.pyplot.figure", "glob.glob", "natsort.natsorted", "pandas.to_numeric", "pandas.DataFram...
[((4274, 4307), 'pandas.DataFrame', 'pd.DataFrame', ([], {'columns': 'column_list'}), '(columns=column_list)\n', (4286, 4307), True, 'import pandas as pd\n'), ((5395, 5428), 'pandas.DataFrame', 'pd.DataFrame', ([], {'columns': 'column_list'}), '(columns=column_list)\n', (5407, 5428), True, 'import pandas as pd\n'), ((7...
#!/usr/bin/env python3 # -*- coding: UTF-8 -*- """ boot script, do initial stuff here, similar to the setup() function on Arduino """ import esp import gc import network import time # custom packages from be_helpers.led_helper import Led # set clock speed to 240MHz instead of default 160MHz # machine.freq(24000000...
[ "be_helpers.led_helper.Led", "time.sleep", "network.WLAN", "gc.collect", "esp.osdebug" ]
[((354, 371), 'esp.osdebug', 'esp.osdebug', (['None'], {}), '(None)\n', (365, 371), False, 'import esp\n'), ((379, 384), 'be_helpers.led_helper.Led', 'Led', ([], {}), '()\n', (382, 384), False, 'from be_helpers.led_helper import Led\n'), ((443, 471), 'network.WLAN', 'network.WLAN', (['network.STA_IF'], {}), '(network.S...
#!/usr/bin/env python # -*- coding: utf8 -*- import os from functools import reduce from Log import log __title__ = 'Structure runner' __all__ = ['Structure'] __author__ = '<NAME> <<EMAIL>>' class Structure(object): @staticmethod def get_directory_structure(confdir: str) -> dict: """ Create...
[ "Log.log.error", "xml.etree.ElementTree.parse", "configparser.ConfigParser", "functools.reduce", "os.path.join", "yaml.load", "json.load", "os.walk" ]
[((565, 581), 'os.walk', 'os.walk', (['confdir'], {}), '(confdir)\n', (572, 581), False, 'import os\n'), ((998, 1014), 'os.walk', 'os.walk', (['confdir'], {}), '(confdir)\n', (1005, 1014), False, 'import os\n'), ((3732, 3746), 'configparser.ConfigParser', 'ConfigParser', ([], {}), '()\n', (3744, 3746), False, 'from con...
import ray from ray.experimental.workflow import storage from ray.experimental.workflow import workflow_storage def some_func(x): return x + 1 def some_func2(x): return x - 1 def test_raw_storage(): ray.init() workflow_id = test_workflow_storage.__name__ raw_storage = storage.get_global_storag...
[ "ray.shutdown", "ray.experimental.workflow.workflow_storage.WorkflowStorage", "ray.experimental.workflow.storage.get_global_storage", "ray.put", "ray.init", "ray.experimental.workflow.workflow_storage.StepInspectResult" ]
[((217, 227), 'ray.init', 'ray.init', ([], {}), '()\n', (225, 227), False, 'import ray\n'), ((295, 323), 'ray.experimental.workflow.storage.get_global_storage', 'storage.get_global_storage', ([], {}), '()\n', (321, 323), False, 'from ray.experimental.workflow import storage\n'), ((517, 541), 'ray.put', 'ray.put', (['ob...
from django.db import models from django.utils import timezone class BaseCoin(models.Model): date = models.DateField(default=timezone.now) exchange = models.CharField(max_length=100, blank=True, default='') price = models.DecimalField(max_digits=12, decimal_places=4, default=0) high = models.DecimalFi...
[ "django.db.models.DecimalField", "django.db.models.DateField", "django.db.models.CharField" ]
[((106, 144), 'django.db.models.DateField', 'models.DateField', ([], {'default': 'timezone.now'}), '(default=timezone.now)\n', (122, 144), False, 'from django.db import models\n'), ((160, 216), 'django.db.models.CharField', 'models.CharField', ([], {'max_length': '(100)', 'blank': '(True)', 'default': '""""""'}), "(max...
from __future__ import print_function from ghidra.program.model.address import Address import platform import ghidra if (platform.system() != "Java"): from ghidra_builtins import * monitor = ghidra.util.task.TaskMonitor.DUMMY def getFuncInstructions(function): instructions = [] instruction = getFirstIns...
[ "ghidra.app.cmd.disassemble.DisassembleCommand", "ghidra.program.util.ProgramSelection", "ghidra.util.datastruct.ListAccumulator", "ghidra.program.model.data.PointerDataType", "ghidra.util.search.memory.SearchInfo", "platform.system", "ghidra.app.plugin.core.searchmem.SearchData.createSearchData" ]
[((123, 140), 'platform.system', 'platform.system', ([], {}), '()\n', (138, 140), False, 'import platform\n'), ((1047, 1135), 'ghidra.app.plugin.core.searchmem.SearchData.createSearchData', 'ghidra.app.plugin.core.searchmem.SearchData.createSearchData', (['"""search"""', 'values', 'mask'], {}), "('search',\n values,...
import unittest import numpy as np import sklearn_supp.random_coordinates as random_coordinates class TestRandomCoordinateForestClassifier(unittest.TestCase): """These are just some simple sanity checks to make sure we don't get exceptions. """ def test_simple(self): X = [[0], [1]] ...
[ "numpy.array", "numpy.all", "sklearn_supp.random_coordinates.RandomCoordinateForestClassifier" ]
[((352, 420), 'sklearn_supp.random_coordinates.RandomCoordinateForestClassifier', 'random_coordinates.RandomCoordinateForestClassifier', ([], {'n_estimators': '(50)'}), '(n_estimators=50)\n', (403, 420), True, 'import sklearn_supp.random_coordinates as random_coordinates\n'), ((537, 556), 'numpy.all', 'np.all', (['(y_p...
# Author : <NAME> # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # d...
[ "os.listdir", "utils.fetch_action_ids", "pandas.read_csv", "os.path.join", "numpy.array", "numpy.sum", "warnings.simplefilter", "utils.fetch_parsed_recipe" ]
[((824, 886), 'warnings.simplefilter', 'warnings.simplefilter', ([], {'action': '"""ignore"""', 'category': 'FutureWarning'}), "(action='ignore', category=FutureWarning)\n", (845, 886), False, 'import warnings\n'), ((974, 992), 'os.listdir', 'os.listdir', (['folder'], {}), '(folder)\n', (984, 992), False, 'import os\n'...
from flask import Flask, current_app, request from flask_babelex import get_locale from flask_sqlalchemy import SQLAlchemy from flask_transalchemy.model import TranslationMixin class TransAlchemy(object): """Flask-TransAlchemy extension class. :param app: Flask application instance :param db: Flask-SQLA...
[ "flask_babelex.get_locale", "flask.current_app.extensions.get" ]
[((2719, 2762), 'flask.current_app.extensions.get', 'current_app.extensions.get', (['"""babel_alchemy"""'], {}), "('babel_alchemy')\n", (2745, 2762), False, 'from flask import Flask, current_app, request\n'), ((3159, 3202), 'flask.current_app.extensions.get', 'current_app.extensions.get', (['"""babel_alchemy"""'], {}),...
# Copyright 2016, The NIG Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); you may not # use this file except in compliance with the License. You may obtain a copy of # the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable la...
[ "timeit.default_timer", "re.escape" ]
[((1001, 1016), 'timeit.default_timer', 'default_timer', ([], {}), '()\n', (1014, 1016), False, 'from timeit import default_timer\n'), ((1101, 1116), 'timeit.default_timer', 'default_timer', ([], {}), '()\n', (1114, 1116), False, 'from timeit import default_timer\n'), ((1336, 1351), 're.escape', 're.escape', (['char'],...
from __future__ import print_function, absolute_import from random import random from time import sleep import threading import pygame import sys try: from olympe.messages.ardrone3.GPSSettingsState import GPSFixStateChanged from olympe.messages.ardrone3.Piloting import TakeOff, Landing from olympe.messag...
[ "olympe.Drone", "pygame.init", "olympe.messages.ardrone3.Piloting.Landing", "olympe.messages.ardrone3.PilotingState.FlyingStateChanged", "pygame.joystick.Joystick", "time.sleep", "olympe.messages.ardrone3.Piloting.TakeOff", "sys.exit", "pygame.event.pump", "olympe.messages.ardrone3.GPSSettingsStat...
[((4034, 4087), 'olympe.Drone', 'olympe.Drone', (['JoystickTeleop.SIMULATED_IP'], {'loglevel': '(0)'}), '(JoystickTeleop.SIMULATED_IP, loglevel=0)\n', (4046, 4087), False, 'import olympe\n'), ((4239, 4250), 'sys.exit', 'sys.exit', (['(0)'], {}), '(0)\n', (4247, 4250), False, 'import sys\n'), ((499, 510), 'sys.exit', 's...
import argparse import csv import glob import json import os import re import time from math import ceil from shutil import copyfile, rmtree from urllib.parse import urlparse import requests from pyproc import Lpse, __version__ from pyproc.helpers import DetilDownloader from urllib3 import disable_warnings from urll...
[ "csv.DictWriter", "pyproc.helpers.DetilDownloader", "pathlib.Path.home", "time.sleep", "argparse.ArgumentParser", "pyproc.Lpse", "csv.reader", "glob.glob", "os.rename", "csv.writer", "os.path.isfile", "urllib3.disable_warnings", "os.path.dirname", "re.findall", "requests.session", "mat...
[((2671, 2709), 'os.makedirs', 'os.makedirs', (['detail_dir'], {'exist_ok': '(True)'}), '(detail_dir, exist_ok=True)\n', (2682, 2709), False, 'import os\n'), ((3397, 3436), 'os.path.join', 'os.path.join', (['folder_name', '"""detil"""', '"""*"""'], {}), "(folder_name, 'detil', '*')\n", (3409, 3436), False, 'import os\n...
import tkinter as tk from tkinter import ttk, messagebox, font, StringVar from tkcalendar import DateEntry from tkcrud.controller.client_controller import ClientController,\ saving_updating, get_clients, window_popup class FormClientRegister(tk.Toplevel, ClientController): def __init__(self, master, tree): ...
[ "tkinter.LabelFrame", "tkinter.Entry", "tkinter.Toplevel.__init__", "tkcalendar.DateEntry", "tkinter.ttk.Scrollbar", "tkcrud.controller.client_controller.get_clients", "tkinter.font.Font", "tkinter.StringVar", "tkinter.Label", "tkinter.ttk.Combobox", "tkcrud.controller.client_controller.window_p...
[((327, 361), 'tkinter.Toplevel.__init__', 'tk.Toplevel.__init__', (['self', 'master'], {}), '(self, master)\n', (347, 361), True, 'import tkinter as tk\n'), ((444, 487), 'tkinter.LabelFrame', 'tk.LabelFrame', (['self'], {'text': '"""Register client"""'}), "(self, text='Register client')\n", (457, 487), True, 'import t...
import sys, os, re, json, pprint, traceback from pathlib import Path from aiohttp import web routes = web.RouteTableDef() async def app(): from web_chains_202105 import directories app = web.Application(middlewares=[exception_middleware]) app.add_routes(routes) app.router.add_static("/js/", path="js",...
[ "web_chains_202105.chart.get_map", "pathlib.Path", "web_chains_202105.table_page.table_page", "web_chains_202105.directories.load", "web_chains_202105.index_page.index_page", "aiohttp.web.Application", "sys.exc_info", "aiohttp.web.RouteTableDef", "web_chains_202105.chain_page.chain_page", "os.getp...
[((103, 122), 'aiohttp.web.RouteTableDef', 'web.RouteTableDef', ([], {}), '()\n', (120, 122), False, 'from aiohttp import web\n'), ((197, 248), 'aiohttp.web.Application', 'web.Application', ([], {'middlewares': '[exception_middleware]'}), '(middlewares=[exception_middleware])\n', (212, 248), False, 'from aiohttp import...
# Copyright 2019 Zuru Tech HK Limited. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable la...
[ "tensorflow.train.Checkpoint", "tensorflow.metrics.Mean", "tensorflow.math.log", "tensorflow.metrics.Accuracy", "tensorflow.keras.layers.Dense", "tensorflow.math.exp", "pathlib.Path", "tensorflow.image.grayscale_to_rgb", "tensorflow.math.reduce_mean", "tensorflow_hub.KerasLayer", "tensorflow.ker...
[((11863, 11889), 'tensorflow.math.exp', 'tf.math.exp', (['kl_divergence'], {}), '(kl_divergence)\n', (11874, 11889), True, 'import tensorflow as tf\n'), ((12157, 12220), 'tensorflow.keras.losses.SparseCategoricalCrossentropy', 'tf.keras.losses.SparseCategoricalCrossentropy', ([], {'from_logits': '(True)'}), '(from_log...
# -*- coding: utf-8 -*- """Database module, including the SQLAlchemy database object and DB-related utilities.""" import json from sqlalchemy.inspection import inspect from sqlalchemy.orm import relationship from .extensions import db # Alias common SQLAlchemy names Column = db.Column relationship = relationship c...
[ "sqlalchemy.inspection.inspect" ]
[((1383, 1396), 'sqlalchemy.inspection.inspect', 'inspect', (['self'], {}), '(self)\n', (1390, 1396), False, 'from sqlalchemy.inspection import inspect\n')]
from owslib.wms import WebMapService from owslib import crs from PIL import Image, ImageEnhance, ImageFilter import cv2 import numpy as np from pyspark import SparkContext from pyproj import Proj c = crs.Crs('EPSG:3857') wms = WebMapService('http://www.ign.es/wms-inspire/pnoa-ma', version='1.3.0') box = 1000 # m? x=...
[ "cv2.imwrite", "cv2.drawContours", "owslib.wms.WebMapService", "cv2.inRange", "owslib.crs.Crs", "cv2.arcLength", "cv2.bitwise_and", "cv2.contourArea", "numpy.array", "cv2.approxPolyDP", "cv2.cvtColor", "pyproj.Proj", "cv2.findContours", "cv2.Canny", "cv2.imread", "cv2.boundingRect" ]
[((201, 221), 'owslib.crs.Crs', 'crs.Crs', (['"""EPSG:3857"""'], {}), "('EPSG:3857')\n", (208, 221), False, 'from owslib import crs\n'), ((228, 299), 'owslib.wms.WebMapService', 'WebMapService', (['"""http://www.ign.es/wms-inspire/pnoa-ma"""'], {'version': '"""1.3.0"""'}), "('http://www.ign.es/wms-inspire/pnoa-ma', ver...
"""Set of basic widgets for BioImageIT Classes ------- BiWebBrowser """ from qtpy.QtWebEngineWidgets import QWebEngineView from qtpy.QtWidgets import (QWidget, QPushButton, QHBoxLayout, QVBoxLayout) class BiWebBrowser(QWidget): def __init__(self, parent: QWidget): super(Bi...
[ "qtpy.QtWidgets.QVBoxLayout", "qtpy.QtWidgets.QWidget", "qtpy.QtWebEngineWidgets.QWebEngineView", "qtpy.QtWidgets.QPushButton", "qtpy.QtWidgets.QHBoxLayout" ]
[((377, 390), 'qtpy.QtWidgets.QVBoxLayout', 'QVBoxLayout', ([], {}), '()\n', (388, 390), False, 'from qtpy.QtWidgets import QWidget, QPushButton, QHBoxLayout, QVBoxLayout\n'), ((581, 601), 'qtpy.QtWebEngineWidgets.QWebEngineView', 'QWebEngineView', (['self'], {}), '(self)\n', (595, 601), False, 'from qtpy.QtWebEngineWi...
import launch import launch_ros def generate_launch_description(): server_node_container = launch_ros.actions.ComposableNodeContainer( node_name='server_node_container', node_namespace='', package='rclcpp_components', node_executable='component_container', ...
[ "launch.LaunchDescription", "launch_ros.descriptions.ComposableNode" ]
[((649, 698), 'launch.LaunchDescription', 'launch.LaunchDescription', (['[server_node_container]'], {}), '([server_node_container])\n', (673, 698), False, 'import launch\n'), ((370, 498), 'launch_ros.descriptions.ComposableNode', 'launch_ros.descriptions.ComposableNode', ([], {'package': '"""hello_world_cpp"""', 'node_...