code stringlengths 22 1.05M | apis listlengths 1 3.31k | extract_api stringlengths 75 3.25M |
|---|---|---|
import logging
import math
import time
from typing import Dict, List
from servo import Servo
class ControllerForPCA9685:
def __init__(self, servos: Dict[object, Servo], chs: Dict[object, int],
pwm_freq: float, init_angles: Dict[object, float] = None):
import Adafruit_PCA9685 as PCA9685
... | [
"math.isclose",
"Adafruit_PCA9685.software_reset",
"Adafruit_PCA9685.PCA9685",
"logging.info",
"RPi.GPIO.setmode"
] | [((965, 989), 'Adafruit_PCA9685.software_reset', 'PCA9685.software_reset', ([], {}), '()\n', (987, 989), True, 'import Adafruit_PCA9685 as PCA9685\n'), ((1013, 1030), 'Adafruit_PCA9685.PCA9685', 'PCA9685.PCA9685', ([], {}), '()\n', (1028, 1030), True, 'import Adafruit_PCA9685 as PCA9685\n'), ((2259, 2283), 'RPi.GPIO.se... |
from django.db import models
from django.utils import timezone
from django.contrib.auth.models import User
# Create your models here.
class Neighbor(models.Model) :
n_name = models.CharField(max_length=35)
n_location = models.CharField(max_length=35)
n_image = models.ImageField(upload_to='n_posts/')
n_title... | [
"django.db.models.TextField",
"django.db.models.ForeignKey",
"django.db.models.DateTimeField",
"django.db.models.ImageField",
"django.db.models.CharField"
] | [((180, 211), 'django.db.models.CharField', 'models.CharField', ([], {'max_length': '(35)'}), '(max_length=35)\n', (196, 211), False, 'from django.db import models\n'), ((227, 258), 'django.db.models.CharField', 'models.CharField', ([], {'max_length': '(35)'}), '(max_length=35)\n', (243, 258), False, 'from django.db im... |
#!/usr/bin/env python3
"""This python program deploys the files needed by the Hillview service
on the machines specified in the configuration file."""
# pylint: disable=invalid-name
from argparse import ArgumentParser
import tempfile
import os.path
from hillviewCommon import ClusterConfiguration, get_config, get_l... | [
"hillviewCommon.get_logger",
"hillviewCommon.get_config",
"argparse.ArgumentParser",
"tempfile.NamedTemporaryFile"
] | [((336, 356), 'hillviewCommon.get_logger', 'get_logger', (['"""deploy"""'], {}), "('deploy')\n", (346, 356), False, 'from hillviewCommon import ClusterConfiguration, get_config, get_logger\n'), ((2641, 2692), 'tempfile.NamedTemporaryFile', 'tempfile.NamedTemporaryFile', ([], {'mode': '"""w"""', 'delete': '(False)'}), "... |
import urllib3
from beautifulsoup4 import beautifulsoup4 as bs
def get_historical_data(name, number_of_days):
data = []
url = "https://finance.yahoo.com/quote/" + name + "/history/"
rows = bs(urllib3.urlopen(url).read()).findAll('table')[0].tbody.findAll('tr')
for each_row in rows:
divs = each_row.findAll('td')... | [
"urllib3.urlopen"
] | [((196, 216), 'urllib3.urlopen', 'urllib3.urlopen', (['url'], {}), '(url)\n', (211, 216), False, 'import urllib3\n')] |
#!/usr/bin/python
import os, glob, hashlib, pickle, argparse, shutil, ntpath
import os, glob, hashlib, pickle, argparse, shutil, multiprocessing, signal, sys
from multiprocessing import Pool
from functools import partial
######################### Classes ##############################
class AndroidDensity:
def __i... | [
"os.path.exists",
"ntpath.basename",
"pickle.dump",
"hashlib.md5",
"argparse.ArgumentParser",
"os.makedirs",
"os.listdir",
"os.path.splitext",
"pickle.load",
"multiprocessing.cpu_count",
"shutil.rmtree",
"os.rmdir",
"functools.partial",
"multiprocessing.Pool",
"os.path.basename",
"sys.... | [((1696, 1723), 'multiprocessing.cpu_count', 'multiprocessing.cpu_count', ([], {}), '()\n', (1721, 1723), False, 'import os, glob, hashlib, pickle, argparse, shutil, multiprocessing, signal, sys\n'), ((2480, 2552), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {'description': '"""Seamless mobile assets man... |
import os
from djangular import utils
from django.test import SimpleTestCase
class SiteAndPathUtilsTest(SimpleTestCase):
site_utils = utils.SiteAndPathUtils()
def test_djangular_root(self):
current_dir = os.path.dirname(os.path.abspath(__file__))
djangular_dir = os.path.dirname(current_dir)... | [
"os.path.abspath",
"os.path.dirname",
"djangular.utils.SiteAndPathUtils"
] | [((142, 166), 'djangular.utils.SiteAndPathUtils', 'utils.SiteAndPathUtils', ([], {}), '()\n', (164, 166), False, 'from djangular import utils\n'), ((292, 320), 'os.path.dirname', 'os.path.dirname', (['current_dir'], {}), '(current_dir)\n', (307, 320), False, 'import os\n'), ((241, 266), 'os.path.abspath', 'os.path.absp... |
from django.conf.urls import patterns, include, url
from django.contrib import admin
from django.conf.urls.static import static
from userModule.views import home
from userModule.views import userSettings
from userModule.views import logout
from groupModule.views import createGroup
from groupModule.views import group
fr... | [
"django.conf.urls.static.static",
"django.conf.urls.include",
"django.contrib.staticfiles.urls.staticfiles_urlpatterns",
"django.conf.urls.url"
] | [((1632, 1657), 'django.contrib.staticfiles.urls.staticfiles_urlpatterns', 'staticfiles_urlpatterns', ([], {}), '()\n', (1655, 1657), False, 'from django.contrib.staticfiles.urls import staticfiles_urlpatterns\n'), ((1673, 1734), 'django.conf.urls.static.static', 'static', (['settings.MEDIA_URL'], {'document_root': 'se... |
# License: Apache 2.0. See LICENSE file in root directory.
# Copyright(c) 2021 Intel Corporation. All Rights Reserved.
# test:device L500*
# test:device D400*
import pyrealsense2 as rs
from rspy.stopwatch import Stopwatch
from rspy import test, log
import time
import platform
# Start depth + color streams and measur... | [
"rspy.test.finish",
"rspy.log.f",
"rspy.stopwatch.Stopwatch",
"rspy.test.check",
"rspy.test.print_results_and_exit",
"time.sleep",
"platform.system",
"rspy.test.find_first_device_or_exit"
] | [((1623, 1636), 'time.sleep', 'time.sleep', (['(3)'], {}), '(3)\n', (1633, 1636), False, 'import time\n'), ((1844, 1855), 'rspy.stopwatch.Stopwatch', 'Stopwatch', ([], {}), '()\n', (1853, 1855), False, 'from rspy.stopwatch import Stopwatch\n'), ((1862, 1894), 'rspy.test.find_first_device_or_exit', 'test.find_first_devi... |
from tensorflow.keras import activations, initializers, regularizers, constraints
from tensorflow.keras import backend as K
from tensorflow.keras.layers import Layer
from spektral.layers import ops
from spektral.utils import localpooling_filter
class GraphConv(Layer):
r"""
A graph convolutional layer (GCN) a... | [
"tensorflow.keras.constraints.get",
"tensorflow.keras.regularizers.serialize",
"tensorflow.keras.activations.get",
"spektral.layers.ops.filter_dot",
"tensorflow.keras.initializers.serialize",
"tensorflow.keras.backend.bias_add",
"spektral.utils.localpooling_filter",
"spektral.layers.ops.dot",
"tenso... | [((2188, 2215), 'tensorflow.keras.activations.get', 'activations.get', (['activation'], {}), '(activation)\n', (2203, 2215), False, 'from tensorflow.keras import activations, initializers, regularizers, constraints\n'), ((2283, 2319), 'tensorflow.keras.initializers.get', 'initializers.get', (['kernel_initializer'], {})... |
from django.contrib import admin
from .models import Device,Metric
class DeviceAdmin(admin.ModelAdmin):
list_display = ["name", "token","user", "created_date"]
search_fields = ["name", "token"]
list_filter = ("user",)
class MetricAdmin(admin.ModelAdmin):
list_display = ["device","temperature", "humid... | [
"django.contrib.admin.site.register"
] | [((404, 444), 'django.contrib.admin.site.register', 'admin.site.register', (['Device', 'DeviceAdmin'], {}), '(Device, DeviceAdmin)\n', (423, 444), False, 'from django.contrib import admin\n'), ((445, 485), 'django.contrib.admin.site.register', 'admin.site.register', (['Metric', 'MetricAdmin'], {}), '(Metric, MetricAdmi... |
from django.urls import reverse
from pytest import mark
from core.models import *
urls = [reverse(name) for name in ["core:index", "core:arts"]]
@mark.parametrize("url", urls)
@mark.django_db
def test_nsfw_filter(url, django_user_model, client):
target = django_user_model.objects.create(username="bob", password... | [
"pytest.mark.parametrize",
"django.urls.reverse"
] | [((150, 179), 'pytest.mark.parametrize', 'mark.parametrize', (['"""url"""', 'urls'], {}), "('url', urls)\n", (166, 179), False, 'from pytest import mark\n'), ((92, 105), 'django.urls.reverse', 'reverse', (['name'], {}), '(name)\n', (99, 105), False, 'from django.urls import reverse\n')] |
"""This script *replaces* the processlog.csh shell script. It has
more sophisticated logic to add the etc test name to the pysynphot
command as another keyword-value pair.
Modified to parse the modified line, then write a dictionary that will
be read by the gencases tool.
"""
import sys,re, pickle
from pysynphot impo... | [
"gencases.line2ktuple",
"pickle.dump",
"re.search"
] | [((1136, 1155), 'pickle.dump', 'pickle.dump', (['d', 'out'], {}), '(d, out)\n', (1147, 1155), False, 'import sys, re, pickle\n'), ((598, 631), 're.search', 're.search', (['"""\'(?P<name>.*)\'"""', 'line'], {}), '("\'(?P<name>.*)\'", line)\n', (607, 631), False, 'import sys, re, pickle\n'), ((915, 940), 'gencases.line2k... |
import teek
def on_click():
print("You clicked me!")
window = teek.Window()
button = teek.Button(window, "Click me", command=on_click)
button.pack()
window.on_delete_window.connect(teek.quit)
teek.run()
| [
"teek.run",
"teek.Window",
"teek.Button"
] | [((70, 83), 'teek.Window', 'teek.Window', ([], {}), '()\n', (81, 83), False, 'import teek\n'), ((93, 142), 'teek.Button', 'teek.Button', (['window', '"""Click me"""'], {'command': 'on_click'}), "(window, 'Click me', command=on_click)\n", (104, 142), False, 'import teek\n'), ((200, 210), 'teek.run', 'teek.run', ([], {})... |
# -*- coding: utf-8 -*-
"""
Created on Wed Dec 4 13:54:27 2019
@author: eric.qian
"""
from collections import Counter
from itertools import combinations
# Numeric hand rankings, higher integer value is stronger hand
HAND_RANKINGS = {0: 'High Card',
1: 'One Pair',
2: 'T... | [
"collections.Counter",
"itertools.combinations"
] | [((2383, 2401), 'collections.Counter', 'Counter', (['card_vals'], {}), '(card_vals)\n', (2390, 2401), False, 'from collections import Counter\n'), ((6462, 6484), 'itertools.combinations', 'combinations', (['cards', 'n'], {}), '(cards, n)\n', (6474, 6484), False, 'from itertools import combinations\n')] |
import sublime
import sublime_plugin
import datetime
import re
from pathlib import Path
import os
import fnmatch
import OrgExtended.orgparse.loader as loader
import OrgExtended.orgparse.node as node
import OrgExtended.orgparse.date as orgdate
from OrgExtended.orgparse.sublimenode import *
import OrgExtended.orgutil.... | [
"logging.getLogger",
"re.compile",
"OrgExtended.orgparse.date.OrgDateFreeFloating.from_str",
"OrgExtended.orgparse.date.OrgDate.format_clock",
"sublime.Region",
"OrgExtended.pymitter.Get",
"OrgExtended.orgproperties.UpdateProperty",
"sublime.active_window",
"OrgExtended.orglinks.CreateLink",
"OrgE... | [((961, 988), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (978, 988), False, 'import logging\n'), ((1694, 1732), 're.compile', 're.compile', (['"""^\\\\s*CLOSED:\\\\s*\\\\[.*\\\\]"""'], {}), "('^\\\\s*CLOSED:\\\\s*\\\\[.*\\\\]')\n", (1704, 1732), False, 'import re\n'), ((4792, 4856), '... |
from nt import DirEntry
import os
import self as self
from keras import Model as KerasModel
class Model(object):
"""
Base class for the different models in this project
"""
_files: dict = {}
""" The files that is used to parse the model """
_raw_content: str = ''
""" The files contents ... | [
"self._parse_directory",
"self._get_assets_full_path",
"self.compile_model",
"self._model.load_weights",
"os.scandir",
"self._raw_content.lower",
"os.path.dirname",
"self._model.summary",
"self._model.compile",
"self._parse_file",
"self._files.items"
] | [((1510, 1543), 'self._model.load_weights', 'self._model.load_weights', (['weights'], {}), '(weights)\n', (1534, 1543), True, 'import self as self\n'), ((1552, 1572), 'self.compile_model', 'self.compile_model', ([], {}), '()\n', (1570, 1572), True, 'import self as self\n'), ((1870, 1940), 'self._model.compile', 'self._... |
# coding=utf-8
import re
import traceback
from pesto_common.config.configer import Configer
from pesto_common.log.logger_factory import LoggerFactory
from pesto_orm.core.base import db_config
from pesto_orm.core.executor import ExecutorFactory
from pesto_orm.core.model import BaseModel
from pesto_orm.core.repository i... | [
"traceback.format_exc",
"pesto_orm.core.executor.ExecutorFactory.get_executor",
"re.match",
"pesto_common.log.logger_factory.LoggerFactory.get_logger",
"pesto_common.config.configer.Configer.get"
] | [((401, 449), 'pesto_common.log.logger_factory.LoggerFactory.get_logger', 'LoggerFactory.get_logger', (['"""dialect.mysql.domain"""'], {}), "('dialect.mysql.domain')\n", (425, 449), False, 'from pesto_common.log.logger_factory import LoggerFactory\n'), ((879, 902), 'pesto_common.config.configer.Configer.get', 'Configer... |
import os
import re
import unittest
import shutil
import subprocess
import logging
import zipfile
from io import BytesIO
from ..experiment.file_storage_driver import FileStorageDriver
from ..experiment.experiment_reader import ExperimentReader
from ..experiment.experiment_writer import ExperimentWriter
class Experim... | [
"os.path.exists",
"os.listdir",
"zipfile.ZipFile",
"subprocess.Popen",
"io.BytesIO",
"os.path.join",
"re.match",
"shutil.rmtree",
"logging.info"
] | [((635, 807), 'subprocess.Popen', 'subprocess.Popen', (["['surround', 'init', '-p', 'test_project', '-d', 'test_description', '-w', 'no'\n ]"], {'cwd': '"""temporary"""', 'stdout': 'subprocess.PIPE', 'stderr': 'subprocess.PIPE'}), "(['surround', 'init', '-p', 'test_project', '-d',\n 'test_description', '-w', 'no'... |
from __future__ import print_function
import os
from argparse import ArgumentParser, Namespace
from zope.interface import implementer
from quickmail.commands import ICommand
from quickmail.utils.misc import quick_mail_dir, heavy_tick, quick_mail_template_dir
@implementer(ICommand)
class ClearCommand:
def add_arg... | [
"os.path.exists",
"zope.interface.implementer",
"os.listdir",
"os.remove"
] | [((263, 284), 'zope.interface.implementer', 'implementer', (['ICommand'], {}), '(ICommand)\n', (274, 284), False, 'from zope.interface import implementer\n'), ((858, 888), 'os.path.exists', 'os.path.exists', (['quick_mail_dir'], {}), '(quick_mail_dir)\n', (872, 888), False, 'import os\n'), ((1567, 1608), 'os.remove', '... |
import os
import pytest
from bank_bot.banking_system import Database
class MockMessage(object):
def __init__(self, from_who, chat_id, message_text):
self.json = {"from": {"id": from_who}, "chat": {"id": chat_id}}
self.text = message_text
@pytest.fixture
def database():
test_file_path = "test_d... | [
"bank_bot.banking_system.Database",
"os.remove"
] | [((347, 371), 'bank_bot.banking_system.Database', 'Database', (['test_file_path'], {}), '(test_file_path)\n', (355, 371), False, 'from bank_bot.banking_system import Database\n'), ((428, 453), 'os.remove', 'os.remove', (['test_file_path'], {}), '(test_file_path)\n', (437, 453), False, 'import os\n')] |
from typing import *
from dataclasses import asdict
from transformers import BertModel
from torch.nn.utils.rnn import pad_sequence
from stud.models.ner_classifier import NERClassifier
from stud.models.polarity_classifier import PolarityClassifier
from stud.constants import LOGGER_TRAIN_LOSS, LOGGER_VALID_LOSS, LOGGER_T... | [
"stud.models.ner_classifier.NERClassifier.load_from_checkpoint",
"stud.utils.aggregate_polarities",
"torch.nn.Dropout",
"stud.utils.merge_layers",
"torch.nn.CrossEntropyLoss",
"dataclasses.asdict",
"stud.utils.vocab_tokens_startswith",
"torch.nn.LSTM",
"stud.torch_utils.batch_scatter_mean",
"stud.... | [((4402, 4417), 'torch.no_grad', 'torch.no_grad', ([], {}), '()\n', (4415, 4417), False, 'import torch\n'), ((8295, 8310), 'torch.no_grad', 'torch.no_grad', ([], {}), '()\n', (8308, 8310), False, 'import torch\n'), ((711, 786), 'stud.models.ner_classifier.NERClassifier.load_from_checkpoint', 'NERClassifier.load_from_ch... |
import argparse
import json
import os.path
import random
import re
import time
import cv2
import numpy as np
import requests
from PIL import Image
from config import config
class GermanLicensePlateImagesGenerator:
def __init__(self, output):
self.output = output
self.COUNTRY_MARKS = np.asarray([... | [
"PIL.Image.fromarray",
"random.choice",
"argparse.ArgumentParser",
"re.compile",
"requests.get",
"random.seed",
"time.sleep",
"cv2.imdecode",
"cv2.cvtColor",
"numpy.fromstring",
"random.randint"
] | [((4263, 4288), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {}), '()\n', (4286, 4288), False, 'import argparse\n'), ((819, 832), 'random.seed', 'random.seed', ([], {}), '()\n', (830, 832), False, 'import random\n'), ((1470, 1503), 'random.choice', 'random.choice', (['self.COUNTRY_MARKS'], {}), '(self.COU... |
# -*- coding:utf8 -*-
import os
from eth_account import Account
import sha3
import base64
import hashlib
from ecdsa import SigningKey, SECP256k1
DID_TYPE = ["weid"]
def create_privkey():
return os.urandom(32)
def create_ecdsa_privkey():
return SigningKey.generate(curve=SECP256k1)
def create_random_weid():
... | [
"eth_account.Account.create",
"os.urandom",
"base64.b64encode",
"ecdsa.SigningKey.generate",
"base64.b64decode",
"sha3.keccak_256",
"eth_account.Account.privateKeyToAccount",
"ecdsa.SigningKey.from_string"
] | [((199, 213), 'os.urandom', 'os.urandom', (['(32)'], {}), '(32)\n', (209, 213), False, 'import os\n'), ((254, 290), 'ecdsa.SigningKey.generate', 'SigningKey.generate', ([], {'curve': 'SECP256k1'}), '(curve=SECP256k1)\n', (273, 290), False, 'from ecdsa import SigningKey, SECP256k1\n'), ((3070, 3087), 'sha3.keccak_256', ... |
import os
from .abstract_command import AbstractCommand
from ..services.state_utils import StateUtils
from ..services.state import StateHolder
from ..services.command_handler import CommandHandler
from ..services.console_logger import ColorPrint
class Start(AbstractCommand):
command = ["start", "up"]
args = ... | [
"os.getcwd"
] | [((1590, 1601), 'os.getcwd', 'os.getcwd', ([], {}), '()\n', (1599, 1601), False, 'import os\n')] |
from netfilterqueue import NetfilterQueue
import socket
from os import system
import scapy.all as scapy
import scapy_http.http
from urllib.parse import urlparse, parse_qs
REQUEST_LOGIN = 0
REQUEST_PROTECT = 1
class NetGuardMITM:
def __init__(self, ip, queue_number=1):
self.ip = ip
self.queue_nu... | [
"urllib.parse.parse_qs",
"os.system",
"urllib.parse.urlparse",
"netfilterqueue.NetfilterQueue"
] | [((2749, 2771), 'urllib.parse.urlparse', 'urlparse', (['request.Path'], {}), '(request.Path)\n', (2757, 2771), False, 'from urllib.parse import urlparse, parse_qs\n'), ((2792, 2809), 'urllib.parse.parse_qs', 'parse_qs', (['o.query'], {}), '(o.query)\n', (2800, 2809), False, 'from urllib.parse import urlparse, parse_qs\... |
"""Config options for WDRAX.
Gets settings from environment variables or .env/settings.ini file.
"""
import pathlib
from decouple import AutoConfig
BASE_DIR = pathlib.Path(__name__).absolute().parent
config = AutoConfig(search_path=str(BASE_DIR)) # pylint: disable=invalid-name
LOG_LEVEL = config('LOG_LEVEL', defa... | [
"pathlib.Path"
] | [((163, 185), 'pathlib.Path', 'pathlib.Path', (['__name__'], {}), '(__name__)\n', (175, 185), False, 'import pathlib\n')] |
# BSD 3-Clause License
#
# Copyright (c) 2019, Elasticsearch BV
# All rights reserved.
#
# Redistribution and use in source and binary forms, with or without
# modification, are permitted provided that the following conditions are met:
#
# * Redistributions of source code must retain the above copyright notice, t... | [
"elasticapm.contrib.zerorpc.Middleware",
"os.path.join",
"pytest.mark.skipif",
"pytest.importorskip",
"tempfile.mkdtemp",
"pytest.raises",
"shutil.rmtree"
] | [((1734, 1764), 'pytest.importorskip', 'pytest.importorskip', (['"""zerorpc"""'], {}), "('zerorpc')\n", (1753, 1764), False, 'import pytest\n'), ((1774, 1803), 'pytest.importorskip', 'pytest.importorskip', (['"""gevent"""'], {}), "('gevent')\n", (1793, 1803), False, 'import pytest\n'), ((1901, 1975), 'pytest.mark.skipi... |
from bs4 import BeautifulSoup as Soup
import re
import json
import sys
import shutil
import tempfile
import os
import subprocess
from pathlib import Path
from docassemble.base.util import log, path_and_mimetype, validation_error, DADict, DAList, Individual, value, force_ask, space_to_underscore
__all__ = ['run_automa... | [
"subprocess.check_output",
"tempfile.TemporaryDirectory",
"docassemble.base.util.value",
"docassemble.base.util.validation_error",
"json.dumps",
"os.path.join",
"docassemble.base.util.force_ask",
"docassemble.base.util.path_and_mimetype",
"os.path.isfile",
"docassemble.base.util.space_to_underscor... | [((1530, 1577), 're.search', 're.search', (['"""[^A-Za-z\\\\\' 0-9\\\\_\\\\-\\\\n\\\\r]"""', 'text'], {}), '("[^A-Za-z\\\\\' 0-9\\\\_\\\\-\\\\n\\\\r]", text)\n', (1539, 1577), False, 'import re\n'), ((1781, 1811), 'docassemble.base.util.space_to_underscore', 'space_to_underscore', (['base_name'], {}), '(base_name)\n', ... |
#! /usr/bin/env python3
import sys
import argparse
from graphenetools import gt
def create_parser():
parser = argparse.ArgumentParser(description="Plot graphene lattice and C1/3 phase corresponding to printed command line arguments for uniaxially strained graphene (for use with QMC software located at https://code... | [
"graphenetools.gt.plt.style.context",
"graphenetools.gt.c_one_third_commensurate_command_plot",
"graphenetools.gt.c_one_third_commensurate_command",
"argparse.ArgumentParser"
] | [((115, 404), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {'description': '"""Plot graphene lattice and C1/3 phase corresponding to printed command line arguments for uniaxially strained graphene (for use with QMC software located at https://code.delmaestro.org)"""', 'formatter_class': 'argparse.Argument... |
import psycopg2
import re
import json
from MedTAG_sket_dock_App.models import *
import os
import pandas as pd
import numpy
from psycopg2.extensions import register_adapter, AsIs
def addapt_numpy_float64(numpy_float64):
return AsIs(numpy_float64)
def addapt_numpy_int64(numpy_int64):
return AsIs(numpy_int64)
regi... | [
"pandas.read_csv",
"psycopg2.extensions.AsIs",
"django.db.transaction.atomic",
"pandas.notnull",
"psycopg2.extensions.register_adapter"
] | [((316, 369), 'psycopg2.extensions.register_adapter', 'register_adapter', (['numpy.float64', 'addapt_numpy_float64'], {}), '(numpy.float64, addapt_numpy_float64)\n', (332, 369), False, 'from psycopg2.extensions import register_adapter, AsIs\n'), ((370, 419), 'psycopg2.extensions.register_adapter', 'register_adapter', (... |
import sys
sys.path.append("/home/ly/workspace/mmsa")
seed = 1938
import numpy as np
import torch
from torch import nn
from torch import optim
np.random.seed(seed)
torch.manual_seed(seed)
torch.cuda.manual_seed(seed)
torch.cuda.manual_seed_all(seed)
from models.bigru_rcnn_gate import *
from utils.train import *
from t... | [
"torch.cuda.manual_seed_all",
"torch.manual_seed",
"torch.nn.CrossEntropyLoss",
"numpy.random.seed",
"torch.cuda.manual_seed",
"sys.path.append"
] | [((11, 53), 'sys.path.append', 'sys.path.append', (['"""/home/ly/workspace/mmsa"""'], {}), "('/home/ly/workspace/mmsa')\n", (26, 53), False, 'import sys\n'), ((144, 164), 'numpy.random.seed', 'np.random.seed', (['seed'], {}), '(seed)\n', (158, 164), True, 'import numpy as np\n'), ((165, 188), 'torch.manual_seed', 'torc... |
#!/usr/bin/env python
# -*- coding: UTF-8 -*-
########################################################################
# GNU General Public License v3.0
# GNU GPLv3
# Copyright (c) 2019, <NAME>
#
# This program is free software: you can redistribute it and/or modify
# it under the terms of the GNU General Public Licen... | [
"numpy.array",
"platform.node"
] | [((1150, 1213), 'numpy.array', 'np.array', (["['caffe', 'tensorflow', 'pytorch', 'keras', 'caffe2']"], {}), "(['caffe', 'tensorflow', 'pytorch', 'keras', 'caffe2'])\n", (1158, 1213), True, 'import numpy as np\n'), ((1271, 1323), 'numpy.array', 'np.array', (['[2048, 2048, 1000, 1024, 1000, 2048, 2048]'], {}), '([2048, 2... |
import json
import numpy as np
import os
import tensorflow as tf
from keras.models import Sequential
from keras.layers import Dense, Dropout, Embedding, Flatten, Activation, BatchNormalization
from sklearn.model_selection import train_test_split
class SuggestionModeler(object):
"""
A collection of functions ... | [
"os.path.exists",
"random.shuffle",
"sklearn.model_selection.train_test_split",
"tensorflow.Session",
"keras.models.Sequential",
"os.chdir",
"numpy.zeros",
"keras.layers.Dense",
"keras.layers.Dropout",
"tensorflow.get_default_graph"
] | [((5754, 5768), 'os.chdir', 'os.chdir', (['""".."""'], {}), "('..')\n", (5762, 5768), False, 'import os\n'), ((490, 502), 'tensorflow.Session', 'tf.Session', ([], {}), '()\n', (500, 502), True, 'import tensorflow as tf\n'), ((524, 546), 'tensorflow.get_default_graph', 'tf.get_default_graph', ([], {}), '()\n', (544, 546... |
##################################################
#
# Tests for model.py
#
#
#
#
#
#
#
#
#
#
#
##################################################
from boole.core.model import *
from boole.core.language import clear_default_language
from nose.tools import *
def is_prime(x):
if x == 0 or x == 1:
return Fa... | [
"boole.core.language.clear_default_language"
] | [((579, 603), 'boole.core.language.clear_default_language', 'clear_default_language', ([], {}), '()\n', (601, 603), False, 'from boole.core.language import clear_default_language\n'), ((4325, 4349), 'boole.core.language.clear_default_language', 'clear_default_language', ([], {}), '()\n', (4347, 4349), False, 'from bool... |
#!/usr/bin/env python3
"""
Host the setup function.
"""
import pathlib
import setuptools
from cushead import info
def setup() -> None:
"""
Execute the setup.
"""
assets_path = pathlib.Path(info.PACKAGE_NAME) / "console" / "assets" / "images"
templates_path = pathlib.Path(info.PACKAGE_NAME) / "ge... | [
"pathlib.Path"
] | [((196, 227), 'pathlib.Path', 'pathlib.Path', (['info.PACKAGE_NAME'], {}), '(info.PACKAGE_NAME)\n', (208, 227), False, 'import pathlib\n'), ((283, 314), 'pathlib.Path', 'pathlib.Path', (['info.PACKAGE_NAME'], {}), '(info.PACKAGE_NAME)\n', (295, 314), False, 'import pathlib\n'), ((1480, 1505), 'pathlib.Path', 'pathlib.P... |
import os, sys
import re
import zipfile
import requests
import warnings
import logging
import pandas as pd
import numpy as np
from stat import S_IREAD, S_IRGRP, S_IROTH
import getpass
import pymysql
# Code by <NAME> (<EMAIL>), 2016-2017
class LoadData():
"""
This class is inherited by the Data class, and cont... | [
"os.path.exists",
"os.listdir",
"zipfile.ZipFile",
"re.compile",
"os.path.join",
"pymysql.connect",
"requests.get",
"getpass.getpass",
"os.chmod",
"pandas.to_numeric",
"sys.stdin.isatty",
"pandas.concat"
] | [((1685, 1757), 're.compile', 're.compile', (['"""(\\\\d{4})\\\\s*=\\\\s*(https?:\\\\/\\\\/.+\\\\.(dat|zip|csv|txt))\\\\s*"""'], {}), "('(\\\\d{4})\\\\s*=\\\\s*(https?:\\\\/\\\\/.+\\\\.(dat|zip|csv|txt))\\\\s*')\n", (1695, 1757), False, 'import re\n'), ((1771, 1787), 're.compile', 're.compile', (['"""^#"""'], {}), "('^... |
# Copyright (c) 2021 <NAME>(<EMAIL>)
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in w... | [
"argparse.ArgumentParser",
"torch.jit.load",
"onnxruntime.InferenceSession",
"yaml.load",
"torch.tensor",
"torch.randn",
"torch.onnx.export"
] | [((683, 742), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {'description': '"""export to onnx model"""'}), "(description='export to onnx model')\n", (706, 742), False, 'import argparse\n'), ((1318, 1348), 'torch.jit.load', 'torch.jit.load', (['args.jit_model'], {}), '(args.jit_model)\n', (1332, 1348), Fal... |
from maya import cmds
import copy
# TODO: Find a way to have different naming for different production.
# Maybe handle it in the rig directly?
class BaseName(object):
"""
This class handle the naming of object.
Store a name as a list of 'tokens'
When resolved, the tokens are joinned using a 'separator... | [
"copy.copy",
"maya.cmds.objExists"
] | [((2244, 2266), 'copy.copy', 'copy.copy', (['self.tokens'], {}), '(self.tokens)\n', (2253, 2266), False, 'import copy\n'), ((4201, 4221), 'maya.cmds.objExists', 'cmds.objExists', (['name'], {}), '(name)\n', (4215, 4221), False, 'from maya import cmds\n')] |
# Copyright 2022 Tiernan8r
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, so... | [
"qcp.matrices.DefaultMatrix"
] | [((1301, 1323), 'qcp.matrices.DefaultMatrix', 'DefaultMatrix', (['entries'], {}), '(entries)\n', (1314, 1323), False, 'from qcp.matrices import SPARSE, DefaultMatrix, Matrix\n')] |
import math
import numpy as np
import torch
from mock import patch
from core.train_utils import compute_angular_error, compute_angular_error_xyz_arr, spherical2cartesial
def test_spherical2cartesial():
spherical = torch.Tensor([
[0, 0],
[math.pi / 2, 0],
[-math.pi / 2, 0],
[0, ma... | [
"core.train_utils.compute_angular_error",
"mock.patch",
"torch.mean",
"core.train_utils.compute_angular_error_xyz_arr",
"torch.Tensor",
"math.sqrt",
"numpy.array",
"core.train_utils.spherical2cartesial"
] | [((740, 776), 'mock.patch', 'patch', (['"""core.train_utils.epsilon"""', '(0)'], {}), "('core.train_utils.epsilon', 0)\n", (745, 776), False, 'from mock import patch\n'), ((1504, 1540), 'mock.patch', 'patch', (['"""core.train_utils.epsilon"""', '(0)'], {}), "('core.train_utils.epsilon', 0)\n", (1509, 1540), False, 'fro... |
import csv
import re
from os import makedirs
from os.path import abspath, basename, dirname, isdir, join
def generate_csv(path, fields, rows, quote_empty=False):
path = abspath(path)
name = basename(path)
name = re.sub('py$', 'csv', name)
cases = join(dirname(dirname(path)), 'cases')
if not isdir(... | [
"csv.DictWriter",
"os.makedirs",
"csv.writer",
"os.path.join",
"os.path.dirname",
"os.path.isdir",
"os.path.basename",
"os.path.abspath",
"re.sub"
] | [((175, 188), 'os.path.abspath', 'abspath', (['path'], {}), '(path)\n', (182, 188), False, 'from os.path import abspath, basename, dirname, isdir, join\n'), ((200, 214), 'os.path.basename', 'basename', (['path'], {}), '(path)\n', (208, 214), False, 'from os.path import abspath, basename, dirname, isdir, join\n'), ((226... |
from datetime import datetime
import traceback
import boto3
from botocore.exceptions import ClientError
from ...config import config
from ...log import log
class Submitter():
def __init__(self, event):
self.event = event
def find_instance(self, instance_id, mac_address): # pylint: disable=R0201
... | [
"traceback.format_exc",
"boto3.resource",
"boto3.client",
"datetime.datetime.now"
] | [((511, 530), 'boto3.client', 'boto3.client', (['"""ec2"""'], {}), "('ec2')\n", (523, 530), False, 'import boto3\n'), ((676, 717), 'boto3.resource', 'boto3.resource', (['"""ec2"""'], {'region_name': 'region'}), "('ec2', region_name=region)\n", (690, 717), False, 'import boto3\n'), ((1420, 1442), 'traceback.format_exc',... |
import pytest
from buyer import serializers
@pytest.mark.django_db
def test_buyer_deserialization():
data = {
'email': '<EMAIL>',
'name': '<NAME>',
'sector': 'AEROSPACE',
'company_name': 'Example corp',
'country': 'China',
}
serializer = serializers.BuyerSerialize... | [
"buyer.serializers.BuyerSerializer"
] | [((294, 332), 'buyer.serializers.BuyerSerializer', 'serializers.BuyerSerializer', ([], {'data': 'data'}), '(data=data)\n', (321, 332), False, 'from buyer import serializers\n')] |
#1d_tests.py
import autograd.numpy as np
import autograd.scipy.stats.norm as norm
from copy import copy
import datetime as dt
from matplotlib import pyplot as plt
import seaborn as sns
import sys; sys.path.append('..')
from mcmc import langevin, MALA, RK_langevin, RWMH, HMC
def bimodal_logprob(z):
return (np.log(n... | [
"autograd.numpy.sin",
"seaborn.distplot",
"autograd.numpy.linspace",
"autograd.scipy.stats.norm.logpdf",
"datetime.datetime.now",
"autograd.numpy.random.randn",
"copy.copy",
"sys.path.append",
"matplotlib.pyplot.show"
] | [((200, 221), 'sys.path.append', 'sys.path.append', (['""".."""'], {}), "('..')\n", (215, 221), False, 'import sys\n'), ((467, 496), 'autograd.numpy.random.randn', 'np.random.randn', (['n_samples', '(1)'], {}), '(n_samples, 1)\n', (482, 496), True, 'import autograd.numpy as np\n'), ((570, 587), 'datetime.datetime.now',... |
import qrcode
# data example
data = "www.google.com"
# file name
file_name = "qrcode.png"
# generate qr code
img = qrcode.make(data=data)
# save generated qr code as img
img.save(file_name)
| [
"qrcode.make"
] | [((118, 140), 'qrcode.make', 'qrcode.make', ([], {'data': 'data'}), '(data=data)\n', (129, 140), False, 'import qrcode\n')] |
"""
Low level miscilanious calls
"""
import UserDict
import types
import random
import datetime
import pprint
import re
import unicodedata
import logging
log = logging.getLogger(__name__)
now_override = None
def now():
"""
A passthough to get now()
We can override this so that automated tests can fake ... | [
"logging.getLogger",
"datetime.datetime.now",
"datetime.date.today",
"datetime.date"
] | [((163, 190), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (180, 190), False, 'import logging\n'), ((409, 432), 'datetime.datetime.now', 'datetime.datetime.now', ([], {}), '()\n', (430, 432), False, 'import datetime\n'), ((3686, 3707), 'datetime.date.today', 'datetime.date.today', ([], ... |
import os
from pydantic import BaseSettings
class AppSettings(BaseSettings):
debug: bool = False
time_zone: str = "Asia/Shanghai"
logger_level: str = "INFO"
logger_formatter: str = "%(asctime)s [%(name)s] %(funcName)s[line:%(lineno)d] %(levelname)-7s: %(message)s"
secret_key: str = "1@3$5^7*9)"
... | [
"os.getenv"
] | [((1078, 1102), 'os.getenv', 'os.getenv', (['"""FASTAPI_ENV"""'], {}), "('FASTAPI_ENV')\n", (1087, 1102), False, 'import os\n'), ((374, 402), 'os.getenv', 'os.getenv', (['"""POSTGRESQL_PORT"""'], {}), "('POSTGRESQL_PORT')\n", (383, 402), False, 'import os\n'), ((432, 459), 'os.getenv', 'os.getenv', (['"""REDIS_PASSWORD... |
# coding: utf-8
"""
Layered Insight Assessment, Compliance, Witness & Control
LI Assessment & Compliance performs static vulnerability analysis, license and package compliance. LI Witness provides deep insight and analytics into containerized applications. Control provides dynamic runtime security and analyti... | [
"six.iteritems"
] | [((3539, 3568), 'six.iteritems', 'iteritems', (['self.swagger_types'], {}), '(self.swagger_types)\n', (3548, 3568), False, 'from six import iteritems\n')] |
'''
Different utils used by plotters.
'''
import mpl_toolkits.basemap as bm
import matplotlib.pyplot as pl
from matplotlib import colors, cm
import scipy as sp
# Make spectral plot
def my_psd(x):
P,f=pl.psd(x); pl.clf()
T=2./f; ind= T>=12.
pl.plot(T[ind]/12,P[ind]);
ax=pl.gca();
ax.set_xscale('lo... | [
"matplotlib.colors.normalize",
"matplotlib.pyplot.gca",
"matplotlib.pyplot.psd",
"matplotlib.pyplot.clf",
"matplotlib.pyplot.plot",
"scipy.ma.maximum",
"scipy.ma.minimum",
"matplotlib.cm.get_cmap"
] | [((206, 215), 'matplotlib.pyplot.psd', 'pl.psd', (['x'], {}), '(x)\n', (212, 215), True, 'import matplotlib.pyplot as pl\n'), ((217, 225), 'matplotlib.pyplot.clf', 'pl.clf', ([], {}), '()\n', (223, 225), True, 'import matplotlib.pyplot as pl\n'), ((254, 282), 'matplotlib.pyplot.plot', 'pl.plot', (['(T[ind] / 12)', 'P[i... |
#!/usr/bin/env python
"""
Lyrebird Voice Changer
Simple and powerful voice changer for Linux, written in GTK 3
(c) Charlotte 2020
"""
import sys
import re
from setuptools import setup
version_regex = r'__version__ = ["\']([^"\']*)["\']'
with open('app/__init__.py', 'r') as f:
text = f.read()
match = re.search... | [
"setuptools.setup",
"re.search"
] | [((1102, 1124), 'setuptools.setup', 'setup', ([], {}), '(**extra_options)\n', (1107, 1124), False, 'from setuptools import setup\n'), ((311, 341), 're.search', 're.search', (['version_regex', 'text'], {}), '(version_regex, text)\n', (320, 341), False, 'import re\n')] |
from django.db import models
from django.utils.translation import ugettext_lazy as _
from helusers.models import AbstractUser
class User(AbstractUser):
is_official = models.BooleanField(verbose_name=_("official"), default=False)
class Meta:
verbose_name = _("user")
verbose_name_plural = _("us... | [
"django.utils.translation.ugettext_lazy"
] | [((275, 284), 'django.utils.translation.ugettext_lazy', '_', (['"""user"""'], {}), "('user')\n", (276, 284), True, 'from django.utils.translation import ugettext_lazy as _\n'), ((315, 325), 'django.utils.translation.ugettext_lazy', '_', (['"""users"""'], {}), "('users')\n", (316, 325), True, 'from django.utils.translat... |
#!/usr/bin/env python
# coding: utf-8
# In[1]:
import numpy as np
import pandas as pd
import glob,os,csv,re,math
import shutil, time
from astropy.io import ascii
import matplotlib.pyplot as plt
# Load all data files:
psdir ='/Users/maryumsayeed/Desktop/HuberNess/mlearning/powerspectrum/'
hrdir ='/Users/maryu... | [
"pandas.read_csv",
"numpy.where",
"math.isnan",
"numpy.array",
"numpy.concatenate",
"numpy.loadtxt",
"astropy.io.ascii.read",
"glob.glob",
"re.search"
] | [((790, 821), 'glob.glob', 'glob.glob', (["(pande_dir + '*.fits')"], {}), "(pande_dir + '*.fits')\n", (799, 821), False, 'import glob, os, csv, re, math\n'), ((835, 864), 'glob.glob', 'glob.glob', (["(ast_dir + '*.fits')"], {}), "(ast_dir + '*.fits')\n", (844, 864), False, 'import glob, os, csv, re, math\n'), ((1184, 1... |
#!/usr/bin/python
# -*- coding: utf-8 -*-
# thumbor imaging service
# https://github.com/TechLaProvence/lp_mongodb
# Licensed under the MIT license:
# http://www.opensource.org/licenses/mit-license
# Copyright (c) 2011 figarocms <EMAIL>
from tornado.concurrent import return_future
from lp_mongodb.storages.mongo_sto... | [
"lp_mongodb.storages.mongo_storage.Storage"
] | [((473, 489), 'lp_mongodb.storages.mongo_storage.Storage', 'Storage', (['context'], {}), '(context)\n', (480, 489), False, 'from lp_mongodb.storages.mongo_storage import Storage\n')] |
import numpy as np
from tabulate import tabulate
import matplotlib.pyplot as plt
import Page_Rank_Utils as pru
def detectedConverged(y,x,epsilon):
C = set()
N = set()
for i in range(len(y)):
if abs(y[i] - x[i])/abs(x[i]) < epsilon:
C.add(i)
else:
N.add(i)
return... | [
"Page_Rank_Utils.stochastic_transition_matrix",
"numpy.zeros"
] | [((591, 640), 'Page_Rank_Utils.stochastic_transition_matrix', 'pru.stochastic_transition_matrix', (['G', 'weight', '(True)'], {}), '(G, weight, True)\n', (623, 640), True, 'import Page_Rank_Utils as pru\n'), ((733, 744), 'numpy.zeros', 'np.zeros', (['n'], {}), '(n)\n', (741, 744), True, 'import numpy as np\n')] |
from troyka_led_matrix import TroykaLedMatrix
from urandom import getrandbits
import time
matrix = TroykaLedMatrix()
while True:
matrix.draw_pixel(getrandbits(3), getrandbits(3))
matrix.clear_pixel(getrandbits(3), getrandbits(3))
time.sleep_ms(50)
| [
"time.sleep_ms",
"urandom.getrandbits",
"troyka_led_matrix.TroykaLedMatrix"
] | [((104, 121), 'troyka_led_matrix.TroykaLedMatrix', 'TroykaLedMatrix', ([], {}), '()\n', (119, 121), False, 'from troyka_led_matrix import TroykaLedMatrix\n'), ((253, 270), 'time.sleep_ms', 'time.sleep_ms', (['(50)'], {}), '(50)\n', (266, 270), False, 'import time\n'), ((160, 174), 'urandom.getrandbits', 'getrandbits', ... |
import math
import itertools
class Vector:
"""
Generic vector operations.
"""
def _apply(self,op, other):
pairwise = None
if type(other) is Vector:
pairwise = zip(self.vals, other.vals)
else:
pairwise = zip(self.vals, [other for _ ... | [
"math.pow",
"math.sqrt",
"itertools.starmap"
] | [((907, 923), 'math.sqrt', 'math.sqrt', (['total'], {}), '(total)\n', (916, 923), False, 'import math\n'), ((359, 390), 'itertools.starmap', 'itertools.starmap', (['op', 'pairwise'], {}), '(op, pairwise)\n', (376, 390), False, 'import itertools\n'), ((863, 877), 'math.pow', 'math.pow', (['a', '(2)'], {}), '(a, 2)\n', (... |
import numpy as np
import os
os.environ['TF_FORCE_GPU_ALLOW_GROWTH']='true'
import datetime
from sacred import Experiment
from sacred.observers import FileStorageObserver
from datasets.ucf101 import UCF101Dataset
from algorithms.kerasvideoonenet import KerasVideoOneNet
from algorithms.kerasvideoonenet_admm import Keras... | [
"datasets.ucf101.UCF101Dataset",
"sacred.observers.FileStorageObserver.create",
"algorithms.numpyvideowaveletsparsity_admm.NumpyVideoWaveletSparsityADMM",
"sacred.Experiment",
"algorithms.kerasvideoonenet_admm.KerasVideoOneNetADMM",
"datetime.datetime.now",
"os.path.basename",
"os.getpid",
"algorith... | [((475, 491), 'sacred.Experiment', 'Experiment', (['name'], {}), '(name)\n', (485, 491), False, 'from sacred import Experiment\n'), ((497, 520), 'datetime.datetime.now', 'datetime.datetime.now', ([], {}), '()\n', (518, 520), False, 'import datetime\n'), ((752, 791), 'sacred.observers.FileStorageObserver.create', 'FileS... |
#tiersweekly.py
from fantasyfootball import tiers
from fantasyfootball import fantasypros as fp
from fantasyfootball import config
from fantasyfootball import ffcalculator
from fantasyfootball.config import FIGURE_DIR
from sklearn.cluster import KMeans
from sklearn.mixture import GaussianMixture
from matplotlib import... | [
"matplotlib.pyplot.ylabel",
"fantasyfootball.tiers.make_clustering_viz",
"numpy.array_split",
"matplotlib.pyplot.annotate",
"matplotlib.style.use",
"matplotlib.lines.Line2D",
"collections.OrderedDict.fromkeys",
"matplotlib.pyplot.xlabel",
"matplotlib.pyplot.scatter",
"sklearn.mixture.GaussianMixtu... | [((2721, 2758), 'fantasyfootball.fantasypros.create_fantasy_pros_ecr_df', 'fp.create_fantasy_pros_ecr_df', (['league'], {}), '(league)\n', (2750, 2758), True, 'from fantasyfootball import fantasypros as fp\n'), ((3148, 3160), 'datetime.date.today', 'date.today', ([], {}), '()\n', (3158, 3160), False, 'from datetime imp... |
"""Pauses the execution."""
import time
from dodo_commands.framework.decorator_utils import uses_decorator
class Decorator:
def is_used(self, config, command_name, decorator_name):
return uses_decorator(config, command_name, decorator_name)
def add_arguments(self, parser): # override
parser... | [
"dodo_commands.framework.decorator_utils.uses_decorator",
"time.sleep"
] | [((203, 255), 'dodo_commands.framework.decorator_utils.uses_decorator', 'uses_decorator', (['config', 'command_name', 'decorator_name'], {}), '(config, command_name, decorator_name)\n', (217, 255), False, 'from dodo_commands.framework.decorator_utils import uses_decorator\n'), ((579, 624), 'time.sleep', 'time.sleep', (... |
import logging
from typing import List
from .AnaplanRequest import AnaplanRequest
from .User import User
from .ModelDetails import ModelDetails
logger = logging.getLogger(__name__)
class Model(User):
def get_models_url(self) -> AnaplanRequest:
"""Get list of all Anaplan model for the specified user.
:return: O... | [
"logging.getLogger"
] | [((154, 181), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (171, 181), False, 'import logging\n')] |
# Hangman game
#
# -----------------------------------
# Helper code
# You don't need to understand this helper code,
# but you will have to know how to use the functions
# (so be sure to read the docstrings!)
import random
import string
WORDLIST_FILENAME = "/Users/deedeebanh/Documents/MITx_6.00.1.x/ProblemSet3/word... | [
"random.choice"
] | [((951, 974), 'random.choice', 'random.choice', (['wordlist'], {}), '(wordlist)\n', (964, 974), False, 'import random\n')] |
#!/usr/bin/python3
import argparse
from datetime import datetime
import json
from PyQt5.QtGui import *
from PyQt5.QtWidgets import *
from PyQt5 import uic
import os
import sys
from weather import Weather
BASE_PATH = os.path.dirname(os.path.abspath(__file__))
DISP_SIZE = (640, 384)
WHITE = 0xffffffff
BLACK = 0xff000... | [
"argparse.ArgumentParser",
"PyQt5.uic.loadUi",
"json.load",
"datetime.datetime.now",
"weather.Weather",
"epd7in5.EPD",
"os.path.abspath"
] | [((234, 259), 'os.path.abspath', 'os.path.abspath', (['__file__'], {}), '(__file__)\n', (249, 259), False, 'import os\n'), ((413, 438), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {}), '()\n', (436, 438), False, 'import argparse\n'), ((1677, 1755), 'weather.Weather', 'Weather', ([], {'api_key': "config['... |
from sklearn.metrics import r2_score
y_true = [3, -0.5, 2, 7]
y_pred = [2.5, 0.0, 2, 8]
r2=r2_score(y_true, y_pred)
print(r2)
y_true = [5,6,7,8]
y_pred = [-100,524,-1,3]
r2=r2_score(y_true, y_pred)
print(r2)
r2_ | [
"sklearn.metrics.r2_score"
] | [((94, 118), 'sklearn.metrics.r2_score', 'r2_score', (['y_true', 'y_pred'], {}), '(y_true, y_pred)\n', (102, 118), False, 'from sklearn.metrics import r2_score\n'), ((177, 201), 'sklearn.metrics.r2_score', 'r2_score', (['y_true', 'y_pred'], {}), '(y_true, y_pred)\n', (185, 201), False, 'from sklearn.metrics import r2_s... |
import pytest
from ssz.sedes import Bitvector
def test_bitvector_instantiation_bound():
with pytest.raises(ValueError):
bit_count = 0
Bitvector(bit_count)
| [
"ssz.sedes.Bitvector",
"pytest.raises"
] | [((100, 125), 'pytest.raises', 'pytest.raises', (['ValueError'], {}), '(ValueError)\n', (113, 125), False, 'import pytest\n'), ((157, 177), 'ssz.sedes.Bitvector', 'Bitvector', (['bit_count'], {}), '(bit_count)\n', (166, 177), False, 'from ssz.sedes import Bitvector\n')] |
# Seguridad Informatica
# ejercicio de encriptacion
# <NAME> (42487)
from Crypto.Cipher import AES
from Crypto.Random import get_random_bytes
with open('key.bin', 'rb') as k:
key = k.read()
with open('vector.bin', 'rb') as v:
init_vector = v.read()
cipher = AES.new(key, AES.MODE_CBC, init_ve... | [
"Crypto.Cipher.AES.new"
] | [((285, 324), 'Crypto.Cipher.AES.new', 'AES.new', (['key', 'AES.MODE_CBC', 'init_vector'], {}), '(key, AES.MODE_CBC, init_vector)\n', (292, 324), False, 'from Crypto.Cipher import AES\n')] |
from setuptools import setup
from setuptools import find_packages
required_packages = [
'beautifulsoup4',
'cssselect',
'duckling',
'feedfinder2',
'feedparser',
'idna',
'jieba3k',
'JPype1',
'Logbook',
'lxml',
'newspaper3k',
'nltk',
'Pillow',
'PyQt5',
'python-d... | [
"setuptools.find_packages"
] | [((725, 740), 'setuptools.find_packages', 'find_packages', ([], {}), '()\n', (738, 740), False, 'from setuptools import find_packages\n')] |
from base64 import b64encode, b64decode
import binascii
from datetime import datetime
from uuid import UUID
from jsonschema._format import FormatChecker
from jsonschema.exceptions import FormatError
from six import binary_type, text_type, integer_types
DATETIME_HAS_STRICT_RFC3339 = False
DATETIME_HAS_ISODATE = False
... | [
"uuid.UUID",
"datetime.datetime.strptime",
"base64.b64decode",
"jsonschema.exceptions.FormatError",
"strict_rfc3339.validate_rfc3339",
"isodate.parse_datetime",
"six.text_type"
] | [((2494, 2533), 'datetime.datetime.strptime', 'datetime.strptime', (['instance', '"""%Y-%m-%d"""'], {}), "(instance, '%Y-%m-%d')\n", (2511, 2533), False, 'from datetime import datetime\n'), ((2134, 2175), 'strict_rfc3339.validate_rfc3339', 'strict_rfc3339.validate_rfc3339', (['instance'], {}), '(instance)\n', (2165, 21... |
#!/usr/bin/python3
import argparse
import socket
import base64
import binascii
# 'argparse' is a very useful library for building python tools that are easy
# to use from the command line. It greatly simplifies the input validation
# and "usage" prompts which really help when trying to debug your own code.
# parser =... | [
"socket.socket",
"socket.makefile",
"base64.b64encode",
"base64.b64decode",
"socket.connect"
] | [((680, 729), 'socket.socket', 'socket.socket', (['socket.AF_INET', 'socket.SOCK_STREAM'], {}), '(socket.AF_INET, socket.SOCK_STREAM)\n', (693, 729), False, 'import socket\n'), ((826, 852), 'socket.connect', 'socket.connect', (['(ip, port)'], {}), '((ip, port))\n', (840, 852), False, 'import socket\n'), ((992, 1009), '... |
import itertools
import os
import shutil
import numpy as np
import gym
from gym import spaces
import robosuite
from robosuite.controllers import load_controller_config
import robosuite.utils.macros as macros
import imageio, tqdm
from her import HERReplayBuffer
from tianshou.data import Batch
macros.SIMULATION_TIMES... | [
"robosuite.controllers.load_controller_config",
"numpy.full_like",
"os.makedirs",
"imageio.imwrite",
"gym.spaces.Box",
"numpy.array",
"numpy.random.seed",
"numpy.concatenate",
"shutil.rmtree",
"numpy.set_printoptions"
] | [((331, 365), 'numpy.set_printoptions', 'np.set_printoptions', ([], {'suppress': '(True)'}), '(suppress=True)\n', (350, 365), True, 'import numpy as np\n'), ((4148, 4171), 'shutil.rmtree', 'shutil.rmtree', (['"""render"""'], {}), "('render')\n", (4161, 4171), False, 'import shutil\n'), ((4176, 4197), 'os.makedirs', 'os... |
import unittest
import acpc_python_client as acpc
from tools.constants import Action
from weak_agents.action_tilted_agent import create_agent_strategy, create_agent_strategy_from_trained_strategy, TiltType
from tools.io_util import read_strategy_from_file
from evaluation.exploitability import Exploitability
from tool... | [
"tools.io_util.read_strategy_from_file",
"weak_agents.action_tilted_agent.create_agent_strategy",
"tools.game_utils.is_correct_strategy",
"acpc_python_client.read_game_file",
"tools.game_utils.is_strategies_equal",
"evaluation.exploitability.Exploitability",
"weak_agents.action_tilted_agent.create_agent... | [((616, 761), 'weak_agents.action_tilted_agent.create_agent_strategy', 'create_agent_strategy', (['KUHN_POKER_GAME_FILE_PATH', 'Action.RAISE', 'TiltType.ADD', '(0.2)'], {'cfr_iterations': '(20)', 'cfr_weight_delay': '(2)', 'show_progress': '(False)'}), '(KUHN_POKER_GAME_FILE_PATH, Action.RAISE, TiltType.ADD,\n 0.2, ... |
import functools
import operator
from collections.abc import Iterable
from typing import overload, Union, TypeVar
T = TypeVar('T')
S = TypeVar('S') # <1>
@overload
def sum(it: Iterable[T]) -> Union[T, int]: ... # <2>
@overload
def sum(it: Iterable[T], /, start: S) -> Union[T, S]: ... # <3>
def sum(it, /, start=0):... | [
"functools.reduce",
"typing.TypeVar"
] | [((119, 131), 'typing.TypeVar', 'TypeVar', (['"""T"""'], {}), "('T')\n", (126, 131), False, 'from typing import overload, Union, TypeVar\n'), ((136, 148), 'typing.TypeVar', 'TypeVar', (['"""S"""'], {}), "('S')\n", (143, 148), False, 'from typing import overload, Union, TypeVar\n'), ((339, 380), 'functools.reduce', 'fun... |
from __future__ import annotations
import math
from collections import deque
from typing import Optional, Callable
import numpy as np
import pygame
from chess.const import PieceType, PieceColour, Piece, CastlingType, Move, \
PIECE_INDICES, init_zobrist, MoveFlags, GameState
from chess.utils import load_image, lo... | [
"chess.utils.load_image",
"collections.deque",
"numpy.where",
"pygame.sprite.Group",
"chess.const.Move",
"chess.const.Piece.empty",
"math.copysign",
"chess.const.MoveFlags",
"pygame.draw.rect",
"pygame.Color",
"chess.utils.load_font",
"chess.const.Piece",
"chess.const.init_zobrist"
] | [((1993, 2007), 'chess.const.init_zobrist', 'init_zobrist', ([], {}), '()\n', (2005, 2007), False, 'from chess.const import PieceType, PieceColour, Piece, CastlingType, Move, PIECE_INDICES, init_zobrist, MoveFlags, GameState\n'), ((2064, 2090), 'pygame.Color', 'pygame.Color', (['light_colour'], {}), '(light_colour)\n',... |
from conans import ConanFile, tools
import os
class MsysBaseInstallerConan(ConanFile):
name = "msys-base_installer"
version = "2013072300"
license = "http://www.mingw.org/license"
url = "http://github.com/danimtb/conan-msys-installer"
settings = "os", "compiler"
build_policy = "missing"
de... | [
"os.path.join",
"conans.tools.download",
"os.unlink"
] | [((4018, 4061), 'conans.tools.download', 'tools.download', (['files[util_name]', 'util_name'], {}), '(files[util_name], util_name)\n', (4032, 4061), False, 'from conans import ConanFile, tools\n'), ((4163, 4183), 'os.unlink', 'os.unlink', (['util_name'], {}), '(util_name)\n', (4172, 4183), False, 'import os\n'), ((4196... |
"""Tests for 'cloudflare-gh-pages-dns' hook."""
import contextlib
import io
import os
import pytest
from hooks.cf_gh_pages_dns_records import check_cloudflare_gh_pages_dns_records
@pytest.mark.skipif(
not os.environ.get("CF_API_KEY"),
reason=(
"Cloudflare user API key defined in 'CF_API_KEY' enviro... | [
"os.environ.get",
"pytest.param",
"pytest.mark.parametrize",
"contextlib.redirect_stderr",
"io.StringIO",
"hooks.cf_gh_pages_dns_records.check_cloudflare_gh_pages_dns_records"
] | [((365, 451), 'pytest.mark.parametrize', 'pytest.mark.parametrize', (['"""quiet"""', '(True, False)'], {'ids': "('quiet=True', 'quiet=False')"}), "('quiet', (True, False), ids=('quiet=True',\n 'quiet=False'))\n", (388, 451), False, 'import pytest\n'), ((1275, 1288), 'io.StringIO', 'io.StringIO', ([], {}), '()\n', (1... |
import os
os.environ['PREFECT__LOGGING__LEVEL'] = 'DEBUG'
os.environ['DJANGO_ALLOW_ASYNC_UNSAFE'] = 'true'
from prefect import flow, task
import numpy as np
import pandas as pd
from django_pandas.io import read_frame
import helpers
@task
def insert_session(session_id):
from django_connect import connect
con... | [
"helpers.spike_count",
"db.models.TrialSpikeCounts",
"db.models.StimulusPresentation.objects.filter",
"prefect.flow",
"db.models.StimulusPresentation",
"db.models.UnitSpikeTimes.objects.filter",
"django_pandas.io.read_frame",
"django_connect.connect",
"helpers.get_session",
"db.models.Unit.objects... | [((3224, 3243), 'prefect.flow', 'flow', ([], {'name': '"""spikes"""'}), "(name='spikes')\n", (3228, 3243), False, 'from prefect import flow, task\n'), ((317, 326), 'django_connect.connect', 'connect', ([], {}), '()\n', (324, 326), False, 'from django_connect import connect\n'), ((368, 399), 'helpers.get_session', 'help... |
from sqlalchemy import Column, ForeignKey, Integer, DateTime
from sqlalchemy.orm import relationship
from ..database import Base
class Artist_album(Base):
__tablename__ = "artists_albums"
id = Column(Integer, primary_key=True, index=True)
artist_id = Column(Integer, ForeignKey("artists.id"))
album_id... | [
"sqlalchemy.orm.relationship",
"sqlalchemy.ForeignKey",
"sqlalchemy.Column"
] | [((204, 249), 'sqlalchemy.Column', 'Column', (['Integer'], {'primary_key': '(True)', 'index': '(True)'}), '(Integer, primary_key=True, index=True)\n', (210, 249), False, 'from sqlalchemy import Column, ForeignKey, Integer, DateTime\n'), ((381, 413), 'sqlalchemy.Column', 'Column', (['DateTime'], {'nullable': '(False)'})... |
from flask import request
from firebase_admin import auth
import requests
from join import firebase
import ast
import time
def get_join():
token = request.headers['authorization']
decoded_token = auth.verify_id_token(token)
email = decoded_token['firebase']['identities']['email'][0]
tags... | [
"ast.literal_eval",
"firebase_admin.auth.verify_id_token",
"requests.get"
] | [((207, 234), 'firebase_admin.auth.verify_id_token', 'auth.verify_id_token', (['token'], {}), '(token)\n', (227, 234), False, 'from firebase_admin import auth\n'), ((332, 431), 'requests.get', 'requests.get', (['"""http://tags:7000/api/users/tags"""'], {'headers': "{'authorization': token}", 'verify': '(False)'}), "('h... |
#!/usr/bin/env python3.8
# coding=utf-8
"""
Simple Lambda Handler
"""
from lbz.dev.server import MyDevServer
from lbz.dev.test import Client
from lbz.exceptions import LambdaFWException
from lbz.resource import Resource
from lbz.response import Response
from lbz.router import add_route
class HelloWorld(Resource):
... | [
"lbz.exceptions.LambdaFWException",
"lbz.router.add_route",
"lbz.response.Response",
"lbz.dev.test.Client",
"lbz.dev.server.MyDevServer"
] | [((322, 350), 'lbz.router.add_route', 'add_route', (['"""/"""'], {'method': '"""GET"""'}), "('/', method='GET')\n", (331, 350), False, 'from lbz.router import add_route\n'), ((1053, 1092), 'lbz.dev.server.MyDevServer', 'MyDevServer', ([], {'acls': 'HelloWorld', 'port': '(8001)'}), '(acls=HelloWorld, port=8001)\n', (106... |
# -*- coding: utf-8 -*-
"""
Created on Thu May 3 08:04:22 2018
@author: af5u13
"""
import numpy as np
import os
from .geometry_functions import deg2rad, sph2cart
from loudspeakerconfig import createArrayConfigFile
def createArrayConfigFromSofa( sofaFile, xmlFile = None, lspLabels = None, twoDSetup = False, virtua... | [
"os.path.exists",
"numpy.asarray",
"h5py.File",
"loudspeakerconfig.createArrayConfigFile",
"os.path.basename"
] | [((1586, 1605), 'h5py.File', 'h5py.File', (['sofaFile'], {}), '(sofaFile)\n', (1595, 1605), False, 'import h5py\n'), ((2157, 2324), 'loudspeakerconfig.createArrayConfigFile', 'createArrayConfigFile', (['xmlFile', 'posCart.T'], {'loudspeakerLabels': 'lspLabels', 'twoDconfig': 'twoDSetup', 'sphericalPositions': '(True)',... |
# from unittest import skip
from django.conf import settings
from django.contrib import auth
from django.urls import reverse
from django.test import TestCase
from plok.models import Blog, Article
from .ext_test_case import ExtTestCase
class BlogList(TestCase):
url_name = 'plok:blog_list'
def test_reverse_blo... | [
"plok.models.Blog.objects.create",
"django.contrib.auth.get_user_model",
"plok.models.Article.objects.create",
"plok.models.Article.objects.all",
"plok.models.Blog.objects.all",
"django.urls.reverse"
] | [((688, 773), 'plok.models.Blog.objects.create', 'Blog.objects.create', ([], {'created_by': 'creator', 'name': '"""test_blog_1"""', 'title': '"""Test blog 1"""'}), "(created_by=creator, name='test_blog_1', title='Test blog 1'\n )\n", (707, 773), False, 'from plok.models import Blog, Article\n'), ((785, 870), 'plok.m... |
# -*- coding: utf-8 -*-
"""
Created on 2014-5-14
django 帮助函数
@author: skycrab
@sns_userinfo
def oauth(request):
openid = request.openid
"""
import json
import logging
import base64
from functools import wraps
from django.conf import settings
from django.core.cache import cache
from django.short... | [
"logging.getLogger",
"django.contrib.auth.authenticate",
"json.loads",
"core.redis_number.RedisStat",
"urllib.unquote",
"wx.WeixinHelper.proxy",
"django.contrib.auth.login",
"functools.wraps",
"django.core.urlresolvers.reverse",
"django.shortcuts.redirect",
"time.time"
] | [((825, 853), 'logging.getLogger', 'logging.getLogger', (['"""control"""'], {}), "('control')\n", (842, 853), False, 'import logging\n'), ((1287, 1298), 'functools.wraps', 'wraps', (['func'], {}), '(func)\n', (1292, 1298), False, 'from functools import wraps\n'), ((4835, 4846), 'functools.wraps', 'wraps', (['func'], {}... |
# Copyright (c) 2021 Project CHIP Authors
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in ... | [
"os.path.join",
"enum.auto"
] | [((684, 690), 'enum.auto', 'auto', ([], {}), '()\n', (688, 690), False, 'from enum import Enum, auto\n'), ((702, 708), 'enum.auto', 'auto', ([], {}), '()\n', (706, 708), False, 'from enum import Enum, auto\n'), ((721, 727), 'enum.auto', 'auto', ([], {}), '()\n', (725, 727), False, 'from enum import Enum, auto\n'), ((75... |
# -*- coding: utf-8 -*-
##############################################################################
# Author:QQ173782910
##############################################################################
import logging
from apscheduler.schedulers.background import BlockingScheduler
from RunUse import TradeRun
format ... | [
"logging.basicConfig",
"apscheduler.schedulers.background.BlockingScheduler",
"logging.getLogger",
"RunUse.TradeRun"
] | [((377, 462), 'logging.basicConfig', 'logging.basicConfig', ([], {'level': 'logging.INFO', 'format': 'format', 'filename': '"""log_print.txt"""'}), "(level=logging.INFO, format=format, filename='log_print.txt'\n )\n", (396, 462), False, 'import logging\n'), ((467, 493), 'logging.getLogger', 'logging.getLogger', (['"... |
# -*- coding: utf-8 -*-
from __future__ import unicode_literals
from django.db import migrations, models
class Migration(migrations.Migration):
dependencies = [
('wiki_images', '0001_initial'),
]
operations = [
migrations.AlterModelTable(
name='image',
table='wik... | [
"django.db.migrations.AlterModelTable"
] | [((244, 311), 'django.db.migrations.AlterModelTable', 'migrations.AlterModelTable', ([], {'name': '"""image"""', 'table': '"""wiki_images_image"""'}), "(name='image', table='wiki_images_image')\n", (270, 311), False, 'from django.db import migrations, models\n'), ((356, 444), 'django.db.migrations.AlterModelTable', 'mi... |
# coding: utf-8
from modules.controller import Controller
LED_PIN = 23 # pin of led
if __name__ == "__main__":
controller = Controller(LED_PIN)
controller.run(50, 1)
print("Led Start for a second with 50% power")
| [
"modules.controller.Controller"
] | [((132, 151), 'modules.controller.Controller', 'Controller', (['LED_PIN'], {}), '(LED_PIN)\n', (142, 151), False, 'from modules.controller import Controller\n')] |
from .base import BaseFactory
from launchkey import LAUNCHKEY_PRODUCTION
from launchkey.clients import ServiceClient
class ServiceFactory(BaseFactory):
"""Factory for creating clients when representing a LaunchKey Service Profile"""
def __init__(self, service_id, private_key, url=LAUNCHKEY_PRODUCTION, testin... | [
"launchkey.clients.ServiceClient"
] | [((1369, 1416), 'launchkey.clients.ServiceClient', 'ServiceClient', (['self._issuer_id', 'self._transport'], {}), '(self._issuer_id, self._transport)\n', (1382, 1416), False, 'from launchkey.clients import ServiceClient\n')] |
""" User api endpoints """
from flask import request
from flask_restplus import Resource
from ..util.dto import UserDto
from ..service.user import save_new_user
api = UserDto.api
_user = UserDto.user
@api.route('/users')
class Users(Resource):
"""
User resource for the API
"""
@api.doc('Create a n... | [
"flask.request.get_json"
] | [((502, 520), 'flask.request.get_json', 'request.get_json', ([], {}), '()\n', (518, 520), False, 'from flask import request\n')] |
import os
import zipfile
from typing import List, Tuple, Dict
import numpy as np
import pandas as pd
import requests
import structlog
import matplotlib.pyplot as plt
from sklearn.ensemble import RandomForestClassifier
from sklearn.feature_extraction.text import TfidfVectorizer
from sklearn.model_selection import train... | [
"zipfile.ZipFile",
"pandas.read_csv",
"sklearn.neighbors.KNeighborsClassifier",
"numpy.array",
"tensorflow.keras.layers.Dense",
"pandas.read_excel",
"tensorflow.keras.losses.CategoricalCrossentropy",
"os.path.exists",
"sklearn.model_selection.train_test_split",
"tensorflow.keras.layers.ReLU",
"s... | [((548, 601), 'matplotlib.pyplot.rcParams.update', 'plt.rcParams.update', (["{'figure.figsize': (16.0, 12.0)}"], {}), "({'figure.figsize': (16.0, 12.0)})\n", (567, 601), True, 'import matplotlib.pyplot as plt\n'), ((612, 642), 'structlog.get_logger', 'structlog.get_logger', (['__file__'], {}), '(__file__)\n', (632, 642... |
# vim: set et nosi ai ts=2 sts=2 sw=2:
# coding: utf-8
from __future__ import absolute_import, print_function, unicode_literals
import unittest
from schwa import dr
class Node(dr.Ann):
label = dr.Field()
class Doc(dr.Doc):
store = dr.Store(Node)
class Test(unittest.TestCase):
def _test_example(self, doc):
... | [
"schwa.dr.Store",
"schwa.dr.Field"
] | [((198, 208), 'schwa.dr.Field', 'dr.Field', ([], {}), '()\n', (206, 208), False, 'from schwa import dr\n'), ((240, 254), 'schwa.dr.Store', 'dr.Store', (['Node'], {}), '(Node)\n', (248, 254), False, 'from schwa import dr\n')] |
#!/usr/bin/env python
import os,sys
sys.path.insert(1, os.path.join(sys.path[0], '..'))
import argparse
from multiagent.environment import MultiAgentEnv
import multiagent.scenarios as scenarios
import numpy as np
import keras.backend.tensorflow_backend as backend
from keras.models import Sequential
from keras.layers i... | [
"keras.layers.Conv2D",
"numpy.array",
"keras.layers.Activation",
"keras.layers.Dense",
"tensorflow.set_random_seed",
"multiagent.environment.MultiAgentEnv",
"collections.deque",
"argparse.ArgumentParser",
"numpy.random.random",
"numpy.max",
"os.path.isdir",
"numpy.random.seed",
"keras.optimi... | [((55, 86), 'os.path.join', 'os.path.join', (['sys.path[0]', '""".."""'], {}), "(sys.path[0], '..')\n", (67, 86), False, 'import os, sys\n'), ((647, 688), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {'description': 'None'}), '(description=None)\n', (670, 688), False, 'import argparse\n'), ((1017, 1144), ... |
#! python3
# -*- encoding: utf-8 -*-
from job import celery
@celery.task()
def add(x, y):
return x+y
| [
"job.celery.task"
] | [((65, 78), 'job.celery.task', 'celery.task', ([], {}), '()\n', (76, 78), False, 'from job import celery\n')] |
from pathlib import Path
from unittest.mock import call
import pytest
from dbt_sugar.core.clients.dbt import DbtProfile
from dbt_sugar.core.config.config import DbtSugarConfig
from dbt_sugar.core.flags import FlagParser
from dbt_sugar.core.main import parser
from dbt_sugar.core.task.audit import AuditTask
from dbt_su... | [
"dbt_sugar.core.task.audit.AuditTask",
"pathlib.Path",
"unittest.mock.call",
"pytest.mark.datafiles",
"pytest.param",
"dbt_sugar.core.flags.FlagParser",
"dbt_sugar.core.config.config.DbtSugarConfig"
] | [((1852, 1886), 'pytest.mark.datafiles', 'pytest.mark.datafiles', (['FIXTURE_DIR'], {}), '(FIXTURE_DIR)\n', (1873, 1886), False, 'import pytest\n'), ((2607, 2641), 'pytest.mark.datafiles', 'pytest.mark.datafiles', (['FIXTURE_DIR'], {}), '(FIXTURE_DIR)\n', (2628, 2641), False, 'import pytest\n'), ((3481, 3515), 'pytest.... |
"""Functionality to evaluate contents of the ast"""
from functools import singledispatch, wraps
import operator
from typing import Any, Union
from altair_transform.utils import ast, Parser
__all__ = ['evaljs']
def evaljs(expression: Union[str, ast.Expr], namespace: dict = None) -> Any:
"""Evaluate a javascript ... | [
"altair_transform.utils.ast.Global",
"altair_transform.utils.Parser",
"functools.wraps"
] | [((3350, 3361), 'functools.wraps', 'wraps', (['func'], {}), '(func)\n', (3355, 3361), False, 'from functools import singledispatch, wraps\n'), ((417, 425), 'altair_transform.utils.Parser', 'Parser', ([], {}), '()\n', (423, 425), False, 'from altair_transform.utils import ast, Parser\n'), ((2547, 2569), 'altair_transfor... |
#!/usr/bin/env python
# -- coding: UTF-8 --
"""
Magnet API NLP Sample
Copyright 2018, Klangoo Inc.
"""
from klangooclient.MagnetAPIClient import MagnetAPIClient
ENDPOINT = 'https://nlp.klangoo.com/Service.svc'
CALK = 'enter your calk here'
SECRET_KEY = 'enter your secret key here'
client = MagnetAPIClient(END... | [
"klangooclient.MagnetAPIClient.MagnetAPIClient"
] | [((301, 344), 'klangooclient.MagnetAPIClient.MagnetAPIClient', 'MagnetAPIClient', (['ENDPOINT', 'CALK', 'SECRET_KEY'], {}), '(ENDPOINT, CALK, SECRET_KEY)\n', (316, 344), False, 'from klangooclient.MagnetAPIClient import MagnetAPIClient\n')] |
"""log_redaction_cli
Usage:
log_redaction_cli.py --tarfile <tarfile> --working-dir <working-dir> --output-dir <output-dir>
log_redaction_cli.py (-h | --help)
log_redaction_cli.py --version
Options:
-h --help Pass in a string: example command: python log_redaction_cli.py --tarfile "test/files/test_output.t... | [
"docopt.docopt"
] | [((580, 612), 'docopt.docopt', 'docopt', (['__doc__'], {'version': '"""0.0.9"""'}), "(__doc__, version='0.0.9')\n", (586, 612), False, 'from docopt import docopt\n')] |
import torch
import torch.nn as nn
from torch.nn.utils.rnn import pack_padded_sequence, pad_packed_sequence
import numpy as np
import torch.nn.functional as F
from attention import AdditiveAttention
class Encoder(nn.Module):
"""Encoder bi-GRU"""
def __init__(self, input_dim, char_embed_dim,
e... | [
"torch.nn.functional.softmax",
"torch.nn.Dropout",
"torch.ones",
"numpy.random.random",
"torch.stack",
"attention.AdditiveAttention",
"torch.argmax",
"torch.cat",
"torch.nn.utils.rnn.pack_padded_sequence",
"torch.nn.Linear",
"torch.nn.Embedding.from_pretrained",
"torch.nn.utils.rnn.pad_packed_... | [((822, 891), 'torch.nn.Embedding', 'nn.Embedding', (['input_dim', 'char_embed_dim'], {'padding_idx': 'char_padding_idx'}), '(input_dim, char_embed_dim, padding_idx=char_padding_idx)\n', (834, 891), True, 'import torch.nn as nn\n'), ((1507, 1738), 'torch.nn.GRU', 'nn.GRU', ([], {'input_size': '(char_embed_dim + morph_e... |
"""CLI entry points of faddr."""
import argparse
import pathlib
import sys
from faddr import logger
from faddr.config import load_config
from faddr.rancid import RancidDir
from faddr.database import Database
def parse_args_db():
"""Parsing CMD keys."""
parser = argparse.ArgumentParser(argument_default=argpa... | [
"argparse.ArgumentParser",
"pathlib.Path",
"faddr.logger.error",
"faddr.logger.debug",
"sys.exit",
"faddr.config.load_config",
"faddr.rancid.RancidDir"
] | [((274, 333), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {'argument_default': 'argparse.SUPPRESS'}), '(argument_default=argparse.SUPPRESS)\n', (297, 333), False, 'import argparse\n'), ((1095, 1138), 'faddr.logger.debug', 'logger.debug', (['f"""Arguments from CMD: {args}"""'], {}), "(f'Arguments from CMD... |
import pandas as pd
from matplotlib import dates as mdates
from timeseries import plot_ts
from timeseries.plotting import ax_settings
def main():
# .csv available to download at:
# https://www.investing.com/currencies/gbp-usd-historical-data
gpd_usd_data = pd.read_csv("../data/GBP_USD Historical Data_mon... | [
"pandas.read_csv",
"matplotlib.dates.DateFormatter",
"timeseries.plotting.ax_settings",
"timeseries.plot_ts",
"matplotlib.dates.YearLocator",
"pandas.to_datetime"
] | [((272, 330), 'pandas.read_csv', 'pd.read_csv', (['"""../data/GBP_USD Historical Data_monthly.csv"""'], {}), "('../data/GBP_USD Historical Data_monthly.csv')\n", (283, 330), True, 'import pandas as pd\n'), ((454, 503), 'pandas.to_datetime', 'pd.to_datetime', (['gpd_usd_data.Date'], {'format': '"""%b %y"""'}), "(gpd_usd... |
from sklearn import metrics
from prettytable import PrettyTable
def evaluate_clustering(name, X, true_labels, pred_labels):
homogeneity = metrics.homogeneity_score(true_labels, pred_labels)
completeness = metrics.completeness_score(true_labels, pred_labels)
v_measure = metrics.v_measure_score(true_labels,... | [
"prettytable.PrettyTable",
"sklearn.metrics.homogeneity_score",
"sklearn.metrics.adjusted_rand_score",
"sklearn.metrics.completeness_score",
"sklearn.metrics.v_measure_score",
"sklearn.metrics.fowlkes_mallows_score",
"sklearn.metrics.normalized_mutual_info_score"
] | [((144, 195), 'sklearn.metrics.homogeneity_score', 'metrics.homogeneity_score', (['true_labels', 'pred_labels'], {}), '(true_labels, pred_labels)\n', (169, 195), False, 'from sklearn import metrics\n'), ((215, 267), 'sklearn.metrics.completeness_score', 'metrics.completeness_score', (['true_labels', 'pred_labels'], {})... |
import os
import pymysql
from dotenv import load_dotenv
from sqlalchemy import create_engine
from sqlalchemy.ext.declarative import declarative_base
from sqlalchemy.orm import relationship
from sqlalchemy import Column, BigInteger, String, Float, ForeignKey
def load_engine():
"""Create the database engine"""
... | [
"sqlalchemy.orm.relationship",
"sqlalchemy.create_engine",
"sqlalchemy.ForeignKey",
"dotenv.load_dotenv",
"sqlalchemy.String",
"sqlalchemy.ext.declarative.declarative_base",
"sqlalchemy.Column"
] | [((609, 627), 'sqlalchemy.ext.declarative.declarative_base', 'declarative_base', ([], {}), '()\n', (625, 627), False, 'from sqlalchemy.ext.declarative import declarative_base\n'), ((321, 334), 'dotenv.load_dotenv', 'load_dotenv', ([], {}), '()\n', (332, 334), False, 'from dotenv import load_dotenv\n'), ((506, 581), 'sq... |