code stringlengths 22 1.05M | apis listlengths 1 3.31k | extract_api stringlengths 75 3.25M |
|---|---|---|
""""""
# Standard library modules.
# Third party modules.
import pytest
import sqlalchemy
# Local modules.
import dataclasses_sql
# Globals and constants variables.
@pytest.fixture
def metadata():
engine = sqlalchemy.create_engine("sqlite:///:memory:")
return sqlalchemy.MetaData(engine)
def test_delete_... | [
"pytest.approx",
"dataclasses_sql.update",
"sqlalchemy.create_engine",
"dataclasses_sql.insert",
"sqlalchemy.MetaData",
"pytest.raises",
"dataclasses_sql.delete"
] | [((216, 262), 'sqlalchemy.create_engine', 'sqlalchemy.create_engine', (['"""sqlite:///:memory:"""'], {}), "('sqlite:///:memory:')\n", (240, 262), False, 'import sqlalchemy\n'), ((274, 301), 'sqlalchemy.MetaData', 'sqlalchemy.MetaData', (['engine'], {}), '(engine)\n', (293, 301), False, 'import sqlalchemy\n'), ((565, 60... |
#!/usr/bin/env python
# -*- coding: utf-8 -*-
from __future__ import annotations
import queue
from typing import TYPE_CHECKING, TypeVar
T = TypeVar("T")
if TYPE_CHECKING:
SimpleQueue = queue.SimpleQueue
else:
class FakeGenericMeta(type):
def __getitem__(self, item):
return self
clas... | [
"typing.TypeVar"
] | [((142, 154), 'typing.TypeVar', 'TypeVar', (['"""T"""'], {}), "('T')\n", (149, 154), False, 'from typing import TYPE_CHECKING, TypeVar\n')] |
from django.db import models
class events(models.Model):
id = models.AutoField(primary_key=True)
note_id = models.BigIntegerField(null=True, blank=True)
tweet_id = models.BigIntegerField()
type = models.IntegerField(null=True, blank=True)
timestamp = models.DateTimeField()
from_user = models.Ch... | [
"django.db.models.IntegerField",
"django.db.models.AutoField",
"django.db.models.BigIntegerField",
"django.db.models.DateTimeField",
"django.db.models.CharField"
] | [((67, 101), 'django.db.models.AutoField', 'models.AutoField', ([], {'primary_key': '(True)'}), '(primary_key=True)\n', (83, 101), False, 'from django.db import models\n'), ((116, 161), 'django.db.models.BigIntegerField', 'models.BigIntegerField', ([], {'null': '(True)', 'blank': '(True)'}), '(null=True, blank=True)\n'... |
#! /usr/bin/python3
"""
__Version__: 0.1
__Author__: <NAME>
Data: 15/02/2020
Description: Sample scrip for scan host ports with only buit-in functions
This code just works with addresses of v4 family.
Python 3.x
"""
# Import modules
import socket
import sys
import errno
import os
import argparse
import ipaddress
# M... | [
"socket.socket",
"os.strerror",
"argparse.ArgumentParser",
"sys.exit"
] | [((2123, 2193), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {'description': '"""Scan ports TCP\nVersion: 0.1"""'}), '(description="""Scan ports TCP\nVersion: 0.1""")\n', (2146, 2193), False, 'import argparse\n'), ((1144, 1193), 'socket.socket', 'socket.socket', (['socket.AF_INET', 'socket.SOCK_STREAM'], ... |
#print_hello_friend.py
from datetime import datetime
print(datetime.now())
print("G'day Mate!")
| [
"datetime.datetime.now"
] | [((61, 75), 'datetime.datetime.now', 'datetime.now', ([], {}), '()\n', (73, 75), False, 'from datetime import datetime\n')] |
from urllib import request
def download_from_url(url, filename):
request.urlretrieve(url, filename) | [
"urllib.request.urlretrieve"
] | [((70, 104), 'urllib.request.urlretrieve', 'request.urlretrieve', (['url', 'filename'], {}), '(url, filename)\n', (89, 104), False, 'from urllib import request\n')] |
from typing import List
from led import Led
from transitions import AbstractTransition
class Sudden(AbstractTransition):
def __init__(self, red: float, green: float, blue: float) -> None:
super().__init__()
self.target = Led(red, green, blue)
@AbstractTransition.brightness.setter
def b... | [
"transitions.AbstractTransition.brightness.fset",
"led.Led"
] | [((246, 267), 'led.Led', 'Led', (['red', 'green', 'blue'], {}), '(red, green, blue)\n', (249, 267), False, 'from led import Led\n'), ((401, 453), 'transitions.AbstractTransition.brightness.fset', 'AbstractTransition.brightness.fset', (['self', 'brightness'], {}), '(self, brightness)\n', (435, 453), False, 'from transit... |
import operator as op
import itertools
with open("input.txt") as file:
data = file.read()
shop = """Weapons: Cost Damage Armor
Dagger 8 4 0
Shortsword 10 5 0
Warhammer 25 6 0
Longsword 40 7 0
Greataxe 74 8 0
Armor: Cost Damage Armor... | [
"itertools.chain",
"operator.itemgetter",
"itertools.combinations"
] | [((1917, 1948), 'itertools.chain', 'itertools.chain', (['[None]', 'armors'], {}), '([None], armors)\n', (1932, 1948), False, 'import itertools\n'), ((1982, 2014), 'itertools.combinations', 'itertools.combinations', (['rings', '(2)'], {}), '(rings, 2)\n', (2004, 2014), False, 'import itertools\n'), ((1549, 1576), 'opera... |
import sys
import os
from PyQt5.QtWidgets import QApplication, QMainWindow, QMessageBox, QDialog
from PyQt5.QtCore import pyqtSignal
from mainUi import Ui_Form
from sortUi import sortUi
from functools import partial
import csvIssue
class MyMainForm(QMainWindow, Ui_Form):
def __init__(self, parent=None):
s... | [
"PyQt5.QtCore.pyqtSignal",
"csvIssue.write_spending",
"PyQt5.QtWidgets.QMessageBox.critical",
"csvIssue.delete_last_line",
"csvIssue.init_first_classifier",
"csvIssue.delete_classifier",
"PyQt5.QtWidgets.QMessageBox.information",
"PyQt5.QtWidgets.QMessageBox.question",
"csvIssue.get_last_line",
"f... | [((2625, 2641), 'PyQt5.QtCore.pyqtSignal', 'pyqtSignal', (['list'], {}), '(list)\n', (2635, 2641), False, 'from PyQt5.QtCore import pyqtSignal\n'), ((5585, 5607), 'PyQt5.QtWidgets.QApplication', 'QApplication', (['sys.argv'], {}), '(sys.argv)\n', (5597, 5607), False, 'from PyQt5.QtWidgets import QApplication, QMainWind... |
import unittest
import uuid
from . import user_util
class TestUtilFuncs(unittest.TestCase):
def test_hash_and_verify_password(self):
passwords = [str(uuid.uuid4()) for i in range(10)]
for pw in passwords:
self.assertTrue(
user_util.verify_password(pw, user_util.hash_p... | [
"uuid.uuid4"
] | [((166, 178), 'uuid.uuid4', 'uuid.uuid4', ([], {}), '()\n', (176, 178), False, 'import uuid\n')] |
#!/usr/bin/env python
from prometheus_client import start_http_server, Summary
import random
import argparse
import time
from prometheus_client import Counter
from prometheus_client import Gauge
from prometheus_client import Summary
from prometheus_client import Histogram
import sys
import time
import json
import date... | [
"socket.gethostbyname",
"urllib.parse.urlparse",
"argparse.ArgumentParser",
"datetime.datetime.utcnow",
"logging.debug",
"threading.RLock",
"time.sleep",
"json.load",
"logging.getLevelName",
"prometheus_client.core.REGISTRY.register",
"watchdog.observers.Observer",
"logging.info",
"logging.e... | [((1218, 1235), 'threading.RLock', 'threading.RLock', ([], {}), '()\n', (1233, 1235), False, 'import threading\n'), ((18169, 18179), 'watchdog.observers.Observer', 'Observer', ([], {}), '()\n', (18177, 18179), False, 'from watchdog.observers import Observer\n'), ((18270, 18317), 'prometheus_client.core.REGISTRY.registe... |
#!/usr/bin/env pythonw
import numpy as np
import matplotlib.pyplot as plt
def flip_coins(flips = 1000000, bins=100):
# Uninformative prior
prior = np.ones(bins, dtype='float')/bins
likelihood_heads = np.arange(bins)/float(bins)
likelihood_tails = 1-likelihood_heads
flips = np.random.choice(a=[True... | [
"numpy.ones",
"numpy.random.choice",
"matplotlib.pyplot.legend",
"numpy.sum",
"numpy.arange",
"matplotlib.pyplot.show"
] | [((954, 996), 'matplotlib.pyplot.legend', 'plt.legend', (['[10, 100, 1000, 10000, 100000]'], {}), '([10, 100, 1000, 10000, 100000])\n', (964, 996), True, 'import matplotlib.pyplot as plt\n'), ((997, 1007), 'matplotlib.pyplot.show', 'plt.show', ([], {}), '()\n', (1005, 1007), True, 'import matplotlib.pyplot as plt\n'), ... |
#%%
from fireworks import PyTorch_Model, Message, HookedPassThroughPipe, Experiment
from fireworks.toolbox import ShufflerPipe, TensorPipe, BatchingPipe, FunctionPipe
from fireworks.toolbox.preprocessing import train_test_split
from fireworks.extensions import IgniteJunction
from fireworks.core import PyTorch_Model
im... | [
"torch.nn.CrossEntropyLoss",
"torch.max",
"fireworks.toolbox.preprocessing.train_test_split",
"torch.cuda.is_available",
"torchvision.datasets.mnist.FashionMNIST",
"matplotlib.pyplot.imshow",
"fireworks.Experiment",
"fireworks.toolbox.BatchingPipe",
"fireworks.Message.from_objects",
"fireworks.Mes... | [((801, 846), 'fireworks.Experiment', 'Experiment', (['env_name'], {'description': 'description'}), '(env_name, description=description)\n', (811, 846), False, 'from fireworks import PyTorch_Model, Message, HookedPassThroughPipe, Experiment\n'), ((864, 896), 'os.environ.get', 'env.get', (['"""MNIST_DIR"""', '"""./MNIST... |
import os
import shutil
video_files = ['.webm', '.mkv', '.vob', '.gif', '.avi', '.amv', '.mp4',]
audio_files = ['.aif','.cda', '.mid', '.mp3', '.mpa', '.ogg', ]
image_files = ['.tif', '.tiff', '.bmp', '.jpg', '.jpeg', '.gif', '.png', '.eps', '.raw', '.cr2', '.nef', '.orf', '.sr2', '.ico']
setup_files = ['.... | [
"os.path.exists",
"os.listdir",
"shutil.move",
"os.path.splitext",
"os.mkdir"
] | [((822, 838), 'os.listdir', 'os.listdir', (['path'], {}), '(path)\n', (832, 838), False, 'import os\n'), ((716, 736), 'os.path.exists', 'os.path.exists', (['path'], {}), '(path)\n', (730, 736), False, 'import os\n'), ((1089, 1136), 'shutil.move', 'shutil.move', (['(path + file_name)', "(path + 'audio\\\\')"], {}), "(pa... |
from pathlib import Path
'''
This script creates a new html that has placed the javascript code inline to make a standalone html
'''
src = Path.cwd() / 'coldcard_address_generator_html.html'
dest = Path.cwd() / 'coldcard_address_generator_html_standalone.html'
dest2 = Path.cwd() / 'index.html' # for github pages
... | [
"pathlib.Path.cwd"
] | [((142, 152), 'pathlib.Path.cwd', 'Path.cwd', ([], {}), '()\n', (150, 152), False, 'from pathlib import Path\n'), ((201, 211), 'pathlib.Path.cwd', 'Path.cwd', ([], {}), '()\n', (209, 211), False, 'from pathlib import Path\n'), ((272, 282), 'pathlib.Path.cwd', 'Path.cwd', ([], {}), '()\n', (280, 282), False, 'from pathl... |
#!/usr/bin/env python
import argparse
import logging
from csv import DictReader
from datetime import date, datetime
from pathlib import Path
logger = logging.getLogger(__name__)
header = """V042
ATaxTool Donations 0.4
D{date:%Y-%m-%d}
^
"""
record_layout_1 = """TD
N280
C1
L1
${amount:0.2f}
X{payee} ({ein})
^
"""
r... | [
"logging.getLogger",
"csv.DictReader",
"argparse.ArgumentParser",
"pathlib.Path",
"datetime.datetime.now",
"datetime.date.today"
] | [((152, 179), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (169, 179), False, 'import logging\n'), ((1772, 1797), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {}), '()\n', (1795, 1797), False, 'import argparse\n'), ((436, 452), 'pathlib.Path', 'Path', (['args.input'], {}),... |
import sys, numpy
from sklearn.feature_extraction.text import TfidfVectorizer
from sklearn.ensemble import RandomForestClassifier
from sklearn.naive_bayes import MultinomialNB
#0=drama,1=comedy,2=animated,3=action/adventure
def random_forest_class(raw_test_set):
x_train=[]
y_train=[]
count=0
vectorize... | [
"sklearn.feature_extraction.text.TfidfVectorizer",
"sklearn.naive_bayes.MultinomialNB",
"sklearn.ensemble.RandomForestClassifier"
] | [((324, 397), 'sklearn.feature_extraction.text.TfidfVectorizer', 'TfidfVectorizer', ([], {'analyzer': '"""word"""', 'stop_words': '"""english"""', 'max_features': '(1024)'}), "(analyzer='word', stop_words='english', max_features=1024)\n", (339, 397), False, 'from sklearn.feature_extraction.text import TfidfVectorizer\n... |
from flask import Blueprint, jsonify, request, session
from pymongo import DESCENDING
from api.views import users
from api import collection_users
@users.route("/api/register", methods=["POST"])
def user_register():
if request.method == "POST":
data = request.get_json()
email = data['e... | [
"api.views.users.route",
"api.collection_users.find_one",
"api.collection_users.update_one",
"api.collection_users.find",
"flask.session.get",
"flask.request.get_json",
"api.collection_users.insert_one",
"flask.session.clear",
"flask.jsonify"
] | [((158, 204), 'api.views.users.route', 'users.route', (['"""/api/register"""'], {'methods': "['POST']"}), "('/api/register', methods=['POST'])\n", (169, 204), False, 'from api.views import users\n'), ((1850, 1900), 'api.views.users.route', 'users.route', (['"""/api/login"""'], {'methods': "['GET', 'POST']"}), "('/api/l... |
import time
import os
import pyglet
from gtts import gTTS
from pydub import AudioSegment
import traceback
def play_text(*txts):
try:
sounds = []
fnames = []
for i, s in enumerate(txts):
g = gTTS(text=s, lang='en')
fname = 'voice{}.mp3'.format(i)
with open... | [
"pydub.AudioSegment.from_mp3",
"os.path.join",
"pyglet.media.load",
"time.sleep",
"gtts.gTTS",
"traceback.print_exc",
"time.time",
"os.remove"
] | [((687, 698), 'time.time', 'time.time', ([], {}), '()\n', (696, 698), False, 'import time\n'), ((1188, 1229), 'pyglet.media.load', 'pyglet.media.load', (['fname'], {'streaming': '(False)'}), '(fname, streaming=False)\n', (1205, 1229), False, 'import pyglet\n'), ((1309, 1320), 'time.time', 'time.time', ([], {}), '()\n',... |
import numpy as np
import pandas as pd
def batch_df2batch(df, evaluate_ids=(), n_obs=-1, tform=np.eye(3), is_vehicles_evaluated=False):
"""
Convert dataframe to SGAN input
:param df:
:param evaluate_ids:
:param n_obs: number of timesteps observed
:param tform:
:param is_vehicles_evaluat... | [
"numpy.eye",
"numpy.unique",
"numpy.ones",
"numpy.sort",
"numpy.stack",
"numpy.zeros",
"numpy.isnan",
"numpy.vstack",
"numpy.zeros_like",
"numpy.arange"
] | [((97, 106), 'numpy.eye', 'np.eye', (['(3)'], {}), '(3)\n', (103, 106), True, 'import numpy as np\n'), ((707, 761), 'numpy.zeros', 'np.zeros', (['(n_obs, agent_ids.size, 2)'], {'dtype': 'np.float32'}), '((n_obs, agent_ids.size, 2), dtype=np.float32)\n', (715, 761), True, 'import numpy as np\n'), ((775, 796), 'numpy.zer... |
"""
base.py -- client for the base Rinnai API
"""
import datetime, json, logging, time
import requests
from rinnaicontrolr.aws_srp import AWSSRP
LOGGER = logging.getLogger('rinnaicontrolr')
from rinnaicontrolr.const import (
POOL_ID,
CLIENT_ID,
POOL_REGION,
GRAPHQL_ENDPOINT,
SHADOW_ENDPOINT,
... | [
"logging.getLogger",
"requests.post",
"rinnaicontrolr.aws_srp.AWSSRP",
"time.time"
] | [((157, 192), 'logging.getLogger', 'logging.getLogger', (['"""rinnaicontrolr"""'], {}), "('rinnaicontrolr')\n", (174, 192), False, 'import datetime, json, logging, time\n'), ((1083, 1094), 'time.time', 'time.time', ([], {}), '()\n', (1092, 1094), False, 'import datetime, json, logging, time\n'), ((1468, 1589), 'rinnaic... |
# Generated by Django 2.1.3 on 2019-02-12 19:18
from django.db import migrations, models
class Migration(migrations.Migration):
dependencies = [
('profile', '0002_auto_20180126_1900'),
]
operations = [
migrations.CreateModel(
name='GlobalAlert',
fields=[
... | [
"django.db.models.AutoField",
"django.db.models.TextField",
"django.db.models.BooleanField"
] | [((335, 428), 'django.db.models.AutoField', 'models.AutoField', ([], {'auto_created': '(True)', 'primary_key': '(True)', 'serialize': '(False)', 'verbose_name': '"""ID"""'}), "(auto_created=True, primary_key=True, serialize=False,\n verbose_name='ID')\n", (351, 428), False, 'from django.db import migrations, models\... |
import click
from agent.pipeline.validators import elastic_query, jdbc_query
from agent import source
class BaseValidator:
@staticmethod
def validate(pipeline):
pass
class ElasticValidator(BaseValidator):
@staticmethod
def validate(pipeline):
with open(pipeline.config['query_file'])... | [
"agent.pipeline.validators.jdbc_query.get_errors",
"click.ClickException",
"agent.pipeline.validators.elastic_query.get_errors"
] | [((373, 475), 'agent.pipeline.validators.elastic_query.get_errors', 'elastic_query.get_errors', (['query', 'pipeline.source.config[source.ElasticSource.CONFIG_OFFSET_FIELD]'], {}), '(query, pipeline.source.config[source.ElasticSource\n .CONFIG_OFFSET_FIELD])\n', (397, 475), False, 'from agent.pipeline.validators imp... |
import numpy as npy
def convert(num):
if num < 0:
# num = -num
num *= 1024
# num += 32768
num = int(num - 0.5)
num = 65535 + num
n_str = str(hex(num))[2:]
if len(n_str) == 1:
n_str = 'fff' + n_str
elif len(n_str) == 2:
n_str = ... | [
"numpy.load"
] | [((849, 879), 'numpy.load', 'npy.load', (['"""dense_kernel_0.npy"""'], {}), "('dense_kernel_0.npy')\n", (857, 879), True, 'import numpy as npy\n'), ((1175, 1203), 'numpy.load', 'npy.load', (['"""dense_bias_0.npy"""'], {}), "('dense_bias_0.npy')\n", (1183, 1203), True, 'import numpy as npy\n'), ((1330, 1362), 'numpy.loa... |
import os
import shutil
from dockerspawner import DockerSpawner
class IllumiDeskDockerSpawner(DockerSpawner):
"""
Custom DockerSpawner which assigns a user notebook image
based on the user's role. This spawner requires:
1. That the `Authenticator.enable_auth_state = True`
2. That the user's ... | [
"os.path.exists",
"os.path.join",
"os.environ.get",
"os.chmod",
"os.mkdir"
] | [((2563, 2594), 'os.path.join', 'os.path.join', (['"""/home"""', 'username'], {}), "('/home', username)\n", (2575, 2594), False, 'import os\n'), ((913, 952), 'os.environ.get', 'os.environ.get', (['"""DOCKER_STANDARD_IMAGE"""'], {}), "('DOCKER_STANDARD_IMAGE')\n", (927, 952), False, 'import os\n'), ((2610, 2635), 'os.pa... |
from django.db import models
from app_asset.models import Host
# Create your models here.
class Project(models.Model):
project_name = models.CharField(max_length=32,unique=True)
project_msg = models.CharField(max_length=64,null=True)
def __unicode__(self):
return self.project_name
class GitCode(... | [
"django.db.models.DateTimeField",
"django.db.models.TextField",
"django.db.models.CharField",
"django.db.models.ForeignKey"
] | [((141, 185), 'django.db.models.CharField', 'models.CharField', ([], {'max_length': '(32)', 'unique': '(True)'}), '(max_length=32, unique=True)\n', (157, 185), False, 'from django.db import models\n'), ((203, 245), 'django.db.models.CharField', 'models.CharField', ([], {'max_length': '(64)', 'null': '(True)'}), '(max_l... |
#/usr/bin/env python
import sys
import logging
logger = logging.getLogger('utility_to_osm.ssr2.git_diff')
import utility_to_osm.file_util as file_util
from osmapis_stedsnr import OSMstedsnr
if __name__ == '__main__':
logging.basicConfig(level=logging.DEBUG)
# diff is called by git with 7 parameters:
... | [
"logging.getLogger",
"osmapis_stedsnr.OSMstedsnr.from_xml",
"utility_to_osm.file_util.read_file",
"logging.basicConfig"
] | [((56, 105), 'logging.getLogger', 'logging.getLogger', (['"""utility_to_osm.ssr2.git_diff"""'], {}), "('utility_to_osm.ssr2.git_diff')\n", (73, 105), False, 'import logging\n'), ((223, 263), 'logging.basicConfig', 'logging.basicConfig', ([], {'level': 'logging.DEBUG'}), '(level=logging.DEBUG)\n', (242, 263), False, 'im... |
#!/usr/bin/env python3
import uuid
from passlib.hash import pbkdf2_sha512
password = input('Enter password: ')
password_parts = pbkdf2_sha512.encrypt(password, salt_size=32).split('$')
password = password_parts[4]
salt = password_parts[3]
def convert_b64(input):
return input.replace('.', '+') + '='
print('Passw... | [
"passlib.hash.pbkdf2_sha512.encrypt"
] | [((131, 176), 'passlib.hash.pbkdf2_sha512.encrypt', 'pbkdf2_sha512.encrypt', (['password'], {'salt_size': '(32)'}), '(password, salt_size=32)\n', (152, 176), False, 'from passlib.hash import pbkdf2_sha512\n')] |
import yaml
import os
import subprocess
import coloredlogs
import logging
import psutil
import shutil
import hashlib
import uuid
import fileinput
import requests
from nvc import __appname__
from dotenv import load_dotenv
import git
app_root = os.path.dirname(os.path.abspath(__file__))
app_home = os.path.expanduser("~... | [
"requests.post",
"git.Repo.clone_from",
"yaml.load",
"psutil.virtual_memory",
"logging.info",
"logging.error",
"logging.warn",
"os.path.isdir",
"os.path.expanduser",
"yaml.dump",
"requests.get",
"uuid.uuid4",
"os.path.isfile",
"shutil.copyfile",
"coloredlogs.install",
"psutil.disk_usag... | [((299, 322), 'os.path.expanduser', 'os.path.expanduser', (['"""~"""'], {}), "('~')\n", (317, 322), False, 'import os\n'), ((333, 344), 'os.getcwd', 'os.getcwd', ([], {}), '()\n', (342, 344), False, 'import os\n'), ((261, 286), 'os.path.abspath', 'os.path.abspath', (['__file__'], {}), '(__file__)\n', (276, 286), False,... |
#!/usr/bin/env python
#-*- coding:utf-8; mode:python; indent-tabs-mode: nil; c-basic-offset: 2; tab-width: 2 -*-
from collections import namedtuple
import os.path as path
from bes.testing.program_unit_test import program_unit_test
from bes.fs.file_util import file_util
from bes.system.host import host
class test_cli... | [
"os.path.join",
"bes.testing.program_unit_test.program_unit_test.main"
] | [((3103, 3127), 'bes.testing.program_unit_test.program_unit_test.main', 'program_unit_test.main', ([], {}), '()\n', (3125, 3127), False, 'from bes.testing.program_unit_test import program_unit_test\n'), ((2393, 2421), 'os.path.join', 'path.join', (['tmp', '"""kitchen.py"""'], {}), "(tmp, 'kitchen.py')\n", (2402, 2421),... |
import sys
sys.path.append('../../Data Structures')
from stack import Stack
def isPalindrome(linkedList={}):
'''
Solution 1 - Hash map
Complexity Analysis
O(n) time | O(n) space
Check if a linked list is a palindrome
dict: linkedList
return: True if its palindrome
'''
# Graceful... | [
"stack.Stack",
"sys.path.append"
] | [((11, 51), 'sys.path.append', 'sys.path.append', (['"""../../Data Structures"""'], {}), "('../../Data Structures')\n", (26, 51), False, 'import sys\n'), ((2671, 2678), 'stack.Stack', 'Stack', ([], {}), '()\n', (2676, 2678), False, 'from stack import Stack\n')] |
"""
Class Report: Part 4 of the Sprint Challenge
- Generate random Product list, and get an Inventory Report on that list
"""
from random import randint, sample, uniform
from acme import Product
ADJECTIVES = ['Awesome', 'Shiny', 'Impressive', 'Portable', 'Improved']
NOUNS = ['Anvil', 'Catapult', 'Disguise', 'Mousetra... | [
"acme.Product",
"random.uniform",
"random.sample",
"random.randint"
] | [((479, 500), 'random.sample', 'sample', (['ADJECTIVES', '(1)'], {}), '(ADJECTIVES, 1)\n', (485, 500), False, 'from random import randint, sample, uniform\n'), ((516, 532), 'random.sample', 'sample', (['NOUNS', '(1)'], {}), '(NOUNS, 1)\n', (522, 532), False, 'from random import randint, sample, uniform\n'), ((644, 659)... |
"""
The Flaskee is an Open Source project for Microservices.
Develop By <NAME> | https://nadeengamage.com | <EMAIL>
"""
from werkzeug.serving import run_simple
from werkzeug.middleware.dispatcher import DispatcherMiddleware
from flaskee import api
app = api.create_app()
application = DispatcherMiddleware(app)
if _... | [
"werkzeug.serving.run_simple",
"flaskee.api.create_app",
"werkzeug.middleware.dispatcher.DispatcherMiddleware"
] | [((257, 273), 'flaskee.api.create_app', 'api.create_app', ([], {}), '()\n', (271, 273), False, 'from flaskee import api\n'), ((289, 314), 'werkzeug.middleware.dispatcher.DispatcherMiddleware', 'DispatcherMiddleware', (['app'], {}), '(app)\n', (309, 314), False, 'from werkzeug.middleware.dispatcher import DispatcherMidd... |
from django.urls import path, include, re_path
from django.views.generic import FormView
from . import views
from .views import news, faq, resources, group, user, puzzles, submission, score_challenge, metrics
import publications.views as plist
news_patterns = [
path('', news.List.as_view(), name="news_list"),
... | [
"django.urls.re_path",
"django.urls.path",
"django.urls.include"
] | [((2178, 2294), 'django.urls.re_path', 're_path', (['"""^unconfirmed/(?P<pk>\\\\d+)/user_confirmed/$"""', 'user.UnconfirmedList.user_confirm'], {'name': '"""user_confirmed"""'}), "('^unconfirmed/(?P<pk>\\\\d+)/user_confirmed/$', user.UnconfirmedList.\n user_confirm, name='user_confirmed')\n", (2185, 2294), False, 'f... |
import numpy as np
import queue
import cv2
import os
import datetime
SIZE = 32
SCALE = 0.007874015748031496
def quantized_np(array,scale,data_width=8):
quantized_array= np.round(array/scale)
quantized_array = np.maximum(quantized_array, -2**(data_width-1))
quantized_array = np.minimum(quantized_array, 2**... | [
"os.listdir",
"numpy.reshape",
"numpy.minimum",
"cv2.resize",
"numpy.size",
"os.path.join",
"queue.Queue",
"datetime.datetime.now",
"numpy.maximum",
"cv2.imread",
"numpy.round"
] | [((175, 198), 'numpy.round', 'np.round', (['(array / scale)'], {}), '(array / scale)\n', (183, 198), True, 'import numpy as np\n'), ((219, 270), 'numpy.maximum', 'np.maximum', (['quantized_array', '(-2 ** (data_width - 1))'], {}), '(quantized_array, -2 ** (data_width - 1))\n', (229, 270), True, 'import numpy as np\n'),... |
"""Packager for cloud environment."""
from setuptools import setup, find_packages
setup(
name='preprocess',
version='1.0.0',
packages=find_packages(),
install_requires=[
'tensorflow',
'numpy',
],
)
| [
"setuptools.find_packages"
] | [((147, 162), 'setuptools.find_packages', 'find_packages', ([], {}), '()\n', (160, 162), False, 'from setuptools import setup, find_packages\n')] |
import numpy as np
import time
def max_subsequence_sum(sequence):
max_sum = 0
for i in range(0, len(sequence)):
for j in range(i, len(sequence)):
this_sum = 0
for k in range(i, j+1):
this_sum += sequence[k]
if this_sum > max_sum:
... | [
"numpy.random.randint",
"time.time"
] | [((372, 417), 'numpy.random.randint', 'np.random.randint', (['(-100000)', '(100000)'], {'size': '(1000)'}), '(-100000, 100000, size=1000)\n', (389, 417), True, 'import numpy as np\n'), ((424, 435), 'time.time', 'time.time', ([], {}), '()\n', (433, 435), False, 'import time\n'), ((476, 487), 'time.time', 'time.time', ([... |
import pygame
from tools import render_text
from graph_object import GraphObject
class Edge (GraphObject):
def __init__(self, v1, v2, weight=0, width=1, color=(0, 0, 0)):
super().__init__()
self.v1, self.v2 = v1, v2
self.__weight, self.__weight_surface = 0, None
self.set_weight(we... | [
"pygame.draw.line"
] | [((766, 837), 'pygame.draw.line', 'pygame.draw.line', (['surface', 'self.color', 'self.pos1', 'self.pos2', 'self.width'], {}), '(surface, self.color, self.pos1, self.pos2, self.width)\n', (782, 837), False, 'import pygame\n')] |
from py_profiler import profiler, profiling_service
@profiler('hello')
def hello():
print('hello')
class Foo:
@profiler('Food.some_thing')
def some_thing(self):
print('some_thing')
@profiler()
def method_2(self):
print('method_2')
raise Exception('aaaa')
if __name__ =... | [
"py_profiler.profiling_service.as_table",
"py_profiler.profiler"
] | [((55, 72), 'py_profiler.profiler', 'profiler', (['"""hello"""'], {}), "('hello')\n", (63, 72), False, 'from py_profiler import profiler, profiling_service\n'), ((124, 151), 'py_profiler.profiler', 'profiler', (['"""Food.some_thing"""'], {}), "('Food.some_thing')\n", (132, 151), False, 'from py_profiler import profiler... |
# -*- coding: utf-8 -*- #
# Copyright 2021 Google LLC. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless requir... | [
"googlecloudsdk.core.log.status.Print",
"googlecloudsdk.command_lib.storage.tasks.task_executor.execute_tasks",
"googlecloudsdk.command_lib.storage.tasks.task_graph_executor.multiprocessing_context.Queue",
"googlecloudsdk.command_lib.storage.flags.add_precondition_flags",
"googlecloudsdk.command_lib.storage... | [((4001, 4037), 'googlecloudsdk.command_lib.storage.flags.add_precondition_flags', 'flags.add_precondition_flags', (['parser'], {}), '(parser)\n', (4029, 4037), False, 'from googlecloudsdk.command_lib.storage import flags\n'), ((4466, 4635), 'googlecloudsdk.command_lib.storage.name_expansion.NameExpansionIterator', 'na... |
import sqlite3
import sys
import datetime
import os
# day of month to switch to new database
change_day = 1
c_added_text_entities = [["sensor.hitachi_relay", "sensor.netatmo_relay"],
["sensor.cooling_target_temp", "sensor.heating_target_temp"]]
user = "pipacsba"
server_ip = "192.16... | [
"sqlite3.connect",
"datetime.datetime.strptime",
"os.rename",
"os.chdir",
"datetime.datetime.now",
"os.system",
"datetime.timedelta"
] | [((486, 504), 'os.chdir', 'os.chdir', (['work_dir'], {}), '(work_dir)\n', (494, 504), False, 'import os\n'), ((701, 723), 'os.system', 'os.system', (['scp_command'], {}), '(scp_command)\n', (710, 723), False, 'import os\n'), ((6526, 6554), 'os.system', 'os.system', (["('rm ' + file_from)"], {}), "('rm ' + file_from)\n"... |
#! /usr/bin/env python
# -*- coding: utf-8 -*-
####################
import time
import logging
import indigo
from os.path import exists
import paho.mqtt.client as mqtt
################################################################################
class MQTTBroker(object):
def __init__(self, device):
s... | [
"logging.getLogger",
"os.path.exists",
"indigo.server.getInstallFolderPath",
"indigo.activePlugin.processReceivedMessage"
] | [((333, 371), 'logging.getLogger', 'logging.getLogger', (['"""Plugin.MQTTBroker"""'], {}), "('Plugin.MQTTBroker')\n", (350, 371), False, 'import logging\n'), ((6149, 6226), 'indigo.activePlugin.processReceivedMessage', 'indigo.activePlugin.processReceivedMessage', (['self.deviceID', 'msg.topic', 'payload'], {}), '(self... |
#!/usr/bin/env python
# pylint: disable=too-many-locals,arguments-differ,unused-import
import tensorflow as tf
from tensorflow.keras.layers import (BatchNormalization, Dense, Dropout,
Flatten, MaxPooling2D, SpatialDropout2D,
add)
from tensorflo... | [
"deep_model_blocks.ConvBlock",
"tensorflow.image.resize_nearest_neighbor",
"deep_model_blocks.DeconvBlock",
"tensorflow.nn.relu",
"tensorflow.logging.info",
"tensorflow.keras.layers.Dropout",
"tensorflow.keras.layers.BatchNormalization",
"deep_model_blocks.ResizeBlock",
"tensorflow.concat",
"tenso... | [((862, 913), 'tensorflow.keras.layers.Dense', 'tf.keras.layers.Dense', (['(15 * 20 * 64)'], {'use_bias': '(False)'}), '(15 * 20 * 64, use_bias=False)\n', (883, 913), True, 'import tensorflow as tf\n'), ((936, 972), 'tensorflow.keras.layers.BatchNormalization', 'tf.keras.layers.BatchNormalization', ([], {}), '()\n', (9... |
#!/usr/bin/env python
# Licensed under a 3-clause BSD style license - see LICENSE.rst
from setuptools import setup, find_packages
setup(
use_scm_version={
"version_scheme": "post-release",
"local_scheme": "dirty-tag",
},
setup_requires=["setuptools_scm"],
packages=find_packages("src"),... | [
"setuptools.find_packages"
] | [((299, 319), 'setuptools.find_packages', 'find_packages', (['"""src"""'], {}), "('src')\n", (312, 319), False, 'from setuptools import setup, find_packages\n')] |
import numpy as np
from pyFAI.multi_geometry import MultiGeometry
from pyFAI.ext import splitBBox
def inpaint_saxs(imgs, ais, masks):
"""
Inpaint the 2D image collected by the pixel detector to remove artifacts in later data reduction
Parameters:
-----------
:param imgs: List of 2D image in pixel... | [
"numpy.mean",
"numpy.shape",
"numpy.ones_like",
"numpy.sqrt",
"numpy.asarray",
"numpy.ma.masked_where",
"numpy.max",
"numpy.deg2rad",
"numpy.linspace",
"numpy.arctan2",
"numpy.min",
"numpy.meshgrid",
"pyFAI.multi_geometry.MultiGeometry",
"numpy.rad2deg",
"numpy.ma.masked_array"
] | [((1972, 2108), 'pyFAI.multi_geometry.MultiGeometry', 'MultiGeometry', (['ais'], {'unit': '"""q_A^-1"""', 'radial_range': 'radial_range', 'azimuth_range': 'azimuth_range', 'wavelength': 'None', 'empty': '(0.0)', 'chi_disc': '(180)'}), "(ais, unit='q_A^-1', radial_range=radial_range, azimuth_range=\n azimuth_range, w... |
import numpy as np
from typing import Tuple
import plotly.io
from IMLearn.metalearners.adaboost import AdaBoost
from IMLearn.learners.classifiers import DecisionStump
from IMLearn.metrics import accuracy
from utils import *
import plotly.graph_objects as go
from plotly.subplots import make_subplots
plotly.io.rendere... | [
"numpy.ones",
"numpy.random.rand",
"plotly.graph_objects.Layout",
"IMLearn.metalearners.adaboost.AdaBoost",
"numpy.max",
"numpy.array",
"numpy.sum",
"numpy.random.seed"
] | [((3011, 3047), 'numpy.array', 'np.array', (["['circle', 'x', 'diamond']"], {}), "(['circle', 'x', 'diamond'])\n", (3019, 3047), True, 'import numpy as np\n'), ((6267, 6284), 'numpy.random.seed', 'np.random.seed', (['(0)'], {}), '(0)\n', (6281, 6284), True, 'import numpy as np\n'), ((1056, 1066), 'numpy.ones', 'np.ones... |
# import packages
from airflow import DAG
from airflow.operators.bash_operator import BashOperator
from airflow.operators.python_operator import PythonOperator
from datetime import datetime, timedelta
from external_func import random_date
def start_print():
print('\nDAG starting...\n')
def end_print():
prin... | [
"datetime.datetime",
"datetime.timedelta",
"airflow.operators.python_operator.PythonOperator",
"airflow.operators.bash_operator.BashOperator"
] | [((957, 1015), 'airflow.operators.bash_operator.BashOperator', 'BashOperator', ([], {'task_id': '"""Date"""', 'bash_command': '"""date"""', 'dag': 'dag'}), "(task_id='Date', bash_command='date', dag=dag)\n", (969, 1015), False, 'from airflow.operators.bash_operator import BashOperator\n'), ((1035, 1103), 'airflow.opera... |
import math
z = [1.0,1 ,1, 1.0]
z_exp = [math.exp(i) for i in z]
print([round(i, 2) for i in z_exp])
sum_z_exp = sum(z_exp)
print(round(sum_z_exp, 2))
softmax = [round(i / sum_z_exp, 3) for i in z_exp]
print(softmax)
| [
"math.exp"
] | [((42, 53), 'math.exp', 'math.exp', (['i'], {}), '(i)\n', (50, 53), False, 'import math\n')] |
import logging
logger = logging.getLogger(__name__)
logger.debug("Loaded " + __name__)
from jsonrpcserver import methods
from .exceptions import *
from .influxdb_api import *
from .meta import *
| [
"logging.getLogger"
] | [((24, 51), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (41, 51), False, 'import logging\n')] |
import os
import pandas as pd
import datetime
from genderperformr import GenderPerformr
from agreementr import Agreementr
from politenessr import Politenessr
from supportr import Supportr
import enchant
import requests
import json
from googleapiclient import discovery
from enchant.checker import SpellChecker
from encha... | [
"supportr.Supportr",
"nltk.download",
"datetime.datetime.strptime",
"politenessr.Politenessr",
"genderperformr.GenderPerformr",
"time.strftime",
"enchant.tokenize.get_tokenizer",
"nltk.tokenize.word_tokenize",
"googleapiclient.discovery.build",
"agreementr.Agreementr",
"time.localtime",
"encha... | [((417, 439), 'nltk.download', 'nltk.download', (['"""punkt"""'], {}), "('punkt')\n", (430, 439), False, 'import nltk\n'), ((477, 496), 'nltk.tokenize.word_tokenize', 'word_tokenize', (['text'], {}), '(text)\n', (490, 496), False, 'from nltk.tokenize import word_tokenize\n'), ((1341, 1375), 'time.localtime', 'time.loca... |
from PyQt5 import QtGui, QtCore, QtWidgets
from PyQt5.uic import loadUi
import sys
class End(QtWidgets.QDialog):
def __init__(self, winners):
super(QtWidgets.QDialog, self).__init__()
loadUi("End/end.ui", self)
msg = f"""
<p style="text-align: center; font-size: 30px;"><em>... | [
"PyQt5.uic.loadUi",
"PyQt5.QtWidgets.QStackedWidget",
"PyQt5.QtWidgets.QApplication"
] | [((766, 798), 'PyQt5.QtWidgets.QApplication', 'QtWidgets.QApplication', (['sys.argv'], {}), '(sys.argv)\n', (788, 798), False, 'from PyQt5 import QtGui, QtCore, QtWidgets\n'), ((812, 838), 'PyQt5.QtWidgets.QStackedWidget', 'QtWidgets.QStackedWidget', ([], {}), '()\n', (836, 838), False, 'from PyQt5 import QtGui, QtCore... |
import os
srcfile = 'DocTools~/assetgraph_from_gdoc.md'
pnglist = 'DocTools~/order.txt'
dstfile = 'Documentation~/assetgraph.md'
num = 1
if os.path.exists(dstfile):
os.remove(dstfile)
with open(srcfile) as f:
doc = f.read()
f.close()
with open(pnglist) as fpng:
while True:
pnglist = fpng.readline()
if n... | [
"os.path.exists",
"os.remove"
] | [((143, 166), 'os.path.exists', 'os.path.exists', (['dstfile'], {}), '(dstfile)\n', (157, 166), False, 'import os\n'), ((169, 187), 'os.remove', 'os.remove', (['dstfile'], {}), '(dstfile)\n', (178, 187), False, 'import os\n')] |
import os
# isort: off
# This import has to come before the CheckmateClient import or the functional
# tests break.
# See https://github.com/gevent/gevent/issues/1016
import pywb.apps.frontendapp # pylint:disable=unused-import
# isort: on
import httpretty as httpretty_
import pytest
import webtest
from tests.conft... | [
"viahtml.app.Application",
"httpretty.disable",
"httpretty.enable",
"webtest.TestApp",
"tests.conftest.environment_variables",
"os.environ.update",
"tests.simple_server.serve_content",
"pytest.fixture",
"httpretty.reset"
] | [((438, 469), 'pytest.fixture', 'pytest.fixture', ([], {'scope': '"""session"""'}), "(scope='session')\n", (452, 469), False, 'import pytest\n'), ((608, 639), 'pytest.fixture', 'pytest.fixture', ([], {'scope': '"""session"""'}), "(scope='session')\n", (622, 639), False, 'import pytest\n'), ((1019, 1064), 'pytest.fixtur... |
# -*- coding: utf-8 -*-
#
# Author: <NAME> <<EMAIL>>
#
# Setup the SMRT module
from __future__ import print_function, absolute_import, division
from distutils.command.clean import clean
# from setuptools import setup # DO NOT use setuptools!!!!!!
import shutil
import os
import sys
if sys.version_info[0] < 3:
imp... | [
"os.path.exists",
"distutils.command.clean.clean.run",
"distutils.core.setup",
"os.path.join",
"numpy.distutils.misc_util.Configuration",
"os.path.splitext",
"os.path.dirname",
"os.unlink",
"shutil.rmtree",
"os.walk"
] | [((3093, 3138), 'numpy.distutils.misc_util.Configuration', 'Configuration', (['None', 'parent_package', 'top_path'], {}), '(None, parent_package, top_path)\n', (3106, 3138), False, 'from numpy.distutils.misc_util import Configuration\n'), ((5720, 5737), 'distutils.core.setup', 'setup', ([], {}), '(**metadata)\n', (5725... |
import numpy
import re
with open('/home/johnny/Documents/navigate_building/source/assets/images_with_plaques.txt') as f:
LIST_OF_POSITIVES = f.read().split('\n')
class ImageDetectionMetadata():
headers = ['label', 'parsed_text', 'found_contour_area', 'ref_contour_area', 'source_image_location', 'image', 'ima... | [
"re.findall"
] | [((1650, 1704), 're.findall', 're.findall', (['"""-[0-9,a-z]*-"""', 'self.source_image_location'], {}), "('-[0-9,a-z]*-', self.source_image_location)\n", (1660, 1704), False, 'import re\n')] |
# INFO : ini merupakan copy source code dari repo one4ubot, dan sudah mendapatkan izin dari pemilik.
# INFO : This is a copy of the source code from the One4ubot repo, and has the permission of the owner.
try:
from userbot.modules.sql_helper import SESSION, BASE
except ImportError:
raise AttributeError
from sq... | [
"userbot.modules.sql_helper.SESSION.commit",
"userbot.modules.sql_helper.SESSION.close",
"userbot.modules.sql_helper.SESSION.delete",
"userbot.modules.sql_helper.SESSION.query",
"userbot.modules.sql_helper.SESSION.add",
"sqlalchemy.String"
] | [((765, 783), 'userbot.modules.sql_helper.SESSION.add', 'SESSION.add', (['adder'], {}), '(adder)\n', (776, 783), False, 'from userbot.modules.sql_helper import SESSION, BASE\n'), ((788, 804), 'userbot.modules.sql_helper.SESSION.commit', 'SESSION.commit', ([], {}), '()\n', (802, 804), False, 'from userbot.modules.sql_he... |
#!/usr/bin/env python3
"""Switch variable case.
A function that takes camel cased strings (i.e. ThisIsCamelCased),
and converts them to snake case (i.e. this_is_camel_cased).
"""
import re
def snake_case(input_str: str, camel_case=False) -> str:
"""
Turn camel case into snake case.
:param input_str: ... | [
"re.sub"
] | [((584, 626), 're.sub', 're.sub', (['regex', 'substitution', 'input_str[1:]'], {}), '(regex, substitution, input_str[1:])\n', (590, 626), False, 'import re\n')] |
import numpy as np
import tensorflow as tf
import unittest
hungarian_module = tf.load_op_library("hungarian.so")
class HungarianTests(unittest.TestCase):
def test_min_weighted_bp_cover_1(self):
W = np.array([[3, 2, 2], [1, 2, 0], [2, 2, 1]])
M, c_0, c_1 = hungarian_module.hungarian(W)
with tf.Session()... | [
"tensorflow.load_op_library",
"numpy.round",
"tensorflow.Session",
"numpy.array",
"unittest.TextTestRunner",
"unittest.TestLoader"
] | [((78, 112), 'tensorflow.load_op_library', 'tf.load_op_library', (['"""hungarian.so"""'], {}), "('hungarian.so')\n", (96, 112), True, 'import tensorflow as tf\n'), ((207, 250), 'numpy.array', 'np.array', (['[[3, 2, 2], [1, 2, 0], [2, 2, 1]]'], {}), '([[3, 2, 2], [1, 2, 0], [2, 2, 1]])\n', (215, 250), True, 'import nump... |
import numpy as np
import matplotlib.pyplot as plt
import seaborn as sns
import xarray as xr
sns.set()
def plot_range(xlabel, ylabel, title, x, values):
"""x and values should have the same size"""
plt.plot(x, values, 'r-', linewidth=2)
plt.gcf().set_size_inches(8, 2)
plt.title(title)
plt.xlabel(... | [
"seaborn.set",
"matplotlib.pyplot.ylabel",
"matplotlib.pyplot.gcf",
"matplotlib.pyplot.xlabel",
"matplotlib.pyplot.plot",
"matplotlib.pyplot.figure",
"matplotlib.pyplot.title",
"xarray.open_dataset",
"numpy.arange",
"matplotlib.pyplot.show"
] | [((93, 102), 'seaborn.set', 'sns.set', ([], {}), '()\n', (100, 102), True, 'import seaborn as sns\n'), ((209, 247), 'matplotlib.pyplot.plot', 'plt.plot', (['x', 'values', '"""r-"""'], {'linewidth': '(2)'}), "(x, values, 'r-', linewidth=2)\n", (217, 247), True, 'import matplotlib.pyplot as plt\n'), ((288, 304), 'matplot... |
import pandas as pd
import numpy as np
import math
from nltk.stem.snowball import SnowballStemmer
def add_prefix(prefix, series):
"""
Returns a pandas series that adds a prefix to a string
:param prefix: str
:return: pd.Series
"""
if type(prefix) != str:
raise TypeError(prefix + ' is n... | [
"nltk.stem.snowball.SnowballStemmer",
"math.isnan"
] | [((3737, 3763), 'nltk.stem.snowball.SnowballStemmer', 'SnowballStemmer', (['"""english"""'], {}), "('english')\n", (3752, 3763), False, 'from nltk.stem.snowball import SnowballStemmer\n'), ((2990, 3003), 'math.isnan', 'math.isnan', (['x'], {}), '(x)\n', (3000, 3003), False, 'import math\n')] |
import os
from pocketsphinx import AudioFile
from pocketsphinx import Pocketsphinx
from src import util
test_video = os.environ['DATA_PATH'] + "/other/sphinx_test_video/beachball.mp4"
test_audio = os.environ['DATA_PATH'] + "/other/sphinx_test_audio/interview.wav"
fps = 100 # default
audio_file = AudioFile(audio_fil... | [
"pocketsphinx.AudioFile"
] | [((301, 344), 'pocketsphinx.AudioFile', 'AudioFile', ([], {'audio_file': 'test_audio', 'frate': '(100)'}), '(audio_file=test_audio, frate=100)\n', (310, 344), False, 'from pocketsphinx import AudioFile\n')] |
from mmdet.apis import init_detector, inference_detector, show_result
import mmcv
import os
import argparse
import numpy as np
from tqdm import tqdm
parser = argparse.ArgumentParser(description='Test different models')
parser.add_argument('--epoch', type=str, default="latest", help='dataset version')
parser... | [
"os.listdir",
"argparse.ArgumentParser",
"mmdet.apis.init_detector",
"tqdm.tqdm",
"os.path.join",
"os.path.split",
"mmdet.apis.inference_detector"
] | [((166, 226), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {'description': '"""Test different models"""'}), "(description='Test different models')\n", (189, 226), False, 'import argparse\n'), ((1216, 1274), 'mmdet.apis.init_detector', 'init_detector', (['config_file', 'checkpoint_file'], {'device': '"""cu... |
from cartodb_services.refactor.storage.redis_connection_config import RedisMetadataConnectionConfigBuilder
from cartodb_services.refactor.storage.redis_connection import RedisConnectionBuilder
from cartodb_services.refactor.storage.redis_config import RedisUserConfigStorageBuilder
class UserConfigBackendFactory(object... | [
"cartodb_services.refactor.storage.redis_config.RedisUserConfigStorageBuilder",
"cartodb_services.refactor.storage.redis_connection.RedisConnectionBuilder",
"cartodb_services.refactor.storage.redis_connection_config.RedisMetadataConnectionConfigBuilder"
] | [((927, 992), 'cartodb_services.refactor.storage.redis_connection_config.RedisMetadataConnectionConfigBuilder', 'RedisMetadataConnectionConfigBuilder', (['self._server_config_backend'], {}), '(self._server_config_backend)\n', (963, 992), False, 'from cartodb_services.refactor.storage.redis_connection_config import Redi... |
import os
import random
from flask import current_app
def save_file(form_file, folder_name):
random_hex = random.token_hex(8)
_, f_ext = os.path.splitext(form_file.filename)
file_fn = random_hex + f_ext
file_path = os.path.join(current_app.root_path, 'static', folder_name, file_fn)
form_file.save(... | [
"os.path.join",
"os.path.splitext",
"random.token_hex"
] | [((112, 131), 'random.token_hex', 'random.token_hex', (['(8)'], {}), '(8)\n', (128, 131), False, 'import random\n'), ((147, 183), 'os.path.splitext', 'os.path.splitext', (['form_file.filename'], {}), '(form_file.filename)\n', (163, 183), False, 'import os\n'), ((233, 300), 'os.path.join', 'os.path.join', (['current_app... |
# -*- coding: utf-8 -*-
# Copyright 2020 EMBL - European Bioinformatics Institute
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless requ... | [
"rest_framework.routers.DefaultRouter"
] | [((815, 858), 'rest_framework.routers.DefaultRouter', 'routers.DefaultRouter', ([], {'trailing_slash': '(False)'}), '(trailing_slash=False)\n', (836, 858), False, 'from rest_framework import routers\n'), ((7304, 7347), 'rest_framework.routers.DefaultRouter', 'routers.DefaultRouter', ([], {'trailing_slash': '(False)'}),... |
# -*- coding: utf-8 -*-
"""
Created on Mon Nov 28 10:47:38 2016
@author: ahefny
Policies are BLIND to the representation of states, which could be (1) observation,
(2) original latent state or (3) predictive state.
Policies takes the "state" dimension x_dim, the number of actions/dim of action as input.
"""
impo... | [
"numpy.sin",
"numpy.prod",
"numpy.dot"
] | [((2094, 2116), 'numpy.dot', 'np.dot', (['self._K', 'state'], {}), '(self._K, state)\n', (2100, 2116), True, 'import numpy as np\n'), ((1598, 1629), 'numpy.prod', 'np.prod', (['(self._high - self._low)'], {}), '(self._high - self._low)\n', (1605, 1629), True, 'import numpy as np\n'), ((2612, 2657), 'numpy.sin', 'np.sin... |
import os
import sys
from datetime import datetime
from subprocess import run
name = datetime.utcnow().strftime("%Y%m%d-%H%M%S.md")
try:
# -t stands for "topic"
topic_index = sys.argv.index('-t')
path = os.path.join(sys.argv[topic_index + 1], name)
except:
path = name
run(['code', path])
| [
"sys.argv.index",
"subprocess.run",
"os.path.join",
"datetime.datetime.utcnow"
] | [((287, 306), 'subprocess.run', 'run', (["['code', path]"], {}), "(['code', path])\n", (290, 306), False, 'from subprocess import run\n'), ((184, 204), 'sys.argv.index', 'sys.argv.index', (['"""-t"""'], {}), "('-t')\n", (198, 204), False, 'import sys\n'), ((216, 261), 'os.path.join', 'os.path.join', (['sys.argv[topic_i... |
# -*- coding: utf-8 -*-
"""module for Superfetch."""
import os, sys
import time
from datetime import datetime, timedelta
from modules import logger
from modules import manager
from modules import interface
from modules.windows_superfetch import sfexport2
from dfvfs.lib import definitions as dfvfs_definitions
class S... | [
"os.path.exists",
"modules.windows_superfetch.sfexport2.main",
"os.mkdir",
"os.path.abspath",
"modules.manager.ModulesManager.RegisterModule",
"os.remove"
] | [((3350, 3408), 'modules.manager.ModulesManager.RegisterModule', 'manager.ModulesManager.RegisterModule', (['SUPERFETCHConnector'], {}), '(SUPERFETCHConnector)\n', (3387, 3408), False, 'from modules import manager\n'), ((3142, 3184), 'os.remove', 'os.remove', (['(output_path + os.sep + fileName)'], {}), '(output_path +... |
from rest_framework.generics import get_object_or_404
from rest_framework.response import Response
from rest_framework.status import HTTP_400_BAD_REQUEST, HTTP_200_OK
from river.models import Function
from river_admin.views import get, post, put, delete
from river_admin.views.serializers import UpdateFunctionDto, Crea... | [
"river_admin.views.get",
"river.models.Function.objects.all",
"river_admin.views.serializers.CreateFunctionDto",
"river_admin.views.put",
"rest_framework.response.Response",
"river_admin.views.serializers.UpdateFunctionDto",
"river_admin.views.delete",
"river_admin.views.post",
"river_admin.views.se... | [((350, 385), 'river_admin.views.get', 'get', (['"""^function/get/(?P<pk>\\\\w+)/$"""'], {}), "('^function/get/(?P<pk>\\\\w+)/$')\n", (353, 385), False, 'from river_admin.views import get, post, put, delete\n'), ((546, 569), 'river_admin.views.get', 'get', (['"""^function/list/$"""'], {}), "('^function/list/$')\n", (54... |
import torch
import numpy as np
from torch.utils.data import Dataset
import torchvision.transforms as transforms
import skimage.io as io
from path import Path
import cv2
import torch.nn.functional as F
class ETH_LFB(Dataset):
def __init__(self, configs):
"""
dataset for eth local feature benchmark
... | [
"torch.from_numpy",
"numpy.array",
"path.Path",
"cv2.SIFT_create",
"skimage.io.imread",
"cv2.cvtColor",
"torchvision.transforms.Normalize",
"torchvision.transforms.ToTensor"
] | [((752, 769), 'cv2.SIFT_create', 'cv2.SIFT_create', ([], {}), '()\n', (767, 769), False, 'import cv2\n'), ((786, 817), 'path.Path', 'Path', (["self.configs['data_path']"], {}), "(self.configs['data_path'])\n", (790, 817), False, 'from path import Path\n'), ((1075, 1089), 'skimage.io.imread', 'io.imread', (['imf'], {}),... |
"""This file contains functions for processing image"""
import cv2
import math
import copy
import numpy as np
import matplotlib.pyplot as plt
def binarize_image(image):
"""Binarize image pixel values to 0 and 255."""
unique_values = np.unique(image)
if len(unique_values) == 2:
if (un... | [
"matplotlib.pyplot.imshow",
"cv2.imwrite",
"numpy.unique",
"numpy.ones",
"numpy.where",
"numpy.zeros_like",
"numpy.count_nonzero",
"numpy.sum",
"numpy.stack",
"numpy.array",
"copy.deepcopy",
"cv2.imread",
"matplotlib.pyplot.show"
] | [((255, 271), 'numpy.unique', 'np.unique', (['image'], {}), '(image)\n', (264, 271), True, 'import numpy as np\n'), ((580, 599), 'cv2.imread', 'cv2.imread', (['path', '(0)'], {}), '(path, 0)\n', (590, 599), False, 'import cv2\n'), ((692, 708), 'cv2.imread', 'cv2.imread', (['path'], {}), '(path)\n', (702, 708), False, '... |
import logging
from logging.config import dictConfig
import dbnd
from dbnd.testing.helpers import run_dbnd_subprocess__with_home
from dbnd_airflow_contrib.dbnd_airflow_default_logger import DEFAULT_LOGGING_CONFIG
class TestDbndAirflowLogging(object):
def test_dbnd_airflow_logging_conifg(self):
# we imp... | [
"logging.config.dictConfig",
"logging.info",
"logging.error"
] | [((919, 942), 'logging.config.dictConfig', 'dictConfig', (['dbnd_config'], {}), '(dbnd_config)\n', (929, 942), False, 'from logging.config import dictConfig\n'), ((947, 984), 'logging.info', 'logging.info', (['"""test_can_be_loaded OK"""'], {}), "('test_can_be_loaded OK')\n", (959, 984), False, 'import logging\n'), ((7... |
#!/usr/bin/env python
# coding: utf-8
# In[1]:
#import bibliotek
from keras.applications.resnet50 import ResNet50, decode_predictions,preprocess_input
from keras.preprocessing import image
import numpy as np
import requests
from io import BytesIO
from PIL import Image
# In[2]:
#podbranie modelu ResNet50
model... | [
"keras.preprocessing.image.img_to_array",
"keras.applications.resnet50.decode_predictions",
"io.BytesIO",
"requests.get",
"numpy.expand_dims",
"keras.applications.resnet50.ResNet50"
] | [((323, 351), 'keras.applications.resnet50.ResNet50', 'ResNet50', ([], {'weights': '"""imagenet"""'}), "(weights='imagenet')\n", (331, 351), False, 'from keras.applications.resnet50 import ResNet50, decode_predictions, preprocess_input\n'), ((1204, 1225), 'requests.get', 'requests.get', (['url_img'], {}), '(url_img)\n'... |
"""
Test all kinds of errors.
@author <EMAIL>
"""
import pytest
from trafficgenerator.tgn_utils import ApiType
from xenavalkyrie.xena_app import init_xena
from xenavalkyrie.xena_object import XenaAttributeError
from .test_base import TestXenaBase
class TestXenaErrors(TestXenaBase):
def setup(self):
su... | [
"pytest.raises"
] | [((656, 680), 'pytest.raises', 'pytest.raises', (['Exception'], {}), '(Exception)\n', (669, 680), False, 'import pytest\n'), ((1415, 1448), 'pytest.raises', 'pytest.raises', (['XenaAttributeError'], {}), '(XenaAttributeError)\n', (1428, 1448), False, 'import pytest\n'), ((1519, 1552), 'pytest.raises', 'pytest.raises', ... |
import time
from datetime import datetime as dt
import colorama
from colorama import Fore, Back, Style
import socket
import os
import sys
def CS(X):
time.sleep(X)
os.system("clear")
def socknames():
myHostName = socket.gethostname()
myIP = socket.gethostbyname(myHostName)
print("\033[35m[\033... | [
"socket.gethostbyname",
"time.sleep",
"datetime.datetime.now",
"os.system",
"socket.gethostname"
] | [((159, 172), 'time.sleep', 'time.sleep', (['X'], {}), '(X)\n', (169, 172), False, 'import time\n'), ((177, 195), 'os.system', 'os.system', (['"""clear"""'], {}), "('clear')\n", (186, 195), False, 'import os\n'), ((231, 251), 'socket.gethostname', 'socket.gethostname', ([], {}), '()\n', (249, 251), False, 'import socke... |
import json
import pandas as pd
from AnalysisModule.prepare.diagram import BuildingUnit
from AnalysisModule.routines.util import read_jsonfile
"""
4 ways to deal with strange bus
A - exclude bu, keep crystals
A' - merge bu, keep crystals
B - exclude crystals
C - keep all
note 2020/11/24:
- all A are modified to... | [
"pandas.DataFrame.from_records",
"json.loads",
"AnalysisModule.routines.util.read_jsonfile",
"pandas.read_csv"
] | [((2992, 3034), 'pandas.DataFrame.from_records', 'pd.DataFrame.from_records', (['curated_records'], {}), '(curated_records)\n', (3017, 3034), True, 'import pandas as pd\n'), ((3229, 3259), 'AnalysisModule.routines.util.read_jsonfile', 'read_jsonfile', (['"""3_bulist.json"""'], {}), "('3_bulist.json')\n", (3242, 3259), ... |
# flake8: noqa
# -*- coding: utf-8 -*-
import datetime
from south.db import db
from south.v2 import SchemaMigration
from django.db import models
class Migration(SchemaMigration):
def forwards(self, orm):
# Adding model 'Milestone'
db.create_table('roadmap_milestone', (
('id', self.gf(... | [
"south.db.db.send_create_signal",
"south.db.db.delete_table"
] | [((472, 519), 'south.db.db.send_create_signal', 'db.send_create_signal', (['"""roadmap"""', "['Milestone']"], {}), "('roadmap', ['Milestone'])\n", (493, 519), False, 'from south.db import db\n'), ((1010, 1068), 'south.db.db.send_create_signal', 'db.send_create_signal', (['"""roadmap"""', "['MilestoneTranslation']"], {}... |
from base import CQPartsTest
from base import testlabel
# units under test
from cqparts_fasteners.fasteners.nutbolt import NutAndBoltFastener
# ---------- Test Assembly ----------
import cadquery
import cqparts
from partslib.basic import Box
from cqparts import constraint
from cqparts.utils import CoordSystem
class... | [
"partslib.basic.Box",
"cqparts.utils.CoordSystem",
"cqparts_fasteners.fasteners.nutbolt.NutAndBoltFastener",
"cqparts.constraint.Coincident",
"cadquery.Vector",
"cqparts.constraint.Fixed"
] | [((403, 438), 'partslib.basic.Box', 'Box', ([], {'length': '(20)', 'width': '(20)', 'height': '(12)'}), '(length=20, width=20, height=12)\n', (406, 438), False, 'from partslib.basic import Box\n'), ((453, 488), 'partslib.basic.Box', 'Box', ([], {'length': '(18)', 'width': '(18)', 'height': '(18)'}), '(length=18, width=... |
# Generated by Django 3.0.8 on 2021-02-20 13:51
import ckeditor.fields
from django.db import migrations, models
import django.db.models.deletion
class Migration(migrations.Migration):
dependencies = [
('eventsapi', '0001_initial'),
]
operations = [
migrations.CreateModel(
na... | [
"django.db.models.OneToOneField",
"django.db.models.DateField",
"django.db.models.TextField",
"django.db.models.ForeignKey",
"django.db.models.AutoField"
] | [((382, 475), 'django.db.models.AutoField', 'models.AutoField', ([], {'auto_created': '(True)', 'primary_key': '(True)', 'serialize': '(False)', 'verbose_name': '"""ID"""'}), "(auto_created=True, primary_key=True, serialize=False,\n verbose_name='ID')\n", (398, 475), False, 'from django.db import migrations, models\... |
"""
This file is for initializing the uWSGI application. It defines a post-fork command in
order to avoid an issue in uWSGI that prevents MongoEngine from lazily creating
mongo connections
"""
from uwsgidecorators import postfork
from mongomail.rest_app import app, db, connection
@postfork
def init_db():
db.init... | [
"mongomail.rest_app.connection.get_api_keys",
"mongomail.rest_app.db.init_app",
"mongomail.rest_app.connection.generate_api_key"
] | [((313, 329), 'mongomail.rest_app.db.init_app', 'db.init_app', (['app'], {}), '(app)\n', (324, 329), False, 'from mongomail.rest_app import app, db, connection\n'), ((370, 395), 'mongomail.rest_app.connection.get_api_keys', 'connection.get_api_keys', ([], {}), '()\n', (393, 395), False, 'from mongomail.rest_app import ... |
import cupy as cp
def get_batches(X,y,batch_size,seed,shuffle):
cp.random.seed(seed)
m = X.shape[0]
mini_batches = []
# 第一步打乱X,Y
if shuffle:
permutation = cp.random.permutation(m) # 返回一个长度为m的list,里面的值为0到m-1
shuffled_X = X[permutation]
shuffled_y = y[permutati... | [
"cupy.random.seed",
"cupy.random.permutation"
] | [((78, 98), 'cupy.random.seed', 'cp.random.seed', (['seed'], {}), '(seed)\n', (92, 98), True, 'import cupy as cp\n'), ((198, 222), 'cupy.random.permutation', 'cp.random.permutation', (['m'], {}), '(m)\n', (219, 222), True, 'import cupy as cp\n')] |
# -*- coding: utf-8 -*-
"""
Plot comparisons of IHME projections to actual data for US states.
IHME data per IHME:
https://covid19.healthdata.org/united-states-of-america
IHME data stored here in the "..\data\ihme" directory for each release
that was obtained.
State-level data per Covid trackin... | [
"read_data.get_data_ctrack",
"os.path.join",
"read_data.format_date_ihme",
"numpy.diff",
"read_data.get_data_ihme",
"numpy.array",
"scipy.signal.medfilt",
"datetime.date.today",
"matplotlib.pyplot.subplots"
] | [((2462, 2474), 'datetime.date.today', 'date.today', ([], {}), '()\n', (2472, 2474), False, 'from datetime import date\n'), ((2506, 2543), 'read_data.get_data_ctrack', 'get_data_ctrack', (['state', 'data_filename'], {}), '(state, data_filename)\n', (2521, 2543), False, 'from read_data import get_data_ctrack, get_data_i... |
from __future__ import absolute_import
from __future__ import division
from __future__ import print_function
import math
import sys
import os.path
from PIL import Image
import logging
import numpy as np
from .base_analyzer import BaseAnnotator
if os.environ.get('PYTORCH_MODE',False):
import dvalib.crnn.utils as ut... | [
"tensorflow.python.training.saver.Saver",
"tensorflow.multiply",
"tensorflow.gfile.FastGFile",
"dvalib.crnn.dataset.resizeNormalize",
"torch.cuda.is_available",
"tensorflow.gfile.GFile",
"logging.info",
"tensorflow.Graph",
"tensorflow.placeholder",
"tensorflow.image.resize_bilinear",
"tensorflow... | [((470, 519), 'logging.info', 'logging.info', (['"""In pytorch mode, not importing TF"""'], {}), "('In pytorch mode, not importing TF')\n", (482, 519), False, 'import logging\n'), ((983, 1009), 'tensorflow.expand_dims', 'tf.expand_dims', (['image', '[0]'], {}), '(image, [0])\n', (997, 1009), True, 'import tensorflow as... |
from datetime import datetime
from typing import Callable
def add(a: int, b: int) -> int:
return a + b
def subtract(a: int, b: int) -> int:
return a - b
def calculate(operation: Callable[[int, int], int],
a: int, b: int) -> int:
"""Demonstration of first class citizen"""
return opera... | [
"datetime.datetime.now"
] | [((705, 719), 'datetime.datetime.now', 'datetime.now', ([], {}), '()\n', (717, 719), False, 'from datetime import datetime\n')] |
import collections
from puzzle.constraints import solution_constraints
from puzzle.steps import generate_solutions
from spec.mamba import *
_SOLUTIONS = collections.OrderedDict((
('early_low', 0.1),
('early_high', 1.0),
('after_early_high', 0.9),
('mid_low', 0.2),
('late_mid', 0.5),
('late_low', 0.3),
(... | [
"collections.OrderedDict",
"puzzle.constraints.solution_constraints.SolutionConstraints",
"puzzle.steps.generate_solutions.GenerateSolutions"
] | [((155, 334), 'collections.OrderedDict', 'collections.OrderedDict', (["(('early_low', 0.1), ('early_high', 1.0), ('after_early_high', 0.9), (\n 'mid_low', 0.2), ('late_mid', 0.5), ('late_low', 0.3), ('late_high', 0.8))"], {}), "((('early_low', 0.1), ('early_high', 1.0), (\n 'after_early_high', 0.9), ('mid_low', 0... |
from torch.utils.data import Dataset
from torchvision.transforms.functional import to_tensor
from utils.image.processor import ImagePreprocessor, colorFormats
from PIL import Image
import glob, random
class ImageData(Dataset):
def __init__(
self, srcPath, crop=True, cropSize=96, colorFromat="RGB", process... | [
"torchvision.transforms.functional.to_tensor",
"PIL.Image.open",
"glob.glob",
"utils.image.processor.ImagePreprocessor"
] | [((396, 438), 'glob.glob', 'glob.glob', (["(srcPath + '/**')"], {'recursive': '(True)'}), "(srcPath + '/**', recursive=True)\n", (405, 438), False, 'import glob, random\n'), ((804, 909), 'utils.image.processor.ImagePreprocessor', 'ImagePreprocessor', ([], {'crop': 'crop', 'cropSize': 'cropSize', 'colorFromat': 'colorFr... |
def main():
from sys import stdin
n = int(stdin.readline())
estrelas = list(map(int, stdin.readline().split()))
total_de_carneiros = sum(estrelas)
i = 0
estrelas_atacadas = list()
carneiros_roubados = 0
while 0 <= i < n:
if i not in estrelas_atacadas:
estrelas_atacada... | [
"sys.stdin.readline"
] | [((50, 66), 'sys.stdin.readline', 'stdin.readline', ([], {}), '()\n', (64, 66), False, 'from sys import stdin\n'), ((97, 113), 'sys.stdin.readline', 'stdin.readline', ([], {}), '()\n', (111, 113), False, 'from sys import stdin\n')] |
"""LC-QuAD 2.0: A Large Scale Complex Question Answering Dataset."""
import json
import os
import datasets
logger = datasets.logging.get_logger(__name__)
_CITATION = """\
@inproceedings{dubey2017lc2,
title={LC-QuAD 2.0: A Large Dataset for Complex Question Answering over Wikidata and DBpedia},
a... | [
"json.load",
"datasets.Value",
"os.path.join",
"datasets.logging.get_logger"
] | [((119, 156), 'datasets.logging.get_logger', 'datasets.logging.get_logger', (['__name__'], {}), '(__name__)\n', (146, 156), False, 'import datasets\n'), ((4285, 4297), 'json.load', 'json.load', (['f'], {}), '(f)\n', (4294, 4297), False, 'import json\n'), ((2262, 2286), 'datasets.Value', 'datasets.Value', (['"""string""... |
# Enter your code here. Read input from STDIN. Print output to STDOUT
from typing import List
from itertools import product
def sum_of_square_mod_m( nums:tuple, mod_m ):
return ( sum( map( lambda x: x**2, nums) ) % mod_m )
def maximize_function_value( list_of_list:List , mod_m):
count_of_list = len( list_... | [
"itertools.product"
] | [((368, 390), 'itertools.product', 'product', (['*list_of_list'], {}), '(*list_of_list)\n', (375, 390), False, 'from itertools import product\n')] |
from unittest import TestCase
from tests import get_data
from pytezos.operation.forge import forge_operation_group
class OperationForgingTestooZWSJ(TestCase):
def setUp(self):
self.maxDiff = None
def test_forge_ooZWSJ(self):
expected = get_data(
path='operations/ooZWSJqG... | [
"tests.get_data"
] | [((273, 373), 'tests.get_data', 'get_data', ([], {'path': '"""operations/ooZWSJqGRqi4hirFmqvFinKp9JGU3vqR23Q15tMvDXMnWtX1jne/forged.hex"""'}), "(path=\n 'operations/ooZWSJqGRqi4hirFmqvFinKp9JGU3vqR23Q15tMvDXMnWtX1jne/forged.hex'\n )\n", (281, 373), False, 'from tests import get_data\n'), ((416, 519), 'tests.get_d... |
from machine import SPI, Pin
import tinypico as TinyPICO
import machine
from dotstar import DotStar
import time, random, micropython, gc
import bluetooth
from ble_simple_peripheral import BLESimplePeripheral
# Configure SPI for controlling the DotStar
# Internally we are using software SPI for this as the... | [
"dotstar.DotStar",
"tinypico.set_dotstar_power",
"bluetooth.BLE",
"machine.Timer",
"machine.deepsleep",
"ble_simple_peripheral.BLESimplePeripheral",
"machine.freq",
"machine.reset_cause",
"machine.Pin",
"tinypico.get_battery_charging",
"tinypico.get_battery_voltage"
] | [((526, 558), 'dotstar.DotStar', 'DotStar', (['spi', '(1)'], {'brightness': '(0.25)'}), '(spi, 1, brightness=0.25)\n', (533, 558), False, 'from dotstar import DotStar\n'), ((648, 670), 'machine.freq', 'machine.freq', (['(80000000)'], {}), '(80000000)\n', (660, 670), False, 'import machine\n'), ((1202, 1217), 'bluetooth... |
from feature import Feature
from itertools import product
import numpy as np
import random
class Node:
def __init__(self, K, Cweights, Dweights, seed):
self.K = K
self.seed = seed
self.Kd = int(K*2/3)
self.Kc = int(K*1/3)
self.Cfeatures = [Feature(False, seed) for k in range(... | [
"numpy.random.rand",
"itertools.product",
"feature.Feature",
"numpy.exp",
"numpy.array",
"numpy.random.randint",
"random.random",
"numpy.arange"
] | [((1003, 1019), 'numpy.random.rand', 'np.random.rand', ([], {}), '()\n', (1017, 1019), True, 'import numpy as np\n'), ((1220, 1233), 'numpy.exp', 'np.exp', (['alpha'], {}), '(alpha)\n', (1226, 1233), True, 'import numpy as np\n'), ((1451, 1463), 'numpy.exp', 'np.exp', (['beta'], {}), '(beta)\n', (1457, 1463), True, 'im... |
##−∗−coding : utf−8−∗−
import sqlite3 as lite
import logging
import sys
from collections import OrderedDict
import conf
LOG_FORMAT = "%(levelname)s > Line:%(lineno)s - %(message)s"
logging.basicConfig(filename="debug.log",
level=logging.DEBUG,
format=LOG_FORMAT,
... | [
"logging.basicConfig",
"logging.getLogger",
"collections.OrderedDict",
"sqlite3.connect",
"sys.exit"
] | [((186, 286), 'logging.basicConfig', 'logging.basicConfig', ([], {'filename': '"""debug.log"""', 'level': 'logging.DEBUG', 'format': 'LOG_FORMAT', 'filemode': '"""w"""'}), "(filename='debug.log', level=logging.DEBUG, format=\n LOG_FORMAT, filemode='w')\n", (205, 286), False, 'import logging\n'), ((373, 400), 'loggin... |
from pathlib import Path
from typing import List
from pmfp.utils.fs_utils import get_abs_path
def _find_pypackage(final_path: Path, packs: List[str]) -> None:
has_init = False
for i in final_path.iterdir():
if i.name == "__init__.py":
has_init = True
if not has_init:
return
... | [
"pmfp.utils.fs_utils.get_abs_path"
] | [((648, 669), 'pmfp.utils.fs_utils.get_abs_path', 'get_abs_path', (['to_path'], {}), '(to_path)\n', (660, 669), False, 'from pmfp.utils.fs_utils import get_abs_path\n')] |
import pydot
from grit.decisiontree.handcrafted_trees import scenario_trees
from grit.core.base import get_img_dir
def build_pydot_tree(graph, root, idx='R'):
node = pydot.Node(idx, label=str(root))
graph.add_node(node)
if root.decision is not None:
true_child = build_pydot_tree(graph, root.decis... | [
"pydot.Dot",
"grit.core.base.get_img_dir"
] | [((930, 961), 'pydot.Dot', 'pydot.Dot', ([], {'graph_type': '"""digraph"""'}), "(graph_type='digraph')\n", (939, 961), False, 'import pydot\n'), ((1024, 1037), 'grit.core.base.get_img_dir', 'get_img_dir', ([], {}), '()\n', (1035, 1037), False, 'from grit.core.base import get_img_dir\n')] |
#https://github.com/Newmu/Theano-Tutorials/blob/master/1_linear_regression.py
import theano
from theano import tensor as T
import numpy as np
trX = np.linspace(-1, 1, 101)
trY = 2 * trX + np.random.randn(*trX.shape) * 0.33
X = T.scalar()
Y = T.scalar()
def model(X, w):
return X * w
w = theano.shared(np.asarray... | [
"theano.tensor.nnet.categorical_crossentropy",
"theano.function",
"matplotlib.pyplot.show",
"theano.tensor.dot",
"numpy.asarray",
"numpy.argmax",
"theano.tensor.sqr",
"numpy.linspace",
"theano.tensor.fmatrix",
"theano.tensor.argmax",
"theano.tensor.scalar",
"numpy.random.randn",
"fuel.datase... | [((150, 173), 'numpy.linspace', 'np.linspace', (['(-1)', '(1)', '(101)'], {}), '(-1, 1, 101)\n', (161, 173), True, 'import numpy as np\n'), ((230, 240), 'theano.tensor.scalar', 'T.scalar', ([], {}), '()\n', (238, 240), True, 'from theano import tensor as T\n'), ((245, 255), 'theano.tensor.scalar', 'T.scalar', ([], {}),... |
"""General functions for working with observations.
"""
import collections
import os
from shapely.geometry import Point
import pandas as pd
from gisutils import df2shp, project
from mfsetup.obs import make_obsname
from mfsetup.units import convert_length_units, convert_volume_units, convert_time_units
from mapgwm.utils... | [
"mfsetup.units.convert_time_units",
"gisutils.project",
"mapgwm.utils.assign_geographic_obsgroups",
"mapgwm.utils.makedirs",
"pandas.read_csv",
"mfsetup.units.convert_volume_units",
"os.path.splitext",
"os.path.split",
"shapely.geometry.Point",
"mapgwm.utils.cull_data_to_active_area",
"pandas.to... | [((14908, 14938), 'pandas.to_datetime', 'pd.to_datetime', (["df['datetime']"], {}), "(df['datetime'])\n", (14922, 14938), True, 'import pandas as pd\n'), ((15056, 15113), 'gisutils.project', 'project', (['(md.x.values, md.y.values)', 'source_crs', 'dest_crs'], {}), '((md.x.values, md.y.values), source_crs, dest_crs)\n'... |
from django.contrib.auth import views as auth_views
from django.urls import path
from django.urls.base import reverse_lazy
from django.views.decorators.csrf import csrf_exempt
from . import views, webhooks
from .forms.authorization import CosmosPasswordChangeForm, CosmosPasswordResetForm, CosmosSetPasswordForm
app_na... | [
"django.contrib.auth.views.PasswordResetDoneView.as_view",
"django.urls.base.reverse_lazy",
"django.contrib.auth.views.PasswordResetCompleteView.as_view",
"django.contrib.auth.views.LogoutView.as_view",
"django.contrib.auth.views.PasswordChangeDoneView.as_view",
"django.urls.path"
] | [((1604, 1656), 'django.urls.path', 'path', (['"""profile/"""', 'views.profile'], {'name': '"""user_profile"""'}), "('profile/', views.profile, name='user_profile')\n", (1608, 1656), False, 'from django.urls import path\n'), ((1662, 1711), 'django.urls.path', 'path', (['"""delete/"""', 'views.delete'], {'name': '"""use... |
import math
import torch
from torch import nn
from torch.nn import CrossEntropyLoss
from custom_model.CXLNetModel import PosNetModel
from transformers import XLNetLMHeadModel, XLNetPreTrainedModel, XLNetModel, Conv1D
from transformers.modeling_utils import prune_conv1d_layer
from transformers.modeling_xlnet import XL... | [
"torch.triu",
"torch.nn.Dropout",
"transformers.Conv1D",
"torch.nn.CrossEntropyLoss",
"torch.sin",
"torch.pow",
"torch.cos",
"torch.arange",
"torch.tril",
"torch.eye",
"torch.nn.LayerNorm",
"transformers.modeling_utils.prune_conv1d_layer",
"torch.matmul",
"torch.nn.Embedding",
"torch.ein... | [((999, 1046), 'torch.nn.Embedding', 'nn.Embedding', (['config.vocab_size', 'config.d_model'], {}), '(config.vocab_size, config.d_model)\n', (1011, 1046), False, 'from torch import nn\n'), ((1314, 1340), 'torch.nn.Dropout', 'nn.Dropout', (['config.dropout'], {}), '(config.dropout)\n', (1324, 1340), False, 'from torch i... |
# -*- coding: utf-8 -*-
"""Make the double periodic shear test grid"""
import matplotlib.pyplot as plt
from configparser import ConfigParser
import numpy as np
import sys
import os
sys.path.append(os.path.abspath("../../.."))
from pycato import *
# Make the empty grid
domain = make_uniform_grid(
n_cells=(256, 25... | [
"numpy.sin",
"numpy.tanh",
"os.path.abspath",
"matplotlib.pyplot.subplots",
"matplotlib.pyplot.show"
] | [((1174, 1221), 'matplotlib.pyplot.subplots', 'plt.subplots', ([], {'figsize': '(18, 8)', 'nrows': '(1)', 'ncols': '(2)'}), '(figsize=(18, 8), nrows=1, ncols=2)\n', (1186, 1221), True, 'import matplotlib.pyplot as plt\n'), ((1735, 1745), 'matplotlib.pyplot.show', 'plt.show', ([], {}), '()\n', (1743, 1745), True, 'impor... |