code
stringlengths
22
1.05M
apis
listlengths
1
3.31k
extract_api
stringlengths
75
3.25M
"""""" # Standard library modules. # Third party modules. import pytest import sqlalchemy # Local modules. import dataclasses_sql # Globals and constants variables. @pytest.fixture def metadata(): engine = sqlalchemy.create_engine("sqlite:///:memory:") return sqlalchemy.MetaData(engine) def test_delete_...
[ "pytest.approx", "dataclasses_sql.update", "sqlalchemy.create_engine", "dataclasses_sql.insert", "sqlalchemy.MetaData", "pytest.raises", "dataclasses_sql.delete" ]
[((216, 262), 'sqlalchemy.create_engine', 'sqlalchemy.create_engine', (['"""sqlite:///:memory:"""'], {}), "('sqlite:///:memory:')\n", (240, 262), False, 'import sqlalchemy\n'), ((274, 301), 'sqlalchemy.MetaData', 'sqlalchemy.MetaData', (['engine'], {}), '(engine)\n', (293, 301), False, 'import sqlalchemy\n'), ((565, 60...
#!/usr/bin/env python # -*- coding: utf-8 -*- from __future__ import annotations import queue from typing import TYPE_CHECKING, TypeVar T = TypeVar("T") if TYPE_CHECKING: SimpleQueue = queue.SimpleQueue else: class FakeGenericMeta(type): def __getitem__(self, item): return self clas...
[ "typing.TypeVar" ]
[((142, 154), 'typing.TypeVar', 'TypeVar', (['"""T"""'], {}), "('T')\n", (149, 154), False, 'from typing import TYPE_CHECKING, TypeVar\n')]
from django.db import models class events(models.Model): id = models.AutoField(primary_key=True) note_id = models.BigIntegerField(null=True, blank=True) tweet_id = models.BigIntegerField() type = models.IntegerField(null=True, blank=True) timestamp = models.DateTimeField() from_user = models.Ch...
[ "django.db.models.IntegerField", "django.db.models.AutoField", "django.db.models.BigIntegerField", "django.db.models.DateTimeField", "django.db.models.CharField" ]
[((67, 101), 'django.db.models.AutoField', 'models.AutoField', ([], {'primary_key': '(True)'}), '(primary_key=True)\n', (83, 101), False, 'from django.db import models\n'), ((116, 161), 'django.db.models.BigIntegerField', 'models.BigIntegerField', ([], {'null': '(True)', 'blank': '(True)'}), '(null=True, blank=True)\n'...
#! /usr/bin/python3 """ __Version__: 0.1 __Author__: <NAME> Data: 15/02/2020 Description: Sample scrip for scan host ports with only buit-in functions This code just works with addresses of v4 family. Python 3.x """ # Import modules import socket import sys import errno import os import argparse import ipaddress # M...
[ "socket.socket", "os.strerror", "argparse.ArgumentParser", "sys.exit" ]
[((2123, 2193), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {'description': '"""Scan ports TCP\nVersion: 0.1"""'}), '(description="""Scan ports TCP\nVersion: 0.1""")\n', (2146, 2193), False, 'import argparse\n'), ((1144, 1193), 'socket.socket', 'socket.socket', (['socket.AF_INET', 'socket.SOCK_STREAM'], ...
#print_hello_friend.py from datetime import datetime print(datetime.now()) print("G'day Mate!")
[ "datetime.datetime.now" ]
[((61, 75), 'datetime.datetime.now', 'datetime.now', ([], {}), '()\n', (73, 75), False, 'from datetime import datetime\n')]
from urllib import request def download_from_url(url, filename): request.urlretrieve(url, filename)
[ "urllib.request.urlretrieve" ]
[((70, 104), 'urllib.request.urlretrieve', 'request.urlretrieve', (['url', 'filename'], {}), '(url, filename)\n', (89, 104), False, 'from urllib import request\n')]
from typing import List from led import Led from transitions import AbstractTransition class Sudden(AbstractTransition): def __init__(self, red: float, green: float, blue: float) -> None: super().__init__() self.target = Led(red, green, blue) @AbstractTransition.brightness.setter def b...
[ "transitions.AbstractTransition.brightness.fset", "led.Led" ]
[((246, 267), 'led.Led', 'Led', (['red', 'green', 'blue'], {}), '(red, green, blue)\n', (249, 267), False, 'from led import Led\n'), ((401, 453), 'transitions.AbstractTransition.brightness.fset', 'AbstractTransition.brightness.fset', (['self', 'brightness'], {}), '(self, brightness)\n', (435, 453), False, 'from transit...
import operator as op import itertools with open("input.txt") as file: data = file.read() shop = """Weapons: Cost Damage Armor Dagger 8 4 0 Shortsword 10 5 0 Warhammer 25 6 0 Longsword 40 7 0 Greataxe 74 8 0 Armor: Cost Damage Armor...
[ "itertools.chain", "operator.itemgetter", "itertools.combinations" ]
[((1917, 1948), 'itertools.chain', 'itertools.chain', (['[None]', 'armors'], {}), '([None], armors)\n', (1932, 1948), False, 'import itertools\n'), ((1982, 2014), 'itertools.combinations', 'itertools.combinations', (['rings', '(2)'], {}), '(rings, 2)\n', (2004, 2014), False, 'import itertools\n'), ((1549, 1576), 'opera...
import sys import os from PyQt5.QtWidgets import QApplication, QMainWindow, QMessageBox, QDialog from PyQt5.QtCore import pyqtSignal from mainUi import Ui_Form from sortUi import sortUi from functools import partial import csvIssue class MyMainForm(QMainWindow, Ui_Form): def __init__(self, parent=None): s...
[ "PyQt5.QtCore.pyqtSignal", "csvIssue.write_spending", "PyQt5.QtWidgets.QMessageBox.critical", "csvIssue.delete_last_line", "csvIssue.init_first_classifier", "csvIssue.delete_classifier", "PyQt5.QtWidgets.QMessageBox.information", "PyQt5.QtWidgets.QMessageBox.question", "csvIssue.get_last_line", "f...
[((2625, 2641), 'PyQt5.QtCore.pyqtSignal', 'pyqtSignal', (['list'], {}), '(list)\n', (2635, 2641), False, 'from PyQt5.QtCore import pyqtSignal\n'), ((5585, 5607), 'PyQt5.QtWidgets.QApplication', 'QApplication', (['sys.argv'], {}), '(sys.argv)\n', (5597, 5607), False, 'from PyQt5.QtWidgets import QApplication, QMainWind...
import unittest import uuid from . import user_util class TestUtilFuncs(unittest.TestCase): def test_hash_and_verify_password(self): passwords = [str(uuid.uuid4()) for i in range(10)] for pw in passwords: self.assertTrue( user_util.verify_password(pw, user_util.hash_p...
[ "uuid.uuid4" ]
[((166, 178), 'uuid.uuid4', 'uuid.uuid4', ([], {}), '()\n', (176, 178), False, 'import uuid\n')]
#!/usr/bin/env python from prometheus_client import start_http_server, Summary import random import argparse import time from prometheus_client import Counter from prometheus_client import Gauge from prometheus_client import Summary from prometheus_client import Histogram import sys import time import json import date...
[ "socket.gethostbyname", "urllib.parse.urlparse", "argparse.ArgumentParser", "datetime.datetime.utcnow", "logging.debug", "threading.RLock", "time.sleep", "json.load", "logging.getLevelName", "prometheus_client.core.REGISTRY.register", "watchdog.observers.Observer", "logging.info", "logging.e...
[((1218, 1235), 'threading.RLock', 'threading.RLock', ([], {}), '()\n', (1233, 1235), False, 'import threading\n'), ((18169, 18179), 'watchdog.observers.Observer', 'Observer', ([], {}), '()\n', (18177, 18179), False, 'from watchdog.observers import Observer\n'), ((18270, 18317), 'prometheus_client.core.REGISTRY.registe...
#!/usr/bin/env pythonw import numpy as np import matplotlib.pyplot as plt def flip_coins(flips = 1000000, bins=100): # Uninformative prior prior = np.ones(bins, dtype='float')/bins likelihood_heads = np.arange(bins)/float(bins) likelihood_tails = 1-likelihood_heads flips = np.random.choice(a=[True...
[ "numpy.ones", "numpy.random.choice", "matplotlib.pyplot.legend", "numpy.sum", "numpy.arange", "matplotlib.pyplot.show" ]
[((954, 996), 'matplotlib.pyplot.legend', 'plt.legend', (['[10, 100, 1000, 10000, 100000]'], {}), '([10, 100, 1000, 10000, 100000])\n', (964, 996), True, 'import matplotlib.pyplot as plt\n'), ((997, 1007), 'matplotlib.pyplot.show', 'plt.show', ([], {}), '()\n', (1005, 1007), True, 'import matplotlib.pyplot as plt\n'), ...
#%% from fireworks import PyTorch_Model, Message, HookedPassThroughPipe, Experiment from fireworks.toolbox import ShufflerPipe, TensorPipe, BatchingPipe, FunctionPipe from fireworks.toolbox.preprocessing import train_test_split from fireworks.extensions import IgniteJunction from fireworks.core import PyTorch_Model im...
[ "torch.nn.CrossEntropyLoss", "torch.max", "fireworks.toolbox.preprocessing.train_test_split", "torch.cuda.is_available", "torchvision.datasets.mnist.FashionMNIST", "matplotlib.pyplot.imshow", "fireworks.Experiment", "fireworks.toolbox.BatchingPipe", "fireworks.Message.from_objects", "fireworks.Mes...
[((801, 846), 'fireworks.Experiment', 'Experiment', (['env_name'], {'description': 'description'}), '(env_name, description=description)\n', (811, 846), False, 'from fireworks import PyTorch_Model, Message, HookedPassThroughPipe, Experiment\n'), ((864, 896), 'os.environ.get', 'env.get', (['"""MNIST_DIR"""', '"""./MNIST...
import os import shutil video_files = ['.webm', '.mkv', '.vob', '.gif', '.avi', '.amv', '.mp4',] audio_files = ['.aif','.cda', '.mid', '.mp3', '.mpa', '.ogg', ] image_files = ['.tif', '.tiff', '.bmp', '.jpg', '.jpeg', '.gif', '.png', '.eps', '.raw', '.cr2', '.nef', '.orf', '.sr2', '.ico'] setup_files = ['....
[ "os.path.exists", "os.listdir", "shutil.move", "os.path.splitext", "os.mkdir" ]
[((822, 838), 'os.listdir', 'os.listdir', (['path'], {}), '(path)\n', (832, 838), False, 'import os\n'), ((716, 736), 'os.path.exists', 'os.path.exists', (['path'], {}), '(path)\n', (730, 736), False, 'import os\n'), ((1089, 1136), 'shutil.move', 'shutil.move', (['(path + file_name)', "(path + 'audio\\\\')"], {}), "(pa...
from pathlib import Path ''' This script creates a new html that has placed the javascript code inline to make a standalone html ''' src = Path.cwd() / 'coldcard_address_generator_html.html' dest = Path.cwd() / 'coldcard_address_generator_html_standalone.html' dest2 = Path.cwd() / 'index.html' # for github pages ...
[ "pathlib.Path.cwd" ]
[((142, 152), 'pathlib.Path.cwd', 'Path.cwd', ([], {}), '()\n', (150, 152), False, 'from pathlib import Path\n'), ((201, 211), 'pathlib.Path.cwd', 'Path.cwd', ([], {}), '()\n', (209, 211), False, 'from pathlib import Path\n'), ((272, 282), 'pathlib.Path.cwd', 'Path.cwd', ([], {}), '()\n', (280, 282), False, 'from pathl...
#!/usr/bin/env python import argparse import logging from csv import DictReader from datetime import date, datetime from pathlib import Path logger = logging.getLogger(__name__) header = """V042 ATaxTool Donations 0.4 D{date:%Y-%m-%d} ^ """ record_layout_1 = """TD N280 C1 L1 ${amount:0.2f} X{payee} ({ein}) ^ """ r...
[ "logging.getLogger", "csv.DictReader", "argparse.ArgumentParser", "pathlib.Path", "datetime.datetime.now", "datetime.date.today" ]
[((152, 179), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (169, 179), False, 'import logging\n'), ((1772, 1797), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {}), '()\n', (1795, 1797), False, 'import argparse\n'), ((436, 452), 'pathlib.Path', 'Path', (['args.input'], {}),...
import sys, numpy from sklearn.feature_extraction.text import TfidfVectorizer from sklearn.ensemble import RandomForestClassifier from sklearn.naive_bayes import MultinomialNB #0=drama,1=comedy,2=animated,3=action/adventure def random_forest_class(raw_test_set): x_train=[] y_train=[] count=0 vectorize...
[ "sklearn.feature_extraction.text.TfidfVectorizer", "sklearn.naive_bayes.MultinomialNB", "sklearn.ensemble.RandomForestClassifier" ]
[((324, 397), 'sklearn.feature_extraction.text.TfidfVectorizer', 'TfidfVectorizer', ([], {'analyzer': '"""word"""', 'stop_words': '"""english"""', 'max_features': '(1024)'}), "(analyzer='word', stop_words='english', max_features=1024)\n", (339, 397), False, 'from sklearn.feature_extraction.text import TfidfVectorizer\n...
from flask import Blueprint, jsonify, request, session from pymongo import DESCENDING from api.views import users from api import collection_users @users.route("/api/register", methods=["POST"]) def user_register(): if request.method == "POST": data = request.get_json() email = data['e...
[ "api.views.users.route", "api.collection_users.find_one", "api.collection_users.update_one", "api.collection_users.find", "flask.session.get", "flask.request.get_json", "api.collection_users.insert_one", "flask.session.clear", "flask.jsonify" ]
[((158, 204), 'api.views.users.route', 'users.route', (['"""/api/register"""'], {'methods': "['POST']"}), "('/api/register', methods=['POST'])\n", (169, 204), False, 'from api.views import users\n'), ((1850, 1900), 'api.views.users.route', 'users.route', (['"""/api/login"""'], {'methods': "['GET', 'POST']"}), "('/api/l...
import time import os import pyglet from gtts import gTTS from pydub import AudioSegment import traceback def play_text(*txts): try: sounds = [] fnames = [] for i, s in enumerate(txts): g = gTTS(text=s, lang='en') fname = 'voice{}.mp3'.format(i) with open...
[ "pydub.AudioSegment.from_mp3", "os.path.join", "pyglet.media.load", "time.sleep", "gtts.gTTS", "traceback.print_exc", "time.time", "os.remove" ]
[((687, 698), 'time.time', 'time.time', ([], {}), '()\n', (696, 698), False, 'import time\n'), ((1188, 1229), 'pyglet.media.load', 'pyglet.media.load', (['fname'], {'streaming': '(False)'}), '(fname, streaming=False)\n', (1205, 1229), False, 'import pyglet\n'), ((1309, 1320), 'time.time', 'time.time', ([], {}), '()\n',...
import numpy as np import pandas as pd def batch_df2batch(df, evaluate_ids=(), n_obs=-1, tform=np.eye(3), is_vehicles_evaluated=False): """ Convert dataframe to SGAN input :param df: :param evaluate_ids: :param n_obs: number of timesteps observed :param tform: :param is_vehicles_evaluat...
[ "numpy.eye", "numpy.unique", "numpy.ones", "numpy.sort", "numpy.stack", "numpy.zeros", "numpy.isnan", "numpy.vstack", "numpy.zeros_like", "numpy.arange" ]
[((97, 106), 'numpy.eye', 'np.eye', (['(3)'], {}), '(3)\n', (103, 106), True, 'import numpy as np\n'), ((707, 761), 'numpy.zeros', 'np.zeros', (['(n_obs, agent_ids.size, 2)'], {'dtype': 'np.float32'}), '((n_obs, agent_ids.size, 2), dtype=np.float32)\n', (715, 761), True, 'import numpy as np\n'), ((775, 796), 'numpy.zer...
""" base.py -- client for the base Rinnai API """ import datetime, json, logging, time import requests from rinnaicontrolr.aws_srp import AWSSRP LOGGER = logging.getLogger('rinnaicontrolr') from rinnaicontrolr.const import ( POOL_ID, CLIENT_ID, POOL_REGION, GRAPHQL_ENDPOINT, SHADOW_ENDPOINT, ...
[ "logging.getLogger", "requests.post", "rinnaicontrolr.aws_srp.AWSSRP", "time.time" ]
[((157, 192), 'logging.getLogger', 'logging.getLogger', (['"""rinnaicontrolr"""'], {}), "('rinnaicontrolr')\n", (174, 192), False, 'import datetime, json, logging, time\n'), ((1083, 1094), 'time.time', 'time.time', ([], {}), '()\n', (1092, 1094), False, 'import datetime, json, logging, time\n'), ((1468, 1589), 'rinnaic...
# Generated by Django 2.1.3 on 2019-02-12 19:18 from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('profile', '0002_auto_20180126_1900'), ] operations = [ migrations.CreateModel( name='GlobalAlert', fields=[ ...
[ "django.db.models.AutoField", "django.db.models.TextField", "django.db.models.BooleanField" ]
[((335, 428), 'django.db.models.AutoField', 'models.AutoField', ([], {'auto_created': '(True)', 'primary_key': '(True)', 'serialize': '(False)', 'verbose_name': '"""ID"""'}), "(auto_created=True, primary_key=True, serialize=False,\n verbose_name='ID')\n", (351, 428), False, 'from django.db import migrations, models\...
import click from agent.pipeline.validators import elastic_query, jdbc_query from agent import source class BaseValidator: @staticmethod def validate(pipeline): pass class ElasticValidator(BaseValidator): @staticmethod def validate(pipeline): with open(pipeline.config['query_file'])...
[ "agent.pipeline.validators.jdbc_query.get_errors", "click.ClickException", "agent.pipeline.validators.elastic_query.get_errors" ]
[((373, 475), 'agent.pipeline.validators.elastic_query.get_errors', 'elastic_query.get_errors', (['query', 'pipeline.source.config[source.ElasticSource.CONFIG_OFFSET_FIELD]'], {}), '(query, pipeline.source.config[source.ElasticSource\n .CONFIG_OFFSET_FIELD])\n', (397, 475), False, 'from agent.pipeline.validators imp...
import numpy as npy def convert(num): if num < 0: # num = -num num *= 1024 # num += 32768 num = int(num - 0.5) num = 65535 + num n_str = str(hex(num))[2:] if len(n_str) == 1: n_str = 'fff' + n_str elif len(n_str) == 2: n_str = ...
[ "numpy.load" ]
[((849, 879), 'numpy.load', 'npy.load', (['"""dense_kernel_0.npy"""'], {}), "('dense_kernel_0.npy')\n", (857, 879), True, 'import numpy as npy\n'), ((1175, 1203), 'numpy.load', 'npy.load', (['"""dense_bias_0.npy"""'], {}), "('dense_bias_0.npy')\n", (1183, 1203), True, 'import numpy as npy\n'), ((1330, 1362), 'numpy.loa...
import os import shutil from dockerspawner import DockerSpawner class IllumiDeskDockerSpawner(DockerSpawner): """ Custom DockerSpawner which assigns a user notebook image based on the user's role. This spawner requires: 1. That the `Authenticator.enable_auth_state = True` 2. That the user's ...
[ "os.path.exists", "os.path.join", "os.environ.get", "os.chmod", "os.mkdir" ]
[((2563, 2594), 'os.path.join', 'os.path.join', (['"""/home"""', 'username'], {}), "('/home', username)\n", (2575, 2594), False, 'import os\n'), ((913, 952), 'os.environ.get', 'os.environ.get', (['"""DOCKER_STANDARD_IMAGE"""'], {}), "('DOCKER_STANDARD_IMAGE')\n", (927, 952), False, 'import os\n'), ((2610, 2635), 'os.pa...
from django.db import models from app_asset.models import Host # Create your models here. class Project(models.Model): project_name = models.CharField(max_length=32,unique=True) project_msg = models.CharField(max_length=64,null=True) def __unicode__(self): return self.project_name class GitCode(...
[ "django.db.models.DateTimeField", "django.db.models.TextField", "django.db.models.CharField", "django.db.models.ForeignKey" ]
[((141, 185), 'django.db.models.CharField', 'models.CharField', ([], {'max_length': '(32)', 'unique': '(True)'}), '(max_length=32, unique=True)\n', (157, 185), False, 'from django.db import models\n'), ((203, 245), 'django.db.models.CharField', 'models.CharField', ([], {'max_length': '(64)', 'null': '(True)'}), '(max_l...
#/usr/bin/env python import sys import logging logger = logging.getLogger('utility_to_osm.ssr2.git_diff') import utility_to_osm.file_util as file_util from osmapis_stedsnr import OSMstedsnr if __name__ == '__main__': logging.basicConfig(level=logging.DEBUG) # diff is called by git with 7 parameters: ...
[ "logging.getLogger", "osmapis_stedsnr.OSMstedsnr.from_xml", "utility_to_osm.file_util.read_file", "logging.basicConfig" ]
[((56, 105), 'logging.getLogger', 'logging.getLogger', (['"""utility_to_osm.ssr2.git_diff"""'], {}), "('utility_to_osm.ssr2.git_diff')\n", (73, 105), False, 'import logging\n'), ((223, 263), 'logging.basicConfig', 'logging.basicConfig', ([], {'level': 'logging.DEBUG'}), '(level=logging.DEBUG)\n', (242, 263), False, 'im...
#!/usr/bin/env python3 import uuid from passlib.hash import pbkdf2_sha512 password = input('Enter password: ') password_parts = pbkdf2_sha512.encrypt(password, salt_size=32).split('$') password = password_parts[4] salt = password_parts[3] def convert_b64(input): return input.replace('.', '+') + '=' print('Passw...
[ "passlib.hash.pbkdf2_sha512.encrypt" ]
[((131, 176), 'passlib.hash.pbkdf2_sha512.encrypt', 'pbkdf2_sha512.encrypt', (['password'], {'salt_size': '(32)'}), '(password, salt_size=32)\n', (152, 176), False, 'from passlib.hash import pbkdf2_sha512\n')]
import yaml import os import subprocess import coloredlogs import logging import psutil import shutil import hashlib import uuid import fileinput import requests from nvc import __appname__ from dotenv import load_dotenv import git app_root = os.path.dirname(os.path.abspath(__file__)) app_home = os.path.expanduser("~...
[ "requests.post", "git.Repo.clone_from", "yaml.load", "psutil.virtual_memory", "logging.info", "logging.error", "logging.warn", "os.path.isdir", "os.path.expanduser", "yaml.dump", "requests.get", "uuid.uuid4", "os.path.isfile", "shutil.copyfile", "coloredlogs.install", "psutil.disk_usag...
[((299, 322), 'os.path.expanduser', 'os.path.expanduser', (['"""~"""'], {}), "('~')\n", (317, 322), False, 'import os\n'), ((333, 344), 'os.getcwd', 'os.getcwd', ([], {}), '()\n', (342, 344), False, 'import os\n'), ((261, 286), 'os.path.abspath', 'os.path.abspath', (['__file__'], {}), '(__file__)\n', (276, 286), False,...
#!/usr/bin/env python #-*- coding:utf-8; mode:python; indent-tabs-mode: nil; c-basic-offset: 2; tab-width: 2 -*- from collections import namedtuple import os.path as path from bes.testing.program_unit_test import program_unit_test from bes.fs.file_util import file_util from bes.system.host import host class test_cli...
[ "os.path.join", "bes.testing.program_unit_test.program_unit_test.main" ]
[((3103, 3127), 'bes.testing.program_unit_test.program_unit_test.main', 'program_unit_test.main', ([], {}), '()\n', (3125, 3127), False, 'from bes.testing.program_unit_test import program_unit_test\n'), ((2393, 2421), 'os.path.join', 'path.join', (['tmp', '"""kitchen.py"""'], {}), "(tmp, 'kitchen.py')\n", (2402, 2421),...
import sys sys.path.append('../../Data Structures') from stack import Stack def isPalindrome(linkedList={}): ''' Solution 1 - Hash map Complexity Analysis O(n) time | O(n) space Check if a linked list is a palindrome dict: linkedList return: True if its palindrome ''' # Graceful...
[ "stack.Stack", "sys.path.append" ]
[((11, 51), 'sys.path.append', 'sys.path.append', (['"""../../Data Structures"""'], {}), "('../../Data Structures')\n", (26, 51), False, 'import sys\n'), ((2671, 2678), 'stack.Stack', 'Stack', ([], {}), '()\n', (2676, 2678), False, 'from stack import Stack\n')]
""" Class Report: Part 4 of the Sprint Challenge - Generate random Product list, and get an Inventory Report on that list """ from random import randint, sample, uniform from acme import Product ADJECTIVES = ['Awesome', 'Shiny', 'Impressive', 'Portable', 'Improved'] NOUNS = ['Anvil', 'Catapult', 'Disguise', 'Mousetra...
[ "acme.Product", "random.uniform", "random.sample", "random.randint" ]
[((479, 500), 'random.sample', 'sample', (['ADJECTIVES', '(1)'], {}), '(ADJECTIVES, 1)\n', (485, 500), False, 'from random import randint, sample, uniform\n'), ((516, 532), 'random.sample', 'sample', (['NOUNS', '(1)'], {}), '(NOUNS, 1)\n', (522, 532), False, 'from random import randint, sample, uniform\n'), ((644, 659)...
""" The Flaskee is an Open Source project for Microservices. Develop By <NAME> | https://nadeengamage.com | <EMAIL> """ from werkzeug.serving import run_simple from werkzeug.middleware.dispatcher import DispatcherMiddleware from flaskee import api app = api.create_app() application = DispatcherMiddleware(app) if _...
[ "werkzeug.serving.run_simple", "flaskee.api.create_app", "werkzeug.middleware.dispatcher.DispatcherMiddleware" ]
[((257, 273), 'flaskee.api.create_app', 'api.create_app', ([], {}), '()\n', (271, 273), False, 'from flaskee import api\n'), ((289, 314), 'werkzeug.middleware.dispatcher.DispatcherMiddleware', 'DispatcherMiddleware', (['app'], {}), '(app)\n', (309, 314), False, 'from werkzeug.middleware.dispatcher import DispatcherMidd...
from django.urls import path, include, re_path from django.views.generic import FormView from . import views from .views import news, faq, resources, group, user, puzzles, submission, score_challenge, metrics import publications.views as plist news_patterns = [ path('', news.List.as_view(), name="news_list"), ...
[ "django.urls.re_path", "django.urls.path", "django.urls.include" ]
[((2178, 2294), 'django.urls.re_path', 're_path', (['"""^unconfirmed/(?P<pk>\\\\d+)/user_confirmed/$"""', 'user.UnconfirmedList.user_confirm'], {'name': '"""user_confirmed"""'}), "('^unconfirmed/(?P<pk>\\\\d+)/user_confirmed/$', user.UnconfirmedList.\n user_confirm, name='user_confirmed')\n", (2185, 2294), False, 'f...
import numpy as np import queue import cv2 import os import datetime SIZE = 32 SCALE = 0.007874015748031496 def quantized_np(array,scale,data_width=8): quantized_array= np.round(array/scale) quantized_array = np.maximum(quantized_array, -2**(data_width-1)) quantized_array = np.minimum(quantized_array, 2**...
[ "os.listdir", "numpy.reshape", "numpy.minimum", "cv2.resize", "numpy.size", "os.path.join", "queue.Queue", "datetime.datetime.now", "numpy.maximum", "cv2.imread", "numpy.round" ]
[((175, 198), 'numpy.round', 'np.round', (['(array / scale)'], {}), '(array / scale)\n', (183, 198), True, 'import numpy as np\n'), ((219, 270), 'numpy.maximum', 'np.maximum', (['quantized_array', '(-2 ** (data_width - 1))'], {}), '(quantized_array, -2 ** (data_width - 1))\n', (229, 270), True, 'import numpy as np\n'),...
"""Packager for cloud environment.""" from setuptools import setup, find_packages setup( name='preprocess', version='1.0.0', packages=find_packages(), install_requires=[ 'tensorflow', 'numpy', ], )
[ "setuptools.find_packages" ]
[((147, 162), 'setuptools.find_packages', 'find_packages', ([], {}), '()\n', (160, 162), False, 'from setuptools import setup, find_packages\n')]
import numpy as np import time def max_subsequence_sum(sequence): max_sum = 0 for i in range(0, len(sequence)): for j in range(i, len(sequence)): this_sum = 0 for k in range(i, j+1): this_sum += sequence[k] if this_sum > max_sum: ...
[ "numpy.random.randint", "time.time" ]
[((372, 417), 'numpy.random.randint', 'np.random.randint', (['(-100000)', '(100000)'], {'size': '(1000)'}), '(-100000, 100000, size=1000)\n', (389, 417), True, 'import numpy as np\n'), ((424, 435), 'time.time', 'time.time', ([], {}), '()\n', (433, 435), False, 'import time\n'), ((476, 487), 'time.time', 'time.time', ([...
import pygame from tools import render_text from graph_object import GraphObject class Edge (GraphObject): def __init__(self, v1, v2, weight=0, width=1, color=(0, 0, 0)): super().__init__() self.v1, self.v2 = v1, v2 self.__weight, self.__weight_surface = 0, None self.set_weight(we...
[ "pygame.draw.line" ]
[((766, 837), 'pygame.draw.line', 'pygame.draw.line', (['surface', 'self.color', 'self.pos1', 'self.pos2', 'self.width'], {}), '(surface, self.color, self.pos1, self.pos2, self.width)\n', (782, 837), False, 'import pygame\n')]
from py_profiler import profiler, profiling_service @profiler('hello') def hello(): print('hello') class Foo: @profiler('Food.some_thing') def some_thing(self): print('some_thing') @profiler() def method_2(self): print('method_2') raise Exception('aaaa') if __name__ =...
[ "py_profiler.profiling_service.as_table", "py_profiler.profiler" ]
[((55, 72), 'py_profiler.profiler', 'profiler', (['"""hello"""'], {}), "('hello')\n", (63, 72), False, 'from py_profiler import profiler, profiling_service\n'), ((124, 151), 'py_profiler.profiler', 'profiler', (['"""Food.some_thing"""'], {}), "('Food.some_thing')\n", (132, 151), False, 'from py_profiler import profiler...
# -*- coding: utf-8 -*- # # Copyright 2021 Google LLC. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless requir...
[ "googlecloudsdk.core.log.status.Print", "googlecloudsdk.command_lib.storage.tasks.task_executor.execute_tasks", "googlecloudsdk.command_lib.storage.tasks.task_graph_executor.multiprocessing_context.Queue", "googlecloudsdk.command_lib.storage.flags.add_precondition_flags", "googlecloudsdk.command_lib.storage...
[((4001, 4037), 'googlecloudsdk.command_lib.storage.flags.add_precondition_flags', 'flags.add_precondition_flags', (['parser'], {}), '(parser)\n', (4029, 4037), False, 'from googlecloudsdk.command_lib.storage import flags\n'), ((4466, 4635), 'googlecloudsdk.command_lib.storage.name_expansion.NameExpansionIterator', 'na...
import sqlite3 import sys import datetime import os # day of month to switch to new database change_day = 1 c_added_text_entities = [["sensor.hitachi_relay", "sensor.netatmo_relay"], ["sensor.cooling_target_temp", "sensor.heating_target_temp"]] user = "pipacsba" server_ip = "192.16...
[ "sqlite3.connect", "datetime.datetime.strptime", "os.rename", "os.chdir", "datetime.datetime.now", "os.system", "datetime.timedelta" ]
[((486, 504), 'os.chdir', 'os.chdir', (['work_dir'], {}), '(work_dir)\n', (494, 504), False, 'import os\n'), ((701, 723), 'os.system', 'os.system', (['scp_command'], {}), '(scp_command)\n', (710, 723), False, 'import os\n'), ((6526, 6554), 'os.system', 'os.system', (["('rm ' + file_from)"], {}), "('rm ' + file_from)\n"...
#! /usr/bin/env python # -*- coding: utf-8 -*- #################### import time import logging import indigo from os.path import exists import paho.mqtt.client as mqtt ################################################################################ class MQTTBroker(object): def __init__(self, device): s...
[ "logging.getLogger", "os.path.exists", "indigo.server.getInstallFolderPath", "indigo.activePlugin.processReceivedMessage" ]
[((333, 371), 'logging.getLogger', 'logging.getLogger', (['"""Plugin.MQTTBroker"""'], {}), "('Plugin.MQTTBroker')\n", (350, 371), False, 'import logging\n'), ((6149, 6226), 'indigo.activePlugin.processReceivedMessage', 'indigo.activePlugin.processReceivedMessage', (['self.deviceID', 'msg.topic', 'payload'], {}), '(self...
#!/usr/bin/env python # pylint: disable=too-many-locals,arguments-differ,unused-import import tensorflow as tf from tensorflow.keras.layers import (BatchNormalization, Dense, Dropout, Flatten, MaxPooling2D, SpatialDropout2D, add) from tensorflo...
[ "deep_model_blocks.ConvBlock", "tensorflow.image.resize_nearest_neighbor", "deep_model_blocks.DeconvBlock", "tensorflow.nn.relu", "tensorflow.logging.info", "tensorflow.keras.layers.Dropout", "tensorflow.keras.layers.BatchNormalization", "deep_model_blocks.ResizeBlock", "tensorflow.concat", "tenso...
[((862, 913), 'tensorflow.keras.layers.Dense', 'tf.keras.layers.Dense', (['(15 * 20 * 64)'], {'use_bias': '(False)'}), '(15 * 20 * 64, use_bias=False)\n', (883, 913), True, 'import tensorflow as tf\n'), ((936, 972), 'tensorflow.keras.layers.BatchNormalization', 'tf.keras.layers.BatchNormalization', ([], {}), '()\n', (9...
#!/usr/bin/env python # Licensed under a 3-clause BSD style license - see LICENSE.rst from setuptools import setup, find_packages setup( use_scm_version={ "version_scheme": "post-release", "local_scheme": "dirty-tag", }, setup_requires=["setuptools_scm"], packages=find_packages("src"),...
[ "setuptools.find_packages" ]
[((299, 319), 'setuptools.find_packages', 'find_packages', (['"""src"""'], {}), "('src')\n", (312, 319), False, 'from setuptools import setup, find_packages\n')]
import numpy as np from pyFAI.multi_geometry import MultiGeometry from pyFAI.ext import splitBBox def inpaint_saxs(imgs, ais, masks): """ Inpaint the 2D image collected by the pixel detector to remove artifacts in later data reduction Parameters: ----------- :param imgs: List of 2D image in pixel...
[ "numpy.mean", "numpy.shape", "numpy.ones_like", "numpy.sqrt", "numpy.asarray", "numpy.ma.masked_where", "numpy.max", "numpy.deg2rad", "numpy.linspace", "numpy.arctan2", "numpy.min", "numpy.meshgrid", "pyFAI.multi_geometry.MultiGeometry", "numpy.rad2deg", "numpy.ma.masked_array" ]
[((1972, 2108), 'pyFAI.multi_geometry.MultiGeometry', 'MultiGeometry', (['ais'], {'unit': '"""q_A^-1"""', 'radial_range': 'radial_range', 'azimuth_range': 'azimuth_range', 'wavelength': 'None', 'empty': '(0.0)', 'chi_disc': '(180)'}), "(ais, unit='q_A^-1', radial_range=radial_range, azimuth_range=\n azimuth_range, w...
import numpy as np from typing import Tuple import plotly.io from IMLearn.metalearners.adaboost import AdaBoost from IMLearn.learners.classifiers import DecisionStump from IMLearn.metrics import accuracy from utils import * import plotly.graph_objects as go from plotly.subplots import make_subplots plotly.io.rendere...
[ "numpy.ones", "numpy.random.rand", "plotly.graph_objects.Layout", "IMLearn.metalearners.adaboost.AdaBoost", "numpy.max", "numpy.array", "numpy.sum", "numpy.random.seed" ]
[((3011, 3047), 'numpy.array', 'np.array', (["['circle', 'x', 'diamond']"], {}), "(['circle', 'x', 'diamond'])\n", (3019, 3047), True, 'import numpy as np\n'), ((6267, 6284), 'numpy.random.seed', 'np.random.seed', (['(0)'], {}), '(0)\n', (6281, 6284), True, 'import numpy as np\n'), ((1056, 1066), 'numpy.ones', 'np.ones...
# import packages from airflow import DAG from airflow.operators.bash_operator import BashOperator from airflow.operators.python_operator import PythonOperator from datetime import datetime, timedelta from external_func import random_date def start_print(): print('\nDAG starting...\n') def end_print(): prin...
[ "datetime.datetime", "datetime.timedelta", "airflow.operators.python_operator.PythonOperator", "airflow.operators.bash_operator.BashOperator" ]
[((957, 1015), 'airflow.operators.bash_operator.BashOperator', 'BashOperator', ([], {'task_id': '"""Date"""', 'bash_command': '"""date"""', 'dag': 'dag'}), "(task_id='Date', bash_command='date', dag=dag)\n", (969, 1015), False, 'from airflow.operators.bash_operator import BashOperator\n'), ((1035, 1103), 'airflow.opera...
import math z = [1.0,1 ,1, 1.0] z_exp = [math.exp(i) for i in z] print([round(i, 2) for i in z_exp]) sum_z_exp = sum(z_exp) print(round(sum_z_exp, 2)) softmax = [round(i / sum_z_exp, 3) for i in z_exp] print(softmax)
[ "math.exp" ]
[((42, 53), 'math.exp', 'math.exp', (['i'], {}), '(i)\n', (50, 53), False, 'import math\n')]
import logging logger = logging.getLogger(__name__) logger.debug("Loaded " + __name__) from jsonrpcserver import methods from .exceptions import * from .influxdb_api import * from .meta import *
[ "logging.getLogger" ]
[((24, 51), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (41, 51), False, 'import logging\n')]
import os import pandas as pd import datetime from genderperformr import GenderPerformr from agreementr import Agreementr from politenessr import Politenessr from supportr import Supportr import enchant import requests import json from googleapiclient import discovery from enchant.checker import SpellChecker from encha...
[ "supportr.Supportr", "nltk.download", "datetime.datetime.strptime", "politenessr.Politenessr", "genderperformr.GenderPerformr", "time.strftime", "enchant.tokenize.get_tokenizer", "nltk.tokenize.word_tokenize", "googleapiclient.discovery.build", "agreementr.Agreementr", "time.localtime", "encha...
[((417, 439), 'nltk.download', 'nltk.download', (['"""punkt"""'], {}), "('punkt')\n", (430, 439), False, 'import nltk\n'), ((477, 496), 'nltk.tokenize.word_tokenize', 'word_tokenize', (['text'], {}), '(text)\n', (490, 496), False, 'from nltk.tokenize import word_tokenize\n'), ((1341, 1375), 'time.localtime', 'time.loca...
from PyQt5 import QtGui, QtCore, QtWidgets from PyQt5.uic import loadUi import sys class End(QtWidgets.QDialog): def __init__(self, winners): super(QtWidgets.QDialog, self).__init__() loadUi("End/end.ui", self) msg = f""" <p style="text-align: center; font-size: 30px;"><em>...
[ "PyQt5.uic.loadUi", "PyQt5.QtWidgets.QStackedWidget", "PyQt5.QtWidgets.QApplication" ]
[((766, 798), 'PyQt5.QtWidgets.QApplication', 'QtWidgets.QApplication', (['sys.argv'], {}), '(sys.argv)\n', (788, 798), False, 'from PyQt5 import QtGui, QtCore, QtWidgets\n'), ((812, 838), 'PyQt5.QtWidgets.QStackedWidget', 'QtWidgets.QStackedWidget', ([], {}), '()\n', (836, 838), False, 'from PyQt5 import QtGui, QtCore...
import os srcfile = 'DocTools~/assetgraph_from_gdoc.md' pnglist = 'DocTools~/order.txt' dstfile = 'Documentation~/assetgraph.md' num = 1 if os.path.exists(dstfile): os.remove(dstfile) with open(srcfile) as f: doc = f.read() f.close() with open(pnglist) as fpng: while True: pnglist = fpng.readline() if n...
[ "os.path.exists", "os.remove" ]
[((143, 166), 'os.path.exists', 'os.path.exists', (['dstfile'], {}), '(dstfile)\n', (157, 166), False, 'import os\n'), ((169, 187), 'os.remove', 'os.remove', (['dstfile'], {}), '(dstfile)\n', (178, 187), False, 'import os\n')]
import os # isort: off # This import has to come before the CheckmateClient import or the functional # tests break. # See https://github.com/gevent/gevent/issues/1016 import pywb.apps.frontendapp # pylint:disable=unused-import # isort: on import httpretty as httpretty_ import pytest import webtest from tests.conft...
[ "viahtml.app.Application", "httpretty.disable", "httpretty.enable", "webtest.TestApp", "tests.conftest.environment_variables", "os.environ.update", "tests.simple_server.serve_content", "pytest.fixture", "httpretty.reset" ]
[((438, 469), 'pytest.fixture', 'pytest.fixture', ([], {'scope': '"""session"""'}), "(scope='session')\n", (452, 469), False, 'import pytest\n'), ((608, 639), 'pytest.fixture', 'pytest.fixture', ([], {'scope': '"""session"""'}), "(scope='session')\n", (622, 639), False, 'import pytest\n'), ((1019, 1064), 'pytest.fixtur...
# -*- coding: utf-8 -*- # # Author: <NAME> <<EMAIL>> # # Setup the SMRT module from __future__ import print_function, absolute_import, division from distutils.command.clean import clean # from setuptools import setup # DO NOT use setuptools!!!!!! import shutil import os import sys if sys.version_info[0] < 3: imp...
[ "os.path.exists", "distutils.command.clean.clean.run", "distutils.core.setup", "os.path.join", "numpy.distutils.misc_util.Configuration", "os.path.splitext", "os.path.dirname", "os.unlink", "shutil.rmtree", "os.walk" ]
[((3093, 3138), 'numpy.distutils.misc_util.Configuration', 'Configuration', (['None', 'parent_package', 'top_path'], {}), '(None, parent_package, top_path)\n', (3106, 3138), False, 'from numpy.distutils.misc_util import Configuration\n'), ((5720, 5737), 'distutils.core.setup', 'setup', ([], {}), '(**metadata)\n', (5725...
import numpy import re with open('/home/johnny/Documents/navigate_building/source/assets/images_with_plaques.txt') as f: LIST_OF_POSITIVES = f.read().split('\n') class ImageDetectionMetadata(): headers = ['label', 'parsed_text', 'found_contour_area', 'ref_contour_area', 'source_image_location', 'image', 'ima...
[ "re.findall" ]
[((1650, 1704), 're.findall', 're.findall', (['"""-[0-9,a-z]*-"""', 'self.source_image_location'], {}), "('-[0-9,a-z]*-', self.source_image_location)\n", (1660, 1704), False, 'import re\n')]
# INFO : ini merupakan copy source code dari repo one4ubot, dan sudah mendapatkan izin dari pemilik. # INFO : This is a copy of the source code from the One4ubot repo, and has the permission of the owner. try: from userbot.modules.sql_helper import SESSION, BASE except ImportError: raise AttributeError from sq...
[ "userbot.modules.sql_helper.SESSION.commit", "userbot.modules.sql_helper.SESSION.close", "userbot.modules.sql_helper.SESSION.delete", "userbot.modules.sql_helper.SESSION.query", "userbot.modules.sql_helper.SESSION.add", "sqlalchemy.String" ]
[((765, 783), 'userbot.modules.sql_helper.SESSION.add', 'SESSION.add', (['adder'], {}), '(adder)\n', (776, 783), False, 'from userbot.modules.sql_helper import SESSION, BASE\n'), ((788, 804), 'userbot.modules.sql_helper.SESSION.commit', 'SESSION.commit', ([], {}), '()\n', (802, 804), False, 'from userbot.modules.sql_he...
#!/usr/bin/env python3 """Switch variable case. A function that takes camel cased strings (i.e. ThisIsCamelCased), and converts them to snake case (i.e. this_is_camel_cased). """ import re def snake_case(input_str: str, camel_case=False) -> str: """ Turn camel case into snake case. :param input_str: ...
[ "re.sub" ]
[((584, 626), 're.sub', 're.sub', (['regex', 'substitution', 'input_str[1:]'], {}), '(regex, substitution, input_str[1:])\n', (590, 626), False, 'import re\n')]
import numpy as np import tensorflow as tf import unittest hungarian_module = tf.load_op_library("hungarian.so") class HungarianTests(unittest.TestCase): def test_min_weighted_bp_cover_1(self): W = np.array([[3, 2, 2], [1, 2, 0], [2, 2, 1]]) M, c_0, c_1 = hungarian_module.hungarian(W) with tf.Session()...
[ "tensorflow.load_op_library", "numpy.round", "tensorflow.Session", "numpy.array", "unittest.TextTestRunner", "unittest.TestLoader" ]
[((78, 112), 'tensorflow.load_op_library', 'tf.load_op_library', (['"""hungarian.so"""'], {}), "('hungarian.so')\n", (96, 112), True, 'import tensorflow as tf\n'), ((207, 250), 'numpy.array', 'np.array', (['[[3, 2, 2], [1, 2, 0], [2, 2, 1]]'], {}), '([[3, 2, 2], [1, 2, 0], [2, 2, 1]])\n', (215, 250), True, 'import nump...
import numpy as np import matplotlib.pyplot as plt import seaborn as sns import xarray as xr sns.set() def plot_range(xlabel, ylabel, title, x, values): """x and values should have the same size""" plt.plot(x, values, 'r-', linewidth=2) plt.gcf().set_size_inches(8, 2) plt.title(title) plt.xlabel(...
[ "seaborn.set", "matplotlib.pyplot.ylabel", "matplotlib.pyplot.gcf", "matplotlib.pyplot.xlabel", "matplotlib.pyplot.plot", "matplotlib.pyplot.figure", "matplotlib.pyplot.title", "xarray.open_dataset", "numpy.arange", "matplotlib.pyplot.show" ]
[((93, 102), 'seaborn.set', 'sns.set', ([], {}), '()\n', (100, 102), True, 'import seaborn as sns\n'), ((209, 247), 'matplotlib.pyplot.plot', 'plt.plot', (['x', 'values', '"""r-"""'], {'linewidth': '(2)'}), "(x, values, 'r-', linewidth=2)\n", (217, 247), True, 'import matplotlib.pyplot as plt\n'), ((288, 304), 'matplot...
import pandas as pd import numpy as np import math from nltk.stem.snowball import SnowballStemmer def add_prefix(prefix, series): """ Returns a pandas series that adds a prefix to a string :param prefix: str :return: pd.Series """ if type(prefix) != str: raise TypeError(prefix + ' is n...
[ "nltk.stem.snowball.SnowballStemmer", "math.isnan" ]
[((3737, 3763), 'nltk.stem.snowball.SnowballStemmer', 'SnowballStemmer', (['"""english"""'], {}), "('english')\n", (3752, 3763), False, 'from nltk.stem.snowball import SnowballStemmer\n'), ((2990, 3003), 'math.isnan', 'math.isnan', (['x'], {}), '(x)\n', (3000, 3003), False, 'import math\n')]
import os from pocketsphinx import AudioFile from pocketsphinx import Pocketsphinx from src import util test_video = os.environ['DATA_PATH'] + "/other/sphinx_test_video/beachball.mp4" test_audio = os.environ['DATA_PATH'] + "/other/sphinx_test_audio/interview.wav" fps = 100 # default audio_file = AudioFile(audio_fil...
[ "pocketsphinx.AudioFile" ]
[((301, 344), 'pocketsphinx.AudioFile', 'AudioFile', ([], {'audio_file': 'test_audio', 'frate': '(100)'}), '(audio_file=test_audio, frate=100)\n', (310, 344), False, 'from pocketsphinx import AudioFile\n')]
from mmdet.apis import init_detector, inference_detector, show_result import mmcv import os import argparse import numpy as np from tqdm import tqdm parser = argparse.ArgumentParser(description='Test different models') parser.add_argument('--epoch', type=str, default="latest", help='dataset version') parser...
[ "os.listdir", "argparse.ArgumentParser", "mmdet.apis.init_detector", "tqdm.tqdm", "os.path.join", "os.path.split", "mmdet.apis.inference_detector" ]
[((166, 226), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {'description': '"""Test different models"""'}), "(description='Test different models')\n", (189, 226), False, 'import argparse\n'), ((1216, 1274), 'mmdet.apis.init_detector', 'init_detector', (['config_file', 'checkpoint_file'], {'device': '"""cu...
from cartodb_services.refactor.storage.redis_connection_config import RedisMetadataConnectionConfigBuilder from cartodb_services.refactor.storage.redis_connection import RedisConnectionBuilder from cartodb_services.refactor.storage.redis_config import RedisUserConfigStorageBuilder class UserConfigBackendFactory(object...
[ "cartodb_services.refactor.storage.redis_config.RedisUserConfigStorageBuilder", "cartodb_services.refactor.storage.redis_connection.RedisConnectionBuilder", "cartodb_services.refactor.storage.redis_connection_config.RedisMetadataConnectionConfigBuilder" ]
[((927, 992), 'cartodb_services.refactor.storage.redis_connection_config.RedisMetadataConnectionConfigBuilder', 'RedisMetadataConnectionConfigBuilder', (['self._server_config_backend'], {}), '(self._server_config_backend)\n', (963, 992), False, 'from cartodb_services.refactor.storage.redis_connection_config import Redi...
import os import random from flask import current_app def save_file(form_file, folder_name): random_hex = random.token_hex(8) _, f_ext = os.path.splitext(form_file.filename) file_fn = random_hex + f_ext file_path = os.path.join(current_app.root_path, 'static', folder_name, file_fn) form_file.save(...
[ "os.path.join", "os.path.splitext", "random.token_hex" ]
[((112, 131), 'random.token_hex', 'random.token_hex', (['(8)'], {}), '(8)\n', (128, 131), False, 'import random\n'), ((147, 183), 'os.path.splitext', 'os.path.splitext', (['form_file.filename'], {}), '(form_file.filename)\n', (163, 183), False, 'import os\n'), ((233, 300), 'os.path.join', 'os.path.join', (['current_app...
# -*- coding: utf-8 -*- # Copyright 2020 EMBL - European Bioinformatics Institute # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # http://www.apache.org/licenses/LICENSE-2.0 # # Unless requ...
[ "rest_framework.routers.DefaultRouter" ]
[((815, 858), 'rest_framework.routers.DefaultRouter', 'routers.DefaultRouter', ([], {'trailing_slash': '(False)'}), '(trailing_slash=False)\n', (836, 858), False, 'from rest_framework import routers\n'), ((7304, 7347), 'rest_framework.routers.DefaultRouter', 'routers.DefaultRouter', ([], {'trailing_slash': '(False)'}),...
# -*- coding: utf-8 -*- """ Created on Mon Nov 28 10:47:38 2016 @author: ahefny Policies are BLIND to the representation of states, which could be (1) observation, (2) original latent state or (3) predictive state. Policies takes the "state" dimension x_dim, the number of actions/dim of action as input. """ impo...
[ "numpy.sin", "numpy.prod", "numpy.dot" ]
[((2094, 2116), 'numpy.dot', 'np.dot', (['self._K', 'state'], {}), '(self._K, state)\n', (2100, 2116), True, 'import numpy as np\n'), ((1598, 1629), 'numpy.prod', 'np.prod', (['(self._high - self._low)'], {}), '(self._high - self._low)\n', (1605, 1629), True, 'import numpy as np\n'), ((2612, 2657), 'numpy.sin', 'np.sin...
import os import sys from datetime import datetime from subprocess import run name = datetime.utcnow().strftime("%Y%m%d-%H%M%S.md") try: # -t stands for "topic" topic_index = sys.argv.index('-t') path = os.path.join(sys.argv[topic_index + 1], name) except: path = name run(['code', path])
[ "sys.argv.index", "subprocess.run", "os.path.join", "datetime.datetime.utcnow" ]
[((287, 306), 'subprocess.run', 'run', (["['code', path]"], {}), "(['code', path])\n", (290, 306), False, 'from subprocess import run\n'), ((184, 204), 'sys.argv.index', 'sys.argv.index', (['"""-t"""'], {}), "('-t')\n", (198, 204), False, 'import sys\n'), ((216, 261), 'os.path.join', 'os.path.join', (['sys.argv[topic_i...
# -*- coding: utf-8 -*- """module for Superfetch.""" import os, sys import time from datetime import datetime, timedelta from modules import logger from modules import manager from modules import interface from modules.windows_superfetch import sfexport2 from dfvfs.lib import definitions as dfvfs_definitions class S...
[ "os.path.exists", "modules.windows_superfetch.sfexport2.main", "os.mkdir", "os.path.abspath", "modules.manager.ModulesManager.RegisterModule", "os.remove" ]
[((3350, 3408), 'modules.manager.ModulesManager.RegisterModule', 'manager.ModulesManager.RegisterModule', (['SUPERFETCHConnector'], {}), '(SUPERFETCHConnector)\n', (3387, 3408), False, 'from modules import manager\n'), ((3142, 3184), 'os.remove', 'os.remove', (['(output_path + os.sep + fileName)'], {}), '(output_path +...
from rest_framework.generics import get_object_or_404 from rest_framework.response import Response from rest_framework.status import HTTP_400_BAD_REQUEST, HTTP_200_OK from river.models import Function from river_admin.views import get, post, put, delete from river_admin.views.serializers import UpdateFunctionDto, Crea...
[ "river_admin.views.get", "river.models.Function.objects.all", "river_admin.views.serializers.CreateFunctionDto", "river_admin.views.put", "rest_framework.response.Response", "river_admin.views.serializers.UpdateFunctionDto", "river_admin.views.delete", "river_admin.views.post", "river_admin.views.se...
[((350, 385), 'river_admin.views.get', 'get', (['"""^function/get/(?P<pk>\\\\w+)/$"""'], {}), "('^function/get/(?P<pk>\\\\w+)/$')\n", (353, 385), False, 'from river_admin.views import get, post, put, delete\n'), ((546, 569), 'river_admin.views.get', 'get', (['"""^function/list/$"""'], {}), "('^function/list/$')\n", (54...
import torch import numpy as np from torch.utils.data import Dataset import torchvision.transforms as transforms import skimage.io as io from path import Path import cv2 import torch.nn.functional as F class ETH_LFB(Dataset): def __init__(self, configs): """ dataset for eth local feature benchmark ...
[ "torch.from_numpy", "numpy.array", "path.Path", "cv2.SIFT_create", "skimage.io.imread", "cv2.cvtColor", "torchvision.transforms.Normalize", "torchvision.transforms.ToTensor" ]
[((752, 769), 'cv2.SIFT_create', 'cv2.SIFT_create', ([], {}), '()\n', (767, 769), False, 'import cv2\n'), ((786, 817), 'path.Path', 'Path', (["self.configs['data_path']"], {}), "(self.configs['data_path'])\n", (790, 817), False, 'from path import Path\n'), ((1075, 1089), 'skimage.io.imread', 'io.imread', (['imf'], {}),...
"""This file contains functions for processing image""" import cv2 import math import copy import numpy as np import matplotlib.pyplot as plt def binarize_image(image): """Binarize image pixel values to 0 and 255.""" unique_values = np.unique(image) if len(unique_values) == 2: if (un...
[ "matplotlib.pyplot.imshow", "cv2.imwrite", "numpy.unique", "numpy.ones", "numpy.where", "numpy.zeros_like", "numpy.count_nonzero", "numpy.sum", "numpy.stack", "numpy.array", "copy.deepcopy", "cv2.imread", "matplotlib.pyplot.show" ]
[((255, 271), 'numpy.unique', 'np.unique', (['image'], {}), '(image)\n', (264, 271), True, 'import numpy as np\n'), ((580, 599), 'cv2.imread', 'cv2.imread', (['path', '(0)'], {}), '(path, 0)\n', (590, 599), False, 'import cv2\n'), ((692, 708), 'cv2.imread', 'cv2.imread', (['path'], {}), '(path)\n', (702, 708), False, '...
import logging from logging.config import dictConfig import dbnd from dbnd.testing.helpers import run_dbnd_subprocess__with_home from dbnd_airflow_contrib.dbnd_airflow_default_logger import DEFAULT_LOGGING_CONFIG class TestDbndAirflowLogging(object): def test_dbnd_airflow_logging_conifg(self): # we imp...
[ "logging.config.dictConfig", "logging.info", "logging.error" ]
[((919, 942), 'logging.config.dictConfig', 'dictConfig', (['dbnd_config'], {}), '(dbnd_config)\n', (929, 942), False, 'from logging.config import dictConfig\n'), ((947, 984), 'logging.info', 'logging.info', (['"""test_can_be_loaded OK"""'], {}), "('test_can_be_loaded OK')\n", (959, 984), False, 'import logging\n'), ((7...
#!/usr/bin/env python # coding: utf-8 # In[1]: #import bibliotek from keras.applications.resnet50 import ResNet50, decode_predictions,preprocess_input from keras.preprocessing import image import numpy as np import requests from io import BytesIO from PIL import Image # In[2]: #podbranie modelu ResNet50 model...
[ "keras.preprocessing.image.img_to_array", "keras.applications.resnet50.decode_predictions", "io.BytesIO", "requests.get", "numpy.expand_dims", "keras.applications.resnet50.ResNet50" ]
[((323, 351), 'keras.applications.resnet50.ResNet50', 'ResNet50', ([], {'weights': '"""imagenet"""'}), "(weights='imagenet')\n", (331, 351), False, 'from keras.applications.resnet50 import ResNet50, decode_predictions, preprocess_input\n'), ((1204, 1225), 'requests.get', 'requests.get', (['url_img'], {}), '(url_img)\n'...
""" Test all kinds of errors. @author <EMAIL> """ import pytest from trafficgenerator.tgn_utils import ApiType from xenavalkyrie.xena_app import init_xena from xenavalkyrie.xena_object import XenaAttributeError from .test_base import TestXenaBase class TestXenaErrors(TestXenaBase): def setup(self): su...
[ "pytest.raises" ]
[((656, 680), 'pytest.raises', 'pytest.raises', (['Exception'], {}), '(Exception)\n', (669, 680), False, 'import pytest\n'), ((1415, 1448), 'pytest.raises', 'pytest.raises', (['XenaAttributeError'], {}), '(XenaAttributeError)\n', (1428, 1448), False, 'import pytest\n'), ((1519, 1552), 'pytest.raises', 'pytest.raises', ...
import time from datetime import datetime as dt import colorama from colorama import Fore, Back, Style import socket import os import sys def CS(X): time.sleep(X) os.system("clear") def socknames(): myHostName = socket.gethostname() myIP = socket.gethostbyname(myHostName) print("\033[35m[\033...
[ "socket.gethostbyname", "time.sleep", "datetime.datetime.now", "os.system", "socket.gethostname" ]
[((159, 172), 'time.sleep', 'time.sleep', (['X'], {}), '(X)\n', (169, 172), False, 'import time\n'), ((177, 195), 'os.system', 'os.system', (['"""clear"""'], {}), "('clear')\n", (186, 195), False, 'import os\n'), ((231, 251), 'socket.gethostname', 'socket.gethostname', ([], {}), '()\n', (249, 251), False, 'import socke...
import json import pandas as pd from AnalysisModule.prepare.diagram import BuildingUnit from AnalysisModule.routines.util import read_jsonfile """ 4 ways to deal with strange bus A - exclude bu, keep crystals A' - merge bu, keep crystals B - exclude crystals C - keep all note 2020/11/24: - all A are modified to...
[ "pandas.DataFrame.from_records", "json.loads", "AnalysisModule.routines.util.read_jsonfile", "pandas.read_csv" ]
[((2992, 3034), 'pandas.DataFrame.from_records', 'pd.DataFrame.from_records', (['curated_records'], {}), '(curated_records)\n', (3017, 3034), True, 'import pandas as pd\n'), ((3229, 3259), 'AnalysisModule.routines.util.read_jsonfile', 'read_jsonfile', (['"""3_bulist.json"""'], {}), "('3_bulist.json')\n", (3242, 3259), ...
# flake8: noqa # -*- coding: utf-8 -*- import datetime from south.db import db from south.v2 import SchemaMigration from django.db import models class Migration(SchemaMigration): def forwards(self, orm): # Adding model 'Milestone' db.create_table('roadmap_milestone', ( ('id', self.gf(...
[ "south.db.db.send_create_signal", "south.db.db.delete_table" ]
[((472, 519), 'south.db.db.send_create_signal', 'db.send_create_signal', (['"""roadmap"""', "['Milestone']"], {}), "('roadmap', ['Milestone'])\n", (493, 519), False, 'from south.db import db\n'), ((1010, 1068), 'south.db.db.send_create_signal', 'db.send_create_signal', (['"""roadmap"""', "['MilestoneTranslation']"], {}...
from base import CQPartsTest from base import testlabel # units under test from cqparts_fasteners.fasteners.nutbolt import NutAndBoltFastener # ---------- Test Assembly ---------- import cadquery import cqparts from partslib.basic import Box from cqparts import constraint from cqparts.utils import CoordSystem class...
[ "partslib.basic.Box", "cqparts.utils.CoordSystem", "cqparts_fasteners.fasteners.nutbolt.NutAndBoltFastener", "cqparts.constraint.Coincident", "cadquery.Vector", "cqparts.constraint.Fixed" ]
[((403, 438), 'partslib.basic.Box', 'Box', ([], {'length': '(20)', 'width': '(20)', 'height': '(12)'}), '(length=20, width=20, height=12)\n', (406, 438), False, 'from partslib.basic import Box\n'), ((453, 488), 'partslib.basic.Box', 'Box', ([], {'length': '(18)', 'width': '(18)', 'height': '(18)'}), '(length=18, width=...
# Generated by Django 3.0.8 on 2021-02-20 13:51 import ckeditor.fields from django.db import migrations, models import django.db.models.deletion class Migration(migrations.Migration): dependencies = [ ('eventsapi', '0001_initial'), ] operations = [ migrations.CreateModel( na...
[ "django.db.models.OneToOneField", "django.db.models.DateField", "django.db.models.TextField", "django.db.models.ForeignKey", "django.db.models.AutoField" ]
[((382, 475), 'django.db.models.AutoField', 'models.AutoField', ([], {'auto_created': '(True)', 'primary_key': '(True)', 'serialize': '(False)', 'verbose_name': '"""ID"""'}), "(auto_created=True, primary_key=True, serialize=False,\n verbose_name='ID')\n", (398, 475), False, 'from django.db import migrations, models\...
""" This file is for initializing the uWSGI application. It defines a post-fork command in order to avoid an issue in uWSGI that prevents MongoEngine from lazily creating mongo connections """ from uwsgidecorators import postfork from mongomail.rest_app import app, db, connection @postfork def init_db(): db.init...
[ "mongomail.rest_app.connection.get_api_keys", "mongomail.rest_app.db.init_app", "mongomail.rest_app.connection.generate_api_key" ]
[((313, 329), 'mongomail.rest_app.db.init_app', 'db.init_app', (['app'], {}), '(app)\n', (324, 329), False, 'from mongomail.rest_app import app, db, connection\n'), ((370, 395), 'mongomail.rest_app.connection.get_api_keys', 'connection.get_api_keys', ([], {}), '()\n', (393, 395), False, 'from mongomail.rest_app import ...
import cupy as cp def get_batches(X,y,batch_size,seed,shuffle): cp.random.seed(seed) m = X.shape[0] mini_batches = [] # 第一步打乱X,Y if shuffle: permutation = cp.random.permutation(m) # 返回一个长度为m的list,里面的值为0到m-1 shuffled_X = X[permutation] shuffled_y = y[permutati...
[ "cupy.random.seed", "cupy.random.permutation" ]
[((78, 98), 'cupy.random.seed', 'cp.random.seed', (['seed'], {}), '(seed)\n', (92, 98), True, 'import cupy as cp\n'), ((198, 222), 'cupy.random.permutation', 'cp.random.permutation', (['m'], {}), '(m)\n', (219, 222), True, 'import cupy as cp\n')]
# -*- coding: utf-8 -*- """ Plot comparisons of IHME projections to actual data for US states. IHME data per IHME: https://covid19.healthdata.org/united-states-of-america IHME data stored here in the "..\data\ihme" directory for each release that was obtained. State-level data per Covid trackin...
[ "read_data.get_data_ctrack", "os.path.join", "read_data.format_date_ihme", "numpy.diff", "read_data.get_data_ihme", "numpy.array", "scipy.signal.medfilt", "datetime.date.today", "matplotlib.pyplot.subplots" ]
[((2462, 2474), 'datetime.date.today', 'date.today', ([], {}), '()\n', (2472, 2474), False, 'from datetime import date\n'), ((2506, 2543), 'read_data.get_data_ctrack', 'get_data_ctrack', (['state', 'data_filename'], {}), '(state, data_filename)\n', (2521, 2543), False, 'from read_data import get_data_ctrack, get_data_i...
from __future__ import absolute_import from __future__ import division from __future__ import print_function import math import sys import os.path from PIL import Image import logging import numpy as np from .base_analyzer import BaseAnnotator if os.environ.get('PYTORCH_MODE',False): import dvalib.crnn.utils as ut...
[ "tensorflow.python.training.saver.Saver", "tensorflow.multiply", "tensorflow.gfile.FastGFile", "dvalib.crnn.dataset.resizeNormalize", "torch.cuda.is_available", "tensorflow.gfile.GFile", "logging.info", "tensorflow.Graph", "tensorflow.placeholder", "tensorflow.image.resize_bilinear", "tensorflow...
[((470, 519), 'logging.info', 'logging.info', (['"""In pytorch mode, not importing TF"""'], {}), "('In pytorch mode, not importing TF')\n", (482, 519), False, 'import logging\n'), ((983, 1009), 'tensorflow.expand_dims', 'tf.expand_dims', (['image', '[0]'], {}), '(image, [0])\n', (997, 1009), True, 'import tensorflow as...
from datetime import datetime from typing import Callable def add(a: int, b: int) -> int: return a + b def subtract(a: int, b: int) -> int: return a - b def calculate(operation: Callable[[int, int], int], a: int, b: int) -> int: """Demonstration of first class citizen""" return opera...
[ "datetime.datetime.now" ]
[((705, 719), 'datetime.datetime.now', 'datetime.now', ([], {}), '()\n', (717, 719), False, 'from datetime import datetime\n')]
import collections from puzzle.constraints import solution_constraints from puzzle.steps import generate_solutions from spec.mamba import * _SOLUTIONS = collections.OrderedDict(( ('early_low', 0.1), ('early_high', 1.0), ('after_early_high', 0.9), ('mid_low', 0.2), ('late_mid', 0.5), ('late_low', 0.3), (...
[ "collections.OrderedDict", "puzzle.constraints.solution_constraints.SolutionConstraints", "puzzle.steps.generate_solutions.GenerateSolutions" ]
[((155, 334), 'collections.OrderedDict', 'collections.OrderedDict', (["(('early_low', 0.1), ('early_high', 1.0), ('after_early_high', 0.9), (\n 'mid_low', 0.2), ('late_mid', 0.5), ('late_low', 0.3), ('late_high', 0.8))"], {}), "((('early_low', 0.1), ('early_high', 1.0), (\n 'after_early_high', 0.9), ('mid_low', 0...
from torch.utils.data import Dataset from torchvision.transforms.functional import to_tensor from utils.image.processor import ImagePreprocessor, colorFormats from PIL import Image import glob, random class ImageData(Dataset): def __init__( self, srcPath, crop=True, cropSize=96, colorFromat="RGB", process...
[ "torchvision.transforms.functional.to_tensor", "PIL.Image.open", "glob.glob", "utils.image.processor.ImagePreprocessor" ]
[((396, 438), 'glob.glob', 'glob.glob', (["(srcPath + '/**')"], {'recursive': '(True)'}), "(srcPath + '/**', recursive=True)\n", (405, 438), False, 'import glob, random\n'), ((804, 909), 'utils.image.processor.ImagePreprocessor', 'ImagePreprocessor', ([], {'crop': 'crop', 'cropSize': 'cropSize', 'colorFromat': 'colorFr...
def main(): from sys import stdin n = int(stdin.readline()) estrelas = list(map(int, stdin.readline().split())) total_de_carneiros = sum(estrelas) i = 0 estrelas_atacadas = list() carneiros_roubados = 0 while 0 <= i < n: if i not in estrelas_atacadas: estrelas_atacada...
[ "sys.stdin.readline" ]
[((50, 66), 'sys.stdin.readline', 'stdin.readline', ([], {}), '()\n', (64, 66), False, 'from sys import stdin\n'), ((97, 113), 'sys.stdin.readline', 'stdin.readline', ([], {}), '()\n', (111, 113), False, 'from sys import stdin\n')]
"""LC-QuAD 2.0: A Large Scale Complex Question Answering Dataset.""" import json import os import datasets logger = datasets.logging.get_logger(__name__) _CITATION = """\ @inproceedings{dubey2017lc2, title={LC-QuAD 2.0: A Large Dataset for Complex Question Answering over Wikidata and DBpedia}, a...
[ "json.load", "datasets.Value", "os.path.join", "datasets.logging.get_logger" ]
[((119, 156), 'datasets.logging.get_logger', 'datasets.logging.get_logger', (['__name__'], {}), '(__name__)\n', (146, 156), False, 'import datasets\n'), ((4285, 4297), 'json.load', 'json.load', (['f'], {}), '(f)\n', (4294, 4297), False, 'import json\n'), ((2262, 2286), 'datasets.Value', 'datasets.Value', (['"""string""...
# Enter your code here. Read input from STDIN. Print output to STDOUT from typing import List from itertools import product def sum_of_square_mod_m( nums:tuple, mod_m ): return ( sum( map( lambda x: x**2, nums) ) % mod_m ) def maximize_function_value( list_of_list:List , mod_m): count_of_list = len( list_...
[ "itertools.product" ]
[((368, 390), 'itertools.product', 'product', (['*list_of_list'], {}), '(*list_of_list)\n', (375, 390), False, 'from itertools import product\n')]
from unittest import TestCase from tests import get_data from pytezos.operation.forge import forge_operation_group class OperationForgingTestooZWSJ(TestCase): def setUp(self): self.maxDiff = None def test_forge_ooZWSJ(self): expected = get_data( path='operations/ooZWSJqG...
[ "tests.get_data" ]
[((273, 373), 'tests.get_data', 'get_data', ([], {'path': '"""operations/ooZWSJqGRqi4hirFmqvFinKp9JGU3vqR23Q15tMvDXMnWtX1jne/forged.hex"""'}), "(path=\n 'operations/ooZWSJqGRqi4hirFmqvFinKp9JGU3vqR23Q15tMvDXMnWtX1jne/forged.hex'\n )\n", (281, 373), False, 'from tests import get_data\n'), ((416, 519), 'tests.get_d...
from machine import SPI, Pin import tinypico as TinyPICO import machine from dotstar import DotStar import time, random, micropython, gc import bluetooth from ble_simple_peripheral import BLESimplePeripheral # Configure SPI for controlling the DotStar # Internally we are using software SPI for this as the...
[ "dotstar.DotStar", "tinypico.set_dotstar_power", "bluetooth.BLE", "machine.Timer", "machine.deepsleep", "ble_simple_peripheral.BLESimplePeripheral", "machine.freq", "machine.reset_cause", "machine.Pin", "tinypico.get_battery_charging", "tinypico.get_battery_voltage" ]
[((526, 558), 'dotstar.DotStar', 'DotStar', (['spi', '(1)'], {'brightness': '(0.25)'}), '(spi, 1, brightness=0.25)\n', (533, 558), False, 'from dotstar import DotStar\n'), ((648, 670), 'machine.freq', 'machine.freq', (['(80000000)'], {}), '(80000000)\n', (660, 670), False, 'import machine\n'), ((1202, 1217), 'bluetooth...
from feature import Feature from itertools import product import numpy as np import random class Node: def __init__(self, K, Cweights, Dweights, seed): self.K = K self.seed = seed self.Kd = int(K*2/3) self.Kc = int(K*1/3) self.Cfeatures = [Feature(False, seed) for k in range(...
[ "numpy.random.rand", "itertools.product", "feature.Feature", "numpy.exp", "numpy.array", "numpy.random.randint", "random.random", "numpy.arange" ]
[((1003, 1019), 'numpy.random.rand', 'np.random.rand', ([], {}), '()\n', (1017, 1019), True, 'import numpy as np\n'), ((1220, 1233), 'numpy.exp', 'np.exp', (['alpha'], {}), '(alpha)\n', (1226, 1233), True, 'import numpy as np\n'), ((1451, 1463), 'numpy.exp', 'np.exp', (['beta'], {}), '(beta)\n', (1457, 1463), True, 'im...
##−∗−coding : utf−8−∗− import sqlite3 as lite import logging import sys from collections import OrderedDict import conf LOG_FORMAT = "%(levelname)s > Line:%(lineno)s - %(message)s" logging.basicConfig(filename="debug.log", level=logging.DEBUG, format=LOG_FORMAT, ...
[ "logging.basicConfig", "logging.getLogger", "collections.OrderedDict", "sqlite3.connect", "sys.exit" ]
[((186, 286), 'logging.basicConfig', 'logging.basicConfig', ([], {'filename': '"""debug.log"""', 'level': 'logging.DEBUG', 'format': 'LOG_FORMAT', 'filemode': '"""w"""'}), "(filename='debug.log', level=logging.DEBUG, format=\n LOG_FORMAT, filemode='w')\n", (205, 286), False, 'import logging\n'), ((373, 400), 'loggin...
from pathlib import Path from typing import List from pmfp.utils.fs_utils import get_abs_path def _find_pypackage(final_path: Path, packs: List[str]) -> None: has_init = False for i in final_path.iterdir(): if i.name == "__init__.py": has_init = True if not has_init: return ...
[ "pmfp.utils.fs_utils.get_abs_path" ]
[((648, 669), 'pmfp.utils.fs_utils.get_abs_path', 'get_abs_path', (['to_path'], {}), '(to_path)\n', (660, 669), False, 'from pmfp.utils.fs_utils import get_abs_path\n')]
import pydot from grit.decisiontree.handcrafted_trees import scenario_trees from grit.core.base import get_img_dir def build_pydot_tree(graph, root, idx='R'): node = pydot.Node(idx, label=str(root)) graph.add_node(node) if root.decision is not None: true_child = build_pydot_tree(graph, root.decis...
[ "pydot.Dot", "grit.core.base.get_img_dir" ]
[((930, 961), 'pydot.Dot', 'pydot.Dot', ([], {'graph_type': '"""digraph"""'}), "(graph_type='digraph')\n", (939, 961), False, 'import pydot\n'), ((1024, 1037), 'grit.core.base.get_img_dir', 'get_img_dir', ([], {}), '()\n', (1035, 1037), False, 'from grit.core.base import get_img_dir\n')]
#https://github.com/Newmu/Theano-Tutorials/blob/master/1_linear_regression.py import theano from theano import tensor as T import numpy as np trX = np.linspace(-1, 1, 101) trY = 2 * trX + np.random.randn(*trX.shape) * 0.33 X = T.scalar() Y = T.scalar() def model(X, w): return X * w w = theano.shared(np.asarray...
[ "theano.tensor.nnet.categorical_crossentropy", "theano.function", "matplotlib.pyplot.show", "theano.tensor.dot", "numpy.asarray", "numpy.argmax", "theano.tensor.sqr", "numpy.linspace", "theano.tensor.fmatrix", "theano.tensor.argmax", "theano.tensor.scalar", "numpy.random.randn", "fuel.datase...
[((150, 173), 'numpy.linspace', 'np.linspace', (['(-1)', '(1)', '(101)'], {}), '(-1, 1, 101)\n', (161, 173), True, 'import numpy as np\n'), ((230, 240), 'theano.tensor.scalar', 'T.scalar', ([], {}), '()\n', (238, 240), True, 'from theano import tensor as T\n'), ((245, 255), 'theano.tensor.scalar', 'T.scalar', ([], {}),...
"""General functions for working with observations. """ import collections import os from shapely.geometry import Point import pandas as pd from gisutils import df2shp, project from mfsetup.obs import make_obsname from mfsetup.units import convert_length_units, convert_volume_units, convert_time_units from mapgwm.utils...
[ "mfsetup.units.convert_time_units", "gisutils.project", "mapgwm.utils.assign_geographic_obsgroups", "mapgwm.utils.makedirs", "pandas.read_csv", "mfsetup.units.convert_volume_units", "os.path.splitext", "os.path.split", "shapely.geometry.Point", "mapgwm.utils.cull_data_to_active_area", "pandas.to...
[((14908, 14938), 'pandas.to_datetime', 'pd.to_datetime', (["df['datetime']"], {}), "(df['datetime'])\n", (14922, 14938), True, 'import pandas as pd\n'), ((15056, 15113), 'gisutils.project', 'project', (['(md.x.values, md.y.values)', 'source_crs', 'dest_crs'], {}), '((md.x.values, md.y.values), source_crs, dest_crs)\n'...
from django.contrib.auth import views as auth_views from django.urls import path from django.urls.base import reverse_lazy from django.views.decorators.csrf import csrf_exempt from . import views, webhooks from .forms.authorization import CosmosPasswordChangeForm, CosmosPasswordResetForm, CosmosSetPasswordForm app_na...
[ "django.contrib.auth.views.PasswordResetDoneView.as_view", "django.urls.base.reverse_lazy", "django.contrib.auth.views.PasswordResetCompleteView.as_view", "django.contrib.auth.views.LogoutView.as_view", "django.contrib.auth.views.PasswordChangeDoneView.as_view", "django.urls.path" ]
[((1604, 1656), 'django.urls.path', 'path', (['"""profile/"""', 'views.profile'], {'name': '"""user_profile"""'}), "('profile/', views.profile, name='user_profile')\n", (1608, 1656), False, 'from django.urls import path\n'), ((1662, 1711), 'django.urls.path', 'path', (['"""delete/"""', 'views.delete'], {'name': '"""use...
import math import torch from torch import nn from torch.nn import CrossEntropyLoss from custom_model.CXLNetModel import PosNetModel from transformers import XLNetLMHeadModel, XLNetPreTrainedModel, XLNetModel, Conv1D from transformers.modeling_utils import prune_conv1d_layer from transformers.modeling_xlnet import XL...
[ "torch.triu", "torch.nn.Dropout", "transformers.Conv1D", "torch.nn.CrossEntropyLoss", "torch.sin", "torch.pow", "torch.cos", "torch.arange", "torch.tril", "torch.eye", "torch.nn.LayerNorm", "transformers.modeling_utils.prune_conv1d_layer", "torch.matmul", "torch.nn.Embedding", "torch.ein...
[((999, 1046), 'torch.nn.Embedding', 'nn.Embedding', (['config.vocab_size', 'config.d_model'], {}), '(config.vocab_size, config.d_model)\n', (1011, 1046), False, 'from torch import nn\n'), ((1314, 1340), 'torch.nn.Dropout', 'nn.Dropout', (['config.dropout'], {}), '(config.dropout)\n', (1324, 1340), False, 'from torch i...
# -*- coding: utf-8 -*- """Make the double periodic shear test grid""" import matplotlib.pyplot as plt from configparser import ConfigParser import numpy as np import sys import os sys.path.append(os.path.abspath("../../..")) from pycato import * # Make the empty grid domain = make_uniform_grid( n_cells=(256, 25...
[ "numpy.sin", "numpy.tanh", "os.path.abspath", "matplotlib.pyplot.subplots", "matplotlib.pyplot.show" ]
[((1174, 1221), 'matplotlib.pyplot.subplots', 'plt.subplots', ([], {'figsize': '(18, 8)', 'nrows': '(1)', 'ncols': '(2)'}), '(figsize=(18, 8), nrows=1, ncols=2)\n', (1186, 1221), True, 'import matplotlib.pyplot as plt\n'), ((1735, 1745), 'matplotlib.pyplot.show', 'plt.show', ([], {}), '()\n', (1743, 1745), True, 'impor...