code
stringlengths
22
1.05M
apis
listlengths
1
3.31k
extract_api
stringlengths
75
3.25M
# currScenRecQuery = """ # SELECT * from 'pla_scenario_values_new2' where rid='{rId}' and qno='{qNo}' # """.format(rId=1, qNo=2) # currScenRec = db.fetch_results(currScenRecQuery, single_row=True) # # updateBetaAttrQuery = """ # UPDATE 'pla_scenario...
[ "os.listdir", "pymysql.connect", "db.fetch_results", "db.execute" ]
[((1141, 1180), 'db.fetch_results', 'db.fetch_results', (['"""cuebiq201911"""', 'query'], {}), "('cuebiq201911', query)\n", (1157, 1180), False, 'import db\n'), ((1255, 1351), 'pymysql.connect', 'mysql.connect', ([], {'host': '"""localhost"""', 'database': '"""cuebiq201911"""', 'user': '"""root"""', 'password': '"""<PA...
# -*- coding: utf-8 -*- ''' VKGroups Copyright © 2010-2018 HeaTTheatR Для предложений и вопросов: <<EMAIL>> Данный файл распространяется по услолвиям той же лицензии, что и фреймворк Kivy. ''' from jnius import autoclass, PythonJavaClass, java_method, cast from android.runnable import run_on_ui_thread Toast = au...
[ "jnius.autoclass" ]
[((318, 351), 'jnius.autoclass', 'autoclass', (['"""android.widget.Toast"""'], {}), "('android.widget.Toast')\n", (327, 351), False, 'from jnius import autoclass, PythonJavaClass, java_method, cast\n'), ((362, 406), 'jnius.autoclass', 'autoclass', (['"""org.kivy.android.PythonActivity"""'], {}), "('org.kivy.android.Pyt...
# Copyright 2020 Adap GmbH. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or ag...
[ "numpy.random.choice", "numpy.array", "statistics.mean", "numpy.max" ]
[((4713, 4780), 'numpy.random.choice', 'np.random.choice', (['indices'], {'size': 'sample_size', 'replace': '(False)', 'p': 'probs'}), '(indices, size=sample_size, replace=False, p=probs)\n', (4729, 4780), True, 'import numpy as np\n'), ((4680, 4696), 'numpy.array', 'np.array', (['logits'], {}), '(logits)\n', (4688, 46...
import tensorflow as tf A = tf.constant([4], tf.int32, name='A') B = tf.constant([4], tf.int32, name='B') C = tf.constant([4], tf.int32, name='C') x = tf.placeholder(tf.int32, name='x') with tf.name_scope("Equation_1"): AX2_1 = tf.multiply(A, tf.pow(x,2), name='AX2_1') Bx = tf.multiply(B, tf.pow(x,2), name='...
[ "tensorflow.pow", "tensorflow.Session", "tensorflow.placeholder", "tensorflow.add_n", "tensorflow.name_scope", "tensorflow.constant", "tensorflow.summary.FileWriter" ]
[((29, 65), 'tensorflow.constant', 'tf.constant', (['[4]', 'tf.int32'], {'name': '"""A"""'}), "([4], tf.int32, name='A')\n", (40, 65), True, 'import tensorflow as tf\n'), ((70, 106), 'tensorflow.constant', 'tf.constant', (['[4]', 'tf.int32'], {'name': '"""B"""'}), "([4], tf.int32, name='B')\n", (81, 106), True, 'import...
import os class Config(): SECRET_KEY = os.environ.get('FLASK_SECRET_KEY', 'f&k*n@kfa_H*tf^$)') SSL_DISABLE = False SQLALCHEMY_COMMIT_ON_TEARDOWN = True SQLALCHEMY_RECORD_QUERIES = True SQLALCHEMY_TRACK_MODIFICATIONS = False FLASKY_SLOW_DB_QUERY_TIME = 0.5 FLASKY_PER_PAGE = 10 CKEDITOR_...
[ "os.environ.get" ]
[((45, 100), 'os.environ.get', 'os.environ.get', (['"""FLASK_SECRET_KEY"""', '"""f&k*n@kfa_H*tf^$)"""'], {}), "('FLASK_SECRET_KEY', 'f&k*n@kfa_H*tf^$)')\n", (59, 100), False, 'import os\n')]
import random class Thing: """ This represents any physical object that can appear in an Environment. """ def is_alive(self): """Things that are 'alive' should return true.""" return hasattr(self, 'alive') and self.alive def show_state(self): """Display the agent's int...
[ "random.choice" ]
[((6333, 6362), 'random.choice', 'random.choice', (['[loc_A, loc_B]'], {}), '([loc_A, loc_B])\n', (6346, 6362), False, 'import random\n'), ((5636, 5669), 'random.choice', 'random.choice', (["['Clean', 'Dirty']"], {}), "(['Clean', 'Dirty'])\n", (5649, 5669), False, 'import random\n'), ((5701, 5734), 'random.choice', 'ra...
# -*- coding: utf-8 -*- """ tests.tests_models.test_water_container ~~~~~~~~~~~~~~~~~~~ This script contains tests for the WaterContainer model. """ from pytest import fixture, mark, raises from src.exceptions import CoffeeMachineException from src.exceptions import NotEnoughWater from src.models.container import Wat...
[ "src.models.container.WaterContainer", "pytest.raises" ]
[((646, 690), 'src.models.container.WaterContainer', 'WaterContainer', ([], {'capacity': '_container_capacity'}), '(capacity=_container_capacity)\n', (660, 690), False, 'from src.models.container import WaterContainer\n'), ((893, 915), 'pytest.raises', 'raises', (['NotEnoughWater'], {}), '(NotEnoughWater)\n', (899, 915...
from quicksectx import IntervalTree, Interval import unittest tree = IntervalTree() tree.add(0, 3, 100) tree.add(5, 8, 110) tree.add(6, 10, 120) tree.add(8, 9, 130) tree.add(15, 23, 140) tree.add(19, 20, 150) tree.add(17, 19, 160) tree.add(26, 26, 160) tree.add(25, 30, 160) tree.add(16, 21, 160) print(tree.search(3,15...
[ "unittest.main", "quicksectx.IntervalTree", "quicksectx.Interval" ]
[((70, 84), 'quicksectx.IntervalTree', 'IntervalTree', ([], {}), '()\n', (82, 84), False, 'from quicksectx import IntervalTree, Interval\n'), ((380, 394), 'quicksectx.IntervalTree', 'IntervalTree', ([], {}), '()\n', (392, 394), False, 'from quicksectx import IntervalTree, Interval\n'), ((1161, 1176), 'unittest.main', '...
#!/usr/bin/python # -*- coding: utf-8 -*- # author arrti from ss_admin import app if __name__ == '__main__': app.run()
[ "ss_admin.app.run" ]
[((116, 125), 'ss_admin.app.run', 'app.run', ([], {}), '()\n', (123, 125), False, 'from ss_admin import app\n')]
# Maximum Gap Problem # http://www.zrzahid.com/the-%E2%80%A9maximum%E2%80%A9-gap%E2%80%A9-problem-%E2%80%A9pigeonhole-%E2%80%A9principle%E2%80%A9/ # http://cgm.cs.mcgill.ca/~godfried/teaching/dm-reading-assignments/Maximum-Gap-Problem.pdf # https://stackoverflow.com/questions/10262937/array-maximum-difference-algorithm...
[ "math.floor" ]
[((689, 714), 'math.floor', 'floor', (['((i - minn) / delta)'], {}), '((i - minn) / delta)\n', (694, 714), False, 'from math import floor\n')]
# Copyright (c) Microsoft Corporation. # Licensed under the MIT license. from .__misc__ import __version__, __data_version__ from maro.utils.utils import deploy, check_deployment_status if not check_deployment_status(): deploy()
[ "maro.utils.utils.deploy", "maro.utils.utils.check_deployment_status" ]
[((196, 221), 'maro.utils.utils.check_deployment_status', 'check_deployment_status', ([], {}), '()\n', (219, 221), False, 'from maro.utils.utils import deploy, check_deployment_status\n'), ((227, 235), 'maro.utils.utils.deploy', 'deploy', ([], {}), '()\n', (233, 235), False, 'from maro.utils.utils import deploy, check_...
# -*- coding: UTF-8 -*- """ @author: <NAME> @file : settings.py @time : 2021/09/10 09:20 @desc : """ import configparser import os import PySimpleGUI as sg LANGUAGE_DEF = '中文/英文' LANGUAGE_NAME_KEY_MAP = { '中文/英文': 'ch', '繁体中文': 'ch_tra', '日语': 'japan', '韩语': 'korean', '...
[ "os.path.exists", "configparser.ConfigParser", "PySimpleGUI.Cancel", "PySimpleGUI.Text", "PySimpleGUI.theme", "PySimpleGUI.OK" ]
[((706, 730), 'PySimpleGUI.theme', 'sg.theme', (['"""LightBrown12"""'], {}), "('LightBrown12')\n", (714, 730), True, 'import PySimpleGUI as sg\n'), ((2297, 2324), 'configparser.ConfigParser', 'configparser.ConfigParser', ([], {}), '()\n', (2322, 2324), False, 'import configparser\n'), ((2217, 2244), 'os.path.exists', '...
# Generated by Django 3.1.1 on 2020-09-13 05:19 from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('log_reg_app', '0001_initial'), ] operations = [ migrations.AddField( model_name='user', name='password', ...
[ "django.db.models.CharField" ]
[((327, 373), 'django.db.models.CharField', 'models.CharField', ([], {'default': 'None', 'max_length': '(255)'}), '(default=None, max_length=255)\n', (343, 373), False, 'from django.db import migrations, models\n')]
import math class Prime: is_debug = True prime_list = [] max_ind = 0 max_keep = 0 prev_no = 0 seq_list = [] inter_list = [] def __init__(self, limit): self.max_keep = round(math.sqrt(limit) / math.log(math.sqrt(limit), math.e) * 1.5) def get(self, index): return s...
[ "math.sqrt" ]
[((216, 232), 'math.sqrt', 'math.sqrt', (['limit'], {}), '(limit)\n', (225, 232), False, 'import math\n'), ((244, 260), 'math.sqrt', 'math.sqrt', (['limit'], {}), '(limit)\n', (253, 260), False, 'import math\n')]
import pandas as pd import numpy as np import re from FeatureExtraction import * from CrossValidation import * import warnings from Preprocessing import * from collections import Counter import seaborn as sns import matplotlib.pyplot as plt warnings.filterwarnings(action = 'ignore') #read train file train=pd.read_csv(...
[ "matplotlib.pyplot.savefig", "pandas.read_csv", "pandas.DataFrame", "seaborn.barplot", "pandas.concat", "warnings.filterwarnings", "matplotlib.pyplot.show" ]
[((241, 281), 'warnings.filterwarnings', 'warnings.filterwarnings', ([], {'action': '"""ignore"""'}), "(action='ignore')\n", (264, 281), False, 'import warnings\n'), ((308, 337), 'pandas.read_csv', 'pd.read_csv', (['"""Data/train.csv"""'], {}), "('Data/train.csv')\n", (319, 337), True, 'import pandas as pd\n'), ((360, ...
#to run this script once per day (auto-update the washingtonpost data from GitHub) add this to crontab: # 0 12 * * * python /Users/quran/Dropbox/MusicArtProjects/endsSentenceFunction/data/download.py import os import urllib fullfilename = os.path.join('/Users/quran/Dropbox/MusicArtProjects/endsSentenceFunction/data', ...
[ "urllib.URLopener", "os.path.join" ]
[((240, 358), 'os.path.join', 'os.path.join', (['"""/Users/quran/Dropbox/MusicArtProjects/endsSentenceFunction/data"""', '"""fatal-police-shootings-data.csv"""'], {}), "('/Users/quran/Dropbox/MusicArtProjects/endsSentenceFunction/data',\n 'fatal-police-shootings-data.csv')\n", (252, 358), False, 'import os\n'), ((36...
# -*- coding: utf-8 -*- from __future__ import unicode_literals from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('core', '0001_initial'), ] operations = [ migrations.CreateModel( name='Event', fields=[ ...
[ "django.db.models.DateField", "django.db.models.TextField", "django.db.models.ForeignKey", "django.db.models.BooleanField", "django.db.models.AutoField", "django.db.models.CharField" ]
[((331, 424), 'django.db.models.AutoField', 'models.AutoField', ([], {'verbose_name': '"""ID"""', 'serialize': '(False)', 'auto_created': '(True)', 'primary_key': '(True)'}), "(verbose_name='ID', serialize=False, auto_created=True,\n primary_key=True)\n", (347, 424), False, 'from django.db import migrations, models\...
################################################################################ # # Select a # # Author(s): <NAME> ################################################################################ import pathlib import random from typing import Union, List from enum import Enum import torch as t from torchaudio.bac...
[ "random.randint" ]
[((5239, 5281), 'random.randint', 'random.randint', (['(0)', '(num_possible_chunks - 1)'], {}), '(0, num_possible_chunks - 1)\n', (5253, 5281), False, 'import random\n'), ((4905, 4957), 'random.randint', 'random.randint', (['(0)', '(num_samples - self.chunk_size - 1)'], {}), '(0, num_samples - self.chunk_size - 1)\n', ...
# Copyright (c) 2021 Graphcore Ltd. All rights reserved. import copy import inspect import torch # Do not import any poptorch.* here: it will break the poptorch module from . import _impl from ._logging import logger class ArgsParser: class Args: def __init__(self): self._args = [] ...
[ "copy.copy", "inspect.signature" ]
[((435, 456), 'copy.copy', 'copy.copy', (['self._args'], {}), '(self._args)\n', (444, 456), False, 'import copy\n'), ((2444, 2482), 'inspect.signature', 'inspect.signature', (['model.model.forward'], {}), '(model.model.forward)\n', (2461, 2482), False, 'import inspect\n'), ((2515, 2547), 'inspect.signature', 'inspect.s...
from unicodedata import name from django.shortcuts import render def home_view(request): return render(request, 'home.html') def contact_view(request): return render(request, 'contact.html') def skills_view(request): return render(request, 'skills.html') def experience_view(request): ...
[ "django.shortcuts.render" ]
[((106, 134), 'django.shortcuts.render', 'render', (['request', '"""home.html"""'], {}), "(request, 'home.html')\n", (112, 134), False, 'from django.shortcuts import render\n'), ((178, 209), 'django.shortcuts.render', 'render', (['request', '"""contact.html"""'], {}), "(request, 'contact.html')\n", (184, 209), False, '...
import sys import tkinter import functools def getwin(): window = tkinter.Tk() globals()['window_data'] = window return window def close(): globals()['window_data'].destroy() def popup(): import argoncraft win = getwin() win.title('ArgonCraft - Settings') win.config(bg=argoncraft.get...
[ "argoncraft.gettheme", "tkinter.Tk" ]
[((71, 83), 'tkinter.Tk', 'tkinter.Tk', ([], {}), '()\n', (81, 83), False, 'import tkinter\n'), ((306, 327), 'argoncraft.gettheme', 'argoncraft.gettheme', ([], {}), '()\n', (325, 327), False, 'import argoncraft\n'), ((450, 471), 'argoncraft.gettheme', 'argoncraft.gettheme', ([], {}), '()\n', (469, 471), False, 'import ...
import time import board import pulseio import adafruit_irremote import neopixel # Configure treasure information TREASURE_ID = 1 TRANSMIT_DELAY = 15 # Create NeoPixel object to indicate status pixels = neopixel.NeoPixel(board.NEOPIXEL, 10) # Create a 'pulseio' output, to send infrared signals on the IR transmitter ...
[ "pulseio.PulseOut", "adafruit_irremote.GenericTransmit", "time.sleep", "pulseio.PWMOut", "neopixel.NeoPixel" ]
[((205, 242), 'neopixel.NeoPixel', 'neopixel.NeoPixel', (['board.NEOPIXEL', '(10)'], {}), '(board.NEOPIXEL, 10)\n', (222, 242), False, 'import neopixel\n'), ((334, 398), 'pulseio.PWMOut', 'pulseio.PWMOut', (['board.IR_TX'], {'frequency': '(38000)', 'duty_cycle': '(2 ** 15)'}), '(board.IR_TX, frequency=38000, duty_cycle...
from django.conf.urls import patterns, url urlpatterns = patterns( 'apps.zblog.views', url(r'^$', 'index', name='index'), url(r'^(?P<id>\d+)/(?P<slug>[\w_-]+)', 'detail', name='detail'), url(r'^archives', 'archives', name='archives'), url(r'^tag/(?P<tag_name>.+)', 'tag', name='tag'), ur...
[ "django.conf.urls.url" ]
[((100, 132), 'django.conf.urls.url', 'url', (['"""^$"""', '"""index"""'], {'name': '"""index"""'}), "('^$', 'index', name='index')\n", (103, 132), False, 'from django.conf.urls import patterns, url\n'), ((140, 204), 'django.conf.urls.url', 'url', (['"""^(?P<id>\\\\d+)/(?P<slug>[\\\\w_-]+)"""', '"""detail"""'], {'name'...
from clients.rabbitmq_client import RabbitMQClient import json from settings import APP_QUEUE if __name__ == "__main__": rabbit_mq = RabbitMQClient() payload = { "component": "DOBIE", "message": { "tasks": [ { "label": "95671c903a5b97a9", ...
[ "json.dumps", "clients.rabbitmq_client.RabbitMQClient" ]
[((139, 155), 'clients.rabbitmq_client.RabbitMQClient', 'RabbitMQClient', ([], {}), '()\n', (153, 155), False, 'from clients.rabbitmq_client import RabbitMQClient\n'), ((736, 755), 'json.dumps', 'json.dumps', (['payload'], {}), '(payload)\n', (746, 755), False, 'import json\n')]
import xlrd import xlwt import os class XlsReader: def __init__(self, file: str): self.file_path = file def read_data(self, n:int = 0): try: wb = xlrd.open_workbook(self.file_path) sheet = wb.sheet_by_index(0) result_list = [] for nr in range(n,...
[ "os.path.exists", "xlwt.Workbook", "xlwt.XFStyle", "xlrd.open_workbook", "xlwt.Font", "os.remove" ]
[((748, 768), 'os.path.exists', 'os.path.exists', (['file'], {}), '(file)\n', (762, 768), False, 'import os\n'), ((896, 911), 'xlwt.Workbook', 'xlwt.Workbook', ([], {}), '()\n', (909, 911), False, 'import xlwt\n'), ((979, 990), 'xlwt.Font', 'xlwt.Font', ([], {}), '()\n', (988, 990), False, 'import xlwt\n'), ((1047, 106...
from subprocess import Popen, PIPE import sys import os def osascript(scpt): p = Popen(['osascript', '-'], stdin=PIPE, stdout=PIPE, stderr=PIPE) stdout, stderr = p.communicate(scpt.encode('utf-8')) return stdout, stderr def openTab(): script = f""" tell application "System Events" ...
[ "subprocess.Popen", "os.getcwd" ]
[((87, 150), 'subprocess.Popen', 'Popen', (["['osascript', '-']"], {'stdin': 'PIPE', 'stdout': 'PIPE', 'stderr': 'PIPE'}), "(['osascript', '-'], stdin=PIPE, stdout=PIPE, stderr=PIPE)\n", (92, 150), False, 'from subprocess import Popen, PIPE\n'), ((530, 541), 'os.getcwd', 'os.getcwd', ([], {}), '()\n', (539, 541), False...
# Generated by Django 3.0.6 on 2020-05-22 08:33 from django.conf import settings from django.db import migrations, models import django.db.models.deletion import usuarios.managers class Migration(migrations.Migration): initial = True dependencies = [ ('auth', '0011_update_proxy_permissions'), ...
[ "django.db.models.EmailField", "django.db.models.ForeignKey", "django.db.models.ManyToManyField", "django.db.models.BooleanField", "django.db.models.AutoField", "django.db.models.DateTimeField", "django.db.models.CharField" ]
[((3112, 3215), 'django.db.models.ManyToManyField', 'models.ManyToManyField', ([], {'related_name': '"""tests"""', 'through': '"""usuarios.User_test"""', 'to': '"""cuestionarios.Test"""'}), "(related_name='tests', through='usuarios.User_test',\n to='cuestionarios.Test')\n", (3134, 3215), False, 'from django.db impor...
from flask import Flask, render_template, jsonify from flask_cors import CORS from makeSentence import makeSentence app = Flask(__name__) CORS(app) @app.route("/") def index(): return render_template("index.html", sentence=makeSentence()) @app.route("/api") def api(): return jsonify(sentence=makeSentence()...
[ "makeSentence.makeSentence", "flask_cors.CORS", "flask.Flask" ]
[((123, 138), 'flask.Flask', 'Flask', (['__name__'], {}), '(__name__)\n', (128, 138), False, 'from flask import Flask, render_template, jsonify\n'), ((139, 148), 'flask_cors.CORS', 'CORS', (['app'], {}), '(app)\n', (143, 148), False, 'from flask_cors import CORS\n'), ((230, 244), 'makeSentence.makeSentence', 'makeSente...
import pickle import random import cv2 as cv import numpy as np import torch from torch.utils.data import Dataset from torchvision import transforms from config import pickle_file, num_workers from utils import align_face # Data augmentation and normalization for training # Just normalization for validation data_tra...
[ "cv2.imwrite", "random.sample", "torchvision.transforms.ToPILImage", "torchvision.transforms.ToTensor", "pickle.load", "torchvision.transforms.RandomHorizontalFlip", "torchvision.transforms.Normalize", "torch.utils.data.DataLoader", "cv2.resize", "cv2.imread", "utils.align_face", "numpy.random...
[((1861, 1895), 'random.sample', 'random.sample', (["data['samples']", '(10)'], {}), "(data['samples'], 10)\n", (1874, 1895), False, 'import random\n'), ((2399, 2517), 'torch.utils.data.DataLoader', 'torch.utils.data.DataLoader', (['train_dataset'], {'batch_size': '(256)', 'shuffle': '(True)', 'num_workers': 'num_worke...
import multiprocessing as mp import pandas as pd import dill from functools import wraps __spawn_ctx__ = None def get_mp_ctx(f): def with_ctx(*args, **kws): global __spawn_ctx__ if __spawn_ctx__ is None: __spawn_ctx__ = mp.get_context('spawn') return f(*args, mp_ctx=__spawn_ct...
[ "multiprocessing.get_context", "dill.loads", "functools.wraps", "multiprocessing.cpu_count", "dill.dumps", "pandas.concat" ]
[((477, 498), 'dill.loads', 'dill.loads', (['f_pickled'], {}), '(f_pickled)\n', (487, 498), False, 'import dill\n'), ((713, 727), 'multiprocessing.cpu_count', 'mp.cpu_count', ([], {}), '()\n', (725, 727), True, 'import multiprocessing as mp\n'), ((1295, 1308), 'dill.dumps', 'dill.dumps', (['f'], {}), '(f)\n', (1305, 13...
from collections import defaultdict from typing import List, Any, Tuple, Dict from util.helpers import solution_timer from util.input_helper import read_entire_input data = read_entire_input(2019,6) def parse(data: List[str]) -> Any: orbits = defaultdict(list) orbiting = {} for row in data: a,b=ro...
[ "util.input_helper.read_entire_input", "collections.defaultdict", "util.helpers.solution_timer" ]
[((174, 200), 'util.input_helper.read_entire_input', 'read_entire_input', (['(2019)', '(6)'], {}), '(2019, 6)\n', (191, 200), False, 'from util.input_helper import read_entire_input\n'), ((1618, 1644), 'util.helpers.solution_timer', 'solution_timer', (['(2019)', '(6)', '(1)'], {}), '(2019, 6, 1)\n', (1632, 1644), False...
import socket MAX_PACKET_SIZE = 1024 def start_client(host=socket.gethostname(), port=5000): # Create socket client_socket = socket.socket() # Connect to host and port client_socket.connect((host, port)) # Enter a message from user msg = input("Enter a message: ") while msg.lower().str...
[ "socket.gethostname", "socket.socket" ]
[((62, 82), 'socket.gethostname', 'socket.gethostname', ([], {}), '()\n', (80, 82), False, 'import socket\n'), ((136, 151), 'socket.socket', 'socket.socket', ([], {}), '()\n', (149, 151), False, 'import socket\n')]
from math import factorial def solve(r): if r == 1: return 1 elif r == 2: return 5 elif r == 3: return 15 fatR = factorial(r) fatR_m1 = factorial(r-1) comb_part_one = fatR/(2*factorial(r-2)) comb_part_two = r * (fatR_m1/(2*factorial((r-1)-2))) arranjo = fatR / ...
[ "math.factorial" ]
[((155, 167), 'math.factorial', 'factorial', (['r'], {}), '(r)\n', (164, 167), False, 'from math import factorial\n'), ((182, 198), 'math.factorial', 'factorial', (['(r - 1)'], {}), '(r - 1)\n', (191, 198), False, 'from math import factorial\n'), ((321, 337), 'math.factorial', 'factorial', (['(r - 4)'], {}), '(r - 4)\n...
import unittest from fxphelper import * C_TEST_PRECISION = 5 class TestConversion(unittest.TestCase): def test_conv_unsigned_int(self): x = 3 y = FXPQNumber(0,5,0, float_value = x) self.assertAlmostEqual(y.to_float(), x, C_TEST_PRECISION) def test_conv_signed_int(self): x = -...
[ "unittest.main" ]
[((1004, 1019), 'unittest.main', 'unittest.main', ([], {}), '()\n', (1017, 1019), False, 'import unittest\n')]
from rest_framework import serializers from django.contrib.auth.models import Permission, ContentType, Group class GroupPermissionModelSerializer(serializers.ModelSerializer): class Meta: model = Group fields = '__all__' class PermissionSimpleSerializer(serializers.ModelSerializer): content_...
[ "rest_framework.serializers.StringRelatedField" ]
[((327, 359), 'rest_framework.serializers.StringRelatedField', 'serializers.StringRelatedField', ([], {}), '()\n', (357, 359), False, 'from rest_framework import serializers\n')]
import FWCore.ParameterSet.Config as cms # Remove duplicates from the electron list electronsNoDuplicates = cms.EDFilter("DuplicatedElectronCleaner", ## reco electron input source electronSource = cms.InputTag("gsfElectrons"), )
[ "FWCore.ParameterSet.Config.InputTag" ]
[((207, 235), 'FWCore.ParameterSet.Config.InputTag', 'cms.InputTag', (['"""gsfElectrons"""'], {}), "('gsfElectrons')\n", (219, 235), True, 'import FWCore.ParameterSet.Config as cms\n')]
import json import os class ToolSettings: def __init__(self, repopath, importFacePoses, DEFAULT_DIR_KEY, RECENT_FILES, RECENT_FILES_MAX): self.repopath = repopath self.importFacePoses = importFacePoses self.DEFAULT_DIR_KEY = DEFAULT_DIR_KEY self.RECENT_FILES = RECENT_FILES ...
[ "json.load", "json.dumps", "os.path.split" ]
[((467, 490), 'os.path.split', 'os.path.split', (['__file__'], {}), '(__file__)\n', (480, 490), False, 'import os\n'), ((745, 768), 'os.path.split', 'os.path.split', (['__file__'], {}), '(__file__)\n', (758, 768), False, 'import os\n'), ((608, 630), 'json.load', 'json.load', (['config_file'], {}), '(config_file)\n', (6...
import streamlit as st import pandas as pd import numpy as np DISPLAY_TEXT = "Display Text" DISPLAY_DATA = "Display Data" DISPLAY_CODE = "Display Code" DISPLAY_GRAPHS = "Display Graphs" USE_MEDIA = "Webcam and Microphone" DRAWABLE_CANVAS = "Drawable Canvas" INTERACTIVE_DISPLAY = "User interaction" PROGRESS_AND_STATUS ...
[ "pages.progress_and_status.display_page", "pages.display_text.display_page", "pages.drawable_canvas.display_page", "streamlit.sidebar.markdown", "pages.display_graphs.display_page", "pages.interactive_widgets.display_page", "pages.display_data.display_page", "streamlit.set_page_config", "pages.displ...
[((364, 397), 'streamlit.set_page_config', 'st.set_page_config', ([], {'layout': '"""wide"""'}), "(layout='wide')\n", (382, 397), True, 'import streamlit as st\n'), ((443, 570), 'streamlit.sidebar.radio', 'st.sidebar.radio', (['"""Choose your page:"""', '[DISPLAY_TEXT, DISPLAY_DATA, DISPLAY_CODE, PROGRESS_AND_STATUS,\n...
import requests from bs4 import BeautifulSoup import re from datetime import datetime class PageScrapper: def __init__(self, server): self.url = 'https://czech-craft.eu/server/' + server + '/vote/' def CanVote(self): page = requests.get(self.url) if page.status_code != 200: ...
[ "bs4.BeautifulSoup", "re.findall", "datetime.datetime.strptime", "requests.get" ]
[((251, 273), 'requests.get', 'requests.get', (['self.url'], {}), '(self.url)\n', (263, 273), False, 'import requests\n'), ((348, 390), 'bs4.BeautifulSoup', 'BeautifulSoup', (['page.content', '"""html.parser"""'], {}), "(page.content, 'html.parser')\n", (361, 390), False, 'from bs4 import BeautifulSoup\n'), ((624, 650)...
#First we'll import the os module #This will allow us to create file paths across operating systems import os #Module for reading CSV files import csv #Set path for file csvpath = os.path.join("budget_data.csv") #Open the csv with open(csvpath) as csvfile: csvreader = csv.reader(csvfile, delimiter=",") #Rea...
[ "os.path.join", "csv.reader" ]
[((182, 213), 'os.path.join', 'os.path.join', (['"""budget_data.csv"""'], {}), "('budget_data.csv')\n", (194, 213), False, 'import os\n'), ((276, 310), 'csv.reader', 'csv.reader', (['csvfile'], {'delimiter': '""","""'}), "(csvfile, delimiter=',')\n", (286, 310), False, 'import csv\n')]
#%% import matplotlib.pyplot as plt import cv2 import numpy as np import os import re from pathlib import Path #%% def find_contours(img): # img channels assumed to be RGB img_bw = cv2.cvtColor(img, cv2.COLOR_RGB2GRAY) _, thresh = cv2.threshold(img_bw, 0, 255, cv2.THRESH_BINARY_INV) contours, _ = cv2....
[ "cv2.fillPoly", "os.path.exists", "cv2.imwrite", "os.makedirs", "pathlib.Path", "cv2.threshold", "cv2.bitwise_and", "os.path.join", "os.path.splitext", "cv2.contourArea", "numpy.zeros", "cv2.cvtColor", "cv2.findContours", "cv2.imread", "cv2.boundingRect" ]
[((191, 228), 'cv2.cvtColor', 'cv2.cvtColor', (['img', 'cv2.COLOR_RGB2GRAY'], {}), '(img, cv2.COLOR_RGB2GRAY)\n', (203, 228), False, 'import cv2\n'), ((245, 297), 'cv2.threshold', 'cv2.threshold', (['img_bw', '(0)', '(255)', 'cv2.THRESH_BINARY_INV'], {}), '(img_bw, 0, 255, cv2.THRESH_BINARY_INV)\n', (258, 297), False, ...
from tile_calculator.main import parse_args def test_main(): parser = parse_args(["54", "18", "1", "10"]) assert parser.latitude == 54 assert parser.longitude == 18 assert parser.radius == 1 assert parser.zoom_level == 10 assert not parser.miles assert not parser.total_only
[ "tile_calculator.main.parse_args" ]
[((76, 111), 'tile_calculator.main.parse_args', 'parse_args', (["['54', '18', '1', '10']"], {}), "(['54', '18', '1', '10'])\n", (86, 111), False, 'from tile_calculator.main import parse_args\n')]
#!/usr/bin/env python3 """ Usage: run_clust.py (create|run) <path> [options] [--] [<add_args> ...] Options: Logging: --log <level> Level of stdout logging [default: INFO] --lformat <level> Which formatter to use [default: extended] Experiments: --raw ...
[ "pathlib.Path", "vst.reasonable_logging_setup", "dervo.experiment.run_cluster_experiment", "dervo.snippets.loglevel_int_to_str", "dervo.snippets.find_exp_path", "dervo.experiment.prepare_cluster_experiment", "docopt.docopt" ]
[((871, 932), 'vst.reasonable_logging_setup', 'vst.reasonable_logging_setup', (['loglevel_int', "args['--lformat']"], {}), "(loglevel_int, args['--lformat'])\n", (899, 932), False, 'import vst\n'), ((1543, 1558), 'docopt.docopt', 'docopt', (['__doc__'], {}), '(__doc__)\n', (1549, 1558), False, 'from docopt import docop...
from django.contrib.auth.decorators import login_required from django.contrib.admin.views.decorators import staff_member_required from django.shortcuts import render, redirect, get_object_or_404 from django.http import Http404 from catalog.models import Paper, Person, Dataset, Venue, Comment, Code from catalog.models i...
[ "catalog.forms.PaperForm", "catalog.forms.SearchDatasetsForm", "re.compile", "catalog.models.Comment", "urllib.request.Request", "catalog.forms.SearchVenuesForm", "catalog.forms.PaperConnectionForm", "catalog.forms.PaperImportForm", "catalog.models.ReadingGroup.objects.all", "django.urls.reverse",...
[((5612, 5725), 'django.shortcuts.render', 'render', (['request', '"""paper_authors.html"""', "{'authors': authors, 'paper': paper, 'number_of_authors': num_authors}"], {}), "(request, 'paper_authors.html', {'authors': authors, 'paper': paper,\n 'number_of_authors': num_authors})\n", (5618, 5725), False, 'from djang...
import multiprocessing debug = False bind = "0.0.0.0:5000" pidfile = "gunicorn.pid" workers = multiprocessing.cpu_count()*2 + 1 worker_class = "gevent" # daemon=True 在docker中不需要daemon运行,反而会导致看不到gunicorn输出而增加排查问题的难度
[ "multiprocessing.cpu_count" ]
[((94, 121), 'multiprocessing.cpu_count', 'multiprocessing.cpu_count', ([], {}), '()\n', (119, 121), False, 'import multiprocessing\n')]
from django.apps import AppConfig from django.utils.translation import gettext_lazy as _ class MicroblogConfig(AppConfig): name = 'microblog' verbose_name = _('Blog')
[ "django.utils.translation.gettext_lazy" ]
[((167, 176), 'django.utils.translation.gettext_lazy', '_', (['"""Blog"""'], {}), "('Blog')\n", (168, 176), True, 'from django.utils.translation import gettext_lazy as _\n')]
# Copyright 2014 Google Inc. All Rights Reserved. """Implements the command for SSHing into an instance.""" import getpass from googlecloudsdk.calliope import exceptions from googlecloudsdk.compute.lib import ssh_utils from googlecloudsdk.compute.lib import utils class SSH(ssh_utils.BaseSSHCLICommand): """SSH int...
[ "getpass.getuser", "googlecloudsdk.compute.lib.utils.AddZoneFlag", "googlecloudsdk.compute.lib.ssh_utils.UserHost", "googlecloudsdk.compute.lib.ssh_utils.BaseSSHCLICommand.Args" ]
[((394, 434), 'googlecloudsdk.compute.lib.ssh_utils.BaseSSHCLICommand.Args', 'ssh_utils.BaseSSHCLICommand.Args', (['parser'], {}), '(parser)\n', (426, 434), False, 'from googlecloudsdk.compute.lib import ssh_utils\n'), ((2033, 2118), 'googlecloudsdk.compute.lib.utils.AddZoneFlag', 'utils.AddZoneFlag', (['parser'], {'re...
from sklearn.preprocessing import scale import numpy as np class GradientDescent: learning_rate = 0.01 max_iter = 2000 scale = False new_theta = [] def __init__(self, learning_rate=0.01, max_iter=2000, scale=False): self.learning_rate = learning_rate self.max_ite...
[ "numpy.array", "numpy.dot", "numpy.transpose", "sklearn.preprocessing.scale" ]
[((616, 634), 'numpy.array', 'np.array', (['[zeroes]'], {}), '([zeroes])\n', (624, 634), True, 'import numpy as np\n'), ((467, 475), 'sklearn.preprocessing.scale', 'scale', (['X'], {}), '(X)\n', (472, 475), False, 'from sklearn.preprocessing import scale\n'), ((697, 713), 'numpy.dot', 'np.dot', (['theta', 'X'], {}), '(...
# Copyright (c) 2013 <NAME> # # Licensed under the Apache License, Version 2.0 (the "License"); you may not # use this file except in compliance with the License. You may obtain a copy of # the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, softw...
[ "ratatosk.log.get_logger", "luigi.Parameter", "ratatosk.utils.fullclassname" ]
[((937, 949), 'ratatosk.log.get_logger', 'get_logger', ([], {}), '()\n', (947, 949), False, 'from ratatosk.log import get_logger\n'), ((1332, 1364), 'luigi.Parameter', 'luigi.Parameter', ([], {'default': '"""bgzip"""'}), "(default='bgzip')\n", (1347, 1364), False, 'import luigi\n'), ((1383, 1451), 'luigi.Parameter', 'l...
import os,sys import numpy as np import random import matplotlib.pyplot as plt import seaborn as sns from copy import deepcopy import math import torch import torch.nn as nn from tqdm import tqdm from torch._six import inf import pandas as pd from PIL import Image from sklearn.feature_extraction import image from argum...
[ "copy.deepcopy", "torch.nn.CrossEntropyLoss", "sklearn.feature_extraction.image.mean", "torch.mean", "torch.no_grad", "torch.utils.data.DataLoader", "arguments.get_args", "unicodedata.numeric", "sklearn.feature_extraction.image.size", "torch.nn.functional.softmax" ]
[((348, 358), 'arguments.get_args', 'get_args', ([], {}), '()\n', (356, 358), False, 'from arguments import get_args\n'), ((2982, 2997), 'copy.deepcopy', 'deepcopy', (['model'], {}), '(model)\n', (2990, 2997), False, 'from copy import deepcopy\n'), ((3781, 3846), 'torch.utils.data.DataLoader', 'torch.utils.data.DataLoa...
# -*- coding: utf-8 -*- """ @author: miko """ from sklearn.feature_extraction import DictVectorizer import csv from sklearn import tree from sklearn import preprocessing from sklearn.externals.six import StringIO import numpy as np np.set_printoptions(threshold = 1e6)#设置打印数量的阈值 # Read in the csv file and put feat...
[ "sklearn.preprocessing.LabelBinarizer", "sklearn.feature_extraction.DictVectorizer", "sklearn.tree.DecisionTreeClassifier", "csv.reader", "numpy.set_printoptions" ]
[((236, 276), 'numpy.set_printoptions', 'np.set_printoptions', ([], {'threshold': '(1000000.0)'}), '(threshold=1000000.0)\n', (255, 276), True, 'import numpy as np\n'), ((512, 542), 'csv.reader', 'csv.reader', (['allElectronicsData'], {}), '(allElectronicsData)\n', (522, 542), False, 'import csv\n'), ((1073, 1089), 'sk...
# -*- coding: utf-8 -*- """ Bot behavior """ from irc3 import plugin, event, rfc from irc3.plugins.cron import cron from lxml import html import aiohttp import random import re @plugin class Behaviors(object): """ Defines bot's behavior by scheduling actions or handling channel events """ def...
[ "aiohttp.ClientSession", "random.choice", "re.compile", "lxml.html.fromstring", "irc3.plugins.cron.cron", "irc3.event" ]
[((1210, 1229), 'irc3.plugins.cron.cron', 'cron', (['"""0 9 * * 1-5"""'], {}), "('0 9 * * 1-5')\n", (1214, 1229), False, 'from irc3.plugins.cron import cron\n'), ((1695, 1715), 'irc3.plugins.cron.cron', 'cron', (['"""0 12 * * 1-5"""'], {}), "('0 12 * * 1-5')\n", (1699, 1715), False, 'from irc3.plugins.cron import cron\...
# -*- coding: utf-8 -*- from __future__ import unicode_literals from django.conf import settings from django.core import validators from django.db import models from django.db.models import fields as django_fields from django.db.models.signals import pre_save, post_save from django.dispatch import receiver from django...
[ "django.dispatch.receiver", "django.utils.translation.ugettext_lazy", "django.core.validators._lazy_re_compile" ]
[((561, 698), 'django.utils.translation.ugettext_lazy', '_', (['"""Kurzfassung des Namens für die Adresszeile im Browser. Vorzugsweise englisch, keine Umlaute, nur Bindestrich als Sonderzeichen."""'], {}), "('Kurzfassung des Namens für die Adresszeile im Browser. Vorzugsweise englisch, keine Umlaute, nur Bindestrich al...
import unittest from forestgame.colour import colour_to_hex class ToHexTest(unittest.TestCase): def test_convert_solid_black(self): hex_code = colour_to_hex((0, 0, 0)) self.assertEqual(hex_code, "#000000") def test_convert_solid_whex_codeite(self): hex_code = colour_to_hex((255, 255, 255)) self.a...
[ "forestgame.colour.colour_to_hex" ]
[((151, 175), 'forestgame.colour.colour_to_hex', 'colour_to_hex', (['(0, 0, 0)'], {}), '((0, 0, 0))\n', (164, 175), False, 'from forestgame.colour import colour_to_hex\n'), ((279, 309), 'forestgame.colour.colour_to_hex', 'colour_to_hex', (['(255, 255, 255)'], {}), '((255, 255, 255))\n', (292, 309), False, 'from forestg...
from setuptools import setup NAME = 'XOR_CheckSum_zsd' VERSION = '1.0.7' URL = 'https://github.com/zhangsheng377/XOR_CheckSum_zsd' KEYWORDS = 'XOR checksum string zsd' EMAIL = '<EMAIL>' DESCRIPTION = 'XOR_CheckSum tools' LONG_DESCRIPTION = ''' \>\>\> from XOR_CheckSum import xor_checksum_string \>\>\> xor_checksum_st...
[ "setuptools.setup" ]
[((593, 1119), 'setuptools.setup', 'setup', ([], {'name': 'NAME', 'author': '"""zhangsheng377"""', 'license': '"""MIT"""', 'zip_safe': '(False)', 'url': 'URL', 'version': 'VERSION', 'description': 'DESCRIPTION', 'long_description': 'LONG_DESCRIPTION', 'author_email': 'EMAIL', 'keywords': 'KEYWORDS', 'install_requires':...
from django.shortcuts import redirect def index(request): return redirect('/home/')
[ "django.shortcuts.redirect" ]
[((68, 86), 'django.shortcuts.redirect', 'redirect', (['"""/home/"""'], {}), "('/home/')\n", (76, 86), False, 'from django.shortcuts import redirect\n')]
import os, sys import logging import weakref import pygame from .resourceLoaders import * class ResourceError(Exception): pass _GAME_PATH=calculate_real_path(os.path.dirname(sys.argv[0])) def _BuildLogger(): logger = logging.getLogger(__name__) logger.setLevel(logging.DEBUG) log_format = logging.F...
[ "logging.getLogger", "logging.Formatter", "os.path.dirname", "os.path.isdir", "weakref.ref" ]
[((167, 195), 'os.path.dirname', 'os.path.dirname', (['sys.argv[0]'], {}), '(sys.argv[0])\n', (182, 195), False, 'import os, sys\n'), ((231, 258), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (248, 258), False, 'import logging\n'), ((311, 373), 'logging.Formatter', 'logging.Formatter', ...
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Thu Jan 16 15:57:57 2020 @author: snoone """ ###FTP transfer of most up to date diff file for the GHCND from NOAA/NCEI ##### import ftplib import os os.chdir(r"/gws/nopw/j04/c3s311a_lot2/data/level0/land/daily_data_processing/superghc...
[ "os.chdir", "ftplib.FTP" ]
[((235, 347), 'os.chdir', 'os.chdir', (['"""/gws/nopw/j04/c3s311a_lot2/data/level0/land/daily_data_processing/superghcnd_daily_updates"""'], {}), "(\n '/gws/nopw/j04/c3s311a_lot2/data/level0/land/daily_data_processing/superghcnd_daily_updates'\n )\n", (243, 347), False, 'import os\n'), ((421, 452), 'ftplib.FTP', ...
"""Data generator for image datasets.""" import itertools import random import numpy as np from kaishi.image.util import swap_channel_dimension from kaishi.image import ops def augment_and_label(imobj): """Augment an image with common issues and return the modified image + label vector. Labels at output laye...
[ "itertools.cycle", "kaishi.image.util.swap_channel_dimension", "random.shuffle", "kaishi.image.ops.add_stretching", "numpy.random.random", "random.seed", "numpy.stack", "numpy.zeros", "kaishi.image.ops.add_rotation", "numpy.random.seed", "kaishi.image.ops.extract_patch" ]
[((684, 698), 'numpy.zeros', 'np.zeros', (['(6,)'], {}), '((6,))\n', (692, 698), True, 'import numpy as np\n'), ((966, 984), 'numpy.random.random', 'np.random.random', ([], {}), '()\n', (982, 984), True, 'import numpy as np\n'), ((1422, 1440), 'numpy.random.random', 'np.random.random', ([], {}), '()\n', (1438, 1440), T...
import unittest from nlu import * class TestMarian(unittest.TestCase): # GENERATE NAME SPACE ENTRIES PROGRAMMATICLY?!?? def test_marian_en_to_de(self): pipe = nlu.load('en.translate_to.de',verbose=True) print('QQP') # pipe.print_info() # pipe['t5'].setTask('Answer the question') ...
[ "unittest.main" ]
[((2940, 2955), 'unittest.main', 'unittest.main', ([], {}), '()\n', (2953, 2955), False, 'import unittest\n')]
# Generated by Django 2.2.10 on 2020-04-29 14:01 from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('api', '0024_auto_20200423_1601'), ] operations = [ migrations.AlterField( model_name='imagereport', name='status', ...
[ "django.db.migrations.DeleteModel", "django.db.models.CharField" ]
[((560, 600), 'django.db.migrations.DeleteModel', 'migrations.DeleteModel', ([], {'name': '"""ImageTags"""'}), "(name='ImageTags')\n", (582, 600), False, 'from django.db import migrations, models\n'), ((338, 549), 'django.db.models.CharField', 'models.CharField', ([], {'choices': "[('pending_review', 'pending_review'),...
import pytest from streaming_form_data.validators import MaxSizeValidator, ValidationError def test_max_size_validator_empty_input(): validator = MaxSizeValidator(0) with pytest.raises(ValidationError): validator('x') def test_max_size_validator_normal(): validator = MaxSizeValidator(5) f...
[ "streaming_form_data.validators.MaxSizeValidator", "pytest.raises" ]
[((153, 172), 'streaming_form_data.validators.MaxSizeValidator', 'MaxSizeValidator', (['(0)'], {}), '(0)\n', (169, 172), False, 'from streaming_form_data.validators import MaxSizeValidator, ValidationError\n'), ((294, 313), 'streaming_form_data.validators.MaxSizeValidator', 'MaxSizeValidator', (['(5)'], {}), '(5)\n', (...
import os; import abc; import math; import multiprocessing; import psutil; import numpy as np; import matplotlib.pyplot as plt; from Errors import *; from KMeans import *; from UnaryLinearRegression import *; class _Node: def __init__(self, samplesCount, featureIndex = None, featureValue = None, leftChild = None...
[ "numpy.mat", "numpy.mean", "numpy.random.random", "numpy.random.choice", "numpy.log", "matplotlib.pyplot.plot", "numpy.quantile", "matplotlib.pyplot.figure", "psutil.cpu_count", "numpy.arange", "matplotlib.pyplot.show" ]
[((6468, 6506), 'numpy.quantile', 'np.quantile', (['scores', 'self.__proportion'], {}), '(scores, self.__proportion)\n', (6479, 6506), True, 'import numpy as np\n'), ((11234, 11254), 'numpy.mat', 'np.mat', (['self._curves'], {}), '(self._curves)\n', (11240, 11254), True, 'import numpy as np\n'), ((2021, 2030), 'numpy.l...
import torch def model_to_vector(model, emb_layer_name='input_emb'): """ get the wordvec weight :param model: :param emb_layer_name: :return: """ sd = model.state_dict() return sd[emb_layer_name + '.weight'].cpu().numpy().tolist() def save_embedding(file_name, embeddings, id2word): ...
[ "torch.tensor" ]
[((1010, 1055), 'torch.tensor', 'torch.tensor', (['vali_examples'], {'dtype': 'torch.long'}), '(vali_examples, dtype=torch.long)\n', (1022, 1055), False, 'import torch\n')]
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Mon Oct 15 15:13:29 2018 @author: kazuki.onodera """ import numpy as np import pandas as pd import os, gc from glob import glob from tqdm import tqdm import sys sys.path.append(f'/home/{os.environ.get("USER")}/PythonLibrary') import lgbextension as ex imp...
[ "numpy.clip", "pandas.read_csv", "utils.send_line", "numpy.log", "multiprocessing.cpu_count", "utils.start", "lightgbm.Dataset", "utils.submit", "utils.savefig_sub", "pandas.read_feather", "pandas.DataFrame", "glob.glob", "utils.stop_instance", "numpy.ones", "utils.savefig_imp", "gc.co...
[((392, 413), 'utils.start', 'utils.start', (['__file__'], {}), '(__file__)\n', (403, 413), False, 'import utils\n'), ((627, 650), 'numpy.random.randint', 'np.random.randint', (['(9999)'], {}), '(9999)\n', (644, 650), True, 'import numpy as np\n'), ((2473, 2485), 'gc.collect', 'gc.collect', ([], {}), '()\n', (2483, 248...
#!/usr/bin/env python # -*- coding: utf-8 -*- from setuptools import setup, find_packages import sys, os import tracking setup( name='django-tracking', version=tracking.get_version(), description="Basic visitor tracking and blacklisting for Django", long_description=open('README.rst', 'r').read(), ...
[ "setuptools.find_packages", "tracking.get_version" ]
[((170, 192), 'tracking.get_version', 'tracking.get_version', ([], {}), '()\n', (190, 192), False, 'import tracking\n'), ((572, 587), 'setuptools.find_packages', 'find_packages', ([], {}), '()\n', (585, 587), False, 'from setuptools import setup, find_packages\n')]
#!/usr/bin/python # (c) 2018, NetApp, Inc # GNU General Public License v3.0+ (see COPYING or https://www.gnu.org/licenses/gpl-3.0.txt) from __future__ import absolute_import, division, print_function __metaclass__ = type DOCUMENTATION = """ --- module: update_file_list short_description: Update log file containing c...
[ "ansible.module_utils.basic.AnsibleModule" ]
[((1838, 1937), 'ansible.module_utils.basic.AnsibleModule', 'AnsibleModule', ([], {'argument_spec': 'ansible_options', 'required_if': 'required_if', 'supports_check_mode': '(True)'}), '(argument_spec=ansible_options, required_if=required_if,\n supports_check_mode=True)\n', (1851, 1937), False, 'from ansible.module_u...
import os import subprocess def render_movie(output_filename, frames_dir, frame_rate): subprocess.call([ 'ffmpeg', '-vcodec', 'png', '-framerate', str(frame_rate), '-i', os.path.join(frames_dir, r'frame.%06d.png'), '-pix_fmt', 'yuv420p', '-vcodec', 'libx264', '-crf', '17', '-thread...
[ "os.path.join" ]
[((200, 242), 'os.path.join', 'os.path.join', (['frames_dir', '"""frame.%06d.png"""'], {}), "(frames_dir, 'frame.%06d.png')\n", (212, 242), False, 'import os\n')]
import sys import os import numpy.random from amuse.test import amusetest from amuse.units import units, nbody_system from amuse.ext.boss_bodenheimer import bb79_cloud numpy.random.seed(1234567) class BossBodenheimerTests(amusetest.TestCase): def test1(self): numpy.random.seed(1234) mc=bb79_cloud...
[ "amuse.ext.boss_bodenheimer.bb79_cloud", "amuse.units.nbody_system.nbody_to_si" ]
[((659, 721), 'amuse.units.nbody_system.nbody_to_si', 'nbody_system.nbody_to_si', (['(1.0 | units.MSun)', '(3.2e+16 | units.cm)'], {}), '(1.0 | units.MSun, 3.2e+16 | units.cm)\n', (683, 721), False, 'from amuse.units import units, nbody_system\n'), ((310, 334), 'amuse.ext.boss_bodenheimer.bb79_cloud', 'bb79_cloud', ([]...
#!/usr/bin/env python # Copyright 2018- The Pixie Authors. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicabl...
[ "ssl.create_default_context", "time.sleep", "random.randint", "socket.socket" ]
[((868, 939), 'ssl.create_default_context', 'ssl.create_default_context', (['ssl.Purpose.SERVER_AUTH'], {'cafile': 'server_cert'}), '(ssl.Purpose.SERVER_AUTH, cafile=server_cert)\n', (894, 939), False, 'import ssl\n'), ((1011, 1060), 'socket.socket', 'socket.socket', (['socket.AF_INET', 'socket.SOCK_STREAM'], {}), '(so...
""" Copyright 2012-2019 <NAME> Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed to in writing, software di...
[ "hqlib.utils.format_link_object" ]
[((1893, 1933), 'hqlib.utils.format_link_object', 'utils.format_link_object', (['cve[1]', 'cve[0]'], {}), '(cve[1], cve[0])\n', (1917, 1933), False, 'from hqlib import utils\n')]
# This script goes through OpenSim funcionalties # required for OpenSim-RL import opensim # Settings stepsize = 0.01 # Load existing model model_path = "../osim/models/gait9dof18musc.osim" model = opensim.Model(model_path) model.setUseVisualizer(True) # Create the ball r = 0.000001 ballBody = opensim.B...
[ "opensim.Manager", "opensim.Model", "opensim.Inertia", "opensim.Ellipsoid", "opensim.Vec3" ]
[((208, 233), 'opensim.Model', 'opensim.Model', (['model_path'], {}), '(model_path)\n', (221, 233), False, 'import opensim\n'), ((410, 436), 'opensim.Ellipsoid', 'opensim.Ellipsoid', (['r', 'r', 'r'], {}), '(r, r, r)\n', (427, 436), False, 'import opensim\n'), ((1273, 1295), 'opensim.Manager', 'opensim.Manager', (['mod...
#!/usr/bin/env python import os import sys BASE_APP_DIR = os.path.dirname(os.path.dirname(os.path.abspath(__file__))) sys.path.append(os.path.join(BASE_APP_DIR, 'src/')) sys.path.append(os.path.join(BASE_APP_DIR, 'venv/lib/python3.4/site-packages')) os.environ.setdefault("DJANGO_SETTINGS_MODULE", "app.settings") fr...
[ "os.environ.setdefault", "django.core.wsgi.get_wsgi_application", "os.path.join", "os.path.abspath" ]
[((253, 316), 'os.environ.setdefault', 'os.environ.setdefault', (['"""DJANGO_SETTINGS_MODULE"""', '"""app.settings"""'], {}), "('DJANGO_SETTINGS_MODULE', 'app.settings')\n", (274, 316), False, 'import os\n'), ((382, 404), 'django.core.wsgi.get_wsgi_application', 'get_wsgi_application', ([], {}), '()\n', (402, 404), Fal...
# Generated by Django 3.1.7 on 2021-04-07 15:24 from django.db import migrations, models import django.db.models.deletion class Migration(migrations.Migration): dependencies = [ ('projectroles', '0019_project_public_guest_access'), ('timeline', '0006_update_user_optional'), ] operations...
[ "django.db.models.ForeignKey" ]
[((441, 648), 'django.db.models.ForeignKey', 'models.ForeignKey', ([], {'help_text': '"""Project to which the event belongs (null for no project)"""', 'null': '(True)', 'on_delete': 'django.db.models.deletion.CASCADE', 'related_name': '"""events"""', 'to': '"""projectroles.project"""'}), "(help_text=\n 'Project to w...
from collections import defaultdict def dict_sample_target_iter_concat(sample_target_iter: iter): """Gets an sample target iterator of dict samples. Returns a concatenation of the samples based on each key and the target list. Example: :: sample_target_iter = [({'a': 'a1', 'b': 'b1'},...
[ "collections.defaultdict" ]
[((546, 563), 'collections.defaultdict', 'defaultdict', (['list'], {}), '(list)\n', (557, 563), False, 'from collections import defaultdict\n')]
# coding: utf-8 from __future__ import absolute_import from datetime import date, datetime # noqa: F401 from typing import List, Dict # noqa: F401 from tapi_server.models.base_model_ import Model from tapi_server.models.tapi_common_local_class import TapiCommonLocalClass # noqa: F401,E501 from tapi_server.models....
[ "tapi_server.util.deserialize_model" ]
[((2650, 2683), 'tapi_server.util.deserialize_model', 'util.deserialize_model', (['dikt', 'cls'], {}), '(dikt, cls)\n', (2672, 2683), False, 'from tapi_server import util\n')]
from projex.lazymodule import lazy_import from ..sqliteconnection import SQLiteStatement orb = lazy_import('orb') class CREATE(SQLiteStatement): def __call__(self, model, owner='', includeReferences=True): if issubclass(model, orb.Table): return self._createTable(model, owner, includeReferenc...
[ "projex.lazymodule.lazy_import" ]
[((96, 114), 'projex.lazymodule.lazy_import', 'lazy_import', (['"""orb"""'], {}), "('orb')\n", (107, 114), False, 'from projex.lazymodule import lazy_import\n')]
from raytkTools import RaytkTools # noinspection PyUnreachableCode if False: # noinspection PyUnresolvedReferences from _stubs import * from ..ropEditor.ropEditor import ROPEditor iop.ropEditor = ROPEditor(COMP()) class CreateRopDialog: def __init__(self, ownerComp: 'COMP'): self.ownerComp = ownerComp def _s...
[ "raytkTools.RaytkTools" ]
[((894, 906), 'raytkTools.RaytkTools', 'RaytkTools', ([], {}), '()\n', (904, 906), False, 'from raytkTools import RaytkTools\n')]
""" Test the stream module Does it provide an interface compatible with circus? """ from calendar import timegm from io import StringIO from pytest import fixture from mock import patch from codado import parseDate, py from circusbase import stream @fixture def stdoutStream(pRemoji): ret = stream.EmojiStdout...
[ "mock.patch.object", "circusbase.stream.EmojiStdoutStream", "io.StringIO", "codado.parseDate" ]
[((302, 328), 'circusbase.stream.EmojiStdoutStream', 'stream.EmojiStdoutStream', ([], {}), '()\n', (326, 328), False, 'from circusbase import stream\n'), ((344, 354), 'io.StringIO', 'StringIO', ([], {}), '()\n', (352, 354), False, 'from io import StringIO\n'), ((656, 693), 'codado.parseDate', 'parseDate', (['"""2018-11...
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Thu Jul 19 15:37:36 2018 @author: phoenix """ import glob path = glob.glob('dataset/*') for each in path: print(glob.glob(each+'/*')[0].split('/')[1]) dynamic_dict = { glob.glob(each+'/*')[0].split('/')[1]:glob.glob(each+'/*')[0] for each in path} pr...
[ "glob.glob" ]
[((130, 152), 'glob.glob', 'glob.glob', (['"""dataset/*"""'], {}), "('dataset/*')\n", (139, 152), False, 'import glob\n'), ((276, 298), 'glob.glob', 'glob.glob', (["(each + '/*')"], {}), "(each + '/*')\n", (285, 298), False, 'import glob\n'), ((238, 260), 'glob.glob', 'glob.glob', (["(each + '/*')"], {}), "(each + '/*'...
import typer cli = typer.Typer()
[ "typer.Typer" ]
[((20, 33), 'typer.Typer', 'typer.Typer', ([], {}), '()\n', (31, 33), False, 'import typer\n')]
"""ResNets. With only TensorFlow backend. Follows fchollet's design Author: <NAME> Email : <EMAIL> """ from __future__ import print_function, absolute_import from builtins import range from spiker import log logger = log.get_logger("models-resnet", log.DEBUG) try: from keras.layers import Input from keras...
[ "keras.layers.add", "keras.layers.Input", "builtins.range", "spiker.log.get_logger", "keras.layers.Activation", "keras.models.Model", "keras.regularizers.l2", "keras.layers.GlobalAveragePooling2D", "keras.layers.BatchNormalization" ]
[((222, 264), 'spiker.log.get_logger', 'log.get_logger', (['"""models-resnet"""', 'log.DEBUG'], {}), "('models-resnet', log.DEBUG)\n", (236, 264), False, 'from spiker import log\n'), ((3569, 3587), 'keras.layers.add', 'add', (['[x, shortcut]'], {}), '([x, shortcut])\n', (3572, 3587), False, 'from keras.layers import ad...
from pathlib import Path from openpecha import formatters from openpecha.formatters import PedurmaFormatter from openpecha.serializers import PedurmaSerializer if __name__ == "__main__": opf_path = "./output/opfs/D1111/D1111.opf/" opfs_path = "./output/opfs/" preview_path = "./output/D1111/" # previe...
[ "openpecha.serializers.PedurmaSerializer", "openpecha.formatters.PedurmaFormatter" ]
[((345, 384), 'openpecha.formatters.PedurmaFormatter', 'PedurmaFormatter', ([], {'output_path': 'opfs_path'}), '(output_path=opfs_path)\n', (361, 384), False, 'from openpecha.formatters import PedurmaFormatter\n'), ((466, 493), 'openpecha.serializers.PedurmaSerializer', 'PedurmaSerializer', (['opf_path'], {}), '(opf_pa...
from logmagic.loggerutil import make_logger from unittest import TestCase import unittest import sys from io import StringIO import contextlib import re class ConsoleOutTest(TestCase): def test_debugout(self): # global saved_stdout # Use sys.__stdout__ instead # saved_stdout = sys.stdout ...
[ "contextlib.redirect_stdout", "re.fullmatch", "contextlib.redirect_stderr", "logmagic.loggerutil.make_logger", "unittest.main", "io.StringIO" ]
[((978, 993), 'unittest.main', 'unittest.main', ([], {}), '()\n', (991, 993), False, 'import unittest\n'), ((361, 371), 'io.StringIO', 'StringIO', ([], {}), '()\n', (369, 371), False, 'from io import StringIO\n'), ((740, 890), 're.fullmatch', 're.fullmatch', (['"""[0-9]{4}\\\\-[0-9]{2}\\\\-[0-9]{2} [0-9]{2}:[0-9]{2}:[0...
import logging import re import gzip from pathlib import Path from argparse import ArgumentParser from Bio import AlignIO from make_prg.from_msa import MSA from make_prg.prg_encoder import PrgEncoder, PRG_Ints def load_alignment_file(msa_file: str, alignment_format: str) -> MSA: msa_file = str(msa_file) log...
[ "Bio.AlignIO.read", "logging.debug", "gzip.open", "make_prg.prg_encoder.PrgEncoder", "re.finditer", "logging.info" ]
[((317, 364), 'logging.info', 'logging.info', (['"""Read from MSA file %s"""', 'msa_file'], {}), "('Read from MSA file %s', msa_file)\n", (329, 364), False, 'import logging\n'), ((406, 437), 'logging.debug', 'logging.debug', (['"""MSA is gzipped"""'], {}), "('MSA is gzipped')\n", (419, 437), False, 'import logging\n'),...
from __future__ import print_function import os import warnings import shutil import re import numpy as np import sami from astropy.io import fits class Tester(): """This class handles the testing of the SAMI pipeline. To run it requires additional data: 1) the folder ``sami_ppl_test_data``, i.e. th...
[ "os.path.exists", "re.compile", "numpy.testing.assert_allclose", "shutil.rmtree", "os.path.basename", "astropy.io.fits.open", "warnings.warn", "numpy.all", "os.walk", "sami.manager.Manager" ]
[((5894, 5937), 'sami.manager.Manager', 'sami.manager.Manager', (['output_dir'], {'fast': 'fast'}), '(output_dir, fast=fast)\n', (5914, 5937), False, 'import sami\n'), ((8989, 9006), 'numpy.all', 'np.all', (['all_equal'], {}), '(all_equal)\n', (8995, 9006), True, 'import numpy as np\n'), ((11015, 11032), 're.compile', ...
from __future__ import absolute_import, division, print_function, unicode_literals import csv def csv_filename_to_objects(filename, json_handler): with open(filename, "r") as f: objects = csv_stream_to_objects(f, json_handler=json_handler) return objects def csv_stream_to_objects(stream, json_handl...
[ "csv.DictWriter", "csv.DictReader" ]
[((353, 375), 'csv.DictReader', 'csv.DictReader', (['stream'], {}), '(stream)\n', (367, 375), False, 'import csv\n'), ((1204, 1249), 'csv.DictWriter', 'csv.DictWriter', (['stream'], {'fieldnames': 'fieldnames'}), '(stream, fieldnames=fieldnames)\n', (1218, 1249), False, 'import csv\n')]
import os, re FILE_DIR = os.path.dirname(os.path.abspath(__file__)) README_PATH = os.path.join(FILE_DIR, os.pardir, 'README.md') README_CODE_PATH = os.path.join(FILE_DIR, 'README_commands.py') CODE_INIT = ''' #------------ CODE INIT ------------ # make sure imports from ../src work import os, sys FILE_DIR = os.path.d...
[ "os.path.abspath", "os.path.join", "re.match" ]
[((83, 129), 'os.path.join', 'os.path.join', (['FILE_DIR', 'os.pardir', '"""README.md"""'], {}), "(FILE_DIR, os.pardir, 'README.md')\n", (95, 129), False, 'import os, re\n'), ((149, 193), 'os.path.join', 'os.path.join', (['FILE_DIR', '"""README_commands.py"""'], {}), "(FILE_DIR, 'README_commands.py')\n", (161, 193), Fa...
#!/usr/bin/env python3 import os, re, random import numpy as np from myClass import Theta from functions import grad_loss from hyperParameters import hyperParameters # hyperParameters T, D = hyperParameters.T, hyperParameters.D alpha, epsilon = hyperParameters.alpha, hyperParameters.epsilon moment = hyperParameters.m...
[ "functions.grad_loss", "os.listdir", "os.getcwd", "numpy.array", "numpy.zeros", "re.search" ]
[((359, 370), 'os.getcwd', 'os.getcwd', ([], {}), '()\n', (368, 370), False, 'import os, re, random\n'), ((430, 445), 'os.listdir', 'os.listdir', (['dir'], {}), '(dir)\n', (440, 445), False, 'import os, re, random\n'), ((449, 478), 're.search', 're.search', (['"""_graph.txt"""', 'file'], {}), "('_graph.txt', file)\n", ...
from netmiko import ConnectHandler # our devices info to connect and session log files to save output arista_1 = { 'device_type': 'arista_eos', 'ip': 'your_ip', 'username': 'your_username', 'password': '<PASSWORD>', 'session_log': 'sw1.txt' } arista_2 = { 'device_type': 'arista_eos', 'ip':...
[ "netmiko.ConnectHandler" ]
[((872, 896), 'netmiko.ConnectHandler', 'ConnectHandler', ([], {}), '(**device)\n', (886, 896), False, 'from netmiko import ConnectHandler\n')]
from django.contrib.auth import authenticate from rest_framework import viewsets from rest_framework.authtoken.models import Token from rest_framework.authentication import TokenAuthentication from rest_framework.permissions import IsAuthenticated from rest_framework.response import Response from rest_framework.status ...
[ "rest_framework.response.Response", "rest_framework.authtoken.models.Token.objects.get_or_create" ]
[((784, 822), 'rest_framework.authtoken.models.Token.objects.get_or_create', 'Token.objects.get_or_create', ([], {'user': 'user'}), '(user=user)\n', (811, 822), False, 'from rest_framework.authtoken.models import Token\n'), ((838, 888), 'rest_framework.response.Response', 'Response', (["{'token': token.key}"], {'status...
"""empty message Revision ID: f5a5f9bbd6e2 Revises: <PASSWORD> Create Date: 2018-04-20 19:02:32.403757 """ from alembic import op import sqlalchemy as sa # revision identifiers, used by Alembic. revision = 'f5a5f9bbd6e2' down_revision = '<PASSWORD>' branch_labels = None depends_on = None def upgrade(): op.add...
[ "alembic.op.drop_constraint", "alembic.op.drop_column", "sqlalchemy.Integer", "alembic.op.create_foreign_key" ]
[((404, 503), 'alembic.op.create_foreign_key', 'op.create_foreign_key', (['"""fk_categories_file"""', '"""categories"""', '"""files"""', "['cover_image_id']", "['id']"], {}), "('fk_categories_file', 'categories', 'files', [\n 'cover_image_id'], ['id'])\n", (425, 503), False, 'from alembic import op\n'), ((589, 680),...
from __future__ import annotations import pygame from scripts.engine import world from scripts.engine.component import Position from scripts.engine.core import queries from scripts.engine.core.constants import GameEvent __all__ = ["process_win_condition"] def process_win_condition(): """ Process the win co...
[ "scripts.engine.world.get_entitys_component", "scripts.engine.world.get_player", "pygame.event.post", "pygame.event.Event" ]
[((380, 398), 'scripts.engine.world.get_player', 'world.get_player', ([], {}), '()\n', (396, 398), False, 'from scripts.engine import world\n'), ((416, 461), 'scripts.engine.world.get_entitys_component', 'world.get_entitys_component', (['player', 'Position'], {}), '(player, Position)\n', (443, 461), False, 'from script...
# Generated by Django 3.1.5 on 2021-09-02 21:07 from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('newapp1', '0004_auto_20210830_2333'), ] operations = [ migrations.AddField( model_name='userreview', name='prolific_...
[ "django.db.models.CharField" ]
[((343, 398), 'django.db.models.CharField', 'models.CharField', ([], {'blank': '(True)', 'max_length': '(300)', 'null': '(True)'}), '(blank=True, max_length=300, null=True)\n', (359, 398), False, 'from django.db import migrations, models\n')]
# Time: O(|V| + |E|) # Space: O(|V| + |E|) # 886 # Given a set of N people (numbered 1, 2, ..., N), # we would like to split everyone into two groups of any size. # # Each person may dislike some other people, # and they should not go into the same group. # # Formally, if dislikes[i] = [a, b], # it means it is not a...
[ "collections.deque", "collections.defaultdict" ]
[((1478, 1506), 'collections.defaultdict', 'collections.defaultdict', (['set'], {}), '(set)\n', (1501, 1506), False, 'import collections\n'), ((2220, 2249), 'collections.defaultdict', 'collections.defaultdict', (['list'], {}), '(list)\n', (2243, 2249), False, 'import collections\n'), ((2355, 2376), 'collections.deque',...
import numpy as np import matplotlib.pyplot as plt import matplotlib.animation as animation import random as rdm # temporary variable, only used in testing, replace with actual graph inputs # currently, generates 100 evenly spaced numbers between 0 and pi ps1 = np.linspace(0, 1 * np.pi, 100) # create the graph plot f...
[ "numpy.sin", "numpy.linspace", "matplotlib.pyplot.pause", "matplotlib.pyplot.subplots", "matplotlib.pyplot.show" ]
[((263, 293), 'numpy.linspace', 'np.linspace', (['(0)', '(1 * np.pi)', '(100)'], {}), '(0, 1 * np.pi, 100)\n', (274, 293), True, 'import numpy as np\n'), ((329, 343), 'matplotlib.pyplot.subplots', 'plt.subplots', ([], {}), '()\n', (341, 343), True, 'import matplotlib.pyplot as plt\n'), ((622, 643), 'matplotlib.pyplot.s...
import urllib.request import shutil import requests ... # Download the file from `url` and save it locally under `file_name`: from fasp.loc import sdlDRSClient def download(url): print('in download') file_name = 'test_file.bam' req = requests.get(url) print (req) file = open(file_name, 'wb') for chunk ...
[ "fasp.loc.sdlDRSClient", "requests.get" ]
[((248, 265), 'requests.get', 'requests.get', (['url'], {}), '(url)\n', (260, 265), False, 'import requests\n'), ((713, 769), 'fasp.loc.sdlDRSClient', 'sdlDRSClient', (['"""~/.keys/prj_11218_D17199.ngc"""'], {'debug': '(True)'}), "('~/.keys/prj_11218_D17199.ngc', debug=True)\n", (725, 769), False, 'from fasp.loc import...
import pytest from pettingzoo.test.api_test import api_test from pettingzoo.test.max_cycles_test import max_cycles_test from pettingzoo.test.parallel_test import parallel_api_test from pettingzoo.test.render_test import render_test from pettingzoo.test.seed_test import check_environment_deterministic, seed_test from p...
[ "pytest.mark.parametrize", "pettingzoo.test.state_test.state_test", "pettingzoo.test.api_test.api_test" ]
[((9632, 9709), 'pytest.mark.parametrize', 'pytest.mark.parametrize', (["['name', 'env_module', 'kwargs']", 'parameterized_envs'], {}), "(['name', 'env_module', 'kwargs'], parameterized_envs)\n", (9655, 9709), False, 'import pytest\n'), ((9793, 9807), 'pettingzoo.test.api_test.api_test', 'api_test', (['_env'], {}), '(_...
from flask_wtf import FlaskForm from wtforms import StringField, PasswordField, SubmitField, TextAreaField, SelectField, BooleanField from wtforms.validators import InputRequired, Length, EqualTo, ValidationError, Email, Required from ..models import User from wtforms import ValidationError from flask_wtf.file import F...
[ "wtforms.validators.Email", "wtforms.validators.InputRequired", "flask_wtf.file.FileAllowed", "wtforms.BooleanField", "wtforms.SubmitField", "wtforms.validators.EqualTo", "wtforms.validators.Required", "wtforms.validators.Length", "wtforms.ValidationError" ]
[((1399, 1424), 'wtforms.BooleanField', 'BooleanField', (['"""Subscribe"""'], {}), "('Subscribe')\n", (1411, 1424), False, 'from wtforms import StringField, PasswordField, SubmitField, TextAreaField, SelectField, BooleanField\n'), ((1445, 1467), 'wtforms.SubmitField', 'SubmitField', (['"""Sign up"""'], {}), "('Sign up'...
# # Copyright (c) 2017 nexB Inc. and others. All rights reserved. # http://nexb.com and https://github.com/nexB/scancode-toolkit/ # The ScanCode software is licensed under the Apache License version 2.0. # Data generated with ScanCode require an acknowledgment. # ScanCode is a trademark of nexB Inc. # # You may not use...
[ "textcode.analysis.unicode_text", "chardet.detect", "lxml.etree.XMLParser", "lxml.etree.parse" ]
[((2217, 2294), 'lxml.etree.XMLParser', 'etree.XMLParser', ([], {'recover': '(True)', 'remove_blank_text': '(True)', 'resolve_entities': '(False)'}), '(recover=True, remove_blank_text=True, resolve_entities=False)\n', (2232, 2294), False, 'from lxml import etree\n'), ((2310, 2339), 'lxml.etree.parse', 'etree.parse', ([...