code
stringlengths
22
1.05M
apis
listlengths
1
3.31k
extract_api
stringlengths
75
3.25M
""" HexitecDAQ for Hexitec ODIN control. <NAME>, STFC Detector Systems Software Group """ import logging from functools import partial from tornado.ioloop import IOLoop from odin.adapters.adapter import ApiAdapterRequest from odin.adapters.parameter_tree import ParameterTree, ParameterTreeError from hexitec.Generat...
[ "os.path.exists", "logging.debug", "os.path.join", "logging.info", "logging.warning", "os.getcwd", "h5py.File", "hexitec.GenerateConfigFiles.GenerateConfigFiles", "os.path.isfile", "tornado.ioloop.IOLoop.instance", "odin.adapters.adapter.ApiAdapterRequest", "datetime.datetime.now", "functool...
[((2072, 2083), 'os.getcwd', 'os.getcwd', ([], {}), '()\n', (2081, 2083), False, 'import os\n'), ((7839, 7878), 'logging.debug', 'logging.debug', (['"""Setting up Acquisition"""'], {}), "('Setting up Acquisition')\n", (7852, 7878), False, 'import logging\n'), ((9097, 9225), 'logging.info', 'logging.info', (['"""FRAME S...
from pathlib import Path import pandas as pd flights = pd.concat( pd.read_csv(file, parse_dates=["firstseen", "lastseen", "day"]) for file in Path("../dataset").glob("flightlist_*.csv.gz") ) flights_dropna = flights.dropna(axis=0, subset=['registration', 'typecode','origin','destination', 'altitude_2']) print(...
[ "pandas.read_csv", "pathlib.Path" ]
[((71, 134), 'pandas.read_csv', 'pd.read_csv', (['file'], {'parse_dates': "['firstseen', 'lastseen', 'day']"}), "(file, parse_dates=['firstseen', 'lastseen', 'day'])\n", (82, 134), True, 'import pandas as pd\n'), ((151, 169), 'pathlib.Path', 'Path', (['"""../dataset"""'], {}), "('../dataset')\n", (155, 169), False, 'fr...
# -*- coding: utf-8 -*- # # This code is part of Qiskit. # # (C) Copyright IBM 2019, 2020. # # This code is licensed under the Apache License, Version 2.0. You may # obtain a copy of this license in the LICENSE.txt file in the root directory # of this source tree or at http://www.apache.org/licenses/LICENSE-2.0. # # An...
[ "numpy.abs", "qiskit.ignis.verification.tomography.basis.default_gateset_basis", "qiskit.ignis.verification.tomography.GatesetTomographyFitter", "numpy.sqrt", "qiskit.ignis.verification.tomography.gateset_tomography_circuits", "numpy.nditer", "qiskit.compiler.assemble", "qiskit.extensions.SGate", "n...
[((6227, 6242), 'unittest.main', 'unittest.main', ([], {}), '()\n', (6240, 6242), False, 'import unittest\n'), ((1256, 1289), 'qiskit.Aer.get_backend', 'Aer.get_backend', (['"""qasm_simulator"""'], {}), "('qasm_simulator')\n", (1271, 1289), False, 'from qiskit import Aer\n'), ((1309, 1365), 'qiskit.ignis.verification.t...
#!/usr/bin/env python3 import dataclasses import enum import json import pathlib import subprocess import textwrap from typing import Any, Dict, Iterable, List, Optional, Set, Tuple, TypeVar import inflection import jinja2 from jinja2 import StrictUndefined PROJECT_ROOT = pathlib.Path(__file__).parent.parent API_INP...
[ "jinja2.Environment", "pathlib.Path", "json.load", "textwrap.wrap", "dataclasses.replace", "inflection.camelize", "inflection.singularize", "dataclasses.field", "typing.TypeVar" ]
[((590, 602), 'typing.TypeVar', 'TypeVar', (['"""T"""'], {}), "('T')\n", (597, 602), False, 'from typing import Any, Dict, Iterable, List, Optional, Set, Tuple, TypeVar\n'), ((5875, 5926), 'dataclasses.field', 'dataclasses.field', ([], {'init': '(False)', 'default_factory': 'list'}), '(init=False, default_factory=list)...
import datetime entrada = input("Informe a data: ") datainicial = datetime.datetime.strptime(entrada,"dd/mm/aa")
[ "datetime.datetime.strptime" ]
[((68, 115), 'datetime.datetime.strptime', 'datetime.datetime.strptime', (['entrada', '"""dd/mm/aa"""'], {}), "(entrada, 'dd/mm/aa')\n", (94, 115), False, 'import datetime\n')]
"""Support for the Airzone diagnostics.""" from __future__ import annotations from typing import Any from aioairzone.const import AZD_MAC from homeassistant.components.diagnostics.util import async_redact_data from homeassistant.config_entries import ConfigEntry from homeassistant.core import HomeAssistant from .co...
[ "homeassistant.components.diagnostics.util.async_redact_data" ]
[((702, 749), 'homeassistant.components.diagnostics.util.async_redact_data', 'async_redact_data', (['config_entry.data', 'TO_REDACT'], {}), '(config_entry.data, TO_REDACT)\n', (719, 749), False, 'from homeassistant.components.diagnostics.util import async_redact_data\n'), ((767, 813), 'homeassistant.components.diagnost...
import torch import torch.nn as nn class CNNLayer(nn.Module): """ This layer is callable for 1d convolution and pooling functions with fatten result """ def __init__(self, input_dim, kernel_size=(3, 4, 5), kernel_num=200): """ :par...
[ "torch.cat", "torch.nn.Conv2d" ]
[((960, 983), 'torch.cat', 'torch.cat', (['pooling_x', '(1)'], {}), '(pooling_x, 1)\n', (969, 983), False, 'import torch\n'), ((694, 735), 'torch.nn.Conv2d', 'nn.Conv2d', (['(1)', 'kernel_num', '(ks, input_dim)'], {}), '(1, kernel_num, (ks, input_dim))\n', (703, 735), True, 'import torch.nn as nn\n')]
"""Tests for lookup handler for env.""" # pylint: disable=no-self-use # pyright: basic from __future__ import annotations from typing import TYPE_CHECKING import pytest from runway.lookups.handlers.env import EnvLookup if TYPE_CHECKING: from ...factories import MockRunwayContext ENV_VARS = {"str_val": "test"} ...
[ "pytest.raises", "runway.lookups.handlers.env.EnvLookup.handle" ]
[((562, 613), 'runway.lookups.handlers.env.EnvLookup.handle', 'EnvLookup.handle', (['"""str_val"""'], {'context': 'runway_context'}), "('str_val', context=runway_context)\n", (578, 613), False, 'from runway.lookups.handlers.env import EnvLookup\n'), ((855, 880), 'pytest.raises', 'pytest.raises', (['ValueError'], {}), '...
import pandas as pd from os import listdir from datetime import datetime as dtt import logging import json logger = logging.getLogger(__name__) logger.setLevel(logging.DEBUG) class WalletManager: def __init__(self, client, states): self.binance = client self.states = states self._wallet_h...
[ "logging.getLogger" ]
[((117, 144), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (134, 144), False, 'import logging\n')]
import tensorflow as tf from wacky_rl import losses from wacky_rl import models from itertools import count class SharedNetLoss(losses.WackyLoss): _ids = count(0) def __init__(self, alphas: list =None, sub_models=None, sub_agents=None, logger=None): super().__init__() self.alphas = alphas ...
[ "itertools.count" ]
[((161, 169), 'itertools.count', 'count', (['(0)'], {}), '(0)\n', (166, 169), False, 'from itertools import count\n')]
import cv2 from shapes import Myinit class Circle(Myinit): def __init__(self): super(Circle, self).__init__() self.center = (100,100) self.radius = 50 self.color = (0,0,255) self.thickness = -1 def form_shape(self): self.img = cv2.circle(self.img, se...
[ "cv2.circle", "cv2.waitKey", "cv2.destroyAllWindows", "cv2.imshow" ]
[((297, 359), 'cv2.circle', 'cv2.circle', (['self.img', 'self.center', 'self.radius', 'self.color', '(-1)'], {}), '(self.img, self.center, self.radius, self.color, -1)\n', (307, 359), False, 'import cv2\n'), ((599, 629), 'cv2.imshow', 'cv2.imshow', (['"""Circle"""', 'self.img'], {}), "('Circle', self.img)\n", (609, 629...
""" Unit test script for convert.py """ import unittest import pyeto class TestConvert(unittest.TestCase): def test_celsius2kelvin(self): self.assertEqual(pyeto.celsius2kelvin(0), 273.15) def test_kelvin2celsius(self): self.assertEqual(pyeto.kelvin2celsius(273.15), 0.0) def test_deg2r...
[ "pyeto.celsius2kelvin", "pyeto.deg2rad", "pyeto.kelvin2celsius", "unittest.main", "pyeto.rad2deg" ]
[((1026, 1041), 'unittest.main', 'unittest.main', ([], {}), '()\n', (1039, 1041), False, 'import unittest\n'), ((172, 195), 'pyeto.celsius2kelvin', 'pyeto.celsius2kelvin', (['(0)'], {}), '(0)\n', (192, 195), False, 'import pyeto\n'), ((266, 294), 'pyeto.kelvin2celsius', 'pyeto.kelvin2celsius', (['(273.15)'], {}), '(273...
# important necessary classes and functions import math from node import Node from queue_assignment import Queue from essentials import transition_nodes from heuristic import HeuristicCalc from copy import deepcopy from flatland.envs.rail_env import RailEnv # class for solving single agent using sipp class SIPP: "...
[ "queue_assignment.Queue", "node.Node", "essentials.transition_nodes", "copy.deepcopy" ]
[((4384, 4391), 'queue_assignment.Queue', 'Queue', ([], {}), '()\n', (4389, 4391), False, 'from queue_assignment import Queue\n'), ((4842, 4986), 'node.Node', 'Node', ([], {'loc': 'start', 'direction': 'start_direction', 'goal': 'goal', 'heuristic_func': 'heuristic_calc.calculate_heuristic', 'heuristic_weight': 'heuris...
from scipy.optimize import fsolve import numpy as np from household_dist import HOUSEHOLD_DIST def compute_household_infection_prob(prevalence, household_dist, SAR=0.3741): """ computes the probability that a household is infected given population level prevalence, household size distribution and househol...
[ "scipy.optimize.fsolve", "numpy.sum" ]
[((2099, 2177), 'scipy.optimize.fsolve', 'fsolve', (['match_prevalence', '(0.005)'], {'args': '(target_prevalence, household_dist, SAR)'}), '(match_prevalence, 0.005, args=(target_prevalence, household_dist, SAR))\n', (2105, 2177), False, 'from scipy.optimize import fsolve\n'), ((564, 586), 'numpy.sum', 'np.sum', (['ho...
import torch import torch.nn as nn # from torch.nn import init import functools # from torch.autograd import Variable import numpy as np import pdb ############################################################################### # Functions ##############################################################################...
[ "torch.nn.Sigmoid", "torch.nn.ReLU", "torch.nn.parallel.data_parallel", "numpy.ceil", "torch.nn.Tanh", "torch.nn.LeakyReLU", "torch.nn.Dropout", "torch.nn.Sequential", "torch.nn.ReflectionPad2d", "torch.nn.Conv2d", "torch.cuda.is_available", "functools.partial", "torch.nn.ReplicationPad2d", ...
[((739, 785), 'functools.partial', 'functools.partial', (['nn.BatchNorm2d'], {'affine': '(True)'}), '(nn.BatchNorm2d, affine=True)\n', (756, 785), False, 'import functools\n'), ((1527, 1552), 'torch.cuda.is_available', 'torch.cuda.is_available', ([], {}), '()\n', (1550, 1552), False, 'import torch\n'), ((4480, 4505), '...
# -*- coding: utf-8 -*- """ Created on Sat Apr 13 16:12:34 2019 @author: bosmanjw """ import httpInterface import sys sys.path.append('../') sys.path.append('../shamir_ss') from shamir_ss.reconstruct_value import reconstruct_value import config import asyncio import json class Combiner(object): ...
[ "httpInterface.Requestor", "json.dumps", "httpInterface.Listener", "asyncio.get_event_loop", "sys.path.append", "shamir_ss.reconstruct_value.reconstruct_value" ]
[((129, 151), 'sys.path.append', 'sys.path.append', (['"""../"""'], {}), "('../')\n", (144, 151), False, 'import sys\n'), ((153, 184), 'sys.path.append', 'sys.path.append', (['"""../shamir_ss"""'], {}), "('../shamir_ss')\n", (168, 184), False, 'import sys\n'), ((453, 489), 'httpInterface.Requestor', 'httpInterface.Requ...
"""Main object for getting the PECO outage counter data.""" from __future__ import annotations from typing import Any import aiohttp from .const import * from pydantic import BaseModel class PecoOutageApi: """Main object for getting the PECO outage counter data.""" def __init__(self) -> None: """Init...
[ "aiohttp.ClientSession" ]
[((839, 862), 'aiohttp.ClientSession', 'aiohttp.ClientSession', ([], {}), '()\n', (860, 862), False, 'import aiohttp\n'), ((1527, 1550), 'aiohttp.ClientSession', 'aiohttp.ClientSession', ([], {}), '()\n', (1548, 1550), False, 'import aiohttp\n'), ((3050, 3073), 'aiohttp.ClientSession', 'aiohttp.ClientSession', ([], {})...
import cv2 import numpy as np import torch from matplotlib import pyplot as plt from dataset import mydataset_PCA from torch.utils.data import DataLoader from tqdm import tqdm import pickle import sklearn.decomposition as dc import argparse parser = argparse.ArgumentParser(description='Process some integers.') parser....
[ "numpy.mean", "argparse.ArgumentParser", "sklearn.decomposition.PCA", "tqdm.tqdm", "numpy.argsort", "numpy.array", "numpy.sum", "numpy.concatenate", "torch.utils.data.DataLoader", "dataset.mydataset_PCA" ]
[((251, 312), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {'description': '"""Process some integers."""'}), "(description='Process some integers.')\n", (274, 312), False, 'import argparse\n'), ((688, 788), 'dataset.mydataset_PCA', 'mydataset_PCA', ([], {'image_dir': 'image_dir', 'text_path': 'text_path',...
import environ env = environ.Env( DEBUG=(bool, True) )
[ "environ.Env" ]
[((22, 53), 'environ.Env', 'environ.Env', ([], {'DEBUG': '(bool, True)'}), '(DEBUG=(bool, True))\n', (33, 53), False, 'import environ\n')]
#coding: utf-8 import requests import AdvancedHTMLParser import json import datetime import html from unidecode import unidecode # Placeholder, will be replaced by reference to main cfg object # This is only to satisfy builtin vs code verifier try: cfg = None cfg.teachermap_filename = None except: pass print("SEMI...
[ "datetime.datetime.strptime", "AdvancedHTMLParser.AdvancedHTMLParser", "html.unescape", "requests.get" ]
[((390, 441), 'requests.get', 'requests.get', (['"""http://www.zseil.edu.pl/zastepstwa/"""'], {}), "('http://www.zseil.edu.pl/zastepstwa/')\n", (402, 441), False, 'import requests\n'), ((519, 558), 'AdvancedHTMLParser.AdvancedHTMLParser', 'AdvancedHTMLParser.AdvancedHTMLParser', ([], {}), '()\n', (556, 558), False, 'im...
#import sqlite3 from flask import (Blueprint, flash, g, redirect, render_template, request, session, url_for) import functools from setting import Config import hashlib bp = Blueprint('login', __name__, url_prefix='/web') @bp.route('/login', methods=('GET', 'POST')) def login(): # 登录 if re...
[ "flask.render_template", "flask.flash", "flask.session.get", "functools.wraps", "flask.url_for", "flask.session.clear", "flask.Blueprint", "setting.Config.PASSWORD.encode" ]
[((194, 241), 'flask.Blueprint', 'Blueprint', (['"""login"""', '__name__'], {'url_prefix': '"""/web"""'}), "('login', __name__, url_prefix='/web')\n", (203, 241), False, 'from flask import Blueprint, flash, g, redirect, render_template, request, session, url_for\n'), ((984, 1017), 'flask.render_template', 'render_templ...
from pfwrapper import PyPatternFinder from collections import namedtuple Pattern = namedtuple('Pattern', ['base', 'sens', 'pBase', 'pAll', 'pDX', 'pD_X', 'pX', 'pDY', 'pD_Y', 'pY', 'pDXY', 'pD_XY', 'pXY']) DivergentPa...
[ "pfwrapper.PyPatternFinder", "collections.namedtuple" ]
[((84, 210), 'collections.namedtuple', 'namedtuple', (['"""Pattern"""', "['base', 'sens', 'pBase', 'pAll', 'pDX', 'pD_X', 'pX', 'pDY', 'pD_Y', 'pY',\n 'pDXY', 'pD_XY', 'pXY']"], {}), "('Pattern', ['base', 'sens', 'pBase', 'pAll', 'pDX', 'pD_X', 'pX',\n 'pDY', 'pD_Y', 'pY', 'pDXY', 'pD_XY', 'pXY'])\n", (94, 210), ...
#!/usr/bin/env python import gtk, sys, os mynumber="5555555555" calnum=0 fakenum=0 def press(widget): num = entry.get_text() pnum = widget.get_label() entry.set_text(num + pnum) def send_press(widget): print("Dialing: " + entry.get_text()) def add_call(widget): callnum = entry.get_text() entry.set_text("") d...
[ "gtk.Table", "gtk.Entry", "gtk.Button", "gtk.Window", "gtk.main_quit", "gtk.VBox", "gtk.HBox", "gtk.main" ]
[((396, 408), 'gtk.Window', 'gtk.Window', ([], {}), '()\n', (406, 408), False, 'import gtk, sys, os\n'), ((466, 476), 'gtk.VBox', 'gtk.VBox', ([], {}), '()\n', (474, 476), False, 'import gtk, sys, os\n'), ((499, 510), 'gtk.Entry', 'gtk.Entry', ([], {}), '()\n', (508, 510), False, 'import gtk, sys, os\n'), ((549, 574), ...
from simple_rl.mdp.oomdp.OOMDPStateClass import OOMDPState class TrenchOOMDPState(OOMDPState): ''' Class for Trench World States ''' def __init__(self, objects): OOMDPState.__init__(self, objects=objects) def get_agent_x(self): return self.objects["agent"][0]["x"] def get_agent_y(sel...
[ "simple_rl.mdp.oomdp.OOMDPStateClass.OOMDPState.__init__" ]
[((180, 222), 'simple_rl.mdp.oomdp.OOMDPStateClass.OOMDPState.__init__', 'OOMDPState.__init__', (['self'], {'objects': 'objects'}), '(self, objects=objects)\n', (199, 222), False, 'from simple_rl.mdp.oomdp.OOMDPStateClass import OOMDPState\n')]
import json file = open('buildings.json', 'r') buildings = json.load(file)['data'] output = [] for building in buildlings: temp = {} temp["id"] = building["number"] temp["name"] = {} temp["name"]["value"] = building["name"] output.append(temp)
[ "json.load" ]
[((60, 75), 'json.load', 'json.load', (['file'], {}), '(file)\n', (69, 75), False, 'import json\n')]
import src.library as s_lib # s_lib.update_team_table() # s_lib.test_func() # s_lib.update_result() st = '2018-10-29' end = '2019-05-23' s_lib.update_period_result(st, end)
[ "src.library.update_period_result" ]
[((140, 175), 'src.library.update_period_result', 's_lib.update_period_result', (['st', 'end'], {}), '(st, end)\n', (166, 175), True, 'import src.library as s_lib\n')]
#!/usr/bin/env python import os import sys import argparse import requests sys.path.append(os.path.join(os.path.dirname(__file__), '..')) import messagebird parser = argparse.ArgumentParser() parser.add_argument('--accessKey', help='Access key for MessageBird API.', type=str, required=True) parser.add_argument('--pho...
[ "messagebird.Client", "os.path.dirname", "argparse.ArgumentParser" ]
[((168, 193), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {}), '()\n', (191, 193), False, 'import argparse\n'), ((494, 531), 'messagebird.Client', 'messagebird.Client', (["args['accessKey']"], {}), "(args['accessKey'])\n", (512, 531), False, 'import messagebird\n'), ((105, 130), 'os.path.dirname', 'os.pa...
# -*- coding: utf-8 -*- import logging from django.conf import settings from django_redis import get_redis_connection from uuid import uuid4 logger = logging.getLogger(__name__) class LatestViewedStudies(object): """ Interfaces with Redis to keep a list of latest viewed studies """ def __init__(self, use...
[ "logging.getLogger", "django_redis.get_redis_connection", "uuid.uuid4" ]
[((154, 181), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (171, 181), False, 'import logging\n'), ((461, 508), 'django_redis.get_redis_connection', 'get_redis_connection', (['settings.EDD_LATEST_CACHE'], {}), '(settings.EDD_LATEST_CACHE)\n', (481, 508), False, 'from django_redis import...
#!/usr/bin/env python # Copyright JS Foundation and other contributors, http://js.foundation # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2....
[ "argparse.ArgumentParser", "os.access", "os.path.join", "os.path.isfile", "itertools.chain.from_iterable", "os.path.basename", "sys.exit", "os.path.abspath", "warnings.warn", "csv.reader" ]
[((907, 980), 'os.path.join', 'os.path.join', (['PROJECT_DIR', '"""jerry-core/lit/lit-unicode-conversions.inc.h"""'], {}), "(PROJECT_DIR, 'jerry-core/lit/lit-unicode-conversions.inc.h')\n", (919, 980), False, 'import os\n'), ((802, 827), 'os.path.abspath', 'os.path.abspath', (['__file__'], {}), '(__file__)\n', (817, 82...
# Solution of; # Project Euler Problem 465: Polar polygons # https://projecteuler.net/problem=465 # # The kernel of a polygon is defined by the set of points from which the # entire polygon's boundary is visible. We define a polar polygon as a polygon # for which the origin is strictly contained inside its kernel. F...
[ "timed.caller" ]
[((1483, 1517), 'timed.caller', 'timed.caller', (['dummy', 'n', 'i', 'prob_id'], {}), '(dummy, n, i, prob_id)\n', (1495, 1517), False, 'import timed\n')]
#!/usr/bin/python """ GFS.py - GFS GRIB2 class """ # Stock modules import sys import os import re import logging import traceback import datetime import collections import operator # Local modules import error_codes import utils from dataset import dataset LOG = logging.getLogger(__name__) # gfs.t18z.pgrb2.1p00.f0...
[ "logging.getLogger", "utils.error", "datetime.datetime.strptime" ]
[((267, 294), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (284, 294), False, 'import logging\n'), ((1042, 1085), 'datetime.datetime.strptime', 'datetime.datetime.strptime', (['DTS', '"""%Y%m%d%H"""'], {}), "(DTS, '%Y%m%d%H')\n", (1068, 1085), False, 'import datetime\n'), ((1432, 1477),...
from __future__ import unicode_literals import frappe from frappe.utils import date_diff, nowdate, add_days, format_datetime from datetime import timedelta, datetime def expire_doc_mail_notify(): # stuff to do every 10 minutes today = nowdate() employee_documents = frappe.db.get_list('Employee Documents',...
[ "frappe.db.get_all", "frappe._", "frappe.utils.nowdate", "frappe.get_doc", "datetime.datetime.today", "frappe.db.get_list" ]
[((245, 254), 'frappe.utils.nowdate', 'nowdate', ([], {}), '()\n', (252, 254), False, 'from frappe.utils import date_diff, nowdate, add_days, format_datetime\n'), ((280, 426), 'frappe.db.get_list', 'frappe.db.get_list', (['"""Employee Documents"""'], {'fields': "['parent', 'doc_name', 'reminder_days', 'expire_date']", ...
from flask import Flask from flask import request from flask import Response from handling import Handler import json app = Flask(__name__) INDEX_NAME = 'contacts' PORT = 9200 handler = Handler(INDEX_NAME, port = PORT, wipe_index = True) @app.route('/contact', methods=['GET','POST']) def contact_without_name(): ...
[ "json.dumps", "flask.Response", "handling.Handler", "flask.Flask" ]
[((125, 140), 'flask.Flask', 'Flask', (['__name__'], {}), '(__name__)\n', (130, 140), False, 'from flask import Flask\n'), ((189, 236), 'handling.Handler', 'Handler', (['INDEX_NAME'], {'port': 'PORT', 'wipe_index': '(True)'}), '(INDEX_NAME, port=PORT, wipe_index=True)\n', (196, 236), False, 'from handling import Handle...
from node import Node class Stack: def __init__(self): self.head = None def __str__(self): node = self.head list = [] while node: list.append(node.get_item()) node = node.get_next() return str(list) def is_empty(self): return not self.head def push(self, item): if not self.head: self.hea...
[ "node.Node" ]
[((324, 334), 'node.Node', 'Node', (['item'], {}), '(item)\n', (328, 334), False, 'from node import Node\n'), ((359, 380), 'node.Node', 'Node', (['item', 'self.head'], {}), '(item, self.head)\n', (363, 380), False, 'from node import Node\n')]
import logging import logging.handlers """ Returns a logging handler with a default format """ def get_logging_handler(log_filename): handler = logging.handlers.RotatingFileHandler(log_filename, maxBytes=2*1024*1024, backupCount=5) formatter = logging.Formatter("%(asctime)s - %(levelname)s - %(name)s - %...
[ "logging.Formatter", "logging.handlers.RotatingFileHandler" ]
[((155, 250), 'logging.handlers.RotatingFileHandler', 'logging.handlers.RotatingFileHandler', (['log_filename'], {'maxBytes': '(2 * 1024 * 1024)', 'backupCount': '(5)'}), '(log_filename, maxBytes=2 * 1024 * 1024,\n backupCount=5)\n', (191, 250), False, 'import logging\n'), ((259, 332), 'logging.Formatter', 'logging....
from queue import Queue from typing import TYPE_CHECKING from galaxy_crawler.crawl import Crawler from galaxy_crawler.filters import DefaultFilter from galaxy_crawler.filters.v1 import V1FilterEnum from galaxy_crawler.models.engine import EngineType from galaxy_crawler.models.dependeny_resolver import DependencyResolv...
[ "galaxy_crawler.filters.DefaultFilter", "galaxy_crawler.queries.v1.V1QueryBuilder", "galaxy_crawler.models.engine.EngineType.from_url", "galaxy_crawler.models.engine.EngineType.from_env_var", "galaxy_crawler.utils.mkdir", "galaxy_crawler.queries.v1.V1QueryOrder.choices", "galaxy_crawler.filters.v1.V1Fil...
[((997, 1004), 'queue.Queue', 'Queue', ([], {}), '()\n', (1002, 1004), False, 'from queue import Queue\n'), ((1151, 1168), 'galaxy_crawler.utils.mkdir', 'mkdir', (['output_dir'], {}), '(output_dir)\n', (1156, 1168), False, 'from galaxy_crawler.utils import mkdir\n'), ((1469, 1485), 'galaxy_crawler.queries.v1.V1QueryBui...
import ipywidgets as widgets import ipywidgets from traitlets import Unicode import traitlets from traittypes import Array import logging import numpy as np from .serialize import array_cube_png_serialization, array_serialization from .transferfunction import * import warnings logger = logging.getLogger("ipyvolume") ...
[ "logging.getLogger", "traitlets.Instance", "traitlets.List", "ipywidgets.FloatSlider", "numpy.nanmin", "ipywidgets.HBox", "traittypes.Array", "ipywidgets.register", "numpy.nanmax", "warnings.warn", "ipyvolume.pylab.figure", "traitlets.Unicode", "traitlets.CInt", "ipywidgets.jslink", "tra...
[((289, 319), 'logging.getLogger', 'logging.getLogger', (['"""ipyvolume"""'], {}), "('ipyvolume')\n", (306, 319), False, 'import logging\n'), ((352, 389), 'ipywidgets.register', 'widgets.register', (['"""ipyvolume.Scatter"""'], {}), "('ipyvolume.Scatter')\n", (368, 389), True, 'import ipywidgets as widgets\n'), ((1748,...
# Copyright (c) 2015 <NAME> <<EMAIL>> # Author: <NAME> <<EMAIL>> # # Permission is hereby granted, free of charge, to any person obtaining a copy of # this software and associated documentation files (the "Software"), to deal in # the Software without restriction, including without limitation the rights to # use, copy...
[ "fine_mapping_pipeline.utils.shell.run_command_return_output", "logging.info" ]
[((3643, 3673), 'logging.info', 'logging.info', (['bcftools_command'], {}), '(bcftools_command)\n', (3655, 3673), False, 'import logging\n'), ((3708, 3763), 'fine_mapping_pipeline.utils.shell.run_command_return_output', 'run_command_return_output', (['bcftools_command'], {'shell': '(True)'}), '(bcftools_command, shell=...
from src.requests.search_requests import search from constants.search_bar_constants import GOOGLE_SEARCH_URL def search_testcases(testcases): for testcase in testcases: response = search(GOOGLE_SEARCH_URL, testcase) print(response.url) print(response.ok) print(response.status_code)...
[ "src.requests.search_requests.search" ]
[((194, 229), 'src.requests.search_requests.search', 'search', (['GOOGLE_SEARCH_URL', 'testcase'], {}), '(GOOGLE_SEARCH_URL, testcase)\n', (200, 229), False, 'from src.requests.search_requests import search\n')]
""" ``$ drafttopic fetch_text -h`` :: Fetches text for labelings using a MediaWiki API. Usage: fetch_text --api-host=<url> [--labelings=<path>] [--output=<path>] [--verbose] Options: -h --help Show this documentation. --api-host=<url> ...
[ "logging.getLogger", "itertools.islice", "re.compile", "concurrent.futures.ThreadPoolExecutor", "sys.stderr.flush", "revscoring.utilities.util.dump_observation", "sys.stderr.write", "revscoring.utilities.util.read_observations", "mw.lib.title.normalize", "docopt.docopt", "mwapi.Session" ]
[((1080, 1107), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (1097, 1107), False, 'import logging\n'), ((1122, 1151), 're.compile', 're.compile', (['"""#redirect"""', 're.I'], {}), "('#redirect', re.I)\n", (1132, 1151), False, 'import re\n'), ((1236, 1262), 'docopt.docopt', 'docopt', ([...
import exodus import os from output_suppression import Suppressor import numpy as np def get_nodal_variable_values(filename, varname, step=1): """ Extracts nodal field data from exodus file and returns a numpy array of nodal values """ with Suppressor(): e = exodus.exodus(filename, array_type='...
[ "exodus.copy_mesh", "os.path.isfile", "output_suppression.Suppressor", "exodus.exodus", "exodus.add_variables", "os.remove" ]
[((1166, 1196), 'os.path.isfile', 'os.path.isfile', (['outputfilename'], {}), '(outputfilename)\n', (1180, 1196), False, 'import os\n'), ((1772, 1802), 'os.path.isfile', 'os.path.isfile', (['outputfilename'], {}), '(outputfilename)\n', (1786, 1802), False, 'import os\n'), ((2729, 2759), 'os.path.isfile', 'os.path.isfil...
# coding: utf-8 from __future__ import absolute_import from datetime import date, datetime # noqa: F401 from typing import List, Dict # noqa: F401 from swagger_server.models.base_model_ import Model from swagger_server import util class PreferencesProfilePeople(Model): """NOTE: This class is auto generated b...
[ "swagger_server.util.deserialize_model" ]
[((3116, 3149), 'swagger_server.util.deserialize_model', 'util.deserialize_model', (['dikt', 'cls'], {}), '(dikt, cls)\n', (3138, 3149), False, 'from swagger_server import util\n')]
# pylint: disable=invalid-name import os import platform from flask import Flask from flask import render_template, request, send_from_directory from flask_cors import CORS import pandas as pd from app.ABSA import AspectsBased from app.scrapper import amazonScrapper EXTERNALS_DIR = os.path.realpath(os.path.join(__fil...
[ "flask.render_template", "flask.send_from_directory", "flask_cors.CORS", "flask.Flask", "os.path.join", "platform.system", "app.ABSA.AspectsBased", "app.scrapper.amazonScrapper" ]
[((352, 451), 'flask.Flask', 'Flask', (['__name__'], {'static_folder': '"""./templates/build/static"""', 'template_folder': '"""./templates/build"""'}), "(__name__, static_folder='./templates/build/static', template_folder=\n './templates/build')\n", (357, 451), False, 'from flask import Flask\n'), ((447, 456), 'fla...
from abc import ABC, abstractmethod import logging from typing import Any from ROAR.utilities_module.module import Module from ROAR.utilities_module.vehicle_models import Vehicle, VehicleControl from collections import deque import numpy as np class ROARManiaPlanner(Module): def __init__(self, agent, **kwargs): ...
[ "logging.getLogger", "numpy.sign" ]
[((406, 433), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (423, 433), False, 'import logging\n'), ((3298, 3318), 'numpy.sign', 'np.sign', (['norm_offset'], {}), '(norm_offset)\n', (3305, 3318), True, 'import numpy as np\n')]
""" Defines a set of document formats. (c) 2015 Massachusetts Institute of Technology """ # Native import datetime import hashlib import os # LO-PHI Automation import lophi_automation.database.utils as utils DELIMITER = '_' def analysis_doc_uid(analyzer_name, analyzer_version, sample_uid): """ Ret...
[ "lophi_automation.database.utils.calculate_sha1", "os.path.getsize", "lophi_automation.database.utils.calculate_combined_hash", "datetime.datetime.utcnow", "lophi_automation.database.utils.calculate_md5", "datetime.datetime.now", "lophi_automation.database.utils.calculate_sha512", "lophi_automation.da...
[((1666, 1692), 'datetime.datetime.utcnow', 'datetime.datetime.utcnow', ([], {}), '()\n', (1690, 1692), False, 'import datetime\n'), ((3687, 3713), 'datetime.datetime.utcnow', 'datetime.datetime.utcnow', ([], {}), '()\n', (3711, 3713), False, 'import datetime\n'), ((5277, 5317), 'lophi_automation.database.utils.calcula...
# Member.py import discord from discord.ext import commands from discord.utils import get import os import mysql.connector import isbnlib from isbnlib import * import asyncio ROLE = "Book Worm" def initdb(): return mysql.connector.connect( host = os.getenv('HOST'), user = os.geten...
[ "discord.Colour.green", "discord.ext.commands.command", "os.getenv" ]
[((907, 925), 'discord.ext.commands.command', 'commands.command', ([], {}), '()\n', (923, 925), False, 'from discord.ext import commands\n'), ((2296, 2314), 'discord.ext.commands.command', 'commands.command', ([], {}), '()\n', (2312, 2314), False, 'from discord.ext import commands\n'), ((4911, 4929), 'discord.ext.comma...
from django.contrib.contenttypes.models import ContentType from nautobot.dcim.models import Site from nautobot.extras.choices import CustomFieldTypeChoices from nautobot.extras.jobs import Job from nautobot.extras.models import CustomField class TestCreateSiteWithCustomField(Job): class Meta: name = "Sit...
[ "nautobot.dcim.models.Site.objects.create", "django.contrib.contenttypes.models.ContentType.objects.get_for_model", "nautobot.extras.models.CustomField.objects.create" ]
[((451, 490), 'django.contrib.contenttypes.models.ContentType.objects.get_for_model', 'ContentType.objects.get_for_model', (['Site'], {}), '(Site)\n', (484, 490), False, 'from django.contrib.contenttypes.models import ContentType\n'), ((504, 599), 'nautobot.extras.models.CustomField.objects.create', 'CustomField.object...
from collections import namedtuple import numpy as np import pandas as pd import rdkit.Chem as Chem from tqdm import tqdm from neural_fingerprint import NFPRegressor from neural_fingerprint.chemutils import rf_evaluation from neural_fingerprint.models.ecfp import ECFP max_val = 1000 train_idx = 800 def read_data(m...
[ "sklearn.gaussian_process.GaussianProcessRegressor", "neural_fingerprint.chemutils.rf_evaluation", "numpy.ones", "matplotlib.pyplot.pcolor", "matplotlib.pyplot.colorbar", "rdkit.Chem.MolFromSmiles", "tqdm.tqdm", "neural_fingerprint.NFPRegressor", "sklearn.manifold.TSNE", "numpy.linspace", "numpy...
[((354, 400), 'pandas.read_table', 'pd.read_table', (['"""./data/train.txt"""'], {'header': 'None'}), "('./data/train.txt', header=None)\n", (367, 400), True, 'import pandas as pd\n'), ((414, 464), 'pandas.read_table', 'pd.read_table', (['"""./data/train.logP-SA"""'], {'header': 'None'}), "('./data/train.logP-SA', head...
# # core/urls.py # # Author: <NAME> # from django.conf.urls import include from django.contrib.auth.decorators import login_required from django.urls import path from . import views urlpatterns = [ path('', login_required(views.index), name='index'), path('libraries/', login_required(views.libraries), name=...
[ "django.contrib.auth.decorators.login_required" ]
[((215, 242), 'django.contrib.auth.decorators.login_required', 'login_required', (['views.index'], {}), '(views.index)\n', (229, 242), False, 'from django.contrib.auth.decorators import login_required\n'), ((282, 313), 'django.contrib.auth.decorators.login_required', 'login_required', (['views.libraries'], {}), '(views...
#!/usr/bin/python # -*- coding: utf-8 -*- from collections import namedtuple import queue Item = namedtuple("Item", ['index', 'value', 'weight', 'density']) # Node = namedtuple("Node", ["level", "profit", "bound", "weight"]) class Node(object): def __init__(self, level=-1, profit=0, weight=0, bound=0): ...
[ "queue.Queue", "collections.namedtuple" ]
[((99, 158), 'collections.namedtuple', 'namedtuple', (['"""Item"""', "['index', 'value', 'weight', 'density']"], {}), "('Item', ['index', 'value', 'weight', 'density'])\n", (109, 158), False, 'from collections import namedtuple\n'), ((3468, 3481), 'queue.Queue', 'queue.Queue', ([], {}), '()\n', (3479, 3481), False, 'im...
import os from pathlib import Path def make_backup(c, file): if file.exists() and not file.is_symlink(): file_backup = Path('{}.orig'.format(file)) if not file_backup.exists(): c.run('mv -f {} {}'.format(file, file_backup)) return True return False def symlink(c, sou...
[ "pathlib.Path.cwd", "pathlib.Path" ]
[((367, 377), 'pathlib.Path.cwd', 'Path.cwd', ([], {}), '()\n', (375, 377), False, 'from pathlib import Path\n'), ((546, 563), 'pathlib.Path', 'Path', (['cwd', 'source'], {}), '(cwd, source)\n', (550, 563), False, 'from pathlib import Path\n'), ((404, 421), 'pathlib.Path', 'Path', (['destination'], {}), '(destination)\...
import importlib import os import sys from typing import Any, Callable, Optional from mkdocs.config import Config, config_options from mkdocs.livereload import LiveReloadServer from mkdocs.plugins import BasePlugin from .collectors import ModuleCollector class Plugin(BasePlugin): """ MkDocs Plugin class. ...
[ "mkdocs.config.config_options.Type", "sys.path.append", "importlib.import_module", "os.getcwd" ]
[((1914, 1925), 'os.getcwd', 'os.getcwd', ([], {}), '()\n', (1923, 1925), False, 'import os\n'), ((456, 495), 'mkdocs.config.config_options.Type', 'config_options.Type', (['str'], {'required': '(True)'}), '(str, required=True)\n', (475, 495), False, 'from mkdocs.config import Config, config_options\n'), ((519, 557), 'm...
# _*_ encoding: utf-8 _*_ __author__ = 'nzb' __date__ = '2020/6/6 17:47' __doc__ = "二值化(Thresholding)" import cv2 import numpy as np # 计算: y={0 (if y<128 ) 255 (else) def Grayscale(img): b = img[:,:,0].copy() g = img[:,:,1].copy() r = img[:,:,2].copy() ret = 0.2126 * r + 0.7152 * g + 0.0722 * b ...
[ "cv2.imread", "cv2.waitKey", "cv2.destroyAllWindows", "cv2.imshow" ]
[((579, 602), 'cv2.imshow', 'cv2.imshow', (['"""ret"""', 'img2'], {}), "('ret', img2)\n", (589, 602), False, 'import cv2\n'), ((604, 618), 'cv2.waitKey', 'cv2.waitKey', (['(0)'], {}), '(0)\n', (615, 618), False, 'import cv2\n'), ((619, 642), 'cv2.destroyAllWindows', 'cv2.destroyAllWindows', ([], {}), '()\n', (640, 642)...
"""Test for pretty format table from jsons.""" from datetime import datetime from tabulate import tabulate from ecs_tasks_ops import pretty_json from ecs_tasks_ops import pretty_table def test_tabulate_list_json_empty(): """Test empty data.""" out = pretty_table.tabulate_list_json() assert "No data to s...
[ "ecs_tasks_ops.pretty_json.DateTimeEncoder", "ecs_tasks_ops.pretty_table.tabulate_list_json_keys", "tabulate.tabulate", "ecs_tasks_ops.pretty_table.tabulate_list_json", "datetime.datetime.now", "ecs_tasks_ops.pretty_json.dumps" ]
[((262, 295), 'ecs_tasks_ops.pretty_table.tabulate_list_json', 'pretty_table.tabulate_list_json', ([], {}), '()\n', (293, 295), False, 'from ecs_tasks_ops import pretty_table\n'), ((460, 512), 'ecs_tasks_ops.pretty_table.tabulate_list_json', 'pretty_table.tabulate_list_json', ([], {'empty_msg': 'empty_msg'}), '(empty_m...
import logging import socketserver import threading from robottelemetryservice.infrastructure.serialisation import serializer from robottelemetryservice.infrastructure.configuration import application log=logging.getLogger(__name__) measurement_repository=application.get_measurement_repository() class UDPMeasurementE...
[ "logging.getLogger", "robottelemetryservice.infrastructure.configuration.application.get_measurement_repository" ]
[((206, 233), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (223, 233), False, 'import logging\n'), ((257, 297), 'robottelemetryservice.infrastructure.configuration.application.get_measurement_repository', 'application.get_measurement_repository', ([], {}), '()\n', (295, 297), False, 'fr...
import base64 import os import unittest from unittest.mock import patch from click.testing import CliRunner from keeper_secrets_manager_core.core import SecretsManager from keeper_secrets_manager_core.storage import InMemoryKeyValueStorage from keeper_secrets_manager_core import mock from keeper_secrets_manager_core.mo...
[ "configparser.ConfigParser", "keeper_secrets_manager_core.mock.MockConfig.make_base64", "unittest.main", "unittest.mock.patch", "re.search", "os.path.exists", "keeper_secrets_manager_core.storage.InMemoryKeyValueStorage", "json.dumps", "keeper_secrets_manager_core.mock.Response", "os.unlink", "t...
[((14095, 14110), 'unittest.main', 'unittest.main', ([], {}), '()\n', (14108, 14110), False, 'import unittest\n'), ((653, 664), 'os.getcwd', 'os.getcwd', ([], {}), '()\n', (662, 664), False, 'import os\n'), ((689, 718), 'tempfile.TemporaryDirectory', 'tempfile.TemporaryDirectory', ([], {}), '()\n', (716, 718), False, '...
import os import pytest from unittest import mock from breaking_changes import inspector from tests import mod TEST_ROOT = os.path.dirname(__file__) BASE_ROOT = os.path.abspath(os.path.join(TEST_ROOT, '..')) def relative(subpath: str) -> str: return os.path.abspath(os.path.join(TEST_ROOT, subpath)) def func...
[ "breaking_changes.inspector.public_interface", "breaking_changes.inspector.module_transform", "os.path.join", "os.path.dirname", "pytest.mark.parametrize", "breaking_changes.inspector.function_args" ]
[((127, 152), 'os.path.dirname', 'os.path.dirname', (['__file__'], {}), '(__file__)\n', (142, 152), False, 'import os\n'), ((716, 830), 'pytest.mark.parametrize', 'pytest.mark.parametrize', (["('pkg_name', 'packages')", "[('package1', ['a', 'b']), ('package2', ['a', 'nested.c'])]"], {}), "(('pkg_name', 'packages'), [('...
import logging from logging.handlers import RotatingFileHandler from flask.logging import default_handler from app import app if __name__ == '__main__': handler = RotatingFileHandler(filename='./log/app.log', maxBytes=1048576, backupCount=3) formatter = logging.Formatter(fmt='%(asctime)s - %(name)s[line:%(line...
[ "app.app.run", "logging.Formatter", "logging.handlers.RotatingFileHandler", "app.app.logger.addHandler", "flask.logging.default_handler.setLevel" ]
[((168, 246), 'logging.handlers.RotatingFileHandler', 'RotatingFileHandler', ([], {'filename': '"""./log/app.log"""', 'maxBytes': '(1048576)', 'backupCount': '(3)'}), "(filename='./log/app.log', maxBytes=1048576, backupCount=3)\n", (187, 246), False, 'from logging.handlers import RotatingFileHandler\n'), ((263, 395), '...
#!/usr/bin/env python # manual """ This script allows you to manually control the simulator or Duckiebot using the keyboard arrows. """ import os os.sys.path.append("../../gym-duckietown") import cv2 import sys import argparse import pyglet from pyglet.window import key import numpy as np import gym import gym_duckie...
[ "PIL.Image.fromarray", "pyglet.window.key.KeyStateHandler", "pyglet.clock.schedule_interval", "pyglet.app.run", "NaiveLaneDetection.NaiveLaneDetection", "os.sys.path.append", "cv2.imshow", "numpy.array", "cv2.addWeighted", "cv2.cvtColor", "sys.exit", "cv2.waitKey", "cv2.namedWindow" ]
[((147, 189), 'os.sys.path.append', 'os.sys.path.append', (['"""../../gym-duckietown"""'], {}), "('../../gym-duckietown')\n", (165, 189), False, 'import os\n'), ((1290, 1346), 'numpy.array', 'np.array', (['[[(0, 200), (640, 200), (640, 480), (0, 480)]]'], {}), '([[(0, 200), (640, 200), (640, 480), (0, 480)]])\n', (1298...
from collections import Counter, defaultdict def fizz_buzz_counter(): values = [] for i in range(1, 101): if i % 3 == 0 and i % 5 == 0: values.append("fizzbuzz") elif i % 3 == 0: values.append("fizz") elif i % 5 == 0: values.append("buzz") els...
[ "collections.Counter", "collections.defaultdict" ]
[((368, 383), 'collections.Counter', 'Counter', (['values'], {}), '(values)\n', (375, 383), False, 'from collections import Counter, defaultdict\n'), ((428, 444), 'collections.defaultdict', 'defaultdict', (['int'], {}), '(int)\n', (439, 444), False, 'from collections import Counter, defaultdict\n')]
import base64 import itertools import json import requests from presqt.json_schemas.schema_handlers import schema_validator from presqt.targets.github.utilities import validation_check from presqt.targets.github.utilities.utils.upload_extra_metadata import upload_extra_metadata from presqt.utilities import PresQTError...
[ "itertools.chain", "json.loads", "base64.b64encode", "json.dumps", "presqt.targets.github.utilities.validation_check", "base64.b64decode", "presqt.targets.github.utilities.utils.upload_extra_metadata.upload_extra_metadata", "presqt.json_schemas.schema_handlers.schema_validator" ]
[((742, 765), 'presqt.targets.github.utilities.validation_check', 'validation_check', (['token'], {}), '(token)\n', (758, 765), False, 'from presqt.targets.github.utilities import validation_check\n'), ((1601, 1648), 'base64.b64decode', 'base64.b64decode', (["metadata_file_data['content']"], {}), "(metadata_file_data['...
#/usr/bin/env python #-*- coding: utf-8 -*- #Refer http://www.wooyun.org/bugs/wooyun-2015-0135532 #__Author__ = 上善若水 #_PlugName_ = tianrongxin_sql Plugin #_FileName_ = tianrongxin_sql.py import re def assign(service, arg): if service == "topsec": return True, arg def audit(arg): paylo...
[ "re.findall" ]
[((628, 669), 're.findall', 're.findall', (['"""thrown in <b>(.*?)</b>"""', 'res'], {}), "('thrown in <b>(.*?)</b>', res)\n", (638, 669), False, 'import re\n')]
#!/usr/bin/env python3 import hashlib import re import requests import sys from pprint import pprint as pp from .exceptions import PCloudException from .connection import AbstractPCloudConnection from .pcloudbin import PCloudBinaryConnection PCLOUD_SERVER_SUFFIX = '.pcloud.com' # only allow downloads from pcloud s...
[ "re.sub", "pprint.pprint", "requests.get" ]
[((7559, 7650), 'requests.get', 'requests.get', (['url'], {'stream': '(True)', 'allow_redirects': '(False)', 'timeout': 'self.connection.timeout'}), '(url, stream=True, allow_redirects=False, timeout=self.\n connection.timeout)\n', (7571, 7650), False, 'import requests\n'), ((3743, 3782), 'pprint.pprint', 'pp', (['(...
import math import os import time import numpy as np from torch.utils.tensorboard import SummaryWriter import utils.loss as loss import utils.tensorboard as utb def get_probs(length, exp): probs = (np.arange(1, length + 1) / 100) ** exp last_x = int(0.9 * length) probs[last_x:] = probs[last_x] retur...
[ "numpy.random.normal", "torch.utils.tensorboard.SummaryWriter", "numpy.tile", "math.ceil", "utils.loss.PSNR", "numpy.log", "numpy.max", "numpy.argsort", "numpy.sum", "numpy.linspace", "utils.tensorboard.load_event_accumulator", "numpy.zeros", "numpy.exp", "numpy.transpose", "time.time", ...
[((1385, 1412), 'numpy.linspace', 'np.linspace', (['(0.01)', '(0.3)', '(200)'], {}), '(0.01, 0.3, 200)\n', (1396, 1412), True, 'import numpy as np\n'), ((733, 766), 'utils.tensorboard.load_loglikelihood', 'utb.load_loglikelihood', (['event_acc'], {}), '(event_acc)\n', (755, 766), True, 'import utils.tensorboard as utb\...
""" References: https://fdc.nal.usda.gov/api-guide.html """ import json import decouple import requests from datatrans import utils from datatrans.fooddata import search __all__ = ['send_food_search_api_request', 'send_food_detail_api_request'] def send_food_search_api_request( criteria: search.reques...
[ "json.dumps", "decouple.config", "datatrans.utils.snake_to_camel", "requests.get" ]
[((368, 417), 'decouple.config', 'decouple.config', (['"""DATA_GOV_API_KEY"""', '"""MY_API_KEY"""'], {}), "('DATA_GOV_API_KEY', 'MY_API_KEY')\n", (383, 417), False, 'import decouple\n'), ((1891, 1940), 'decouple.config', 'decouple.config', (['"""DATA_GOV_API_KEY"""', '"""MY_API_KEY"""'], {}), "('DATA_GOV_API_KEY', 'MY_...
#!/usr/bin/python3 # -*- coding: utf-8 -*- # @Time : 2019/5/18 10:20 # @Author : chensw、wangwei # @File : model_fusion.py # @Describe: 标明文件实现的功能 # @Modify : 修改的地方 import pandas as pd # Step3:模型投票融合 # 读取5个模型输出的预测概率文件 predict_1_4 = pd.read_csv('../data/output/predict/prediction_1_4.csv') predict_2...
[ "pandas.merge", "pandas.read_csv" ]
[((253, 309), 'pandas.read_csv', 'pd.read_csv', (['"""../data/output/predict/prediction_1_4.csv"""'], {}), "('../data/output/predict/prediction_1_4.csv')\n", (264, 309), True, 'import pandas as pd\n'), ((325, 381), 'pandas.read_csv', 'pd.read_csv', (['"""../data/output/predict/prediction_2_4.csv"""'], {}), "('../data/o...
import unittest # Run tests without using GPU import os os.environ["CUDA_VISIBLE_DEVICES"] = "-1" os.environ['TEST'] = "1" import sys from pathlib import Path TEST_DIR = str(Path(__file__).parent.resolve()) BASE_DIR = str(Path(__file__).parent.parent.resolve()) sys.path.append(BASE_DIR) from core.filters import Publ...
[ "core.api.IndependentClaimsRequest", "core.api.ForwardCitationsRequest", "core.api.DrawingRequest", "unittest.main", "core.api.CitationsRequest", "core.api.ConceptVectorRequest", "sys.path.append", "core.api.AggregatedCitationsRequest", "core.api.TitleRequest", "core.api.OneClaimRequest", "core....
[((264, 289), 'sys.path.append', 'sys.path.append', (['BASE_DIR'], {}), '(BASE_DIR)\n', (279, 289), False, 'import sys\n'), ((19348, 19363), 'unittest.main', 'unittest.main', ([], {}), '()\n', (19361, 19363), False, 'import unittest\n'), ((2285, 2297), 'core.api.APIRequest', 'APIRequest', ([], {}), '()\n', (2295, 2297)...
#%% Setup from matplotlib import rc rc("font", **{"family": "serif", "serif": ["Computer Modern Roman"], "size": 20}) rc("text", usetex=True) import os import numpy as np import mpmath as mp import matplotlib.pyplot as plt from scipy.special import lambertw from scipy import optimize # path = os.path.expanduser( # ...
[ "matplotlib.pyplot.xticks", "matplotlib.pyplot.ylabel", "scipy.special.lambertw", "matplotlib.pyplot.xlabel", "matplotlib.pyplot.plot", "numpy.log", "matplotlib.pyplot.fill_between", "numpy.linspace", "matplotlib.pyplot.figure", "matplotlib.pyplot.yticks", "matplotlib.rc", "scipy.optimize.root...
[((37, 122), 'matplotlib.rc', 'rc', (['"""font"""'], {}), "('font', **{'family': 'serif', 'serif': ['Computer Modern Roman'], 'size':\n 20})\n", (39, 122), False, 'from matplotlib import rc\n'), ((119, 142), 'matplotlib.rc', 'rc', (['"""text"""'], {'usetex': '(True)'}), "('text', usetex=True)\n", (121, 142), False, ...
#!/usr/bin/env python3 import cudasift import cv2 import numpy as np from profiling import TaggedTimer def main(): sift = cudasift.PyCudaSift(dev_num=0) timr = TaggedTimer() filename = "../data/CY_279b46b9_1575825158217_1575825184058.jpg" # filename = "/home/jfinken/projects/here/sp/jfinken/faiss_gp...
[ "numpy.asarray", "cudasift.i_align_up", "cudasift.PyCudaSift", "profiling.TaggedTimer", "cv2.imread" ]
[((129, 159), 'cudasift.PyCudaSift', 'cudasift.PyCudaSift', ([], {'dev_num': '(0)'}), '(dev_num=0)\n', (148, 159), False, 'import cudasift\n'), ((171, 184), 'profiling.TaggedTimer', 'TaggedTimer', ([], {}), '()\n', (182, 184), False, 'from profiling import TaggedTimer\n'), ((349, 391), 'cv2.imread', 'cv2.imread', (['fi...
# - *- coding: utf- 8 - *- #decorator fonksiyonlar, pythonda fonksiyonlarımızın dinamik olarak ekstra özellik ekledigimiz fonksiyonlarıdr #kod tekrarını engeller #flask django gibi frameworklerde cok kullanılır import time def calculate_time(func): def wrapper(numbers): start = time.time() resul...
[ "time.time" ]
[((294, 305), 'time.time', 'time.time', ([], {}), '()\n', (303, 305), False, 'import time\n'), ((353, 364), 'time.time', 'time.time', ([], {}), '()\n', (362, 364), False, 'import time\n')]
#!/usr/bin/env python """ Perform the Mann-Whitney U test, the Kolmogorov-Smirnov test, and the Student's t-test for the following ensembles: - GPU double precision (reference & control) - CPU double precision - GPU single precision - GPU double precision with additional explicit diffusion Make sure to compile the cp...
[ "numpy.mean", "pickle.dump", "numpy.sqrt", "kolmogorov_smirnov.ks", "numpy.max", "mannwhitneyu.mwu", "numpy.zeros", "numpy.quantile", "numpy.min", "xarray.open_dataset", "numpy.arange" ]
[((2243, 2267), 'xarray.open_dataset', 'xr.open_dataset', (['fname_r'], {}), '(fname_r)\n', (2258, 2267), True, 'import xarray as xr\n'), ((2292, 2316), 'xarray.open_dataset', 'xr.open_dataset', (['fname_c'], {}), '(fname_c)\n', (2307, 2316), True, 'import xarray as xr\n'), ((2343, 2369), 'xarray.open_dataset', 'xr.ope...
#!/usr/bin/env python """Tests for Yara flows.""" import functools import psutil import yara import yara_procdump from grr.client.client_actions import tempfiles from grr.client.client_actions import yara_actions from grr.lib import flags from grr.lib import utils from grr.lib.rdfvalues import client as rdf_client fr...
[ "psutil.NoSuchProcess", "grr.lib.flags.StartMain", "grr.server.aff4.FACTORY.Open", "grr.test_lib.test_lib.main", "grr.test_lib.test_lib.FakeTime", "psutil.Process.old_target", "yara.TimeoutError", "functools.partial", "grr.test_lib.flow_test_lib.TestFlowHelper", "grr.test_lib.action_mocks.ActionMo...
[((9321, 9340), 'grr.test_lib.test_lib.main', 'test_lib.main', (['argv'], {}), '(argv)\n', (9334, 9340), False, 'from grr.test_lib import test_lib\n'), ((9372, 9393), 'grr.lib.flags.StartMain', 'flags.StartMain', (['main'], {}), '(main)\n', (9387, 9393), False, 'from grr.lib import flags\n'), ((1955, 2008), 'psutil.NoS...
import os import csv import pandas as pd import numpy as np import test #gets the path of current working directory CWD = os.getcwd() def Load_Data (csvpath): try: datapath = CWD + csvpath df= pd.read_csv(datapath) except: print("Error with reading csv: ", csvpath) print (df.hea...
[ "pandas.read_csv", "os.getcwd" ]
[((125, 136), 'os.getcwd', 'os.getcwd', ([], {}), '()\n', (134, 136), False, 'import os\n'), ((218, 239), 'pandas.read_csv', 'pd.read_csv', (['datapath'], {}), '(datapath)\n', (229, 239), True, 'import pandas as pd\n'), ((595, 625), 'pandas.read_csv', 'pd.read_csv', (['advantix_raw_path'], {}), '(advantix_raw_path)\n',...
from matplotlib.gridspec import GridSpec import matplotlib.pyplot as plt import numpy as np import matplotlib import math from ._default_matplotlib_figure_dimensions import _default_matplotlib_figure_dimensions def _calculate_nrows(nplots, ncols): return math.ceil(nplots / ncols) def _initialize_plot_with_dimens...
[ "math.ceil", "numpy.ones", "numpy.any", "numpy.array", "matplotlib.pyplot.figure", "matplotlib.gridspec.GridSpec" ]
[((261, 286), 'math.ceil', 'math.ceil', (['(nplots / ncols)'], {}), '(nplots / ncols)\n', (270, 286), False, 'import math\n'), ((980, 1014), 'matplotlib.pyplot.figure', 'plt.figure', ([], {'figsize': 'fig_dimensions'}), '(figsize=fig_dimensions)\n', (990, 1014), True, 'import matplotlib.pyplot as plt\n'), ((1783, 1906)...
from Section01_SOLID.OpenClosePrinciple.BetterFilter import BetterFilter from Section01_SOLID.OpenClosePrinciple.Color import Color from Section01_SOLID.OpenClosePrinciple.ColorSpecification import ColorSpecification from Section01_SOLID.OpenClosePrinciple.Product import Product from Section01_SOLID.OpenClosePrinciple....
[ "Section01_SOLID.OpenClosePrinciple.ColorSpecification.ColorSpecification", "Section01_SOLID.OpenClosePrinciple.Product.Product", "Section01_SOLID.OpenClosePrinciple.ProductFilter.ProductFilter", "Section01_SOLID.OpenClosePrinciple.BetterFilter.BetterFilter", "Section01_SOLID.OpenClosePrinciple.SizeSpecific...
[((783, 824), 'Section01_SOLID.OpenClosePrinciple.Product.Product', 'Product', (['"""Apple"""', 'Color.GREEN', 'Size.SMALL'], {}), "('Apple', Color.GREEN, Size.SMALL)\n", (790, 824), False, 'from Section01_SOLID.OpenClosePrinciple.Product import Product\n'), ((832, 872), 'Section01_SOLID.OpenClosePrinciple.Product.Prod...
# open external links in a new window from django.utils.html import escape from wagtail.contrib.modeladmin.helpers import ButtonHelper from wagtail.contrib.modeladmin.mixins import ThumbnailMixin from wagtail.contrib.modeladmin.options import ModelAdmin, modeladmin_register from wagtail.core import hooks from wagtail....
[ "wagtail.contrib.modeladmin.options.modeladmin_register", "django.utils.html.escape", "wagtail.core.hooks.register" ]
[((982, 1027), 'wagtail.core.hooks.register', 'hooks.register', (['"""register_rich_text_features"""'], {}), "('register_rich_text_features')\n", (996, 1027), False, 'from wagtail.core import hooks\n'), ((1130, 1178), 'wagtail.core.hooks.register', 'hooks.register', (['"""construct_page_listing_buttons"""'], {}), "('co...
import sublime, sublime_plugin import os try: # ST3 from ..apis.core import Core except (ImportError, ValueError): # ST2 from apis.core import Core # Completion class ERBAutocompleteListener(sublime_plugin.EventListener): def on_query_completions(self, view, prefix, locations): core = Core(...
[ "apis.core.Core" ]
[((315, 321), 'apis.core.Core', 'Core', ([], {}), '()\n', (319, 321), False, 'from apis.core import Core\n'), ((1187, 1193), 'apis.core.Core', 'Core', ([], {}), '()\n', (1191, 1193), False, 'from apis.core import Core\n')]
from tkinter import * from tkinter import ttk from Controller import controleBanco # Tela de relatorios em xlsx def TelaRelatorio(tela): # Cria a tela de relatorios telaRelatorio = Toplevel(tela) telaRelatorio.title('RELATORIO') telaRelatorio.geometry('500x500+620+120') telaRelatorio['bg'] = 'gr...
[ "Controller.controleBanco.ControleLerDados", "Controller.controleBanco.BuscaPrimeiro", "Controller.controleBanco.BuscaDadosPeloId", "Controller.controleBanco.BuscaUltimo", "Controller.controleBanco.EscreveNoExcel", "tkinter.ttk.Treeview" ]
[((440, 470), 'Controller.controleBanco.BuscaPrimeiro', 'controleBanco.BuscaPrimeiro', (['(3)'], {}), '(3)\n', (467, 470), False, 'from Controller import controleBanco\n'), ((484, 512), 'Controller.controleBanco.BuscaUltimo', 'controleBanco.BuscaUltimo', (['(3)'], {}), '(3)\n', (509, 512), False, 'from Controller impor...
# -*- coding: utf-8 -*- from runScraper import test, scrapeRecipeUrl from parser import parseIngredients test('https://healthyfitnessmeals.com/artichoke-stuffed-chicken/') # ingredient = parseIngredients( # ["cans refrigerated crescent dinner rolls"]) # print(ingredient) # print(scrapeRecipeUrl)
[ "runScraper.test" ]
[((107, 173), 'runScraper.test', 'test', (['"""https://healthyfitnessmeals.com/artichoke-stuffed-chicken/"""'], {}), "('https://healthyfitnessmeals.com/artichoke-stuffed-chicken/')\n", (111, 173), False, 'from runScraper import test, scrapeRecipeUrl\n')]
# Generated by Django 2.2.13 on 2020-10-21 14:39 from django.db import migrations class Migration(migrations.Migration): dependencies = [ ('houses', '0002_house_user_id'), ] operations = [ migrations.RenameField( model_name='house', old_name='user_id', ...
[ "django.db.migrations.RenameField" ]
[((222, 301), 'django.db.migrations.RenameField', 'migrations.RenameField', ([], {'model_name': '"""house"""', 'old_name': '"""user_id"""', 'new_name': '"""user"""'}), "(model_name='house', old_name='user_id', new_name='user')\n", (244, 301), False, 'from django.db import migrations\n')]
from django.contrib.auth import authenticate, login, logout from django.contrib.auth.decorators import login_required from django.db import IntegrityError from django.http import HttpResponse, HttpResponseRedirect from django.shortcuts import render from django.urls import reverse from .models import User, Listing, Bi...
[ "django.shortcuts.render", "django.contrib.auth.authenticate", "django.contrib.auth.login", "django.urls.reverse", "django.contrib.auth.logout" ]
[((401, 463), 'django.shortcuts.render', 'render', (['request', '"""auctions/index.html"""', "{'listings': listings}"], {}), "(request, 'auctions/index.html', {'listings': listings})\n", (407, 463), False, 'from django.shortcuts import render\n'), ((577, 646), 'django.shortcuts.render', 'render', (['request', '"""aucti...
import sys import logging from face_client import face_client from face_client.api_proxy import FaceApiProxy from face_client.camera_controller import CameraController from face_client.image_displayer import ImageDisplayer def main(argv): client = face_client.FaceClient(CameraController(), ImageDisplayer(), Face...
[ "face_client.api_proxy.FaceApiProxy", "face_client.image_displayer.ImageDisplayer", "logging.exception", "face_client.camera_controller.CameraController", "sys.exit" ]
[((278, 296), 'face_client.camera_controller.CameraController', 'CameraController', ([], {}), '()\n', (294, 296), False, 'from face_client.camera_controller import CameraController\n'), ((298, 314), 'face_client.image_displayer.ImageDisplayer', 'ImageDisplayer', ([], {}), '()\n', (312, 314), False, 'from face_client.im...
""" Categorize mapping between Dutch and English into polysemy profiles Usage: categorize_lemma_to_lemma_polysemy.py --config_path=<config_path> --input_folder=<input_folder> --output_folder=<output_folder> --rbn_pos=<rbn_pos> --fn_pos=<fn_pos> --verbose=<verbose> Options: --config_path=<config_path> --inpu...
[ "os.path.exists", "pickle.dump", "pathlib.Path", "graph_utils.get_paths_from_rbn_to_fn", "datetime.datetime.now", "collections.defaultdict", "load_utils.load_python_module", "networkx.read_gpickle", "docopt.docopt" ]
[((1071, 1183), 'load_utils.load_python_module', 'load_utils.load_python_module', ([], {'module_path': '"""../resources/ODWN_Reader"""', 'module_name': '"""odwn_classes"""', 'verbose': '(1)'}), "(module_path='../resources/ODWN_Reader',\n module_name='odwn_classes', verbose=1)\n", (1100, 1183), False, 'import load_ut...
#!/usr/bin/env python3 """ corrections.py: Script to apply corrections to the images. """ import os from argparse import ArgumentParser from datetime import date, datetime from typing import Optional, Sequence import numpy as np from astropy.io import fits from dresscode.utils import load_config def main(argv: Op...
[ "numpy.abs", "numpy.full_like", "astropy.io.fits.PrimaryHDU", "argparse.ArgumentParser", "dresscode.utils.load_config", "numpy.where", "numpy.std", "numpy.nanmedian", "os.path.isfile", "numpy.sum", "datetime.date", "os.path.basename", "astropy.io.fits.open", "numpy.log1p" ]
[((371, 387), 'argparse.ArgumentParser', 'ArgumentParser', ([], {}), '()\n', (385, 387), False, 'from argparse import ArgumentParser\n'), ((542, 566), 'dresscode.utils.load_config', 'load_config', (['args.config'], {}), '(args.config)\n', (553, 566), False, 'from dresscode.utils import load_config\n'), ((2419, 2438), '...
import pathlib from setuptools import setup, find_packages from pokemontcgsdkasync.config import __version__, __pypi_package_name__, __github_username__, __github_repo_name__ # The directory containing this file HERE = pathlib.Path(__file__).parent # The text of the README file README = (HERE / "README.md").read_te...
[ "setuptools.find_packages", "pathlib.Path" ]
[((222, 244), 'pathlib.Path', 'pathlib.Path', (['__file__'], {}), '(__file__)\n', (234, 244), False, 'import pathlib\n'), ((1325, 1340), 'setuptools.find_packages', 'find_packages', ([], {}), '()\n', (1338, 1340), False, 'from setuptools import setup, find_packages\n')]
#!/usr/bin/env python3 import sys def plot(x, y): print((x, y)) def plot_line_low(x0, y0, x1, y1): dx = x1 - x0 dy = y1 - y0 yi = 1 if dy < 0: yi = -1 dy = -dy D = (2 * dy) - dx y = y0 for x in range(x0, x1): plot(x, y) if D > 0: y = y...
[ "sys.exit" ]
[((3031, 3042), 'sys.exit', 'sys.exit', (['(0)'], {}), '(0)\n', (3039, 3042), False, 'import sys\n')]
# -*- coding: utf-8 -*- """ /*************************************************************************** Name : DB Manager Description : Database manager plugin for QGIS Date : May 23, 2011 copyright : (C) 2011 by <NAME> email : <EMAIL> The content of...
[ "psycopg2.connect", "functools.cmp_to_key", "psycopg2.extensions.register_type", "os.environ.get", "qgis.PyQt.QtCore.QFile", "builtins.str", "builtins.range", "qgis.core.QgsCredentials.instance", "qgis.PyQt.QtCore.QRegExp", "qgis.core.QgsDataSourceUri" ]
[((1443, 1505), 'psycopg2.extensions.register_type', 'psycopg2.extensions.register_type', (['psycopg2.extensions.UNICODE'], {}), '(psycopg2.extensions.UNICODE)\n', (1476, 1505), False, 'import psycopg2\n'), ((1506, 1573), 'psycopg2.extensions.register_type', 'psycopg2.extensions.register_type', (['psycopg2.extensions.U...
#!/usr/bin/python import os from subprocess import call NCMDS = 2 message = str(raw_input('type your message: ')) command1 = "git commit -m '%s'"%(message) push = 'git push origin master' cmd_list = [command1, push] for cmd in cmd_list: call(cmd, shell=True)
[ "subprocess.call" ]
[((250, 271), 'subprocess.call', 'call', (['cmd'], {'shell': '(True)'}), '(cmd, shell=True)\n', (254, 271), False, 'from subprocess import call\n')]
import smtplib from contextlib import contextmanager import pandas as pd, datetime as dt,numpy as np import re, os, ssl import credentials, glob import base64 from templateReport import * # template html of all content of the email from email.mime.multipart import MIMEMultipart from email.mime.base import MIMEBase...
[ "smtplib.SMTP", "pandas.read_csv", "datetime.datetime.now", "email.mime.multipart.MIMEMultipart", "datetime.date.today", "email.mime.text.MIMEText" ]
[((521, 585), 'pandas.read_csv', 'pd.read_csv', (['f"""/home/lorenzo/Quanvas/Dollar Futures {today}.csv"""'], {}), "(f'/home/lorenzo/Quanvas/Dollar Futures {today}.csv')\n", (532, 585), True, 'import pandas as pd, datetime as dt, numpy as np\n'), ((2091, 2141), 'pandas.read_csv', 'pd.read_csv', (['"""/home/lorenzo/Quan...
"""Block.Io API backend. Supports Bitcoin, Dogecoin and Litecoin on `block.io <https://block.io>`_ API. The backend configuration takes following parameters. :param class: Always ``cryptoassets.core.backend.blockio.BlockIo`` :param api_key: block.io API key :param password: <PASSWORD> :param network: one of ``btc...
[ "logging.getLogger", "datetime.datetime.fromtimestamp", "block_io.BlockIo", "decimal.Decimal", "slugify.slugify" ]
[((1573, 1600), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (1590, 1600), False, 'import logging\n'), ((8431, 8441), 'decimal.Decimal', 'Decimal', (['(1)'], {}), '(1)\n', (8438, 8441), False, 'from decimal import Decimal\n'), ((2611, 2636), 'block_io.BlockIo', '_BlockIo', (['api_key', ...
from setuptools import setup setup( name="flex_version", version='1.2.3', url='https://github.com/caesar0301/FlexVersion', author='<NAME>', author_email='<EMAIL>', description='A cute Python library to manipulate version stuff.', license="Apache License, Version 2.0", packages=['flex_ve...
[ "setuptools.setup" ]
[((30, 968), 'setuptools.setup', 'setup', ([], {'name': '"""flex_version"""', 'version': '"""1.2.3"""', 'url': '"""https://github.com/caesar0301/FlexVersion"""', 'author': '"""<NAME>"""', 'author_email': '"""<EMAIL>"""', 'description': '"""A cute Python library to manipulate version stuff."""', 'license': '"""Apache Li...
#!/usr/bin/env python # Copyright 2017-2018 Biomedical Imaging Group Rotterdam, Departments of # Medical Informatics and Radiology, Erasmus MC, Rotterdam, The Netherlands # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obt...
[ "logging.getLogger", "collections.OrderedDict", "radiomics.featureextractor.RadiomicsFeaturesExtractor", "logging.Formatter", "logging.FileHandler" ]
[((3018, 3071), 'radiomics.featureextractor.RadiomicsFeaturesExtractor', 'featureextractor.RadiomicsFeaturesExtractor', ([], {}), '(**kwargs)\n', (3061, 3071), False, 'from radiomics import featureextractor\n'), ((3262, 3292), 'logging.getLogger', 'logging.getLogger', (['"""radiomics"""'], {}), "('radiomics')\n", (3279...
from django import forms from django.contrib.auth.forms import UserCreationForm, AuthenticationForm from .models import User class RegistrationForm(UserCreationForm): class Meta: model = User fields = ('email',) class LoginForm(AuthenticationForm): username = forms.EmailField()
[ "django.forms.EmailField" ]
[((288, 306), 'django.forms.EmailField', 'forms.EmailField', ([], {}), '()\n', (304, 306), False, 'from django import forms\n')]
# -*- coding: utf-8 -*- import re import json import requests import itertools from settings import * import tools # retrieve prices from tarkov-market for each item def crawl_prices_tarkov_market(data): logger.info("Getting JSON from " + CONST_TARKOV_MARKET) try: # escape the 401 first, no Simple...
[ "tools.find_substring", "re.sub", "json.loads", "requests.get" ]
[((3212, 3243), 'requests.get', 'requests.get', (['CONST_LOOT_GOBLIN'], {}), '(CONST_LOOT_GOBLIN)\n', (3224, 3243), False, 'import requests\n'), ((3287, 3305), 'json.loads', 'json.loads', (['r.text'], {}), '(r.text)\n', (3297, 3305), False, 'import json\n'), ((397, 430), 'requests.get', 'requests.get', (['CONST_TARKOV_...
import numpy as np import os from scipy.optimize import least_squares, minimize from scipy.special import fresnel def autophase(S): '''Optimize phase of complex data by maximizing the sum of imaginary over sum of real .. math:: \phi = \\arctan \left( \\frac{\sum_i^N \Im(s_i) }{ \sum_i^N \Re(s_i) } \\...
[ "numpy.sqrt", "numpy.linalg.pinv", "numpy.polyfit", "numpy.log", "numpy.roots", "numpy.array", "numpy.linalg.norm", "numpy.sin", "numpy.imag", "scipy.optimize.least_squares", "numpy.max", "numpy.exp", "numpy.real", "numpy.dot", "scipy.special.fresnel", "numpy.abs", "numpy.eye", "nu...
[((2633, 2666), 'os.path.join', 'os.path.join', (['directory', 'filename'], {}), '(directory, filename)\n', (2645, 2666), False, 'import os\n'), ((2687, 2723), 'numpy.loadtxt', 'np.loadtxt', (['full_path'], {'delimiter': '""","""'}), "(full_path, delimiter=',')\n", (2697, 2723), True, 'import numpy as np\n'), ((2995, 3...
import sys, arrow, time, dateutil from uuid import uuid4 from datetime import datetime sys.path.insert(0, 'controller/') #Import library essentials from sumy.parsers.plaintext import PlaintextParser #We're choosing a plaintext parser here, other parsers available for HTML etc. from sumy.nlp.tokenizers import Tokenize...
[ "sys.path.insert", "textblob.classifiers.NaiveBayesClassifier", "sumy.summarizers.lex_rank.LexRankSummarizer", "dateutil.tz.gettz", "time.strftime", "arrow.get", "sumy.nlp.tokenizers.Tokenizer", "datetime.datetime.today" ]
[((88, 121), 'sys.path.insert', 'sys.path.insert', (['(0)', '"""controller/"""'], {}), "(0, 'controller/')\n", (103, 121), False, 'import sys, arrow, time, dateutil\n'), ((692, 726), 'time.strftime', 'time.strftime', (['"""%d/%m/%y %H:%M:%S"""'], {}), "('%d/%m/%y %H:%M:%S')\n", (705, 726), False, 'import sys, arrow, ti...
''' IoT Bootcamp Trainer: <NAME> Phase 4 - Part 2: Building Web App for Home Automation Code Sample: Controlling LED using an HTML page ''' from flask import Flask, render_template, request from gpiozero import LED from time import sleep app = Flask(__name__) # Default: GPIO4 led = LED(4) @app.route('/') def home()...
[ "flask.render_template", "gpiozero.LED", "flask.Flask" ]
[((245, 260), 'flask.Flask', 'Flask', (['__name__'], {}), '(__name__)\n', (250, 260), False, 'from flask import Flask, render_template, request\n'), ((285, 291), 'gpiozero.LED', 'LED', (['(4)'], {}), '(4)\n', (288, 291), False, 'from gpiozero import LED\n'), ((425, 468), 'flask.render_template', 'render_template', (['"...
import os import cv2 import numpy as np from PIL import Image from IPython.display import Video from IPython.display import display as ds DESTINATION_FOLDER = "results" def check_folder(folder): if not os.path.exists(folder): os.makedirs(folder) def display(images_array, save=False): for im in ...
[ "cv2.rectangle", "os.path.exists", "PIL.Image.fromarray", "cv2.imwrite", "os.makedirs", "IPython.display.Video", "cv2.putText", "cv2.imdecode", "cv2.cvtColor", "numpy.fromstring" ]
[((888, 918), 'numpy.fromstring', 'np.fromstring', (['image', 'np.uint8'], {}), '(image, np.uint8)\n', (901, 918), True, 'import numpy as np\n'), ((934, 971), 'cv2.imdecode', 'cv2.imdecode', (['nparr', 'cv2.IMREAD_COLOR'], {}), '(nparr, cv2.IMREAD_COLOR)\n', (946, 971), False, 'import cv2\n'), ((1482, 1523), 'cv2.cvtCo...
#!/usr/bin/env python # coding=utf-8 import unittest from app.domain.model import User, AnonymousUser, Permission, Role class UserModelTestCase(unittest.TestCase): def test_password_setter(self): u = User(password = '<PASSWORD>') self.assertTrue(u.password_hash is not None) def test_no_passwo...
[ "app.domain.model.User", "app.domain.model.Role.insert_roles", "app.domain.model.AnonymousUser" ]
[((214, 241), 'app.domain.model.User', 'User', ([], {'password': '"""<PASSWORD>"""'}), "(password='<PASSWORD>')\n", (218, 241), False, 'from app.domain.model import User, AnonymousUser, Permission, Role\n'), ((349, 376), 'app.domain.model.User', 'User', ([], {'password': '"""<PASSWORD>"""'}), "(password='<PASSWORD>')\n...
import os import sys import logging from flask import Flask # type: ignore import sentry_sdk from sentry_sdk.integrations.flask import FlaskIntegration from sentry_sdk.integrations.sqlalchemy import SqlalchemyIntegration def create_app(test_config=None): # create and configure the app app = Flask(__name__, ...
[ "sentry_sdk.integrations.flask.FlaskIntegration", "logging.StreamHandler", "flask.Flask", "os.environ.get", "sentry_sdk.integrations.sqlalchemy.SqlalchemyIntegration", "doppelkopf.db.init_app" ]
[((304, 350), 'flask.Flask', 'Flask', (['__name__'], {'instance_relative_config': '(True)'}), '(__name__, instance_relative_config=True)\n', (309, 350), False, 'from flask import Flask\n'), ((1179, 1195), 'doppelkopf.db.init_app', 'db.init_app', (['app'], {}), '(app)\n', (1190, 1195), False, 'from doppelkopf import db\...