code
stringlengths
22
1.05M
apis
listlengths
1
3.31k
extract_api
stringlengths
75
3.25M
# Code generated by `typeddictgen`. DO NOT EDIT. """V2beta2MetricStatusDict generated type.""" from typing import TypedDict from kubernetes_typed.client import V2beta2ExternalMetricStatusDict, V2beta2ObjectMetricStatusDict, V2beta2PodsMetricStatusDict, V2beta2ResourceMetricStatusDict V2beta2MetricStatusDict = TypedDi...
[ "typing.TypedDict" ]
[((313, 557), 'typing.TypedDict', 'TypedDict', (['"""V2beta2MetricStatusDict"""', "{'external': V2beta2ExternalMetricStatusDict, 'object':\n V2beta2ObjectMetricStatusDict, 'pods': V2beta2PodsMetricStatusDict,\n 'resource': V2beta2ResourceMetricStatusDict, 'type': str}"], {'total': '(False)'}), "('V2beta2MetricSta...
#!/usr/bin/env python # Simulate the /robot_driver so that interfaces to it can operate the same on # the real robot and in simulation from __future__ import print_function, division from threading import Lock import rospy import diagnostic_updater from diagnostic_msgs.msg import DiagnosticStatus from fetch_driver_...
[ "fetch_driver_msgs.msg.ChargerState", "rospy.Subscriber", "rospy.is_shutdown", "rospy.init_node", "threading.Lock", "rospy.Service", "diagnostic_updater.Updater", "rospy.Time.now", "power_msgs.msg.BreakerState", "fetch_driver_msgs.msg.JointState", "rospy.Rate", "rospy.Publisher", "std_srvs.s...
[((8125, 8156), 'rospy.init_node', 'rospy.init_node', (['"""robot_driver"""'], {}), "('robot_driver')\n", (8140, 8156), False, 'import rospy\n'), ((1757, 1823), 'power_msgs.msg.BreakerState', 'BreakerState', ([], {'name': '"""arm_breaker"""', 'state': 'BreakerState.STATE_ENABLED'}), "(name='arm_breaker', state=BreakerS...
# Configuration file for the Sphinx documentation builder. # # This file only contains a selection of the most common options. For a full # list see the documentation: # https://www.sphinx-doc.org/en/master/usage/configuration.html # -- Path setup -------------------------------------------------------------- # If ex...
[ "sphinx.ext.apidoc.main", "pathlib.Path" ]
[((2908, 2925), 'sphinx.ext.apidoc.main', 'apidoc.main', (['argv'], {}), '(argv)\n', (2919, 2925), False, 'from sphinx.ext import apidoc\n'), ((711, 725), 'pathlib.Path', 'Path', (['__file__'], {}), '(__file__)\n', (715, 725), False, 'from pathlib import Path\n')]
from random import randint from time import sleep print ('=' * 18) print ('\033[1mSTONE PAPER AND SCISSORS\033[m') print ('=' * 18) print ('I already chose mine now missing you') sleep (1) computer = randint (1, 3) player = int(input('\033[1mChoose between\033[m \033[1;33m1) Stone 2) Paper and 3) Scissors\033[m : '))...
[ "random.randint", "time.sleep" ]
[((180, 188), 'time.sleep', 'sleep', (['(1)'], {}), '(1)\n', (185, 188), False, 'from time import sleep\n'), ((201, 214), 'random.randint', 'randint', (['(1)', '(3)'], {}), '(1, 3)\n', (208, 214), False, 'from random import randint\n'), ((321, 329), 'time.sleep', 'sleep', (['(1)'], {}), '(1)\n', (326, 329), False, 'fro...
# @l2g 1673 python3 # [1673] Find the Most Competitive Subsequence # Difficulty: Medium # https://leetcode.com/problems/find-the-most-competitive-subsequence # # Given an integer array nums and a positive integer k, # return the most competitive subsequence of nums of size k. # An array's subsequence is a resulting seq...
[ "os.path.join" ]
[((1609, 1646), 'os.path.join', 'os.path.join', (['"""tests"""', '"""test_1673.py"""'], {}), "('tests', 'test_1673.py')\n", (1621, 1646), False, 'import os\n')]
#!/usr/bin/env python import os import unittest import sys from test_common import TestCommon sys.path.append(os.path.join(os.path.dirname(__file__), '..', 'library')) from fastly_service import FastlyConfiguration class TestFastlyLoggingS3(TestCommon): @TestCommon.vcr.use_cassette() def test_fastly_s3s(sel...
[ "unittest.main", "os.path.dirname", "test_common.TestCommon.vcr.use_cassette", "fastly_service.FastlyConfiguration" ]
[((263, 292), 'test_common.TestCommon.vcr.use_cassette', 'TestCommon.vcr.use_cassette', ([], {}), '()\n', (290, 292), False, 'from test_common import TestCommon\n'), ((2124, 2153), 'test_common.TestCommon.vcr.use_cassette', 'TestCommon.vcr.use_cassette', ([], {}), '()\n', (2151, 2153), False, 'from test_common import T...
#! /usr/bin/env python3 """ PrepareTreeData.py Prepares the tree data a bit for inclusion into a database. """ import pandas as pd import numpy as np """ Prepare the tree data for inclusion into a database. Args: inname: Input file outname: Output file """ def PrepareTreeData(inname='data/street_trees_2015...
[ "pandas.read_csv" ]
[((387, 419), 'pandas.read_csv', 'pd.read_csv', (['inname'], {'index_col': '(0)'}), '(inname, index_col=0)\n', (398, 419), True, 'import pandas as pd\n')]
#encoding: UTF-8 # Copyright (C) 2016 <NAME> # This file is distributed under the terms of the # MIT License. # See the file `License' in the root directory of the present distribution. """ An earlier and now oblosete implementation of functions for computing the thermal expansion tensor as a function of temperatur...
[ "math.pow", "numpy.array", "math.exp", "numpy.zeros" ]
[((1910, 1922), 'math.exp', 'math.exp', (['(-x)'], {}), '(-x)\n', (1918, 1922), False, 'import math\n'), ((1953, 1967), 'math.pow', 'math.pow', (['x', '(2)'], {}), '(x, 2)\n', (1961, 1967), False, 'import math\n'), ((2399, 2411), 'math.exp', 'math.exp', (['(-x)'], {}), '(-x)\n', (2407, 2411), False, 'import math\n'), (...
from __future__ import division import numpy as np from sklearn.utils import shuffle from sklearn.metrics import * """ Module with different fitness functions implemented to be used by the CRO algorithm. The functions' only argument must be an individual (coral) and return its fitness, a number. The fitness might re...
[ "sklearn.utils.shuffle", "numpy.multiply" ]
[((1584, 1623), 'sklearn.utils.shuffle', 'shuffle', (['X', 'y'], {'random_state': 'random_seed'}), '(X, y, random_state=random_seed)\n', (1591, 1623), False, 'from sklearn.utils import shuffle\n'), ((1633, 1655), 'numpy.multiply', 'np.multiply', (['Xs', 'coral'], {}), '(Xs, coral)\n', (1644, 1655), True, 'import numpy ...
from abc import ABC, abstractmethod from typing import Tuple, Iterable, Any from pygame.rect import Rect class Shape(ABC): """Abstract shape interface.""" @abstractmethod def shape(self) -> Any: pass @abstractmethod def top_left(self) -> Tuple: pass @abstractmethod def t...
[ "pygame.rect.Rect" ]
[((786, 800), 'pygame.rect.Rect', 'Rect', (['position'], {}), '(position)\n', (790, 800), False, 'from pygame.rect import Rect\n')]
import numpy as np from ..local_interpolation import ThirdOrderHermitePolynomialInterpolation from .runge_kutta import AbstractESDIRK, ButcherTableau γ = 0.26 a21 = γ a31 = 0.13 a32 = 0.84033320996790809 a41 = 0.22371961478320505 a42 = 0.47675532319799699 a43 = -0.06470895363112615 a51 = 0.16648564323248321 a52 = 0....
[ "numpy.array" ]
[((1729, 1760), 'numpy.array', 'np.array', (['[0, γ, γ, γ, γ, γ, γ]'], {}), '([0, γ, γ, γ, γ, γ, γ])\n', (1737, 1760), True, 'import numpy as np\n'), ((2016, 2059), 'numpy.array', 'np.array', (['[a71, a72, a73, a74, a75, a76, γ]'], {}), '([a71, a72, a73, a74, a75, a76, γ])\n', (2024, 2059), True, 'import numpy as np\n'...
#MenuTitle: Remove Zero Deltas in Selected Glyphs # -*- coding: utf-8 -*- from __future__ import division, print_function, unicode_literals __doc__=""" Goes through all layers of each selected glyph, and deletes all TT Delta Hints with an offset of zero. Detailed Report in Macro Window. """ def process( Layer ): try:...
[ "traceback.format_exc" ]
[((1433, 1455), 'traceback.format_exc', 'traceback.format_exc', ([], {}), '()\n', (1453, 1455), False, 'import traceback\n')]
#!/usr/bin/env python3 import re import sys from glob import glob from subprocess import run def main(args): assert len(args) >= 1 from_image = args.pop(0) optional = [x for x in map(str.strip, args) if x] optional_used = set() with open("Dockerfile", "w") as fout: print(f"from {from_ima...
[ "subprocess.run", "glob.glob", "re.search" ]
[((1077, 1133), 'subprocess.run', 'run', (["['docker', 'build', '-t', our_tag, '.']"], {'check': '(True)'}), "(['docker', 'build', '-t', our_tag, '.'], check=True)\n", (1080, 1133), False, 'from subprocess import run\n'), ((366, 386), 'glob.glob', 'glob', (['"""*.Dockerfile"""'], {}), "('*.Dockerfile')\n", (370, 386), ...
import torch import torch.nn as nn import torch.nn.functional as F import torchvision class FPAv2(nn.Module): def __init__(self, input_dim, output_dim): super(FPAv2, self).__init__() self.glob = nn.Sequential(nn.AdaptiveAvgPool2d(1), nn.Conv2d(input_dim, output_di...
[ "torch.nn.functional.upsample", "torch.nn.BatchNorm2d", "torch.nn.Sigmoid", "torch.nn.ReLU", "torch.nn.Dropout", "torch.nn.functional.adaptive_avg_pool2d", "torch.nn.Sequential", "torch.nn.Dropout2d", "torch.nn.Conv2d", "torchvision.models.resnet34", "torch.nn.MaxPool2d", "torch.nn.AdaptiveAvg...
[((1596, 1668), 'torch.nn.functional.upsample', 'F.upsample', (['x_glob'], {'scale_factor': '(16)', 'mode': '"""bilinear"""', 'align_corners': '(True)'}), "(x_glob, scale_factor=16, mode='bilinear', align_corners=True)\n", (1606, 1668), True, 'import torch.nn.functional as F\n'), ((1871, 1938), 'torch.nn.functional.ups...
from django.utils.translation import ugettext_lazy as _ SHIPPING_STANDARD = 'Standard' SHIPPING_EXPEDITED = 'Expedited' SHIPPING_PRIORITY = 'Priority' SHIPPING_SPEED_CATEGORIES = ( (SHIPPING_STANDARD, _("Standard")), (SHIPPING_EXPEDITED, _("Expedited")), (SHIPPING_PRIORITY, _("Priority")), ) METHOD_CONS...
[ "django.utils.translation.ugettext_lazy" ]
[((208, 221), 'django.utils.translation.ugettext_lazy', '_', (['"""Standard"""'], {}), "('Standard')\n", (209, 221), True, 'from django.utils.translation import ugettext_lazy as _\n'), ((249, 263), 'django.utils.translation.ugettext_lazy', '_', (['"""Expedited"""'], {}), "('Expedited')\n", (250, 263), True, 'from djang...
import os from tkinter import * from tkinter import filedialog global archivo archivo= [] cadena2 = [] archivo = [] global cadena_inicial cadena_inicial = [] global lista_numeros lista_numeros = [] global intlist intlist= [] global numeros1 numeros1=[] global lista_buscar2 lista_buscar2 =...
[ "os.startfile", "tkinter.filedialog.askopenfilename" ]
[((1844, 1872), 'tkinter.filedialog.askopenfilename', 'filedialog.askopenfilename', ([], {}), '()\n', (1870, 1872), False, 'from tkinter import filedialog\n'), ((14254, 14283), 'os.startfile', 'os.startfile', (['"""reporte0.html"""'], {}), "('reporte0.html')\n", (14266, 14283), False, 'import os\n')]
from pyexocross.hitran.hitran import HITRANLinelist from pyexocross.pyexocross import PyExocross from pyexocross.exomol.exomolbroads import ExomolBroadener import numpy as np from pyexocross.util import create_grid_res, convert_to_wavenumber from pyexocross.writer.hdf5writer import HDF5Writer import matplotlib.pyplot a...
[ "matplotlib.pyplot.show", "pyexocross.exomol.exomolbroads.ExomolBroadener", "pyexocross.hitran.hitran.HITRANLinelist", "matplotlib.pyplot.ylabel", "matplotlib.pyplot.xlabel", "matplotlib.pyplot.plot", "matplotlib.pyplot.figure", "pyexocross.util.create_grid_res", "matplotlib.pyplot.yscale", "matpl...
[((475, 549), 'pyexocross.hitran.hitran.HITRANLinelist', 'HITRANLinelist', (['"""/Users/ahmed/Documents/molecular_data/HITRAN/CH4/CH4.par"""'], {}), "('/Users/ahmed/Documents/molecular_data/HITRAN/CH4/CH4.par')\n", (489, 549), False, 'from pyexocross.hitran.hitran import HITRANLinelist\n'), ((644, 776), 'pyexocross.exo...
from bigflow.workflow import Workflow, hourly_start_time from datetime import datetime from datetime import timedelta class HourlyJob: def __init__(self): self.id = 'hourly_job' def run(self, runtime): print(f'I should process data with timestamps from: {runtime} ' f'to {dateti...
[ "datetime.datetime.strptime", "datetime.timedelta" ]
[((314, 361), 'datetime.datetime.strptime', 'datetime.strptime', (['runtime', '"""%Y-%m-%d %H:%M:%S"""'], {}), "(runtime, '%Y-%m-%d %H:%M:%S')\n", (331, 361), False, 'from datetime import datetime\n'), ((364, 397), 'datetime.timedelta', 'timedelta', ([], {'minutes': '(59)', 'seconds': '(59)'}), '(minutes=59, seconds=59...
#!/usr/local/bin/env python3.7 # -*- coding: utf-8; py-indent-offset:4 -*- ############################################################################### # # The MIT License (MIT) # Copyright (c) 2020 <NAME> # # Permission is hereby granted, free of charge, to any person obtaining a copy # of this software and ass...
[ "charting.Charting", "manip_data.ManipData.pd_tolist" ]
[((1759, 1837), 'charting.Charting', 'cht.Charting', (['self.series', 'self.date_name', 'self.default_data'], {}), '(self.series, self.date_name, self.default_data, **self.indicator)\n', (1771, 1837), True, 'import charting as cht\n'), ((2188, 2235), 'manip_data.ManipData.pd_tolist', 'md.pd_tolist', (['self.trades_trac...
# Generated by Django 3.1.2 on 2021-03-14 11:56 from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('cart', '0022_delete_sendpassword'), ] operations = [ migrations.AlterField( model_name='orderbeta', name='finish_pri...
[ "django.db.models.PositiveIntegerField" ]
[((343, 381), 'django.db.models.PositiveIntegerField', 'models.PositiveIntegerField', ([], {'default': '(0)'}), '(default=0)\n', (370, 381), False, 'from django.db import migrations, models\n')]
import logging.config import os import yaml __location__ = os.path.realpath( os.path.join(os.getcwd(), os.path.dirname(__file__))) def read_logging_config(): with open(os.path.join(__location__, 'logging.cfg'), 'r') as stream: try: logging_config = yaml.safe_load(stream) logg...
[ "os.path.dirname", "os.path.join", "yaml.safe_load", "os.getcwd" ]
[((96, 107), 'os.getcwd', 'os.getcwd', ([], {}), '()\n', (105, 107), False, 'import os\n'), ((109, 134), 'os.path.dirname', 'os.path.dirname', (['__file__'], {}), '(__file__)\n', (124, 134), False, 'import os\n'), ((180, 221), 'os.path.join', 'os.path.join', (['__location__', '"""logging.cfg"""'], {}), "(__location__, ...
import requests import logging from bbot.core import ChatbotEngine, BBotException, BBotCore, BBotExtensionException from engines.dotflow2.chatbot_engine import DotFlow2LoggerAdapter class DotFlow2MSCSSentimentAnalysis(): """ChatScript DotFlow2 function""" def __init__(self, config: dict, dotbot: dict) -> No...
[ "logging.getLogger", "bbot.core.BBotExtensionException", "requests.post" ]
[((1984, 2130), 'requests.post', 'requests.post', (['f"""https://{self.azure_location}.api.cognitive.microsoft.com/text/analytics/v2.0/sentiment"""'], {'json': 'payload', 'headers': 'headers'}), "(\n f'https://{self.azure_location}.api.cognitive.microsoft.com/text/analytics/v2.0/sentiment'\n , json=payload, heade...
import os import pandas as pd from poor_trader import config from poor_trader import trading from poor_trader import systems class CombinedIndicators(trading.TradingSystem): def __init__(self, portfolio, systems_method_list, name='CombinedIndicators'): super(CombinedIndicators, self).__init__(name=name) ...
[ "pandas.read_pickle", "os.path.exists", "poor_trader.systems.run_slsma", "poor_trader.systems.run_dcsma", "poor_trader.systems.run_atr_channel_breakout", "pandas.DataFrame" ]
[((561, 575), 'pandas.DataFrame', 'pd.DataFrame', ([], {}), '()\n', (573, 575), True, 'import pandas as pd\n'), ((650, 676), 'os.path.exists', 'os.path.exists', (['self.fpath'], {}), '(self.fpath)\n', (664, 676), False, 'import os\n'), ((1502, 1516), 'pandas.DataFrame', 'pd.DataFrame', ([], {}), '()\n', (1514, 1516), T...
# The MIT License (MIT) # # Copyright (c) 2018 PyBER # # Permission is hereby granted, free of charge, to any person obtaining a copy # of this software and associated documentation files (the "Software"), to deal # in the Software without restriction, including without limitation the rights # to use, copy, modify, mer...
[ "lib.pyqtgraph.pyqtgraph.Qt.QtGui.QFrame", "lib.pyqtgraph.pyqtgraph.Qt.QtGui.QFormLayout", "lib.pyqtgraph.pyqtgraph.Qt.QtGui.QLineEdit", "os.startfile", "lib.pyqtgraph.pyqtgraph.Qt.QtGui.QScrollArea", "lib.pyqtgraph.pyqtgraph.Qt.QtGui.QFileSystemModel", "lib.pyqtgraph.pyqtgraph.Qt.QtGui.QWidget", "os....
[((13455, 13479), 'lib.pyqtgraph.pyqtgraph.Qt.QtGui.QFileSystemModel', 'QtGui.QFileSystemModel', ([], {}), '()\n', (13477, 13479), False, 'from lib.pyqtgraph.pyqtgraph.Qt import QtCore, QtGui, QtWidgets\n'), ((3260, 3271), 'time.time', 'time.time', ([], {}), '()\n', (3269, 3271), False, 'import time\n'), ((3362, 3389),...
# -*- coding: utf-8 -*- import os import json import torch BASE_DIR = os.path.abspath(os.path.dirname(__file__)) class Config: def __init__(self, json_file): self.config = json.loads(open(json_file).read()) self.device = torch.device("cuda:1" if torch.cuda.is_available() else "cpu") self...
[ "os.path.dirname", "torch.cuda.is_available", "os.path.join", "torch.device" ]
[((89, 114), 'os.path.dirname', 'os.path.dirname', (['__file__'], {}), '(__file__)\n', (104, 114), False, 'import os\n'), ((334, 353), 'torch.device', 'torch.device', (['"""cpu"""'], {}), "('cpu')\n", (346, 353), False, 'import torch\n'), ((832, 904), 'os.path.join', 'os.path.join', (['BASE_DIR', '"""dataset"""', "self...
import unittest import torch import numpy as np from spectralgp.samplers import MeanEllipticalSlice class TestMeanEllipticalSlice(unittest.TestCase): def test_m_ess(self, nsamples=10000): pmean = torch.zeros(2) pmean[0] = -2. prior_dist = torch.distributions.MultivariateNormal(pmean, covar...
[ "numpy.mean", "spectralgp.samplers.MeanEllipticalSlice", "torch.eye", "unittest.main", "numpy.cov", "torch.zeros", "torch.inverse" ]
[((2193, 2208), 'unittest.main', 'unittest.main', ([], {}), '()\n', (2206, 2208), False, 'import unittest\n'), ((210, 224), 'torch.zeros', 'torch.zeros', (['(2)'], {}), '(2)\n', (221, 224), False, 'import torch\n'), ((372, 386), 'torch.zeros', 'torch.zeros', (['(2)'], {}), '(2)\n', (383, 386), False, 'import torch\n'),...
""" The input widgets generally allow entering arbitrary information into a text field or similar. """ from __future__ import absolute_import, division, unicode_literals import ast from base64 import b64decode, b64encode from datetime import datetime from six import string_types import param from bokeh.models.widge...
[ "param.ClassSelector", "param.Number", "param.Dict", "datetime.datetime.strptime", "param.Date", "param.Boolean", "datetime.datetime.strftime", "base64.b64decode", "base64.b64encode", "param.Parameter", "ast.literal_eval", "param.String", "param.Color" ]
[((647, 688), 'param.String', 'param.String', ([], {'default': '""""""', 'allow_None': '(True)'}), "(default='', allow_None=True)\n", (659, 688), False, 'import param\n'), ((708, 732), 'param.String', 'param.String', ([], {'default': '""""""'}), "(default='')\n", (720, 732), False, 'import param\n'), ((810, 836), 'para...
# -*- coding=utf-8 -*- import urllib2 import json import logger import traceback def send(apiUrl,data,method=None): logger.debug("调用内部系统[%s],data[%r]",apiUrl,data) try: data_json = json.dumps(data) headers = {'Content-Type': 'application/json'} # 设置数据为json格式,很重要 request = urllib2.Reques...
[ "urllib2.urlopen", "logger.init_4_debug", "json.dumps", "urllib2.Request", "logger.exception", "logger.debug" ]
[((121, 170), 'logger.debug', 'logger.debug', (['"""调用内部系统[%s],data[%r]"""', 'apiUrl', 'data'], {}), "('调用内部系统[%s],data[%r]', apiUrl, data)\n", (133, 170), False, 'import logger\n'), ((815, 836), 'logger.init_4_debug', 'logger.init_4_debug', ([], {}), '()\n', (834, 836), False, 'import logger\n'), ((198, 214), 'json.du...
from sklearn.datasets import load_wine from sklearn.neighbors import KNeighborsClassifier from sklearn.ensemble import RandomForestClassifier from sklearn.linear_model import LogisticRegression wine = load_wine(as_frame=True) X, y = wine.data, wine.target X_train, X_test, y_train, y_test = train_test_split( X, y...
[ "sklearn.neighbors.KNeighborsClassifier", "sklearn.datasets.load_wine", "sklearn.ensemble.RandomForestClassifier", "sklearn.linear_model.LogisticRegression" ]
[((202, 226), 'sklearn.datasets.load_wine', 'load_wine', ([], {'as_frame': '(True)'}), '(as_frame=True)\n', (211, 226), False, 'from sklearn.datasets import load_wine\n'), ((358, 380), 'sklearn.neighbors.KNeighborsClassifier', 'KNeighborsClassifier', ([], {}), '()\n', (378, 380), False, 'from sklearn.neighbors import K...
# Generated by Django 3.0.3 on 2020-02-25 18:50 from django.conf import settings from django.db import migrations, models import django.db.models.deletion import uuid class Migration(migrations.Migration): initial = True dependencies = [ migrations.swappable_dependency(settings.AUTH_USER_MODEL), ...
[ "django.db.models.UUIDField", "django.db.models.FloatField", "django.db.models.TextField", "django.db.models.ForeignKey", "django.db.models.IntegerField", "django.db.models.DateTimeField", "django.db.models.BooleanField", "django.db.models.AutoField", "django.db.models.URLField", "django.db.migrat...
[((259, 316), 'django.db.migrations.swappable_dependency', 'migrations.swappable_dependency', (['settings.AUTH_USER_MODEL'], {}), '(settings.AUTH_USER_MODEL)\n', (290, 316), False, 'from django.db import migrations, models\n'), ((449, 542), 'django.db.models.AutoField', 'models.AutoField', ([], {'auto_created': '(True)...
""" Summary: - Command-line Interface (CLI) Utilities Module - Python3 Module Functions: - convert_strtime_datetime: Convert human-readable datetime string into a datetime object for conducting time operations. - convert_timedelta: Convert a datetime duration object into human-r...
[ "logging.getLogger", "datetime.timedelta", "datetime.datetime.strptime", "inspect.stack" ]
[((568, 598), 'logging.getLogger', 'logging.getLogger', (['__version__'], {}), '(__version__)\n', (585, 598), False, 'import logging\n'), ((937, 988), 'datetime.datetime.strptime', 'datetime.datetime.strptime', (['dt', '"""%Y-%m-%dT%H:%M:%S"""'], {}), "(dt, '%Y-%m-%dT%H:%M:%S')\n", (963, 988), False, 'import datetime\n...
import json def translate_name(name): with open("name_translator.json", "r") as file: data = json.load(file) return data[name]
[ "json.load" ]
[((107, 122), 'json.load', 'json.load', (['file'], {}), '(file)\n', (116, 122), False, 'import json\n')]
#!/usr/bin/env python import sys import collections listas = list(map(str.split, sys.stdin.readlines())) entrada = [item for sublist in listas for item in sublist] # simplificando as listas, para facilitar contador = collections.Counter() palavras = [] for palavra in entrada: if palavra == "#": break palavr...
[ "collections.Counter", "sys.stdin.readlines" ]
[((219, 240), 'collections.Counter', 'collections.Counter', ([], {}), '()\n', (238, 240), False, 'import collections\n'), ((83, 104), 'sys.stdin.readlines', 'sys.stdin.readlines', ([], {}), '()\n', (102, 104), False, 'import sys\n')]
import torch from viz.visualizer import Visualizer from modules.deepmil import Attention from msidata.dataset_msi_features_with_patients import PreProcessedMSIFeatureDataset from testing.logistic_regression import get_precomputed_dataloader import argparse from experiment import ex from utils import post_config_hook...
[ "utils.post_config_hook", "modules.deepmil.Attention", "torch.cuda.is_available", "argparse.Namespace", "viz.visualizer.Visualizer", "testing.logistic_regression.get_precomputed_dataloader" ]
[((372, 405), 'argparse.Namespace', 'argparse.Namespace', ([], {}), '(**_run.config)\n', (390, 405), False, 'import argparse\n'), ((417, 445), 'utils.post_config_hook', 'post_config_hook', (['args', '_run'], {}), '(args, _run)\n', (433, 445), False, 'from utils import post_config_hook\n'), ((586, 652), 'testing.logisti...
import socket s = socket.socket() host = input(str("Please enter the host address of the sender: ")) port = 8080 s.connect((host, port)) print(f"[+] CONNECTED TO {host}:{port}") filename = input(str("Please enter filename for the incoming file: ")) file = open(filename, 'wb') file_data = s.recv(1024) file.write(fi...
[ "socket.socket" ]
[((19, 34), 'socket.socket', 'socket.socket', ([], {}), '()\n', (32, 34), False, 'import socket\n')]
from django.contrib import admin from .models import Book, Favorite admin.site.register(Book) admin.site.register(Favorite)
[ "django.contrib.admin.site.register" ]
[((71, 96), 'django.contrib.admin.site.register', 'admin.site.register', (['Book'], {}), '(Book)\n', (90, 96), False, 'from django.contrib import admin\n'), ((97, 126), 'django.contrib.admin.site.register', 'admin.site.register', (['Favorite'], {}), '(Favorite)\n', (116, 126), False, 'from django.contrib import admin\n...
from django.core import exceptions from django.http import FileResponse from django.utils.text import format_lazy from django.utils.translation import gettext_lazy as _ from rest_framework import status from rest_framework.decorators import action from rest_framework.exceptions import ValidationError from rest_framewor...
[ "applications.models.SummerVoucher.objects.all", "applications.api.v1.serializers.AttachmentSerializer", "django.utils.translation.gettext_lazy", "rest_framework.decorators.action", "applications.models.Application.objects.all", "rest_framework.response.Response", "django.http.FileResponse", "rest_fra...
[((1161, 1186), 'applications.models.Application.objects.all', 'Application.objects.all', ([], {}), '()\n', (1184, 1186), False, 'from applications.models import Application, SummerVoucher\n'), ((2943, 2970), 'applications.models.SummerVoucher.objects.all', 'SummerVoucher.objects.all', ([], {}), '()\n', (2968, 2970), F...
import json import requests import time from discord_webhook import DiscordWebhook, DiscordEmbed webhook_url = 'https://discordapp.com/api/webhooks/672159508675690497/4UtaClAc7rKMJsEvbR4iYf-Razv4M3ZWtkYDOxBzLfiDzJhI7RSFpoLn6iijBiRcaNOR' webhook = DiscordWebhook(webhook_url) pid = '508214-660' headers = { 'Connecti...
[ "discord_webhook.DiscordEmbed", "requests.post", "discord_webhook.DiscordWebhook", "time.sleep" ]
[((248, 275), 'discord_webhook.DiscordWebhook', 'DiscordWebhook', (['webhook_url'], {}), '(webhook_url)\n', (262, 275), False, 'from discord_webhook import DiscordWebhook, DiscordEmbed\n'), ((1239, 1378), 'requests.post', 'requests.post', (['"""https://2fwotdvm2o-dsn.algolia.net/1/indexes/product_variants_v2/query"""']...
"""Unit tests for the pytai application. License: MIT License Permission is hereby granted, free of charge, to any person obtaining a copy of this software and associated documentation files (the "Software"), to deal in the Software without restriction, including without limitation the rights ...
[ "unittest.main", "xml.etree.ElementTree.Element", "xml.etree.ElementTree.SubElement", "pathlib.Path" ]
[((3883, 3898), 'unittest.main', 'unittest.main', ([], {}), '()\n', (3896, 3898), False, 'import unittest\n'), ((2221, 2262), 'xml.etree.ElementTree.SubElement', 'ET.SubElement', (['parent_handle', '"""node"""'], {}), "(parent_handle, 'node', **d)\n", (2234, 2262), True, 'import xml.etree.ElementTree as ET\n'), ((2102,...
# Copyright (c) 2019 Science and Technology Facilities Council # All rights reserved. # Modifications made as part of the fparser project are distributed # under the following license: # Redistribution and use in source and binary forms, with or without # modification, are permitted provided that the following condi...
[ "fparser.api.get_reader", "pytest.raises", "fparser.two.Fortran2003.Include_Stmt" ]
[((2737, 2753), 'fparser.api.get_reader', 'get_reader', (['line'], {}), '(line)\n', (2747, 2753), False, 'from fparser.api import get_reader\n'), ((2993, 3011), 'fparser.two.Fortran2003.Include_Stmt', 'Include_Stmt', (['line'], {}), '(line)\n', (3005, 3011), False, 'from fparser.two.Fortran2003 import Include_Stmt, Int...
import os, csv, time, shutil from bin import cardbank from bin import builder def add_build(add_cards): '''Build card bank by specifically adding new cards.''' # process new cards to add if not add_cards: return add_card_map = {} for card in add_cards: add_card_map[card["inf"]] = ...
[ "csv.DictWriter", "os.path.exists", "bin.builder.get", "bin.builder.session", "time.sleep", "shutil.copyfile", "bin.builder.build", "bin.cardbank.read" ]
[((385, 427), 'os.path.exists', 'os.path.exists', (['"""bank/card-bank-built.csv"""'], {}), "('bank/card-bank-built.csv')\n", (399, 427), False, 'import os, csv, time, shutil\n'), ((560, 577), 'bin.builder.session', 'builder.session', ([], {}), '()\n', (575, 577), False, 'from bin import builder\n'), ((1739, 1799), 'bi...
from mongoengine import StringField, EmailField, BooleanField from flask.ext.login import UserMixin import requests import json from mongoengine import Document from social.apps.flask_app.me.models import FlaskStorage class User(Document, UserMixin): username = StringField(max_length=200) password = StringFi...
[ "json.loads", "social.apps.flask_app.me.models.FlaskStorage.user.get_social_auth_for_user", "requests.get", "mongoengine.StringField", "mongoengine.EmailField", "mongoengine.BooleanField" ]
[((269, 296), 'mongoengine.StringField', 'StringField', ([], {'max_length': '(200)'}), '(max_length=200)\n', (280, 296), False, 'from mongoengine import StringField, EmailField, BooleanField\n'), ((312, 351), 'mongoengine.StringField', 'StringField', ([], {'max_length': '(200)', 'default': '""""""'}), "(max_length=200,...
""" ******************************************************************************** compas_blender.geometry ******************************************************************************** .. currentmodule:: compas_blender.geometry Object-oriented convenience wrappers for native Blender geometry. .. autosummary:: ...
[ "bpy.ops.wm.redraw_timer" ]
[((1094, 1153), 'bpy.ops.wm.redraw_timer', 'bpy.ops.wm.redraw_timer', ([], {'type': '"""DRAW_WIN_SWAP"""', 'iterations': '(1)'}), "(type='DRAW_WIN_SWAP', iterations=1)\n", (1117, 1153), False, 'import bpy\n')]
# Copyright 2020 Google LLC # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, ...
[ "copy.deepcopy" ]
[((1447, 1481), 'copy.deepcopy', 'deepcopy', (['COLUMN_VALIDATION_CONFIG'], {}), '(COLUMN_VALIDATION_CONFIG)\n', (1455, 1481), False, 'from copy import deepcopy\n')]
import traceback from typing import Any, Callable, Dict, List, Optional, Set, Tuple from chainalytic_icon.common import config, util class ApiBundle(object): """ The interface to external consumers/applications """ def __init__(self, working_dir: str): super(ApiBundle, self).__init__() ...
[ "traceback.format_exc", "chainalytic_icon.common.util.get_child_logger" ]
[((406, 450), 'chainalytic_icon.common.util.get_child_logger', 'util.get_child_logger', (['"""provider.api_bundle"""'], {}), "('provider.api_bundle')\n", (427, 450), False, 'from chainalytic_icon.common import config, util\n'), ((1193, 1215), 'traceback.format_exc', 'traceback.format_exc', ([], {}), '()\n', (1213, 1215...
import copy import torch from utils import helpers from utils.layers import conv, linear, batch_norm def ticketfy(model, split_rate, split_mode="kels"): conv_layers, linear_layers, bn_layers = helpers.get_layers(model) for n, _ in conv_layers: cur_conv = helpers.rgetattr(model, n) helpers.rs...
[ "utils.layers.batch_norm.SplitBatchNorm", "utils.helpers.get_layers", "utils.layers.conv.SplitConv", "copy.deepcopy", "utils.helpers.rgetattr" ]
[((200, 225), 'utils.helpers.get_layers', 'helpers.get_layers', (['model'], {}), '(model)\n', (218, 225), False, 'from utils import helpers\n'), ((2410, 2430), 'copy.deepcopy', 'copy.deepcopy', (['model'], {}), '(model)\n', (2423, 2430), False, 'import copy\n'), ((275, 301), 'utils.helpers.rgetattr', 'helpers.rgetattr'...
from utils import scrape_helper url = "http://www.investopedia.com/terms/1/" links = scrape_helper.get_term_links_from_page(url) print(links)
[ "utils.scrape_helper.get_term_links_from_page" ]
[((87, 130), 'utils.scrape_helper.get_term_links_from_page', 'scrape_helper.get_term_links_from_page', (['url'], {}), '(url)\n', (125, 130), False, 'from utils import scrape_helper\n')]
import requests import bs4 dateList = [] higlist = [] lowlist= [] r = requests.get( 'https://coinmarketcap.com/currencies/bitcoin/historical-data/') soup = bs4.BeautifulSoup(r.text, "lxml") tr = soup.find_all('tr',{'class':'text-right'}) for item in tr: dateList.append(item.find('td', {'class':'text-left'}...
[ "bs4.BeautifulSoup", "requests.get" ]
[((74, 151), 'requests.get', 'requests.get', (['"""https://coinmarketcap.com/currencies/bitcoin/historical-data/"""'], {}), "('https://coinmarketcap.com/currencies/bitcoin/historical-data/')\n", (86, 151), False, 'import requests\n'), ((164, 197), 'bs4.BeautifulSoup', 'bs4.BeautifulSoup', (['r.text', '"""lxml"""'], {})...
import os from colorama import Fore, Style from pathlib import Path DIF = "dif10" ECHO10 = "echo10" UMM_JSON = "umm-json" ROOT_DIR = ( # go up one directory Path(__file__).resolve().parents[1] ) SCHEMAS_BASE_PATH = f"{ROOT_DIR}/schemas" SCHEMAS = { "json": [ "checks", "check_messages", ...
[ "pathlib.Path" ]
[((168, 182), 'pathlib.Path', 'Path', (['__file__'], {}), '(__file__)\n', (172, 182), False, 'from pathlib import Path\n')]
""" Generic publisher for graphana """ import abc import six from decisionengine.framework.modules import Publisher import decisionengine_modules.graphite_client as graphite DEFAULT_GRAPHITE_HOST = 'fermicloud399.fnal.gov' DEFAULT_GRAPHITE_PORT = 2004 DEFAULT_GRAPHITE_CONTEXT = "" @six.add_metaclass(abc.ABCMeta) c...
[ "decisionengine_modules.graphite_client.Graphite", "six.add_metaclass" ]
[((288, 318), 'six.add_metaclass', 'six.add_metaclass', (['abc.ABCMeta'], {}), '(abc.ABCMeta)\n', (305, 318), False, 'import six\n'), ((1405, 1479), 'decisionengine_modules.graphite_client.Graphite', 'graphite.Graphite', ([], {'host': 'self.graphite_host', 'pickle_port': 'self.graphite_port'}), '(host=self.graphite_hos...
from flask import Blueprint api = Blueprint('api', __name__) from . import authentication, videos, shows, users, comments, errors
[ "flask.Blueprint" ]
[((35, 61), 'flask.Blueprint', 'Blueprint', (['"""api"""', '__name__'], {}), "('api', __name__)\n", (44, 61), False, 'from flask import Blueprint\n')]
""" Build SSW scripts from Jinja 2 templates """ import os import datetime import subprocess import tempfile from jinja2 import (Environment as Env, FileSystemLoader, PackageLoader) from scipy.io import readsav from .read_config import defaults from .util import SSWIDLError, ID...
[ "tempfile.TemporaryDirectory", "jinja2.Environment", "subprocess.run", "os.path.isfile", "os.path.dirname", "datetime.datetime.now", "subprocess.call", "os.path.basename", "scipy.io.readsav", "jinja2.PackageLoader" ]
[((1959, 1981), 'os.path.isfile', 'os.path.isfile', (['script'], {}), '(script)\n', (1973, 1981), False, 'import os\n'), ((2170, 2175), 'jinja2.Environment', 'Env', ([], {}), '()\n', (2173, 2175), True, 'from jinja2 import Environment as Env, FileSystemLoader, PackageLoader\n'), ((3869, 3898), 'tempfile.TemporaryDirect...
""" Entity-relation diagram (ERD) GraphViz dot-file generator. Usage: python -m erd db.json -o db.dot Then pass the result to the GraphViz `dot` tool: dot db.dot -T png -o db.png Inspired by: https://github.com/ehne/ERDot """ import argparse import json import re from pathlib import Path from typing import D...
[ "json.loads", "re.match", "argparse.ArgumentParser" ]
[((2412, 2595), 're.match', 're.match', (['"""^(?P<source_name>\\\\w+):(?P<source_fk>\\\\w+) (?P<left_cardinality>[\\\\d\\\\+\\\\*])--(?P<right_cardinality>[\\\\d\\\\+\\\\*]) (?P<dest_name>\\\\w+):(?P<dest_pk>\\\\w+)$"""', 'relation'], {}), "(\n '^(?P<source_name>\\\\w+):(?P<source_fk>\\\\w+) (?P<left_cardinality>[\...
# pylint: disable=no-self-use,invalid-name import unittest import spacy from scispacy.hyponym_detector import HyponymDetector class TestHyponymDetector(unittest.TestCase): def setUp(self): super().setUp() self.nlp = spacy.load("en_core_sci_sm") self.detector = HyponymDetector(self.nlp, ex...
[ "spacy.load", "scispacy.hyponym_detector.HyponymDetector" ]
[((239, 267), 'spacy.load', 'spacy.load', (['"""en_core_sci_sm"""'], {}), "('en_core_sci_sm')\n", (249, 267), False, 'import spacy\n'), ((292, 332), 'scispacy.hyponym_detector.HyponymDetector', 'HyponymDetector', (['self.nlp'], {'extended': '(True)'}), '(self.nlp, extended=True)\n', (307, 332), False, 'from scispacy.hy...
"""Steps up and down""" import calendar import numpy as np from pandas.io.sql import read_sql from pyiem import network from pyiem.plot.use_agg import plt from pyiem.util import get_autoplot_context, get_dbconn PDICT = {'spring': '1 January - 30 June', 'fall': '1 July - 31 December'} def get_description():...
[ "pyiem.network.Table", "pandas.io.sql.read_sql", "pyiem.util.get_dbconn", "numpy.array", "pyiem.plot.use_agg.plt.subplots" ]
[((1072, 1090), 'pyiem.util.get_dbconn', 'get_dbconn', (['"""coop"""'], {}), "('coop')\n", (1082, 1090), False, 'from pyiem.util import get_autoplot_context, get_dbconn\n'), ((1256, 1299), 'pyiem.network.Table', 'network.Table', (["('%sCLIMATE' % (station[:2],))"], {}), "('%sCLIMATE' % (station[:2],))\n", (1269, 1299),...
import struct import os current_pe = None class PE: """Basic PE parsing. Ref: - https://hshrzd.wordpress.com/pe-bear/ - https://blog.kowalczyk.info/articles/pefileformat.html """ X86_64 = 0x8664 X86_32 = 0x14c ARM = 0x1c0 ARM64 ...
[ "os.access" ]
[((1378, 1400), 'os.access', 'os.access', (['pe', 'os.R_OK'], {}), '(pe, os.R_OK)\n', (1387, 1400), False, 'import os\n')]
""" This module is the custom resource used by the MSAM's CloudFormation templates to populate the web bucket with contents of the MSAM web archive. """ # Copyright 2018 Amazon.com, Inc. or its affiliates. All Rights Reserved. # SPDX-License-Identifier: Apache-2.0 import json import os from subprocess import call imp...
[ "boto3.client", "json.dumps", "subprocess.call", "resource_tools.send", "resource_tools.stack_name", "os.walk" ]
[((1786, 1889), 'resource_tools.send', 'resource_tools.send', (['event', 'context', "result['Status']", "result['Data']", "result['PhysicalResourceId']"], {}), "(event, context, result['Status'], result['Data'],\n result['PhysicalResourceId'])\n", (1805, 1889), False, 'import resource_tools\n'), ((2098, 2116), 'boto...
# -*- coding: utf-8 -*- # ------------------------------------------------------------------------------ # # Copyright 2018-2019 Fetch.AI Limited # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the ...
[ "aea.cli.common.logger.exception", "pathlib.Path", "click.option", "aea.helpers.base.load_env_file", "click.echo", "click.Path", "sys.exit", "click.command" ]
[((2790, 2805), 'click.command', 'click.command', ([], {}), '()\n', (2803, 2805), False, 'import click\n'), ((2807, 3032), 'click.option', 'click.option', (['"""--connections"""', '"""connection_ids"""'], {'cls': 'ConnectionsOption', 'required': '(False)', 'default': 'None', 'help': '"""The connection names to use for ...
"""An environment to skip k frames and return a max between the last two.""" import gym import numpy as np class MaxFrameskipEnv(gym.Wrapper): """An environment to skip k frames and return a max between the last two.""" def __init__(self, env, skip: int=4) -> None: """ Initialize a new max fr...
[ "numpy.zeros", "gym.Wrapper.__init__" ]
[((557, 588), 'gym.Wrapper.__init__', 'gym.Wrapper.__init__', (['self', 'env'], {}), '(self, env)\n', (577, 588), False, 'import gym\n'), ((691, 750), 'numpy.zeros', 'np.zeros', (['(2, *env.observation_space.shape)'], {'dtype': 'np.uint8'}), '((2, *env.observation_space.shape), dtype=np.uint8)\n', (699, 750), True, 'im...
"""Script containing the DeepLoco environments.""" import gym import numpy as np import os import sys import cv2 try: sys.path.append(os.path.join(os.environ["TERRAINRL_PATH"], "simAdapter")) import terrainRLSim # noqa: F401 except (KeyError, ImportError, ModuleNotFoundError): pass class BipedalSoccer(g...
[ "numpy.flip", "terrainRLSim.getEnv", "os.path.join", "gym.spaces.Box", "cv2.imshow", "numpy.array", "cv2.waitKey", "gym.make" ]
[((139, 195), 'os.path.join', 'os.path.join', (["os.environ['TERRAINRL_PATH']", '"""simAdapter"""'], {}), "(os.environ['TERRAINRL_PATH'], 'simAdapter')\n", (151, 195), False, 'import os\n'), ((1016, 1077), 'terrainRLSim.getEnv', 'terrainRLSim.getEnv', (['"""PD-Biped3D-HLC-Soccer-v1"""'], {'render': '(False)'}), "('PD-B...
# -*- coding:utf-8 -*- ''' Created on 2015年3月2日 @author: wanhao01 ''' import sys from crawler.minispider import logerror import main reload(sys) sys.setdefaultencoding('utf-8') if __name__ == '__main__': try: main.main() except Exception as exception: logerror("error du...
[ "main.main", "sys.setdefaultencoding" ]
[((164, 195), 'sys.setdefaultencoding', 'sys.setdefaultencoding', (['"""utf-8"""'], {}), "('utf-8')\n", (186, 195), False, 'import sys\n'), ((245, 256), 'main.main', 'main.main', ([], {}), '()\n', (254, 256), False, 'import main\n')]
import types from tf.advanced.app import App MODIFIERS = """ remark folio note ref emph und super special q num den """.strip().split() def fmt_layoutFull(app, n, **kwargs): return app._wrapHtml(n, ("",)) def fmt_layoutRemarks(app, n, **kwargs): return app._wrapHtml(n, ("r",)) def fmt_layoutNotes(ap...
[ "types.MethodType" ]
[((819, 856), 'types.MethodType', 'types.MethodType', (['fmt_layoutFull', 'app'], {}), '(fmt_layoutFull, app)\n', (835, 856), False, 'import types\n'), ((889, 929), 'types.MethodType', 'types.MethodType', (['fmt_layoutRemarks', 'app'], {}), '(fmt_layoutRemarks, app)\n', (905, 929), False, 'import types\n'), ((960, 998)...
import hashlib import logging import os import shutil import traceback from contextlib import closing from pywb.utils.loaders import BlockLoader from webrecorder.rec.storage.base import BaseStorage from webrecorder.rec.storage.storagepaths import add_local_store_prefix, strip_prefix logger = logging.getLogger('wr.io...
[ "logging.getLogger", "hashlib.md5", "pywb.utils.loaders.BlockLoader", "os.path.join", "webrecorder.rec.storage.storagepaths.strip_prefix", "os.path.isfile", "os.path.dirname", "shutil.copyfile", "os.removedirs", "shutil.rmtree", "os.remove" ]
[((296, 322), 'logging.getLogger', 'logging.getLogger', (['"""wr.io"""'], {}), "('wr.io')\n", (313, 322), False, 'import logging\n'), ((875, 916), 'os.path.join', 'os.path.join', (['self.storage_root', 'dir_path'], {}), '(self.storage_root, dir_path)\n', (887, 916), False, 'import os\n'), ((2178, 2204), 'os.path.isfile...
# This file is part of OctoBot (https://github.com/Drakkar-Software/OctoBot) # Copyright (c) 2021 Drakkar-Software, All rights reserved. # # OctoBot is free software; you can redistribute it and/or # modify it under the terms of the GNU General Public License # as published by the Free Software Foundation; either ...
[ "octobot.community.ErrorsUploader", "time.time", "octobot.community.Error" ]
[((915, 926), 'time.time', 'time.time', ([], {}), '()\n', (924, 926), False, 'import time\n'), ((1010, 1074), 'octobot.community.Error', 'community.Error', (['None', 'ERROR_TITLE', 'ERROR_TIME', 'ERROR_METRICS_ID'], {}), '(None, ERROR_TITLE, ERROR_TIME, ERROR_METRICS_ID)\n', (1025, 1074), True, 'import octobot.communit...
import connexion import six from ga4ghtest.models import Plugin # noqa: E501 from ga4ghtest import util from ga4ghtest.core.controllers import plugins_controller as controller def create_plugin( body ): # noqa: E501 """Create a test plugin Add a plugin for testing functionality of an API. # noqa: E501...
[ "connexion.request.get_json", "ga4ghtest.core.controllers.plugins_controller.get_plugins", "ga4ghtest.core.controllers.plugins_controller.create_plugin" ]
[((515, 550), 'ga4ghtest.core.controllers.plugins_controller.create_plugin', 'controller.create_plugin', ([], {'body': 'body'}), '(body=body)\n', (539, 550), True, 'from ga4ghtest.core.controllers import plugins_controller as controller\n'), ((1010, 1075), 'ga4ghtest.core.controllers.plugins_controller.get_plugins', 'c...
import os from split_settings.tools import include, optional ENVIRONMENT = os.getenv('DJANGO_ENV') or 'development' include( # Load environment settings 'base/env.py', optional('local/env.py'), # We can "patch" any settings from local folder env.py file. # Here we should have the order because of de...
[ "os.getenv", "split_settings.tools.optional" ]
[((76, 99), 'os.getenv', 'os.getenv', (['"""DJANGO_ENV"""'], {}), "('DJANGO_ENV')\n", (85, 99), False, 'import os\n'), ((182, 206), 'split_settings.tools.optional', 'optional', (['"""local/env.py"""'], {}), "('local/env.py')\n", (190, 206), False, 'from split_settings.tools import include, optional\n'), ((520, 542), 's...
# -*- coding: utf-8 -*- ''' Utility methods for dictionary ============================== ''' __all__ = ( 'calling_dict_from', 'combine_dict', 'dict_sorted') from itertools import chain from typing import Tuple from builder.utils import assertion def calling_dict_from(calling: (str, dict), n...
[ "builder.utils.assertion.is_bool", "builder.utils.assertion.is_dict", "builder.utils.assertion.is_str" ]
[((818, 838), 'builder.utils.assertion.is_dict', 'assertion.is_dict', (['a'], {}), '(a)\n', (835, 838), False, 'from builder.utils import assertion\n'), ((842, 862), 'builder.utils.assertion.is_dict', 'assertion.is_dict', (['b'], {}), '(b)\n', (859, 862), False, 'from builder.utils import assertion\n'), ((570, 595), 'b...
from flask import Flask from flask_socketio import SocketIO from flask_sqlalchemy import SQLAlchemy UPLOAD_FOLDER = 'uploads' ALLOWED_EXTENSIONS = {'png', 'jpg', 'jpeg', 'gif'} app = Flask(__name__) app.config['SQLALCHEMY_DATABASE_URI'] = 'sqlite:////tmp/dev.db' app.config['UPLOAD_FOLDER'] = UPLOAD_FOLDER app.secret_...
[ "flask_sqlalchemy.SQLAlchemy", "flask_socketio.SocketIO", "flask.Flask" ]
[((185, 200), 'flask.Flask', 'Flask', (['__name__'], {}), '(__name__)\n', (190, 200), False, 'from flask import Flask\n'), ((346, 359), 'flask_socketio.SocketIO', 'SocketIO', (['app'], {}), '(app)\n', (354, 359), False, 'from flask_socketio import SocketIO\n'), ((365, 380), 'flask_sqlalchemy.SQLAlchemy', 'SQLAlchemy', ...
import numpy as np class Loss(): def output_gradient(self): return class MSE(Loss): def __call__(self, predicted, labels): return 0.5 * np.square(predicted - labels) def output_gradient(self, predicted, labels): return predicted - labels class BinaryCrossEntropy(Loss): def _...
[ "numpy.log", "numpy.nan_to_num", "numpy.square" ]
[((512, 581), 'numpy.nan_to_num', 'np.nan_to_num', (['(-(labels / predicted) + (1 - labels) / (1 - predicted))'], {}), '(-(labels / predicted) + (1 - labels) / (1 - predicted))\n', (525, 581), True, 'import numpy as np\n'), ((163, 192), 'numpy.square', 'np.square', (['(predicted - labels)'], {}), '(predicted - labels)\...
""" This file is part of pynadc https://github.com/rmvanhees/pynadc Methods to query the NADC Sciamachy SQLite database Copyright (c) 2012-2021 SRON - Netherlands Institute for Space Research All Rights Reserved License: BSD-3-Clause """ from pathlib import Path import sqlite3 # ------------------------------...
[ "sqlite3.connect", "pathlib.Path" ]
[((1944, 1967), 'sqlite3.connect', 'sqlite3.connect', (['dbname'], {}), '(dbname)\n', (1959, 1967), False, 'import sqlite3\n'), ((7502, 7525), 'sqlite3.connect', 'sqlite3.connect', (['dbname'], {}), '(dbname)\n', (7517, 7525), False, 'import sqlite3\n'), ((2581, 2596), 'pathlib.Path', 'Path', (['*row[:-1]'], {}), '(*ro...
from pycord.discord.ext import commands import pycord.discord as discord from pycord.discord import Embed import requests import json from discord import Embed class Data(commands.Cog): def __init__(self, bot) -> None: self.bot: commands.Bot = bot @commands.command() async def ping(self, ctx): await ctx...
[ "requests.get", "json.loads", "discord.Embed", "pycord.discord.ext.commands.command" ]
[((259, 277), 'pycord.discord.ext.commands.command', 'commands.command', ([], {}), '()\n', (275, 277), False, 'from pycord.discord.ext import commands\n'), ((378, 396), 'pycord.discord.ext.commands.command', 'commands.command', ([], {}), '()\n', (394, 396), False, 'from pycord.discord.ext import commands\n'), ((445, 52...
#!/usr/bin/python3 from SetupRunDirectory import verifyDirectoryFiles, setupRunDirectory from CleanupRunDirectory import cleanUpRunDirectory from RunAssembly import verifyConfigFiles, verifyFastaFiles, runAssembly, initializeAssembler from SaveRun import saveRun import configparser from datetime import datetime from ...
[ "configparser.ConfigParser", "SaveRun.saveRun", "RunAssembly.verifyFastaFiles", "os.path.exists", "SetupRunDirectory.verifyDirectoryFiles", "SetupRunDirectory.setupRunDirectory", "argparse.ArgumentParser", "subprocess.Popen", "CleanupRunDirectory.cleanUpRunDirectory", "os.mkdir", "os.path.dirnam...
[((584, 598), 'datetime.datetime.now', 'datetime.now', ([], {}), '()\n', (596, 598), False, 'from datetime import datetime\n'), ((4709, 4743), 'os.path.abspath', 'os.path.abspath', (['readsSequencePath'], {}), '(readsSequencePath)\n', (4724, 4743), False, 'import os\n'), ((5347, 5395), 'os.path.join', 'os.path.join', (...
""" You work for a retail store that wants to increase sales on Tuesday and Wednesday, which are the store's slowest sales days. On Tuesday and Wednesday, if a customer's subtotal is greater than $50, the store will discount the customer's purchase by 10%. """ # Import the datatime module so that # it can be used in t...
[ "datetime.datetime.now" ]
[((1143, 1157), 'datetime.datetime.now', 'datetime.now', ([], {}), '()\n', (1155, 1157), False, 'from datetime import datetime\n')]
# Copyright (c) 2021, NVIDIA CORPORATION. All rights reserved. # # NVIDIA CORPORATION and its licensors retain all intellectual property # and proprietary rights in and to this software, related documentation # and any modifications thereto. Any use, reproduction, disclosure or # distribution of this software and rel...
[ "numpy.sqrt", "torch.full_like", "torch.nn.functional.softplus", "torch.randn_like", "torch_utils.ops.conv2d_gradfix.no_weight_gradients", "torch.autograd.profiler.record_function", "torch.no_grad", "torch_utils.training_stats.report", "torch.linspace", "torch_utils.misc.ddp_sync", "torch.empty"...
[((1718, 1748), 'torch.zeros', 'torch.zeros', (['[]'], {'device': 'device'}), '([], device=device)\n', (1729, 1748), False, 'import torch\n'), ((2130, 2165), 'torch_utils.misc.ddp_sync', 'misc.ddp_sync', (['self.G_mapping', 'sync'], {}), '(self.G_mapping, sync)\n', (2143, 2165), False, 'from torch_utils import misc, tr...
from BurstPaperWallet.api import brs_api from BurstPaperWallet.api import passphrase_url_transform as transform def initialize(account, old_passphrase, fee=735000): url = "sendMoney&recipient={}&secretPhrase={}&amountNQT=1&feeNQT={}&recipientPublicKey={}&deadline=1440"\ .format(account["reed solomon"], tr...
[ "BurstPaperWallet.api.passphrase_url_transform", "BurstPaperWallet.api.brs_api" ]
[((512, 524), 'BurstPaperWallet.api.brs_api', 'brs_api', (['url'], {}), '(url)\n', (519, 524), False, 'from BurstPaperWallet.api import brs_api\n'), ((318, 343), 'BurstPaperWallet.api.passphrase_url_transform', 'transform', (['old_passphrase'], {}), '(old_passphrase)\n', (327, 343), True, 'from BurstPaperWallet.api imp...
import numpy as np import os import keras from keras import regularizers, losses from keras.models import Sequential, Model from keras.layers import Lambda, Input, Dense, Dropout, Reshape, BatchNormalization, Softmax, Concatenate from keras.utils import plot_model import keras.backend as K class multiVAE: def __init_...
[ "keras.layers.Concatenate", "keras.utils.plot_model", "keras.layers.Lambda", "keras.models.Model", "keras.backend.random_normal", "keras.layers.Dense", "keras.backend.exp" ]
[((1138, 1207), 'keras.backend.random_normal', 'K.random_normal', ([], {'shape': '(32, self.inf_layerSize)', 'mean': '(0.0)', 'stddev': '(1.0)'}), '(shape=(32, self.inf_layerSize), mean=0.0, stddev=1.0)\n', (1153, 1207), True, 'import keras.backend as K\n'), ((3020, 3047), 'keras.models.Model', 'Model', ([], {'inputs':...
import pytest from tests.stub_client import StubHttpClient from coinoxr.requestor import Requestor from coinoxr.response import Response def content(file): return StubHttpClient.json(file)["content"] @pytest.fixture def client(): client = StubHttpClient() client.add_app_id("fake_app_id") client.add...
[ "tests.stub_client.StubHttpClient", "coinoxr.response.Response", "coinoxr.requestor.Requestor", "tests.stub_client.StubHttpClient.json" ]
[((252, 268), 'tests.stub_client.StubHttpClient', 'StubHttpClient', ([], {}), '()\n', (266, 268), False, 'from tests.stub_client import StubHttpClient\n'), ((733, 765), 'coinoxr.requestor.Requestor', 'Requestor', (['"""fake_app_id"""', 'client'], {}), "('fake_app_id', client)\n", (742, 765), False, 'from coinoxr.reques...
#!flask/bin/python # coding=utf-8 from flask import Flask, jsonify app = Flask(__name__) tasks = { "event_id" : "1.9", "introductions" : [ { "title" : "情怀", "details" : "各种无敌, 各种牛人, 各种挑战, 等你来战", "image" : "hello.png", "background_image" : "backgroundImage.png" }, { "title...
[ "flask.jsonify", "flask.Flask" ]
[((74, 89), 'flask.Flask', 'Flask', (['__name__'], {}), '(__name__)\n', (79, 89), False, 'from flask import Flask, jsonify\n'), ((539, 553), 'flask.jsonify', 'jsonify', (['tasks'], {}), '(tasks)\n', (546, 553), False, 'from flask import Flask, jsonify\n')]
# Copyright (c) 2015-2017 <NAME> # License: MIT """ nbrun - Run an Jupyter/IPython notebook, optionally passing arguments. USAGE ----- Copy this file in the folder containing the master notebook used to execute the other notebooks. Then use `run_notebook()` to execute notebooks. """ import time from pathlib import P...
[ "nbconvert.HTMLExporter", "time.ctime", "argparse.ArgumentParser", "pathlib.Path", "nbformat.v4.new_code_cell", "nbconvert.preprocessors.ExecutePreprocessor", "time.time", "nbformat.v4.new_markdown_cell" ]
[((5814, 5833), 'pathlib.Path', 'Path', (['notebook_path'], {}), '(notebook_path)\n', (5818, 5833), False, 'from pathlib import Path\n'), ((6930, 6967), 'nbconvert.preprocessors.ExecutePreprocessor', 'ExecutePreprocessor', ([], {}), '(**execute_kwargs)\n', (6949, 6967), False, 'from nbconvert.preprocessors import Execu...
import numpy as np import pandas as pd import seaborn as sns from nninst.backend.tensorflow.model import AlexNet from nninst.backend.tensorflow.trace.alexnet_imagenet_inter_class_similarity import ( alexnet_imagenet_inter_class_similarity_frequency, ) from nninst.op import Conv2dOp, DenseOp np.random.seed(0) sns....
[ "seaborn.set", "nninst.backend.tensorflow.model.AlexNet.graph", "numpy.eye", "nninst.backend.tensorflow.trace.alexnet_imagenet_inter_class_similarity.alexnet_imagenet_inter_class_similarity_frequency", "numpy.around", "numpy.random.seed", "pandas.DataFrame", "numpy.tri" ]
[((298, 315), 'numpy.random.seed', 'np.random.seed', (['(0)'], {}), '(0)\n', (312, 315), True, 'import numpy as np\n'), ((316, 325), 'seaborn.set', 'sns.set', ([], {}), '()\n', (323, 325), True, 'import seaborn as sns\n'), ((3070, 3184), 'pandas.DataFrame', 'pd.DataFrame', (["{'Same Class': same_class_similarity, 'Diff...
# -*- coding: utf-8 -*- """ <NAME> Computational Biologist Target Sciences GSK <EMAIL> """ import sys import get_generalizable_features import get_merged_features import get_useful_features def main(validation_rep=0, validation_fold=0): print('VALIDATION_REP: {0!s}, VALIDATION_FOLD:{1!s}'.format(validation_r...
[ "get_useful_features.main", "get_merged_features.main", "get_generalizable_features.main" ]
[((418, 482), 'get_generalizable_features.main', 'get_generalizable_features.main', (['validation_rep', 'validation_fold'], {}), '(validation_rep, validation_fold)\n', (449, 482), False, 'import get_generalizable_features\n'), ((544, 601), 'get_merged_features.main', 'get_merged_features.main', (['validation_rep', 'val...
# Generated by Django 3.1.6 on 2021-04-12 08:45 from django.db import migrations class Migration(migrations.Migration): dependencies = [ ('SocialApp', '0002_auto_20210411_2237'), ] operations = [ migrations.DeleteModel( name='RemoteFollow', ), ]
[ "django.db.migrations.DeleteModel" ]
[((229, 272), 'django.db.migrations.DeleteModel', 'migrations.DeleteModel', ([], {'name': '"""RemoteFollow"""'}), "(name='RemoteFollow')\n", (251, 272), False, 'from django.db import migrations\n')]
"""Tests for thread_async_queue.""" from __future__ import annotations import asyncio from concurrent.futures import ThreadPoolExecutor from itertools import chain from typing import List, NamedTuple import pytest from opentrons.protocol_runner.thread_async_queue import ( ThreadAsyncQueue, QueueClosed, ) ...
[ "itertools.chain", "concurrent.futures.ThreadPoolExecutor", "pytest.raises", "asyncio.sleep" ]
[((606, 632), 'pytest.raises', 'pytest.raises', (['QueueClosed'], {}), '(QueueClosed)\n', (619, 632), False, 'import pytest\n'), ((666, 692), 'pytest.raises', 'pytest.raises', (['QueueClosed'], {}), '(QueueClosed)\n', (679, 692), False, 'import pytest\n'), ((770, 796), 'pytest.raises', 'pytest.raises', (['QueueClosed']...
"""Solution for subdag_example.py. Uses a factory function to return a DAG that can be used as the subdag argument to SubDagOperator. Notice that: 1) the SubDAG's dag_id is formatted as parent_dag_id.subdag_task_id 2) the start_date and schedule_interval of the SubDAG are copied from the parent DAG. """ from airflo...
[ "datetime.datetime.min.time", "airflow.operators.dummy_operator.DummyOperator", "airflow.DAG", "datetime.datetime.today", "datetime.timedelta" ]
[((744, 763), 'datetime.datetime.min.time', 'datetime.min.time', ([], {}), '()\n', (761, 763), False, 'from datetime import datetime, timedelta\n'), ((963, 983), 'datetime.timedelta', 'timedelta', ([], {'minutes': '(5)'}), '(minutes=5)\n', (972, 983), False, 'from datetime import datetime, timedelta\n'), ((2142, 2214),...
#!/usr/bin/env python3.8 """ export_blueprints.py Connect to a Nutanix Prism Central instance, grab all Calm blueprints and export them to JSON files. You would need to *heavily* modify this script for use in a production environment so that it contains appropriate error-checking and exception handling....
[ "argparse.ArgumentParser", "getpass.getpass", "apiclient.ApiClient", "urllib3.disable_warnings", "time.localtime", "json.dump" ]
[((1042, 1121), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {'description': '"""Export all Calm blueprints to JSON files"""'}), "(description='Export all Calm blueprints to JSON files')\n", (1065, 1121), False, 'import argparse\n'), ((1997, 2064), 'urllib3.disable_warnings', 'urllib3.disable_warnings', (...
#* pyx509 - Python library for parsing X.509 #* Copyright (C) 2009-2010 CZ.NIC, z.s.p.o. (http://www.nic.cz) #* #* This library is free software; you can redistribute it and/or #* modify it under the terms of the GNU Library General Public #* License as published by the Free Software Foundation; either...
[ "pyasn1.type.univ.OctetString", "pyasn1.type.tag.Tag", "pyasn1.type.useful.UTCTime", "pyasn1.type.univ.Boolean", "pyasn1.type.useful.GeneralizedTime", "pyasn1.type.univ.ObjectIdentifier" ]
[((1270, 1293), 'pyasn1.type.univ.ObjectIdentifier', 'univ.ObjectIdentifier', ([], {}), '()\n', (1291, 1293), False, 'from pyasn1.type import tag, namedtype, univ, useful\n'), ((1345, 1366), 'pyasn1.type.univ.Boolean', 'univ.Boolean', (['"""False"""'], {}), "('False')\n", (1357, 1366), False, 'from pyasn1.type import t...
from math import isqrt is_square = lambda n: isqrt(n) ** 2 == n if n >= 0 else False def is_square_soln(n): pass print(is_square(-1))
[ "math.isqrt" ]
[((46, 54), 'math.isqrt', 'isqrt', (['n'], {}), '(n)\n', (51, 54), False, 'from math import isqrt\n')]
# -*- coding: utf-8 -*- """ TencentBlueKing is pleased to support the open source community by making 蓝鲸智云-用户管理(Bk-User) available. Copyright (C) 2017-2021 THL A29 Limited, a Tencent company. All rights reserved. Licensed under the MIT License (the "License"); you may not use this file except in compliance with the Lic...
[ "logging.getLogger", "rest_framework.serializers.DateTimeField", "rest_framework.serializers.BooleanField", "bkuser_core.categories.loader.register_plugin", "bkuser_core.common.models.is_obj_needed_update", "bkuser_core.user_settings.models.Setting.objects.get_or_create", "yaml.safe_load", "bkuser_cor...
[((1411, 1438), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (1428, 1438), False, 'import logging\n'), ((1566, 1592), 'rest_framework.serializers.BooleanField', 'serializers.BooleanField', ([], {}), '()\n', (1590, 1592), False, 'from rest_framework import serializers\n'), ((1602, 1629),...
from django import template import mistune register = template.Library() @register.filter def markdown(value): markdown = mistune.Markdown() return markdown(value)
[ "mistune.Markdown", "django.template.Library" ]
[((55, 73), 'django.template.Library', 'template.Library', ([], {}), '()\n', (71, 73), False, 'from django import template\n'), ((129, 147), 'mistune.Markdown', 'mistune.Markdown', ([], {}), '()\n', (145, 147), False, 'import mistune\n')]
import json from finite.storage import new_uuid class Unimplemented(Exception): pass class RoleFail(Exception): pass SUPERUSER = '*' """ role used to bypass all permission checks """ ROOT_UUID = '00000000-0000-0000-0000-000000000000' """ parent UUID used to initialize a stream """ DEFAULT_SCHEMA = 'base...
[ "json.dumps" ]
[((1518, 1547), 'json.dumps', 'json.dumps', (["kwargs['payload']"], {}), "(kwargs['payload'])\n", (1528, 1547), False, 'import json\n')]
# This file is part of Indico. # Copyright (C) 2002 - 2019 CERN # # Indico is free software; you can redistribute it and/or # modify it under the terms of the MIT License; see the # LICENSE file for more details. from datetime import datetime, timedelta from uuid import uuid4 import pytest from flask import session f...
[ "mock.MagicMock", "indico.modules.oauth.provider.load_token", "datetime.datetime.utcnow", "indico.modules.oauth.provider.load_client", "uuid.uuid4", "indico.modules.oauth.models.applications.OAuthApplication.find", "pytest.mark.parametrize", "pytest.raises", "pytest.mark.usefixtures", "datetime.ti...
[((1787, 1829), 'pytest.mark.usefixtures', 'pytest.mark.usefixtures', (['"""request_context"""'], {}), "('request_context')\n", (1810, 1829), False, 'import pytest\n'), ((2504, 2546), 'pytest.mark.usefixtures', 'pytest.mark.usefixtures', (['"""request_context"""'], {}), "('request_context')\n", (2527, 2546), False, 'im...
#!/usr/bin/env python3 # Copyright (c) Facebook, Inc. and its affiliates. # # This source code is licensed under the MIT license found in the # LICENSE file in the root directory of this source tree. import unittest from unittest.mock import MagicMock, patch from fbpcs.service.log_cloudwatch import CloudWatchLogServi...
[ "unittest.mock.MagicMock", "fbpcs.service.log_cloudwatch.CloudWatchLogService", "unittest.mock.patch" ]
[((460, 511), 'unittest.mock.patch', 'patch', (['"""fbpcs.gateway.cloudwatch.CloudWatchGateway"""'], {}), "('fbpcs.gateway.cloudwatch.CloudWatchGateway')\n", (465, 511), False, 'from unittest.mock import MagicMock, patch\n'), ((583, 622), 'fbpcs.service.log_cloudwatch.CloudWatchLogService', 'CloudWatchLogService', (['L...
import adsk.core import adsk.fusion import os from ...lib import fusion360utils as futil from ... import config import math app = adsk.core.Application.get() ui = app.userInterface # TODO *** コマンドのID情報を指定します。 *** CMD_ID = f'{config.COMPANY_NAME}_{config.ADDIN_NAME}_Meteor' CMD_NAME = 'メテオ' CMD_Descript...
[ "os.path.abspath" ]
[((903, 928), 'os.path.abspath', 'os.path.abspath', (['__file__'], {}), '(__file__)\n', (918, 928), False, 'import os\n')]
import re import time from datetime import datetime from enum import Enum from fishbase.fish_logger import logger from .elk_connector import Es from ..base.tp_base import TpBase, TestStatus, Conf, VerticalContext from ..base.tp_base import get_params_dict # LogTestPoint class LogESTestPoint(TpBase): # 类的初始化过程 ...
[ "datetime.datetime.utcfromtimestamp", "datetime.datetime.utcnow" ]
[((1053, 1091), 'datetime.datetime.utcfromtimestamp', 'datetime.utcfromtimestamp', (['time_struct'], {}), '(time_struct)\n', (1078, 1091), False, 'from datetime import datetime\n'), ((1637, 1654), 'datetime.datetime.utcnow', 'datetime.utcnow', ([], {}), '()\n', (1652, 1654), False, 'from datetime import datetime\n')]
import math from typing import List import numpy as np import torch a = torch.randn(4, 3,2) print(a) print(torch.argmax(a, -1))
[ "torch.randn", "torch.argmax" ]
[((73, 93), 'torch.randn', 'torch.randn', (['(4)', '(3)', '(2)'], {}), '(4, 3, 2)\n', (84, 93), False, 'import torch\n'), ((108, 127), 'torch.argmax', 'torch.argmax', (['a', '(-1)'], {}), '(a, -1)\n', (120, 127), False, 'import torch\n')]
import sys import json import boto3 from botocore.exceptions import ClientError from . import config def status_instance(instance_id, dry_run=False): ec2 = boto3.client('ec2', region_name='ap-northeast-1', aws_access_key_id=config.AWS_ACCESS_KEY_ID, ...
[ "boto3.client" ]
[((170, 313), 'boto3.client', 'boto3.client', (['"""ec2"""'], {'region_name': '"""ap-northeast-1"""', 'aws_access_key_id': 'config.AWS_ACCESS_KEY_ID', 'aws_secret_access_key': 'config.AWS_SECRET_KEY'}), "('ec2', region_name='ap-northeast-1', aws_access_key_id=config.\n AWS_ACCESS_KEY_ID, aws_secret_access_key=config...
from django import template register = template.Library() @register.filter def has_group(user, name): return user.groups.filter(name=name).exists()
[ "django.template.Library" ]
[((40, 58), 'django.template.Library', 'template.Library', ([], {}), '()\n', (56, 58), False, 'from django import template\n')]
import importlib.util import inspect from skaio import log from skaio.core.publisher import Publisher from skaio.core.base.task import BaseTask from skaio.utils.common import get_loop tasks = ['samples.simple_tasks'] class Scheduler: def start(self): publisher = Publisher() loop = get_loop() ...
[ "skaio.utils.common.get_loop", "inspect.getmembers", "skaio.core.publisher.Publisher", "inspect.isclass", "skaio.log.info" ]
[((278, 289), 'skaio.core.publisher.Publisher', 'Publisher', ([], {}), '()\n', (287, 289), False, 'from skaio.core.publisher import Publisher\n'), ((305, 315), 'skaio.utils.common.get_loop', 'get_loop', ([], {}), '()\n', (313, 315), False, 'from skaio.utils.common import get_loop\n'), ((629, 650), 'inspect.getmembers',...
# Copyright (c) 2020 <NAME> # # This software is released under the MIT License. # https://opensource.org/licenses/MIT from bq_test_kit.interpolators.shell_interpolator import ShellInterpolator def test_interpolate(): si = ShellInterpolator({"LOCAL_KEY": "VALUE"}) result = si.interpolate("Local key has value...
[ "bq_test_kit.interpolators.shell_interpolator.ShellInterpolator" ]
[((230, 271), 'bq_test_kit.interpolators.shell_interpolator.ShellInterpolator', 'ShellInterpolator', (["{'LOCAL_KEY': 'VALUE'}"], {}), "({'LOCAL_KEY': 'VALUE'})\n", (247, 271), False, 'from bq_test_kit.interpolators.shell_interpolator import ShellInterpolator\n')]
# Copyright (c) Microsoft Corporation. All rights reserved. # Licensed under the MIT License. import os, uuid, sys, pickle, shutil, io, logging from azure.storage.filedatalake import DataLakeServiceClient from azure.core._match_conditions import MatchConditions from azure.storage.filedatalake._models import ContentSet...
[ "shutil.make_archive" ]
[((1024, 1127), 'shutil.make_archive', 'shutil.make_archive', (['"""retailai_recommendation_model"""', '"""zip"""', '"""model\\\\retailai_recommendation_model"""'], {}), "('retailai_recommendation_model', 'zip',\n 'model\\\\retailai_recommendation_model')\n", (1043, 1127), False, 'import os, uuid, sys, pickle, shuti...