code stringlengths 22 1.05M | apis listlengths 1 3.31k | extract_api stringlengths 75 3.25M |
|---|---|---|
# Code generated by `typeddictgen`. DO NOT EDIT.
"""V2beta2MetricStatusDict generated type."""
from typing import TypedDict
from kubernetes_typed.client import V2beta2ExternalMetricStatusDict, V2beta2ObjectMetricStatusDict, V2beta2PodsMetricStatusDict, V2beta2ResourceMetricStatusDict
V2beta2MetricStatusDict = TypedDi... | [
"typing.TypedDict"
] | [((313, 557), 'typing.TypedDict', 'TypedDict', (['"""V2beta2MetricStatusDict"""', "{'external': V2beta2ExternalMetricStatusDict, 'object':\n V2beta2ObjectMetricStatusDict, 'pods': V2beta2PodsMetricStatusDict,\n 'resource': V2beta2ResourceMetricStatusDict, 'type': str}"], {'total': '(False)'}), "('V2beta2MetricSta... |
#!/usr/bin/env python
# Simulate the /robot_driver so that interfaces to it can operate the same on
# the real robot and in simulation
from __future__ import print_function, division
from threading import Lock
import rospy
import diagnostic_updater
from diagnostic_msgs.msg import DiagnosticStatus
from fetch_driver_... | [
"fetch_driver_msgs.msg.ChargerState",
"rospy.Subscriber",
"rospy.is_shutdown",
"rospy.init_node",
"threading.Lock",
"rospy.Service",
"diagnostic_updater.Updater",
"rospy.Time.now",
"power_msgs.msg.BreakerState",
"fetch_driver_msgs.msg.JointState",
"rospy.Rate",
"rospy.Publisher",
"std_srvs.s... | [((8125, 8156), 'rospy.init_node', 'rospy.init_node', (['"""robot_driver"""'], {}), "('robot_driver')\n", (8140, 8156), False, 'import rospy\n'), ((1757, 1823), 'power_msgs.msg.BreakerState', 'BreakerState', ([], {'name': '"""arm_breaker"""', 'state': 'BreakerState.STATE_ENABLED'}), "(name='arm_breaker', state=BreakerS... |
# Configuration file for the Sphinx documentation builder.
#
# This file only contains a selection of the most common options. For a full
# list see the documentation:
# https://www.sphinx-doc.org/en/master/usage/configuration.html
# -- Path setup --------------------------------------------------------------
# If ex... | [
"sphinx.ext.apidoc.main",
"pathlib.Path"
] | [((2908, 2925), 'sphinx.ext.apidoc.main', 'apidoc.main', (['argv'], {}), '(argv)\n', (2919, 2925), False, 'from sphinx.ext import apidoc\n'), ((711, 725), 'pathlib.Path', 'Path', (['__file__'], {}), '(__file__)\n', (715, 725), False, 'from pathlib import Path\n')] |
from random import randint
from time import sleep
print ('=' * 18)
print ('\033[1mSTONE PAPER AND SCISSORS\033[m')
print ('=' * 18)
print ('I already chose mine now missing you')
sleep (1)
computer = randint (1, 3)
player = int(input('\033[1mChoose between\033[m \033[1;33m1) Stone 2) Paper and 3) Scissors\033[m : '))... | [
"random.randint",
"time.sleep"
] | [((180, 188), 'time.sleep', 'sleep', (['(1)'], {}), '(1)\n', (185, 188), False, 'from time import sleep\n'), ((201, 214), 'random.randint', 'randint', (['(1)', '(3)'], {}), '(1, 3)\n', (208, 214), False, 'from random import randint\n'), ((321, 329), 'time.sleep', 'sleep', (['(1)'], {}), '(1)\n', (326, 329), False, 'fro... |
# @l2g 1673 python3
# [1673] Find the Most Competitive Subsequence
# Difficulty: Medium
# https://leetcode.com/problems/find-the-most-competitive-subsequence
#
# Given an integer array nums and a positive integer k,
# return the most competitive subsequence of nums of size k.
# An array's subsequence is a resulting seq... | [
"os.path.join"
] | [((1609, 1646), 'os.path.join', 'os.path.join', (['"""tests"""', '"""test_1673.py"""'], {}), "('tests', 'test_1673.py')\n", (1621, 1646), False, 'import os\n')] |
#!/usr/bin/env python
import os
import unittest
import sys
from test_common import TestCommon
sys.path.append(os.path.join(os.path.dirname(__file__), '..', 'library'))
from fastly_service import FastlyConfiguration
class TestFastlyLoggingS3(TestCommon):
@TestCommon.vcr.use_cassette()
def test_fastly_s3s(sel... | [
"unittest.main",
"os.path.dirname",
"test_common.TestCommon.vcr.use_cassette",
"fastly_service.FastlyConfiguration"
] | [((263, 292), 'test_common.TestCommon.vcr.use_cassette', 'TestCommon.vcr.use_cassette', ([], {}), '()\n', (290, 292), False, 'from test_common import TestCommon\n'), ((2124, 2153), 'test_common.TestCommon.vcr.use_cassette', 'TestCommon.vcr.use_cassette', ([], {}), '()\n', (2151, 2153), False, 'from test_common import T... |
#! /usr/bin/env python3
""" PrepareTreeData.py
Prepares the tree data a bit for inclusion into a database.
"""
import pandas as pd
import numpy as np
""" Prepare the tree data for inclusion into a database.
Args:
inname: Input file
outname: Output file
"""
def PrepareTreeData(inname='data/street_trees_2015... | [
"pandas.read_csv"
] | [((387, 419), 'pandas.read_csv', 'pd.read_csv', (['inname'], {'index_col': '(0)'}), '(inname, index_col=0)\n', (398, 419), True, 'import pandas as pd\n')] |
#encoding: UTF-8
# Copyright (C) 2016 <NAME>
# This file is distributed under the terms of the # MIT License.
# See the file `License' in the root directory of the present distribution.
"""
An earlier and now oblosete implementation of functions for computing the
thermal expansion tensor as a function
of temperatur... | [
"math.pow",
"numpy.array",
"math.exp",
"numpy.zeros"
] | [((1910, 1922), 'math.exp', 'math.exp', (['(-x)'], {}), '(-x)\n', (1918, 1922), False, 'import math\n'), ((1953, 1967), 'math.pow', 'math.pow', (['x', '(2)'], {}), '(x, 2)\n', (1961, 1967), False, 'import math\n'), ((2399, 2411), 'math.exp', 'math.exp', (['(-x)'], {}), '(-x)\n', (2407, 2411), False, 'import math\n'), (... |
from __future__ import division
import numpy as np
from sklearn.utils import shuffle
from sklearn.metrics import *
"""
Module with different fitness functions implemented to be used by the CRO algorithm.
The functions' only argument must be an individual (coral) and return its fitness, a number.
The fitness might re... | [
"sklearn.utils.shuffle",
"numpy.multiply"
] | [((1584, 1623), 'sklearn.utils.shuffle', 'shuffle', (['X', 'y'], {'random_state': 'random_seed'}), '(X, y, random_state=random_seed)\n', (1591, 1623), False, 'from sklearn.utils import shuffle\n'), ((1633, 1655), 'numpy.multiply', 'np.multiply', (['Xs', 'coral'], {}), '(Xs, coral)\n', (1644, 1655), True, 'import numpy ... |
from abc import ABC, abstractmethod
from typing import Tuple, Iterable, Any
from pygame.rect import Rect
class Shape(ABC):
"""Abstract shape interface."""
@abstractmethod
def shape(self) -> Any:
pass
@abstractmethod
def top_left(self) -> Tuple:
pass
@abstractmethod
def t... | [
"pygame.rect.Rect"
] | [((786, 800), 'pygame.rect.Rect', 'Rect', (['position'], {}), '(position)\n', (790, 800), False, 'from pygame.rect import Rect\n')] |
import numpy as np
from ..local_interpolation import ThirdOrderHermitePolynomialInterpolation
from .runge_kutta import AbstractESDIRK, ButcherTableau
γ = 0.26
a21 = γ
a31 = 0.13
a32 = 0.84033320996790809
a41 = 0.22371961478320505
a42 = 0.47675532319799699
a43 = -0.06470895363112615
a51 = 0.16648564323248321
a52 = 0.... | [
"numpy.array"
] | [((1729, 1760), 'numpy.array', 'np.array', (['[0, γ, γ, γ, γ, γ, γ]'], {}), '([0, γ, γ, γ, γ, γ, γ])\n', (1737, 1760), True, 'import numpy as np\n'), ((2016, 2059), 'numpy.array', 'np.array', (['[a71, a72, a73, a74, a75, a76, γ]'], {}), '([a71, a72, a73, a74, a75, a76, γ])\n', (2024, 2059), True, 'import numpy as np\n'... |
#MenuTitle: Remove Zero Deltas in Selected Glyphs
# -*- coding: utf-8 -*-
from __future__ import division, print_function, unicode_literals
__doc__="""
Goes through all layers of each selected glyph, and deletes all TT Delta Hints with an offset of zero. Detailed Report in Macro Window.
"""
def process( Layer ):
try:... | [
"traceback.format_exc"
] | [((1433, 1455), 'traceback.format_exc', 'traceback.format_exc', ([], {}), '()\n', (1453, 1455), False, 'import traceback\n')] |
#!/usr/bin/env python3
import re
import sys
from glob import glob
from subprocess import run
def main(args):
assert len(args) >= 1
from_image = args.pop(0)
optional = [x for x in map(str.strip, args) if x]
optional_used = set()
with open("Dockerfile", "w") as fout:
print(f"from {from_ima... | [
"subprocess.run",
"glob.glob",
"re.search"
] | [((1077, 1133), 'subprocess.run', 'run', (["['docker', 'build', '-t', our_tag, '.']"], {'check': '(True)'}), "(['docker', 'build', '-t', our_tag, '.'], check=True)\n", (1080, 1133), False, 'from subprocess import run\n'), ((366, 386), 'glob.glob', 'glob', (['"""*.Dockerfile"""'], {}), "('*.Dockerfile')\n", (370, 386), ... |
import torch
import torch.nn as nn
import torch.nn.functional as F
import torchvision
class FPAv2(nn.Module):
def __init__(self, input_dim, output_dim):
super(FPAv2, self).__init__()
self.glob = nn.Sequential(nn.AdaptiveAvgPool2d(1),
nn.Conv2d(input_dim, output_di... | [
"torch.nn.functional.upsample",
"torch.nn.BatchNorm2d",
"torch.nn.Sigmoid",
"torch.nn.ReLU",
"torch.nn.Dropout",
"torch.nn.functional.adaptive_avg_pool2d",
"torch.nn.Sequential",
"torch.nn.Dropout2d",
"torch.nn.Conv2d",
"torchvision.models.resnet34",
"torch.nn.MaxPool2d",
"torch.nn.AdaptiveAvg... | [((1596, 1668), 'torch.nn.functional.upsample', 'F.upsample', (['x_glob'], {'scale_factor': '(16)', 'mode': '"""bilinear"""', 'align_corners': '(True)'}), "(x_glob, scale_factor=16, mode='bilinear', align_corners=True)\n", (1606, 1668), True, 'import torch.nn.functional as F\n'), ((1871, 1938), 'torch.nn.functional.ups... |
from django.utils.translation import ugettext_lazy as _
SHIPPING_STANDARD = 'Standard'
SHIPPING_EXPEDITED = 'Expedited'
SHIPPING_PRIORITY = 'Priority'
SHIPPING_SPEED_CATEGORIES = (
(SHIPPING_STANDARD, _("Standard")),
(SHIPPING_EXPEDITED, _("Expedited")),
(SHIPPING_PRIORITY, _("Priority")),
)
METHOD_CONS... | [
"django.utils.translation.ugettext_lazy"
] | [((208, 221), 'django.utils.translation.ugettext_lazy', '_', (['"""Standard"""'], {}), "('Standard')\n", (209, 221), True, 'from django.utils.translation import ugettext_lazy as _\n'), ((249, 263), 'django.utils.translation.ugettext_lazy', '_', (['"""Expedited"""'], {}), "('Expedited')\n", (250, 263), True, 'from djang... |
import os
from tkinter import *
from tkinter import filedialog
global archivo
archivo= []
cadena2 = []
archivo = []
global cadena_inicial
cadena_inicial = []
global lista_numeros
lista_numeros = []
global intlist
intlist= []
global numeros1
numeros1=[]
global lista_buscar2
lista_buscar2 =... | [
"os.startfile",
"tkinter.filedialog.askopenfilename"
] | [((1844, 1872), 'tkinter.filedialog.askopenfilename', 'filedialog.askopenfilename', ([], {}), '()\n', (1870, 1872), False, 'from tkinter import filedialog\n'), ((14254, 14283), 'os.startfile', 'os.startfile', (['"""reporte0.html"""'], {}), "('reporte0.html')\n", (14266, 14283), False, 'import os\n')] |
from pyexocross.hitran.hitran import HITRANLinelist
from pyexocross.pyexocross import PyExocross
from pyexocross.exomol.exomolbroads import ExomolBroadener
import numpy as np
from pyexocross.util import create_grid_res, convert_to_wavenumber
from pyexocross.writer.hdf5writer import HDF5Writer
import matplotlib.pyplot a... | [
"matplotlib.pyplot.show",
"pyexocross.exomol.exomolbroads.ExomolBroadener",
"pyexocross.hitran.hitran.HITRANLinelist",
"matplotlib.pyplot.ylabel",
"matplotlib.pyplot.xlabel",
"matplotlib.pyplot.plot",
"matplotlib.pyplot.figure",
"pyexocross.util.create_grid_res",
"matplotlib.pyplot.yscale",
"matpl... | [((475, 549), 'pyexocross.hitran.hitran.HITRANLinelist', 'HITRANLinelist', (['"""/Users/ahmed/Documents/molecular_data/HITRAN/CH4/CH4.par"""'], {}), "('/Users/ahmed/Documents/molecular_data/HITRAN/CH4/CH4.par')\n", (489, 549), False, 'from pyexocross.hitran.hitran import HITRANLinelist\n'), ((644, 776), 'pyexocross.exo... |
from bigflow.workflow import Workflow, hourly_start_time
from datetime import datetime
from datetime import timedelta
class HourlyJob:
def __init__(self):
self.id = 'hourly_job'
def run(self, runtime):
print(f'I should process data with timestamps from: {runtime} '
f'to {dateti... | [
"datetime.datetime.strptime",
"datetime.timedelta"
] | [((314, 361), 'datetime.datetime.strptime', 'datetime.strptime', (['runtime', '"""%Y-%m-%d %H:%M:%S"""'], {}), "(runtime, '%Y-%m-%d %H:%M:%S')\n", (331, 361), False, 'from datetime import datetime\n'), ((364, 397), 'datetime.timedelta', 'timedelta', ([], {'minutes': '(59)', 'seconds': '(59)'}), '(minutes=59, seconds=59... |
#!/usr/local/bin/env python3.7
# -*- coding: utf-8; py-indent-offset:4 -*-
###############################################################################
#
# The MIT License (MIT)
# Copyright (c) 2020 <NAME>
#
# Permission is hereby granted, free of charge, to any person obtaining a copy
# of this software and ass... | [
"charting.Charting",
"manip_data.ManipData.pd_tolist"
] | [((1759, 1837), 'charting.Charting', 'cht.Charting', (['self.series', 'self.date_name', 'self.default_data'], {}), '(self.series, self.date_name, self.default_data, **self.indicator)\n', (1771, 1837), True, 'import charting as cht\n'), ((2188, 2235), 'manip_data.ManipData.pd_tolist', 'md.pd_tolist', (['self.trades_trac... |
# Generated by Django 3.1.2 on 2021-03-14 11:56
from django.db import migrations, models
class Migration(migrations.Migration):
dependencies = [
('cart', '0022_delete_sendpassword'),
]
operations = [
migrations.AlterField(
model_name='orderbeta',
name='finish_pri... | [
"django.db.models.PositiveIntegerField"
] | [((343, 381), 'django.db.models.PositiveIntegerField', 'models.PositiveIntegerField', ([], {'default': '(0)'}), '(default=0)\n', (370, 381), False, 'from django.db import migrations, models\n')] |
import logging.config
import os
import yaml
__location__ = os.path.realpath(
os.path.join(os.getcwd(), os.path.dirname(__file__)))
def read_logging_config():
with open(os.path.join(__location__, 'logging.cfg'), 'r') as stream:
try:
logging_config = yaml.safe_load(stream)
logg... | [
"os.path.dirname",
"os.path.join",
"yaml.safe_load",
"os.getcwd"
] | [((96, 107), 'os.getcwd', 'os.getcwd', ([], {}), '()\n', (105, 107), False, 'import os\n'), ((109, 134), 'os.path.dirname', 'os.path.dirname', (['__file__'], {}), '(__file__)\n', (124, 134), False, 'import os\n'), ((180, 221), 'os.path.join', 'os.path.join', (['__location__', '"""logging.cfg"""'], {}), "(__location__, ... |
import requests
import logging
from bbot.core import ChatbotEngine, BBotException, BBotCore, BBotExtensionException
from engines.dotflow2.chatbot_engine import DotFlow2LoggerAdapter
class DotFlow2MSCSSentimentAnalysis():
"""ChatScript DotFlow2 function"""
def __init__(self, config: dict, dotbot: dict) -> No... | [
"logging.getLogger",
"bbot.core.BBotExtensionException",
"requests.post"
] | [((1984, 2130), 'requests.post', 'requests.post', (['f"""https://{self.azure_location}.api.cognitive.microsoft.com/text/analytics/v2.0/sentiment"""'], {'json': 'payload', 'headers': 'headers'}), "(\n f'https://{self.azure_location}.api.cognitive.microsoft.com/text/analytics/v2.0/sentiment'\n , json=payload, heade... |
import os
import pandas as pd
from poor_trader import config
from poor_trader import trading
from poor_trader import systems
class CombinedIndicators(trading.TradingSystem):
def __init__(self, portfolio, systems_method_list, name='CombinedIndicators'):
super(CombinedIndicators, self).__init__(name=name)
... | [
"pandas.read_pickle",
"os.path.exists",
"poor_trader.systems.run_slsma",
"poor_trader.systems.run_dcsma",
"poor_trader.systems.run_atr_channel_breakout",
"pandas.DataFrame"
] | [((561, 575), 'pandas.DataFrame', 'pd.DataFrame', ([], {}), '()\n', (573, 575), True, 'import pandas as pd\n'), ((650, 676), 'os.path.exists', 'os.path.exists', (['self.fpath'], {}), '(self.fpath)\n', (664, 676), False, 'import os\n'), ((1502, 1516), 'pandas.DataFrame', 'pd.DataFrame', ([], {}), '()\n', (1514, 1516), T... |
# The MIT License (MIT)
#
# Copyright (c) 2018 PyBER
#
# Permission is hereby granted, free of charge, to any person obtaining a copy
# of this software and associated documentation files (the "Software"), to deal
# in the Software without restriction, including without limitation the rights
# to use, copy, modify, mer... | [
"lib.pyqtgraph.pyqtgraph.Qt.QtGui.QFrame",
"lib.pyqtgraph.pyqtgraph.Qt.QtGui.QFormLayout",
"lib.pyqtgraph.pyqtgraph.Qt.QtGui.QLineEdit",
"os.startfile",
"lib.pyqtgraph.pyqtgraph.Qt.QtGui.QScrollArea",
"lib.pyqtgraph.pyqtgraph.Qt.QtGui.QFileSystemModel",
"lib.pyqtgraph.pyqtgraph.Qt.QtGui.QWidget",
"os.... | [((13455, 13479), 'lib.pyqtgraph.pyqtgraph.Qt.QtGui.QFileSystemModel', 'QtGui.QFileSystemModel', ([], {}), '()\n', (13477, 13479), False, 'from lib.pyqtgraph.pyqtgraph.Qt import QtCore, QtGui, QtWidgets\n'), ((3260, 3271), 'time.time', 'time.time', ([], {}), '()\n', (3269, 3271), False, 'import time\n'), ((3362, 3389),... |
# -*- coding: utf-8 -*-
import os
import json
import torch
BASE_DIR = os.path.abspath(os.path.dirname(__file__))
class Config:
def __init__(self, json_file):
self.config = json.loads(open(json_file).read())
self.device = torch.device("cuda:1" if torch.cuda.is_available() else "cpu")
self... | [
"os.path.dirname",
"torch.cuda.is_available",
"os.path.join",
"torch.device"
] | [((89, 114), 'os.path.dirname', 'os.path.dirname', (['__file__'], {}), '(__file__)\n', (104, 114), False, 'import os\n'), ((334, 353), 'torch.device', 'torch.device', (['"""cpu"""'], {}), "('cpu')\n", (346, 353), False, 'import torch\n'), ((832, 904), 'os.path.join', 'os.path.join', (['BASE_DIR', '"""dataset"""', "self... |
import unittest
import torch
import numpy as np
from spectralgp.samplers import MeanEllipticalSlice
class TestMeanEllipticalSlice(unittest.TestCase):
def test_m_ess(self, nsamples=10000):
pmean = torch.zeros(2)
pmean[0] = -2.
prior_dist = torch.distributions.MultivariateNormal(pmean, covar... | [
"numpy.mean",
"spectralgp.samplers.MeanEllipticalSlice",
"torch.eye",
"unittest.main",
"numpy.cov",
"torch.zeros",
"torch.inverse"
] | [((2193, 2208), 'unittest.main', 'unittest.main', ([], {}), '()\n', (2206, 2208), False, 'import unittest\n'), ((210, 224), 'torch.zeros', 'torch.zeros', (['(2)'], {}), '(2)\n', (221, 224), False, 'import torch\n'), ((372, 386), 'torch.zeros', 'torch.zeros', (['(2)'], {}), '(2)\n', (383, 386), False, 'import torch\n'),... |
"""
The input widgets generally allow entering arbitrary information into
a text field or similar.
"""
from __future__ import absolute_import, division, unicode_literals
import ast
from base64 import b64decode, b64encode
from datetime import datetime
from six import string_types
import param
from bokeh.models.widge... | [
"param.ClassSelector",
"param.Number",
"param.Dict",
"datetime.datetime.strptime",
"param.Date",
"param.Boolean",
"datetime.datetime.strftime",
"base64.b64decode",
"base64.b64encode",
"param.Parameter",
"ast.literal_eval",
"param.String",
"param.Color"
] | [((647, 688), 'param.String', 'param.String', ([], {'default': '""""""', 'allow_None': '(True)'}), "(default='', allow_None=True)\n", (659, 688), False, 'import param\n'), ((708, 732), 'param.String', 'param.String', ([], {'default': '""""""'}), "(default='')\n", (720, 732), False, 'import param\n'), ((810, 836), 'para... |
# -*- coding=utf-8 -*-
import urllib2
import json
import logger
import traceback
def send(apiUrl,data,method=None):
logger.debug("调用内部系统[%s],data[%r]",apiUrl,data)
try:
data_json = json.dumps(data)
headers = {'Content-Type': 'application/json'} # 设置数据为json格式,很重要
request = urllib2.Reques... | [
"urllib2.urlopen",
"logger.init_4_debug",
"json.dumps",
"urllib2.Request",
"logger.exception",
"logger.debug"
] | [((121, 170), 'logger.debug', 'logger.debug', (['"""调用内部系统[%s],data[%r]"""', 'apiUrl', 'data'], {}), "('调用内部系统[%s],data[%r]', apiUrl, data)\n", (133, 170), False, 'import logger\n'), ((815, 836), 'logger.init_4_debug', 'logger.init_4_debug', ([], {}), '()\n', (834, 836), False, 'import logger\n'), ((198, 214), 'json.du... |
from sklearn.datasets import load_wine
from sklearn.neighbors import KNeighborsClassifier
from sklearn.ensemble import RandomForestClassifier
from sklearn.linear_model import LogisticRegression
wine = load_wine(as_frame=True)
X, y = wine.data, wine.target
X_train, X_test, y_train, y_test = train_test_split(
X, y... | [
"sklearn.neighbors.KNeighborsClassifier",
"sklearn.datasets.load_wine",
"sklearn.ensemble.RandomForestClassifier",
"sklearn.linear_model.LogisticRegression"
] | [((202, 226), 'sklearn.datasets.load_wine', 'load_wine', ([], {'as_frame': '(True)'}), '(as_frame=True)\n', (211, 226), False, 'from sklearn.datasets import load_wine\n'), ((358, 380), 'sklearn.neighbors.KNeighborsClassifier', 'KNeighborsClassifier', ([], {}), '()\n', (378, 380), False, 'from sklearn.neighbors import K... |
# Generated by Django 3.0.3 on 2020-02-25 18:50
from django.conf import settings
from django.db import migrations, models
import django.db.models.deletion
import uuid
class Migration(migrations.Migration):
initial = True
dependencies = [
migrations.swappable_dependency(settings.AUTH_USER_MODEL),
... | [
"django.db.models.UUIDField",
"django.db.models.FloatField",
"django.db.models.TextField",
"django.db.models.ForeignKey",
"django.db.models.IntegerField",
"django.db.models.DateTimeField",
"django.db.models.BooleanField",
"django.db.models.AutoField",
"django.db.models.URLField",
"django.db.migrat... | [((259, 316), 'django.db.migrations.swappable_dependency', 'migrations.swappable_dependency', (['settings.AUTH_USER_MODEL'], {}), '(settings.AUTH_USER_MODEL)\n', (290, 316), False, 'from django.db import migrations, models\n'), ((449, 542), 'django.db.models.AutoField', 'models.AutoField', ([], {'auto_created': '(True)... |
"""
Summary:
- Command-line Interface (CLI) Utilities Module
- Python3
Module Functions:
- convert_strtime_datetime:
Convert human-readable datetime string into a datetime object for
conducting time operations.
- convert_timedelta:
Convert a datetime duration object into human-r... | [
"logging.getLogger",
"datetime.timedelta",
"datetime.datetime.strptime",
"inspect.stack"
] | [((568, 598), 'logging.getLogger', 'logging.getLogger', (['__version__'], {}), '(__version__)\n', (585, 598), False, 'import logging\n'), ((937, 988), 'datetime.datetime.strptime', 'datetime.datetime.strptime', (['dt', '"""%Y-%m-%dT%H:%M:%S"""'], {}), "(dt, '%Y-%m-%dT%H:%M:%S')\n", (963, 988), False, 'import datetime\n... |
import json
def translate_name(name):
with open("name_translator.json", "r") as file:
data = json.load(file)
return data[name]
| [
"json.load"
] | [((107, 122), 'json.load', 'json.load', (['file'], {}), '(file)\n', (116, 122), False, 'import json\n')] |
#!/usr/bin/env python
import sys
import collections
listas = list(map(str.split, sys.stdin.readlines()))
entrada = [item for sublist in listas for item in sublist] # simplificando as listas, para facilitar
contador = collections.Counter()
palavras = []
for palavra in entrada:
if palavra == "#":
break
palavr... | [
"collections.Counter",
"sys.stdin.readlines"
] | [((219, 240), 'collections.Counter', 'collections.Counter', ([], {}), '()\n', (238, 240), False, 'import collections\n'), ((83, 104), 'sys.stdin.readlines', 'sys.stdin.readlines', ([], {}), '()\n', (102, 104), False, 'import sys\n')] |
import torch
from viz.visualizer import Visualizer
from modules.deepmil import Attention
from msidata.dataset_msi_features_with_patients import PreProcessedMSIFeatureDataset
from testing.logistic_regression import get_precomputed_dataloader
import argparse
from experiment import ex
from utils import post_config_hook... | [
"utils.post_config_hook",
"modules.deepmil.Attention",
"torch.cuda.is_available",
"argparse.Namespace",
"viz.visualizer.Visualizer",
"testing.logistic_regression.get_precomputed_dataloader"
] | [((372, 405), 'argparse.Namespace', 'argparse.Namespace', ([], {}), '(**_run.config)\n', (390, 405), False, 'import argparse\n'), ((417, 445), 'utils.post_config_hook', 'post_config_hook', (['args', '_run'], {}), '(args, _run)\n', (433, 445), False, 'from utils import post_config_hook\n'), ((586, 652), 'testing.logisti... |
import socket
s = socket.socket()
host = input(str("Please enter the host address of the sender: "))
port = 8080
s.connect((host, port))
print(f"[+] CONNECTED TO {host}:{port}")
filename = input(str("Please enter filename for the incoming file: "))
file = open(filename, 'wb')
file_data = s.recv(1024)
file.write(fi... | [
"socket.socket"
] | [((19, 34), 'socket.socket', 'socket.socket', ([], {}), '()\n', (32, 34), False, 'import socket\n')] |
from django.contrib import admin
from .models import Book, Favorite
admin.site.register(Book)
admin.site.register(Favorite) | [
"django.contrib.admin.site.register"
] | [((71, 96), 'django.contrib.admin.site.register', 'admin.site.register', (['Book'], {}), '(Book)\n', (90, 96), False, 'from django.contrib import admin\n'), ((97, 126), 'django.contrib.admin.site.register', 'admin.site.register', (['Favorite'], {}), '(Favorite)\n', (116, 126), False, 'from django.contrib import admin\n... |
from django.core import exceptions
from django.http import FileResponse
from django.utils.text import format_lazy
from django.utils.translation import gettext_lazy as _
from rest_framework import status
from rest_framework.decorators import action
from rest_framework.exceptions import ValidationError
from rest_framewor... | [
"applications.models.SummerVoucher.objects.all",
"applications.api.v1.serializers.AttachmentSerializer",
"django.utils.translation.gettext_lazy",
"rest_framework.decorators.action",
"applications.models.Application.objects.all",
"rest_framework.response.Response",
"django.http.FileResponse",
"rest_fra... | [((1161, 1186), 'applications.models.Application.objects.all', 'Application.objects.all', ([], {}), '()\n', (1184, 1186), False, 'from applications.models import Application, SummerVoucher\n'), ((2943, 2970), 'applications.models.SummerVoucher.objects.all', 'SummerVoucher.objects.all', ([], {}), '()\n', (2968, 2970), F... |
import json
import requests
import time
from discord_webhook import DiscordWebhook, DiscordEmbed
webhook_url = 'https://discordapp.com/api/webhooks/672159508675690497/4UtaClAc7rKMJsEvbR4iYf-Razv4M3ZWtkYDOxBzLfiDzJhI7RSFpoLn6iijBiRcaNOR'
webhook = DiscordWebhook(webhook_url)
pid = '508214-660'
headers = {
'Connecti... | [
"discord_webhook.DiscordEmbed",
"requests.post",
"discord_webhook.DiscordWebhook",
"time.sleep"
] | [((248, 275), 'discord_webhook.DiscordWebhook', 'DiscordWebhook', (['webhook_url'], {}), '(webhook_url)\n', (262, 275), False, 'from discord_webhook import DiscordWebhook, DiscordEmbed\n'), ((1239, 1378), 'requests.post', 'requests.post', (['"""https://2fwotdvm2o-dsn.algolia.net/1/indexes/product_variants_v2/query"""']... |
"""Unit tests for the pytai application.
License:
MIT License
Permission is hereby granted, free of charge, to any person obtaining a copy
of this software and associated documentation files (the "Software"), to deal
in the Software without restriction, including without limitation the rights
... | [
"unittest.main",
"xml.etree.ElementTree.Element",
"xml.etree.ElementTree.SubElement",
"pathlib.Path"
] | [((3883, 3898), 'unittest.main', 'unittest.main', ([], {}), '()\n', (3896, 3898), False, 'import unittest\n'), ((2221, 2262), 'xml.etree.ElementTree.SubElement', 'ET.SubElement', (['parent_handle', '"""node"""'], {}), "(parent_handle, 'node', **d)\n", (2234, 2262), True, 'import xml.etree.ElementTree as ET\n'), ((2102,... |
# Copyright (c) 2019 Science and Technology Facilities Council
# All rights reserved.
# Modifications made as part of the fparser project are distributed
# under the following license:
# Redistribution and use in source and binary forms, with or without
# modification, are permitted provided that the following condi... | [
"fparser.api.get_reader",
"pytest.raises",
"fparser.two.Fortran2003.Include_Stmt"
] | [((2737, 2753), 'fparser.api.get_reader', 'get_reader', (['line'], {}), '(line)\n', (2747, 2753), False, 'from fparser.api import get_reader\n'), ((2993, 3011), 'fparser.two.Fortran2003.Include_Stmt', 'Include_Stmt', (['line'], {}), '(line)\n', (3005, 3011), False, 'from fparser.two.Fortran2003 import Include_Stmt, Int... |
import os, csv, time, shutil
from bin import cardbank
from bin import builder
def add_build(add_cards):
'''Build card bank by specifically adding new cards.'''
# process new cards to add
if not add_cards:
return
add_card_map = {}
for card in add_cards:
add_card_map[card["inf"]] = ... | [
"csv.DictWriter",
"os.path.exists",
"bin.builder.get",
"bin.builder.session",
"time.sleep",
"shutil.copyfile",
"bin.builder.build",
"bin.cardbank.read"
] | [((385, 427), 'os.path.exists', 'os.path.exists', (['"""bank/card-bank-built.csv"""'], {}), "('bank/card-bank-built.csv')\n", (399, 427), False, 'import os, csv, time, shutil\n'), ((560, 577), 'bin.builder.session', 'builder.session', ([], {}), '()\n', (575, 577), False, 'from bin import builder\n'), ((1739, 1799), 'bi... |
from mongoengine import StringField, EmailField, BooleanField
from flask.ext.login import UserMixin
import requests
import json
from mongoengine import Document
from social.apps.flask_app.me.models import FlaskStorage
class User(Document, UserMixin):
username = StringField(max_length=200)
password = StringFi... | [
"json.loads",
"social.apps.flask_app.me.models.FlaskStorage.user.get_social_auth_for_user",
"requests.get",
"mongoengine.StringField",
"mongoengine.EmailField",
"mongoengine.BooleanField"
] | [((269, 296), 'mongoengine.StringField', 'StringField', ([], {'max_length': '(200)'}), '(max_length=200)\n', (280, 296), False, 'from mongoengine import StringField, EmailField, BooleanField\n'), ((312, 351), 'mongoengine.StringField', 'StringField', ([], {'max_length': '(200)', 'default': '""""""'}), "(max_length=200,... |
"""
********************************************************************************
compas_blender.geometry
********************************************************************************
.. currentmodule:: compas_blender.geometry
Object-oriented convenience wrappers for native Blender geometry.
.. autosummary::
... | [
"bpy.ops.wm.redraw_timer"
] | [((1094, 1153), 'bpy.ops.wm.redraw_timer', 'bpy.ops.wm.redraw_timer', ([], {'type': '"""DRAW_WIN_SWAP"""', 'iterations': '(1)'}), "(type='DRAW_WIN_SWAP', iterations=1)\n", (1117, 1153), False, 'import bpy\n')] |
# Copyright 2020 Google LLC
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, ... | [
"copy.deepcopy"
] | [((1447, 1481), 'copy.deepcopy', 'deepcopy', (['COLUMN_VALIDATION_CONFIG'], {}), '(COLUMN_VALIDATION_CONFIG)\n', (1455, 1481), False, 'from copy import deepcopy\n')] |
import traceback
from typing import Any, Callable, Dict, List, Optional, Set, Tuple
from chainalytic_icon.common import config, util
class ApiBundle(object):
"""
The interface to external consumers/applications
"""
def __init__(self, working_dir: str):
super(ApiBundle, self).__init__()
... | [
"traceback.format_exc",
"chainalytic_icon.common.util.get_child_logger"
] | [((406, 450), 'chainalytic_icon.common.util.get_child_logger', 'util.get_child_logger', (['"""provider.api_bundle"""'], {}), "('provider.api_bundle')\n", (427, 450), False, 'from chainalytic_icon.common import config, util\n'), ((1193, 1215), 'traceback.format_exc', 'traceback.format_exc', ([], {}), '()\n', (1213, 1215... |
import copy
import torch
from utils import helpers
from utils.layers import conv, linear, batch_norm
def ticketfy(model, split_rate, split_mode="kels"):
conv_layers, linear_layers, bn_layers = helpers.get_layers(model)
for n, _ in conv_layers:
cur_conv = helpers.rgetattr(model, n)
helpers.rs... | [
"utils.layers.batch_norm.SplitBatchNorm",
"utils.helpers.get_layers",
"utils.layers.conv.SplitConv",
"copy.deepcopy",
"utils.helpers.rgetattr"
] | [((200, 225), 'utils.helpers.get_layers', 'helpers.get_layers', (['model'], {}), '(model)\n', (218, 225), False, 'from utils import helpers\n'), ((2410, 2430), 'copy.deepcopy', 'copy.deepcopy', (['model'], {}), '(model)\n', (2423, 2430), False, 'import copy\n'), ((275, 301), 'utils.helpers.rgetattr', 'helpers.rgetattr'... |
from utils import scrape_helper
url = "http://www.investopedia.com/terms/1/"
links = scrape_helper.get_term_links_from_page(url)
print(links)
| [
"utils.scrape_helper.get_term_links_from_page"
] | [((87, 130), 'utils.scrape_helper.get_term_links_from_page', 'scrape_helper.get_term_links_from_page', (['url'], {}), '(url)\n', (125, 130), False, 'from utils import scrape_helper\n')] |
import requests
import bs4
dateList = []
higlist = []
lowlist= []
r = requests.get(
'https://coinmarketcap.com/currencies/bitcoin/historical-data/')
soup = bs4.BeautifulSoup(r.text, "lxml")
tr = soup.find_all('tr',{'class':'text-right'})
for item in tr:
dateList.append(item.find('td', {'class':'text-left'}... | [
"bs4.BeautifulSoup",
"requests.get"
] | [((74, 151), 'requests.get', 'requests.get', (['"""https://coinmarketcap.com/currencies/bitcoin/historical-data/"""'], {}), "('https://coinmarketcap.com/currencies/bitcoin/historical-data/')\n", (86, 151), False, 'import requests\n'), ((164, 197), 'bs4.BeautifulSoup', 'bs4.BeautifulSoup', (['r.text', '"""lxml"""'], {})... |
import os
from colorama import Fore, Style
from pathlib import Path
DIF = "dif10"
ECHO10 = "echo10"
UMM_JSON = "umm-json"
ROOT_DIR = (
# go up one directory
Path(__file__).resolve().parents[1]
)
SCHEMAS_BASE_PATH = f"{ROOT_DIR}/schemas"
SCHEMAS = {
"json": [
"checks",
"check_messages",
... | [
"pathlib.Path"
] | [((168, 182), 'pathlib.Path', 'Path', (['__file__'], {}), '(__file__)\n', (172, 182), False, 'from pathlib import Path\n')] |
"""
Generic publisher for graphana
"""
import abc
import six
from decisionengine.framework.modules import Publisher
import decisionengine_modules.graphite_client as graphite
DEFAULT_GRAPHITE_HOST = 'fermicloud399.fnal.gov'
DEFAULT_GRAPHITE_PORT = 2004
DEFAULT_GRAPHITE_CONTEXT = ""
@six.add_metaclass(abc.ABCMeta)
c... | [
"decisionengine_modules.graphite_client.Graphite",
"six.add_metaclass"
] | [((288, 318), 'six.add_metaclass', 'six.add_metaclass', (['abc.ABCMeta'], {}), '(abc.ABCMeta)\n', (305, 318), False, 'import six\n'), ((1405, 1479), 'decisionengine_modules.graphite_client.Graphite', 'graphite.Graphite', ([], {'host': 'self.graphite_host', 'pickle_port': 'self.graphite_port'}), '(host=self.graphite_hos... |
from flask import Blueprint
api = Blueprint('api', __name__)
from . import authentication, videos, shows, users, comments, errors | [
"flask.Blueprint"
] | [((35, 61), 'flask.Blueprint', 'Blueprint', (['"""api"""', '__name__'], {}), "('api', __name__)\n", (44, 61), False, 'from flask import Blueprint\n')] |
"""
Build SSW scripts from Jinja 2 templates
"""
import os
import datetime
import subprocess
import tempfile
from jinja2 import (Environment as Env,
FileSystemLoader,
PackageLoader)
from scipy.io import readsav
from .read_config import defaults
from .util import SSWIDLError, ID... | [
"tempfile.TemporaryDirectory",
"jinja2.Environment",
"subprocess.run",
"os.path.isfile",
"os.path.dirname",
"datetime.datetime.now",
"subprocess.call",
"os.path.basename",
"scipy.io.readsav",
"jinja2.PackageLoader"
] | [((1959, 1981), 'os.path.isfile', 'os.path.isfile', (['script'], {}), '(script)\n', (1973, 1981), False, 'import os\n'), ((2170, 2175), 'jinja2.Environment', 'Env', ([], {}), '()\n', (2173, 2175), True, 'from jinja2 import Environment as Env, FileSystemLoader, PackageLoader\n'), ((3869, 3898), 'tempfile.TemporaryDirect... |
"""
Entity-relation diagram (ERD) GraphViz dot-file generator.
Usage:
python -m erd db.json -o db.dot
Then pass the result to the GraphViz `dot` tool:
dot db.dot -T png -o db.png
Inspired by: https://github.com/ehne/ERDot
"""
import argparse
import json
import re
from pathlib import Path
from typing import D... | [
"json.loads",
"re.match",
"argparse.ArgumentParser"
] | [((2412, 2595), 're.match', 're.match', (['"""^(?P<source_name>\\\\w+):(?P<source_fk>\\\\w+) (?P<left_cardinality>[\\\\d\\\\+\\\\*])--(?P<right_cardinality>[\\\\d\\\\+\\\\*]) (?P<dest_name>\\\\w+):(?P<dest_pk>\\\\w+)$"""', 'relation'], {}), "(\n '^(?P<source_name>\\\\w+):(?P<source_fk>\\\\w+) (?P<left_cardinality>[\... |
# pylint: disable=no-self-use,invalid-name
import unittest
import spacy
from scispacy.hyponym_detector import HyponymDetector
class TestHyponymDetector(unittest.TestCase):
def setUp(self):
super().setUp()
self.nlp = spacy.load("en_core_sci_sm")
self.detector = HyponymDetector(self.nlp, ex... | [
"spacy.load",
"scispacy.hyponym_detector.HyponymDetector"
] | [((239, 267), 'spacy.load', 'spacy.load', (['"""en_core_sci_sm"""'], {}), "('en_core_sci_sm')\n", (249, 267), False, 'import spacy\n'), ((292, 332), 'scispacy.hyponym_detector.HyponymDetector', 'HyponymDetector', (['self.nlp'], {'extended': '(True)'}), '(self.nlp, extended=True)\n', (307, 332), False, 'from scispacy.hy... |
"""Steps up and down"""
import calendar
import numpy as np
from pandas.io.sql import read_sql
from pyiem import network
from pyiem.plot.use_agg import plt
from pyiem.util import get_autoplot_context, get_dbconn
PDICT = {'spring': '1 January - 30 June',
'fall': '1 July - 31 December'}
def get_description():... | [
"pyiem.network.Table",
"pandas.io.sql.read_sql",
"pyiem.util.get_dbconn",
"numpy.array",
"pyiem.plot.use_agg.plt.subplots"
] | [((1072, 1090), 'pyiem.util.get_dbconn', 'get_dbconn', (['"""coop"""'], {}), "('coop')\n", (1082, 1090), False, 'from pyiem.util import get_autoplot_context, get_dbconn\n'), ((1256, 1299), 'pyiem.network.Table', 'network.Table', (["('%sCLIMATE' % (station[:2],))"], {}), "('%sCLIMATE' % (station[:2],))\n", (1269, 1299),... |
import struct
import os
current_pe = None
class PE:
"""Basic PE parsing.
Ref:
- https://hshrzd.wordpress.com/pe-bear/
- https://blog.kowalczyk.info/articles/pefileformat.html
"""
X86_64 = 0x8664
X86_32 = 0x14c
ARM = 0x1c0
ARM64 ... | [
"os.access"
] | [((1378, 1400), 'os.access', 'os.access', (['pe', 'os.R_OK'], {}), '(pe, os.R_OK)\n', (1387, 1400), False, 'import os\n')] |
"""
This module is the custom resource used by the MSAM's CloudFormation
templates to populate the web bucket with contents of the MSAM web archive.
"""
# Copyright 2018 Amazon.com, Inc. or its affiliates. All Rights Reserved.
# SPDX-License-Identifier: Apache-2.0
import json
import os
from subprocess import call
imp... | [
"boto3.client",
"json.dumps",
"subprocess.call",
"resource_tools.send",
"resource_tools.stack_name",
"os.walk"
] | [((1786, 1889), 'resource_tools.send', 'resource_tools.send', (['event', 'context', "result['Status']", "result['Data']", "result['PhysicalResourceId']"], {}), "(event, context, result['Status'], result['Data'],\n result['PhysicalResourceId'])\n", (1805, 1889), False, 'import resource_tools\n'), ((2098, 2116), 'boto... |
# -*- coding: utf-8 -*-
# ------------------------------------------------------------------------------
#
# Copyright 2018-2019 Fetch.AI Limited
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the ... | [
"aea.cli.common.logger.exception",
"pathlib.Path",
"click.option",
"aea.helpers.base.load_env_file",
"click.echo",
"click.Path",
"sys.exit",
"click.command"
] | [((2790, 2805), 'click.command', 'click.command', ([], {}), '()\n', (2803, 2805), False, 'import click\n'), ((2807, 3032), 'click.option', 'click.option', (['"""--connections"""', '"""connection_ids"""'], {'cls': 'ConnectionsOption', 'required': '(False)', 'default': 'None', 'help': '"""The connection names to use for ... |
"""An environment to skip k frames and return a max between the last two."""
import gym
import numpy as np
class MaxFrameskipEnv(gym.Wrapper):
"""An environment to skip k frames and return a max between the last two."""
def __init__(self, env, skip: int=4) -> None:
"""
Initialize a new max fr... | [
"numpy.zeros",
"gym.Wrapper.__init__"
] | [((557, 588), 'gym.Wrapper.__init__', 'gym.Wrapper.__init__', (['self', 'env'], {}), '(self, env)\n', (577, 588), False, 'import gym\n'), ((691, 750), 'numpy.zeros', 'np.zeros', (['(2, *env.observation_space.shape)'], {'dtype': 'np.uint8'}), '((2, *env.observation_space.shape), dtype=np.uint8)\n', (699, 750), True, 'im... |
"""Script containing the DeepLoco environments."""
import gym
import numpy as np
import os
import sys
import cv2
try:
sys.path.append(os.path.join(os.environ["TERRAINRL_PATH"], "simAdapter"))
import terrainRLSim # noqa: F401
except (KeyError, ImportError, ModuleNotFoundError):
pass
class BipedalSoccer(g... | [
"numpy.flip",
"terrainRLSim.getEnv",
"os.path.join",
"gym.spaces.Box",
"cv2.imshow",
"numpy.array",
"cv2.waitKey",
"gym.make"
] | [((139, 195), 'os.path.join', 'os.path.join', (["os.environ['TERRAINRL_PATH']", '"""simAdapter"""'], {}), "(os.environ['TERRAINRL_PATH'], 'simAdapter')\n", (151, 195), False, 'import os\n'), ((1016, 1077), 'terrainRLSim.getEnv', 'terrainRLSim.getEnv', (['"""PD-Biped3D-HLC-Soccer-v1"""'], {'render': '(False)'}), "('PD-B... |
# -*- coding:utf-8 -*-
'''
Created on 2015年3月2日
@author: wanhao01
'''
import sys
from crawler.minispider import logerror
import main
reload(sys)
sys.setdefaultencoding('utf-8')
if __name__ == '__main__':
try:
main.main()
except Exception as exception:
logerror("error du... | [
"main.main",
"sys.setdefaultencoding"
] | [((164, 195), 'sys.setdefaultencoding', 'sys.setdefaultencoding', (['"""utf-8"""'], {}), "('utf-8')\n", (186, 195), False, 'import sys\n'), ((245, 256), 'main.main', 'main.main', ([], {}), '()\n', (254, 256), False, 'import main\n')] |
import types
from tf.advanced.app import App
MODIFIERS = """
remark folio note ref emph und super special q num den
""".strip().split()
def fmt_layoutFull(app, n, **kwargs):
return app._wrapHtml(n, ("",))
def fmt_layoutRemarks(app, n, **kwargs):
return app._wrapHtml(n, ("r",))
def fmt_layoutNotes(ap... | [
"types.MethodType"
] | [((819, 856), 'types.MethodType', 'types.MethodType', (['fmt_layoutFull', 'app'], {}), '(fmt_layoutFull, app)\n', (835, 856), False, 'import types\n'), ((889, 929), 'types.MethodType', 'types.MethodType', (['fmt_layoutRemarks', 'app'], {}), '(fmt_layoutRemarks, app)\n', (905, 929), False, 'import types\n'), ((960, 998)... |
import hashlib
import logging
import os
import shutil
import traceback
from contextlib import closing
from pywb.utils.loaders import BlockLoader
from webrecorder.rec.storage.base import BaseStorage
from webrecorder.rec.storage.storagepaths import add_local_store_prefix, strip_prefix
logger = logging.getLogger('wr.io... | [
"logging.getLogger",
"hashlib.md5",
"pywb.utils.loaders.BlockLoader",
"os.path.join",
"webrecorder.rec.storage.storagepaths.strip_prefix",
"os.path.isfile",
"os.path.dirname",
"shutil.copyfile",
"os.removedirs",
"shutil.rmtree",
"os.remove"
] | [((296, 322), 'logging.getLogger', 'logging.getLogger', (['"""wr.io"""'], {}), "('wr.io')\n", (313, 322), False, 'import logging\n'), ((875, 916), 'os.path.join', 'os.path.join', (['self.storage_root', 'dir_path'], {}), '(self.storage_root, dir_path)\n', (887, 916), False, 'import os\n'), ((2178, 2204), 'os.path.isfile... |
# This file is part of OctoBot (https://github.com/Drakkar-Software/OctoBot)
# Copyright (c) 2021 Drakkar-Software, All rights reserved.
#
# OctoBot is free software; you can redistribute it and/or
# modify it under the terms of the GNU General Public License
# as published by the Free Software Foundation; either
... | [
"octobot.community.ErrorsUploader",
"time.time",
"octobot.community.Error"
] | [((915, 926), 'time.time', 'time.time', ([], {}), '()\n', (924, 926), False, 'import time\n'), ((1010, 1074), 'octobot.community.Error', 'community.Error', (['None', 'ERROR_TITLE', 'ERROR_TIME', 'ERROR_METRICS_ID'], {}), '(None, ERROR_TITLE, ERROR_TIME, ERROR_METRICS_ID)\n', (1025, 1074), True, 'import octobot.communit... |
import connexion
import six
from ga4ghtest.models import Plugin # noqa: E501
from ga4ghtest import util
from ga4ghtest.core.controllers import plugins_controller as controller
def create_plugin(
body
): # noqa: E501
"""Create a test plugin
Add a plugin for testing functionality of an API. # noqa: E501... | [
"connexion.request.get_json",
"ga4ghtest.core.controllers.plugins_controller.get_plugins",
"ga4ghtest.core.controllers.plugins_controller.create_plugin"
] | [((515, 550), 'ga4ghtest.core.controllers.plugins_controller.create_plugin', 'controller.create_plugin', ([], {'body': 'body'}), '(body=body)\n', (539, 550), True, 'from ga4ghtest.core.controllers import plugins_controller as controller\n'), ((1010, 1075), 'ga4ghtest.core.controllers.plugins_controller.get_plugins', 'c... |
import os
from split_settings.tools import include, optional
ENVIRONMENT = os.getenv('DJANGO_ENV') or 'development'
include(
# Load environment settings
'base/env.py',
optional('local/env.py'), # We can "patch" any settings from local folder env.py file.
# Here we should have the order because of de... | [
"os.getenv",
"split_settings.tools.optional"
] | [((76, 99), 'os.getenv', 'os.getenv', (['"""DJANGO_ENV"""'], {}), "('DJANGO_ENV')\n", (85, 99), False, 'import os\n'), ((182, 206), 'split_settings.tools.optional', 'optional', (['"""local/env.py"""'], {}), "('local/env.py')\n", (190, 206), False, 'from split_settings.tools import include, optional\n'), ((520, 542), 's... |
# -*- coding: utf-8 -*-
'''
Utility methods for dictionary
==============================
'''
__all__ = (
'calling_dict_from',
'combine_dict',
'dict_sorted')
from itertools import chain
from typing import Tuple
from builder.utils import assertion
def calling_dict_from(calling: (str, dict), n... | [
"builder.utils.assertion.is_bool",
"builder.utils.assertion.is_dict",
"builder.utils.assertion.is_str"
] | [((818, 838), 'builder.utils.assertion.is_dict', 'assertion.is_dict', (['a'], {}), '(a)\n', (835, 838), False, 'from builder.utils import assertion\n'), ((842, 862), 'builder.utils.assertion.is_dict', 'assertion.is_dict', (['b'], {}), '(b)\n', (859, 862), False, 'from builder.utils import assertion\n'), ((570, 595), 'b... |
from flask import Flask
from flask_socketio import SocketIO
from flask_sqlalchemy import SQLAlchemy
UPLOAD_FOLDER = 'uploads'
ALLOWED_EXTENSIONS = {'png', 'jpg', 'jpeg', 'gif'}
app = Flask(__name__)
app.config['SQLALCHEMY_DATABASE_URI'] = 'sqlite:////tmp/dev.db'
app.config['UPLOAD_FOLDER'] = UPLOAD_FOLDER
app.secret_... | [
"flask_sqlalchemy.SQLAlchemy",
"flask_socketio.SocketIO",
"flask.Flask"
] | [((185, 200), 'flask.Flask', 'Flask', (['__name__'], {}), '(__name__)\n', (190, 200), False, 'from flask import Flask\n'), ((346, 359), 'flask_socketio.SocketIO', 'SocketIO', (['app'], {}), '(app)\n', (354, 359), False, 'from flask_socketio import SocketIO\n'), ((365, 380), 'flask_sqlalchemy.SQLAlchemy', 'SQLAlchemy', ... |
import numpy as np
class Loss():
def output_gradient(self):
return
class MSE(Loss):
def __call__(self, predicted, labels):
return 0.5 * np.square(predicted - labels)
def output_gradient(self, predicted, labels):
return predicted - labels
class BinaryCrossEntropy(Loss):
def _... | [
"numpy.log",
"numpy.nan_to_num",
"numpy.square"
] | [((512, 581), 'numpy.nan_to_num', 'np.nan_to_num', (['(-(labels / predicted) + (1 - labels) / (1 - predicted))'], {}), '(-(labels / predicted) + (1 - labels) / (1 - predicted))\n', (525, 581), True, 'import numpy as np\n'), ((163, 192), 'numpy.square', 'np.square', (['(predicted - labels)'], {}), '(predicted - labels)\... |
"""
This file is part of pynadc
https://github.com/rmvanhees/pynadc
Methods to query the NADC Sciamachy SQLite database
Copyright (c) 2012-2021 SRON - Netherlands Institute for Space Research
All Rights Reserved
License: BSD-3-Clause
"""
from pathlib import Path
import sqlite3
# ------------------------------... | [
"sqlite3.connect",
"pathlib.Path"
] | [((1944, 1967), 'sqlite3.connect', 'sqlite3.connect', (['dbname'], {}), '(dbname)\n', (1959, 1967), False, 'import sqlite3\n'), ((7502, 7525), 'sqlite3.connect', 'sqlite3.connect', (['dbname'], {}), '(dbname)\n', (7517, 7525), False, 'import sqlite3\n'), ((2581, 2596), 'pathlib.Path', 'Path', (['*row[:-1]'], {}), '(*ro... |
from pycord.discord.ext import commands
import pycord.discord as discord
from pycord.discord import Embed
import requests
import json
from discord import Embed
class Data(commands.Cog):
def __init__(self, bot) -> None:
self.bot: commands.Bot = bot
@commands.command()
async def ping(self, ctx):
await ctx... | [
"requests.get",
"json.loads",
"discord.Embed",
"pycord.discord.ext.commands.command"
] | [((259, 277), 'pycord.discord.ext.commands.command', 'commands.command', ([], {}), '()\n', (275, 277), False, 'from pycord.discord.ext import commands\n'), ((378, 396), 'pycord.discord.ext.commands.command', 'commands.command', ([], {}), '()\n', (394, 396), False, 'from pycord.discord.ext import commands\n'), ((445, 52... |
#!/usr/bin/python3
from SetupRunDirectory import verifyDirectoryFiles, setupRunDirectory
from CleanupRunDirectory import cleanUpRunDirectory
from RunAssembly import verifyConfigFiles, verifyFastaFiles, runAssembly, initializeAssembler
from SaveRun import saveRun
import configparser
from datetime import datetime
from ... | [
"configparser.ConfigParser",
"SaveRun.saveRun",
"RunAssembly.verifyFastaFiles",
"os.path.exists",
"SetupRunDirectory.verifyDirectoryFiles",
"SetupRunDirectory.setupRunDirectory",
"argparse.ArgumentParser",
"subprocess.Popen",
"CleanupRunDirectory.cleanUpRunDirectory",
"os.mkdir",
"os.path.dirnam... | [((584, 598), 'datetime.datetime.now', 'datetime.now', ([], {}), '()\n', (596, 598), False, 'from datetime import datetime\n'), ((4709, 4743), 'os.path.abspath', 'os.path.abspath', (['readsSequencePath'], {}), '(readsSequencePath)\n', (4724, 4743), False, 'import os\n'), ((5347, 5395), 'os.path.join', 'os.path.join', (... |
"""
You work for a retail store that wants to increase sales on Tuesday and
Wednesday, which are the store's slowest sales days. On Tuesday and
Wednesday, if a customer's subtotal is greater than $50, the store will
discount the customer's purchase by 10%.
"""
# Import the datatime module so that
# it can be used in t... | [
"datetime.datetime.now"
] | [((1143, 1157), 'datetime.datetime.now', 'datetime.now', ([], {}), '()\n', (1155, 1157), False, 'from datetime import datetime\n')] |
# Copyright (c) 2021, NVIDIA CORPORATION. All rights reserved.
#
# NVIDIA CORPORATION and its licensors retain all intellectual property
# and proprietary rights in and to this software, related documentation
# and any modifications thereto. Any use, reproduction, disclosure or
# distribution of this software and rel... | [
"numpy.sqrt",
"torch.full_like",
"torch.nn.functional.softplus",
"torch.randn_like",
"torch_utils.ops.conv2d_gradfix.no_weight_gradients",
"torch.autograd.profiler.record_function",
"torch.no_grad",
"torch_utils.training_stats.report",
"torch.linspace",
"torch_utils.misc.ddp_sync",
"torch.empty"... | [((1718, 1748), 'torch.zeros', 'torch.zeros', (['[]'], {'device': 'device'}), '([], device=device)\n', (1729, 1748), False, 'import torch\n'), ((2130, 2165), 'torch_utils.misc.ddp_sync', 'misc.ddp_sync', (['self.G_mapping', 'sync'], {}), '(self.G_mapping, sync)\n', (2143, 2165), False, 'from torch_utils import misc, tr... |
from BurstPaperWallet.api import brs_api
from BurstPaperWallet.api import passphrase_url_transform as transform
def initialize(account, old_passphrase, fee=735000):
url = "sendMoney&recipient={}&secretPhrase={}&amountNQT=1&feeNQT={}&recipientPublicKey={}&deadline=1440"\
.format(account["reed solomon"], tr... | [
"BurstPaperWallet.api.passphrase_url_transform",
"BurstPaperWallet.api.brs_api"
] | [((512, 524), 'BurstPaperWallet.api.brs_api', 'brs_api', (['url'], {}), '(url)\n', (519, 524), False, 'from BurstPaperWallet.api import brs_api\n'), ((318, 343), 'BurstPaperWallet.api.passphrase_url_transform', 'transform', (['old_passphrase'], {}), '(old_passphrase)\n', (327, 343), True, 'from BurstPaperWallet.api imp... |
import numpy as np
import os
import keras
from keras import regularizers, losses
from keras.models import Sequential, Model
from keras.layers import Lambda, Input, Dense, Dropout, Reshape, BatchNormalization, Softmax, Concatenate
from keras.utils import plot_model
import keras.backend as K
class multiVAE:
def __init_... | [
"keras.layers.Concatenate",
"keras.utils.plot_model",
"keras.layers.Lambda",
"keras.models.Model",
"keras.backend.random_normal",
"keras.layers.Dense",
"keras.backend.exp"
] | [((1138, 1207), 'keras.backend.random_normal', 'K.random_normal', ([], {'shape': '(32, self.inf_layerSize)', 'mean': '(0.0)', 'stddev': '(1.0)'}), '(shape=(32, self.inf_layerSize), mean=0.0, stddev=1.0)\n', (1153, 1207), True, 'import keras.backend as K\n'), ((3020, 3047), 'keras.models.Model', 'Model', ([], {'inputs':... |
import pytest
from tests.stub_client import StubHttpClient
from coinoxr.requestor import Requestor
from coinoxr.response import Response
def content(file):
return StubHttpClient.json(file)["content"]
@pytest.fixture
def client():
client = StubHttpClient()
client.add_app_id("fake_app_id")
client.add... | [
"tests.stub_client.StubHttpClient",
"coinoxr.response.Response",
"coinoxr.requestor.Requestor",
"tests.stub_client.StubHttpClient.json"
] | [((252, 268), 'tests.stub_client.StubHttpClient', 'StubHttpClient', ([], {}), '()\n', (266, 268), False, 'from tests.stub_client import StubHttpClient\n'), ((733, 765), 'coinoxr.requestor.Requestor', 'Requestor', (['"""fake_app_id"""', 'client'], {}), "('fake_app_id', client)\n", (742, 765), False, 'from coinoxr.reques... |
#!flask/bin/python
# coding=utf-8
from flask import Flask, jsonify
app = Flask(__name__)
tasks = {
"event_id" : "1.9",
"introductions" : [
{
"title" : "情怀",
"details" : "各种无敌, 各种牛人, 各种挑战, 等你来战",
"image" : "hello.png",
"background_image" : "backgroundImage.png"
},
{
"title... | [
"flask.jsonify",
"flask.Flask"
] | [((74, 89), 'flask.Flask', 'Flask', (['__name__'], {}), '(__name__)\n', (79, 89), False, 'from flask import Flask, jsonify\n'), ((539, 553), 'flask.jsonify', 'jsonify', (['tasks'], {}), '(tasks)\n', (546, 553), False, 'from flask import Flask, jsonify\n')] |
# Copyright (c) 2015-2017 <NAME>
# License: MIT
"""
nbrun - Run an Jupyter/IPython notebook, optionally passing arguments.
USAGE
-----
Copy this file in the folder containing the master notebook used to
execute the other notebooks. Then use `run_notebook()` to execute
notebooks.
"""
import time
from pathlib import P... | [
"nbconvert.HTMLExporter",
"time.ctime",
"argparse.ArgumentParser",
"pathlib.Path",
"nbformat.v4.new_code_cell",
"nbconvert.preprocessors.ExecutePreprocessor",
"time.time",
"nbformat.v4.new_markdown_cell"
] | [((5814, 5833), 'pathlib.Path', 'Path', (['notebook_path'], {}), '(notebook_path)\n', (5818, 5833), False, 'from pathlib import Path\n'), ((6930, 6967), 'nbconvert.preprocessors.ExecutePreprocessor', 'ExecutePreprocessor', ([], {}), '(**execute_kwargs)\n', (6949, 6967), False, 'from nbconvert.preprocessors import Execu... |
import numpy as np
import pandas as pd
import seaborn as sns
from nninst.backend.tensorflow.model import AlexNet
from nninst.backend.tensorflow.trace.alexnet_imagenet_inter_class_similarity import (
alexnet_imagenet_inter_class_similarity_frequency,
)
from nninst.op import Conv2dOp, DenseOp
np.random.seed(0)
sns.... | [
"seaborn.set",
"nninst.backend.tensorflow.model.AlexNet.graph",
"numpy.eye",
"nninst.backend.tensorflow.trace.alexnet_imagenet_inter_class_similarity.alexnet_imagenet_inter_class_similarity_frequency",
"numpy.around",
"numpy.random.seed",
"pandas.DataFrame",
"numpy.tri"
] | [((298, 315), 'numpy.random.seed', 'np.random.seed', (['(0)'], {}), '(0)\n', (312, 315), True, 'import numpy as np\n'), ((316, 325), 'seaborn.set', 'sns.set', ([], {}), '()\n', (323, 325), True, 'import seaborn as sns\n'), ((3070, 3184), 'pandas.DataFrame', 'pd.DataFrame', (["{'Same Class': same_class_similarity, 'Diff... |
# -*- coding: utf-8 -*-
"""
<NAME>
Computational Biologist
Target Sciences
GSK
<EMAIL>
"""
import sys
import get_generalizable_features
import get_merged_features
import get_useful_features
def main(validation_rep=0, validation_fold=0):
print('VALIDATION_REP: {0!s}, VALIDATION_FOLD:{1!s}'.format(validation_r... | [
"get_useful_features.main",
"get_merged_features.main",
"get_generalizable_features.main"
] | [((418, 482), 'get_generalizable_features.main', 'get_generalizable_features.main', (['validation_rep', 'validation_fold'], {}), '(validation_rep, validation_fold)\n', (449, 482), False, 'import get_generalizable_features\n'), ((544, 601), 'get_merged_features.main', 'get_merged_features.main', (['validation_rep', 'val... |
# Generated by Django 3.1.6 on 2021-04-12 08:45
from django.db import migrations
class Migration(migrations.Migration):
dependencies = [
('SocialApp', '0002_auto_20210411_2237'),
]
operations = [
migrations.DeleteModel(
name='RemoteFollow',
),
]
| [
"django.db.migrations.DeleteModel"
] | [((229, 272), 'django.db.migrations.DeleteModel', 'migrations.DeleteModel', ([], {'name': '"""RemoteFollow"""'}), "(name='RemoteFollow')\n", (251, 272), False, 'from django.db import migrations\n')] |
"""Tests for thread_async_queue."""
from __future__ import annotations
import asyncio
from concurrent.futures import ThreadPoolExecutor
from itertools import chain
from typing import List, NamedTuple
import pytest
from opentrons.protocol_runner.thread_async_queue import (
ThreadAsyncQueue,
QueueClosed,
)
... | [
"itertools.chain",
"concurrent.futures.ThreadPoolExecutor",
"pytest.raises",
"asyncio.sleep"
] | [((606, 632), 'pytest.raises', 'pytest.raises', (['QueueClosed'], {}), '(QueueClosed)\n', (619, 632), False, 'import pytest\n'), ((666, 692), 'pytest.raises', 'pytest.raises', (['QueueClosed'], {}), '(QueueClosed)\n', (679, 692), False, 'import pytest\n'), ((770, 796), 'pytest.raises', 'pytest.raises', (['QueueClosed']... |
"""Solution for subdag_example.py.
Uses a factory function to return a DAG that can be used as the subdag argument
to SubDagOperator. Notice that:
1) the SubDAG's dag_id is formatted as parent_dag_id.subdag_task_id
2) the start_date and schedule_interval of the SubDAG are copied from the parent
DAG.
"""
from airflo... | [
"datetime.datetime.min.time",
"airflow.operators.dummy_operator.DummyOperator",
"airflow.DAG",
"datetime.datetime.today",
"datetime.timedelta"
] | [((744, 763), 'datetime.datetime.min.time', 'datetime.min.time', ([], {}), '()\n', (761, 763), False, 'from datetime import datetime, timedelta\n'), ((963, 983), 'datetime.timedelta', 'timedelta', ([], {'minutes': '(5)'}), '(minutes=5)\n', (972, 983), False, 'from datetime import datetime, timedelta\n'), ((2142, 2214),... |
#!/usr/bin/env python3.8
"""
export_blueprints.py
Connect to a Nutanix Prism Central instance, grab all Calm blueprints and export them to JSON files.
You would need to *heavily* modify this script for use in a production environment so that it contains appropriate error-checking and exception handling.... | [
"argparse.ArgumentParser",
"getpass.getpass",
"apiclient.ApiClient",
"urllib3.disable_warnings",
"time.localtime",
"json.dump"
] | [((1042, 1121), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {'description': '"""Export all Calm blueprints to JSON files"""'}), "(description='Export all Calm blueprints to JSON files')\n", (1065, 1121), False, 'import argparse\n'), ((1997, 2064), 'urllib3.disable_warnings', 'urllib3.disable_warnings', (... |
#* pyx509 - Python library for parsing X.509
#* Copyright (C) 2009-2010 CZ.NIC, z.s.p.o. (http://www.nic.cz)
#*
#* This library is free software; you can redistribute it and/or
#* modify it under the terms of the GNU Library General Public
#* License as published by the Free Software Foundation; either... | [
"pyasn1.type.univ.OctetString",
"pyasn1.type.tag.Tag",
"pyasn1.type.useful.UTCTime",
"pyasn1.type.univ.Boolean",
"pyasn1.type.useful.GeneralizedTime",
"pyasn1.type.univ.ObjectIdentifier"
] | [((1270, 1293), 'pyasn1.type.univ.ObjectIdentifier', 'univ.ObjectIdentifier', ([], {}), '()\n', (1291, 1293), False, 'from pyasn1.type import tag, namedtype, univ, useful\n'), ((1345, 1366), 'pyasn1.type.univ.Boolean', 'univ.Boolean', (['"""False"""'], {}), "('False')\n", (1357, 1366), False, 'from pyasn1.type import t... |
from math import isqrt
is_square = lambda n: isqrt(n) ** 2 == n if n >= 0 else False
def is_square_soln(n):
pass
print(is_square(-1)) | [
"math.isqrt"
] | [((46, 54), 'math.isqrt', 'isqrt', (['n'], {}), '(n)\n', (51, 54), False, 'from math import isqrt\n')] |
# -*- coding: utf-8 -*-
"""
TencentBlueKing is pleased to support the open source community by making 蓝鲸智云-用户管理(Bk-User) available.
Copyright (C) 2017-2021 THL A29 Limited, a Tencent company. All rights reserved.
Licensed under the MIT License (the "License"); you may not use this file except in compliance with the Lic... | [
"logging.getLogger",
"rest_framework.serializers.DateTimeField",
"rest_framework.serializers.BooleanField",
"bkuser_core.categories.loader.register_plugin",
"bkuser_core.common.models.is_obj_needed_update",
"bkuser_core.user_settings.models.Setting.objects.get_or_create",
"yaml.safe_load",
"bkuser_cor... | [((1411, 1438), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (1428, 1438), False, 'import logging\n'), ((1566, 1592), 'rest_framework.serializers.BooleanField', 'serializers.BooleanField', ([], {}), '()\n', (1590, 1592), False, 'from rest_framework import serializers\n'), ((1602, 1629),... |
from django import template
import mistune
register = template.Library()
@register.filter
def markdown(value):
markdown = mistune.Markdown()
return markdown(value)
| [
"mistune.Markdown",
"django.template.Library"
] | [((55, 73), 'django.template.Library', 'template.Library', ([], {}), '()\n', (71, 73), False, 'from django import template\n'), ((129, 147), 'mistune.Markdown', 'mistune.Markdown', ([], {}), '()\n', (145, 147), False, 'import mistune\n')] |
import json
from finite.storage import new_uuid
class Unimplemented(Exception):
pass
class RoleFail(Exception):
pass
SUPERUSER = '*'
""" role used to bypass all permission checks """
ROOT_UUID = '00000000-0000-0000-0000-000000000000'
""" parent UUID used to initialize a stream """
DEFAULT_SCHEMA = 'base... | [
"json.dumps"
] | [((1518, 1547), 'json.dumps', 'json.dumps', (["kwargs['payload']"], {}), "(kwargs['payload'])\n", (1528, 1547), False, 'import json\n')] |
# This file is part of Indico.
# Copyright (C) 2002 - 2019 CERN
#
# Indico is free software; you can redistribute it and/or
# modify it under the terms of the MIT License; see the
# LICENSE file for more details.
from datetime import datetime, timedelta
from uuid import uuid4
import pytest
from flask import session
f... | [
"mock.MagicMock",
"indico.modules.oauth.provider.load_token",
"datetime.datetime.utcnow",
"indico.modules.oauth.provider.load_client",
"uuid.uuid4",
"indico.modules.oauth.models.applications.OAuthApplication.find",
"pytest.mark.parametrize",
"pytest.raises",
"pytest.mark.usefixtures",
"datetime.ti... | [((1787, 1829), 'pytest.mark.usefixtures', 'pytest.mark.usefixtures', (['"""request_context"""'], {}), "('request_context')\n", (1810, 1829), False, 'import pytest\n'), ((2504, 2546), 'pytest.mark.usefixtures', 'pytest.mark.usefixtures', (['"""request_context"""'], {}), "('request_context')\n", (2527, 2546), False, 'im... |
#!/usr/bin/env python3
# Copyright (c) Facebook, Inc. and its affiliates.
#
# This source code is licensed under the MIT license found in the
# LICENSE file in the root directory of this source tree.
import unittest
from unittest.mock import MagicMock, patch
from fbpcs.service.log_cloudwatch import CloudWatchLogServi... | [
"unittest.mock.MagicMock",
"fbpcs.service.log_cloudwatch.CloudWatchLogService",
"unittest.mock.patch"
] | [((460, 511), 'unittest.mock.patch', 'patch', (['"""fbpcs.gateway.cloudwatch.CloudWatchGateway"""'], {}), "('fbpcs.gateway.cloudwatch.CloudWatchGateway')\n", (465, 511), False, 'from unittest.mock import MagicMock, patch\n'), ((583, 622), 'fbpcs.service.log_cloudwatch.CloudWatchLogService', 'CloudWatchLogService', (['L... |
import adsk.core
import adsk.fusion
import os
from ...lib import fusion360utils as futil
from ... import config
import math
app = adsk.core.Application.get()
ui = app.userInterface
# TODO *** コマンドのID情報を指定します。 ***
CMD_ID = f'{config.COMPANY_NAME}_{config.ADDIN_NAME}_Meteor'
CMD_NAME = 'メテオ'
CMD_Descript... | [
"os.path.abspath"
] | [((903, 928), 'os.path.abspath', 'os.path.abspath', (['__file__'], {}), '(__file__)\n', (918, 928), False, 'import os\n')] |
import re
import time
from datetime import datetime
from enum import Enum
from fishbase.fish_logger import logger
from .elk_connector import Es
from ..base.tp_base import TpBase, TestStatus, Conf, VerticalContext
from ..base.tp_base import get_params_dict
# LogTestPoint
class LogESTestPoint(TpBase):
# 类的初始化过程
... | [
"datetime.datetime.utcfromtimestamp",
"datetime.datetime.utcnow"
] | [((1053, 1091), 'datetime.datetime.utcfromtimestamp', 'datetime.utcfromtimestamp', (['time_struct'], {}), '(time_struct)\n', (1078, 1091), False, 'from datetime import datetime\n'), ((1637, 1654), 'datetime.datetime.utcnow', 'datetime.utcnow', ([], {}), '()\n', (1652, 1654), False, 'from datetime import datetime\n')] |
import math
from typing import List
import numpy as np
import torch
a = torch.randn(4, 3,2)
print(a)
print(torch.argmax(a, -1)) | [
"torch.randn",
"torch.argmax"
] | [((73, 93), 'torch.randn', 'torch.randn', (['(4)', '(3)', '(2)'], {}), '(4, 3, 2)\n', (84, 93), False, 'import torch\n'), ((108, 127), 'torch.argmax', 'torch.argmax', (['a', '(-1)'], {}), '(a, -1)\n', (120, 127), False, 'import torch\n')] |
import sys
import json
import boto3
from botocore.exceptions import ClientError
from . import config
def status_instance(instance_id, dry_run=False):
ec2 = boto3.client('ec2',
region_name='ap-northeast-1',
aws_access_key_id=config.AWS_ACCESS_KEY_ID,
... | [
"boto3.client"
] | [((170, 313), 'boto3.client', 'boto3.client', (['"""ec2"""'], {'region_name': '"""ap-northeast-1"""', 'aws_access_key_id': 'config.AWS_ACCESS_KEY_ID', 'aws_secret_access_key': 'config.AWS_SECRET_KEY'}), "('ec2', region_name='ap-northeast-1', aws_access_key_id=config.\n AWS_ACCESS_KEY_ID, aws_secret_access_key=config... |
from django import template
register = template.Library()
@register.filter
def has_group(user, name):
return user.groups.filter(name=name).exists()
| [
"django.template.Library"
] | [((40, 58), 'django.template.Library', 'template.Library', ([], {}), '()\n', (56, 58), False, 'from django import template\n')] |
import importlib.util
import inspect
from skaio import log
from skaio.core.publisher import Publisher
from skaio.core.base.task import BaseTask
from skaio.utils.common import get_loop
tasks = ['samples.simple_tasks']
class Scheduler:
def start(self):
publisher = Publisher()
loop = get_loop()
... | [
"skaio.utils.common.get_loop",
"inspect.getmembers",
"skaio.core.publisher.Publisher",
"inspect.isclass",
"skaio.log.info"
] | [((278, 289), 'skaio.core.publisher.Publisher', 'Publisher', ([], {}), '()\n', (287, 289), False, 'from skaio.core.publisher import Publisher\n'), ((305, 315), 'skaio.utils.common.get_loop', 'get_loop', ([], {}), '()\n', (313, 315), False, 'from skaio.utils.common import get_loop\n'), ((629, 650), 'inspect.getmembers',... |
# Copyright (c) 2020 <NAME>
#
# This software is released under the MIT License.
# https://opensource.org/licenses/MIT
from bq_test_kit.interpolators.shell_interpolator import ShellInterpolator
def test_interpolate():
si = ShellInterpolator({"LOCAL_KEY": "VALUE"})
result = si.interpolate("Local key has value... | [
"bq_test_kit.interpolators.shell_interpolator.ShellInterpolator"
] | [((230, 271), 'bq_test_kit.interpolators.shell_interpolator.ShellInterpolator', 'ShellInterpolator', (["{'LOCAL_KEY': 'VALUE'}"], {}), "({'LOCAL_KEY': 'VALUE'})\n", (247, 271), False, 'from bq_test_kit.interpolators.shell_interpolator import ShellInterpolator\n')] |
# Copyright (c) Microsoft Corporation. All rights reserved.
# Licensed under the MIT License.
import os, uuid, sys, pickle, shutil, io, logging
from azure.storage.filedatalake import DataLakeServiceClient
from azure.core._match_conditions import MatchConditions
from azure.storage.filedatalake._models import ContentSet... | [
"shutil.make_archive"
] | [((1024, 1127), 'shutil.make_archive', 'shutil.make_archive', (['"""retailai_recommendation_model"""', '"""zip"""', '"""model\\\\retailai_recommendation_model"""'], {}), "('retailai_recommendation_model', 'zip',\n 'model\\\\retailai_recommendation_model')\n", (1043, 1127), False, 'import os, uuid, sys, pickle, shuti... |