code
stringlengths
22
1.05M
apis
listlengths
1
3.31k
extract_api
stringlengths
75
3.25M
import tkinter as tk from tkinter import filedialog from urllib.request import urlopen from pathlib import Path from tkinter import ttk import numpy as np import base64 import io import re from src.theme import theme from src.algorithm import blosum from src.utils import RichText def qopen(path:str): '''Opens and...
[ "tkinter.ttk.Button", "src.theme.theme", "numpy.hstack", "pathlib.Path.home", "src.algorithm.blosum", "tkinter.ttk.LabelFrame", "base64.encodestring", "src.utils.RichText", "tkinter.Label", "tkinter.ttk.Entry", "tkinter.ttk.Frame", "tkinter.ttk.Label", "numpy.asarray", "numpy.ndenumerate",...
[((500, 507), 'tkinter.Tk', 'tk.Tk', ([], {}), '()\n', (505, 507), True, 'import tkinter as tk\n'), ((702, 713), 'tkinter.ttk.Style', 'ttk.Style', ([], {}), '()\n', (711, 713), False, 'from tkinter import ttk\n'), ((984, 1019), 'tkinter.ttk.Notebook', 'ttk.Notebook', (['self.root'], {'padding': '(10)'}), '(self.root, p...
import numpy as np import re def apply_colormaps_based_on_mask(mask, data_for_inside_mask, data_for_outside_mask, colormap_inside_mask, colormap_outside_mask): """ Returns the combination of applying two colormaps to two datasets on two mutually exclusive sets of pixels a...
[ "re.fullmatch", "numpy.stack", "numpy.clip" ]
[((2605, 2638), 'numpy.stack', 'np.stack', (['([frame] * depth)'], {'axis': '(2)'}), '([frame] * depth, axis=2)\n', (2613, 2638), True, 'import numpy as np\n'), ((2915, 2968), 're.fullmatch', 're.fullmatch', (['"""[ra][\\\\-\\\\.0-9]+"""', 'mv_thresholdOnValue'], {}), "('[ra][\\\\-\\\\.0-9]+', mv_thresholdOnValue)\n", ...
import roslib;roslib.load_manifest('kobuki_dashboard') import rospy import diagnostic_msgs from rqt_robot_dashboard.dashboard import Dashboard from rqt_robot_dashboard.widgets import ConsoleDashWidget, MenuDashWidget, IconToolButton from python_qt_binding.QtWidgets import QMessageBox, QAction from python_qt_binding.Q...
[ "rqt_robot_dashboard.widgets.ConsoleDashWidget", "rospy.logwarn", "rospy.get_time", "roslib.load_manifest", "python_qt_binding.QtCore.QTimer", "rospy.Subscriber" ]
[((14, 54), 'roslib.load_manifest', 'roslib.load_manifest', (['"""kobuki_dashboard"""'], {}), "('kobuki_dashboard')\n", (34, 54), False, 'import roslib\n'), ((818, 826), 'python_qt_binding.QtCore.QTimer', 'QTimer', ([], {}), '()\n', (824, 826), False, 'from python_qt_binding.QtCore import QSize, QTimer\n'), ((1171, 127...
from __future__ import absolute_import, division, print_function import glob # Format expected by setup.py and doc/source/conf.py: string of form "X.Y.Z" _version_major = 0 _version_minor = 1 _version_micro = '' # use '' for first of series, number for 1 and above _version_extra = 'dev' # _version_extra = '' # Uncom...
[ "glob.glob" ]
[((2104, 2129), 'glob.glob', 'glob.glob', (['"""scripts/*.py"""'], {}), "('scripts/*.py')\n", (2113, 2129), False, 'import glob\n')]
# -*- coding: utf8 from gb.randomkit.random import RNG from gb.samplers import BaseSampler from gb.samplers import CollapsedGibbsSampler from gb.stamps import Timestamps from gb.sloppy import SloppyCounter from numpy.testing import assert_equal import numpy as np def test_get_probability(): d = {} d[0] = ...
[ "numpy.array", "gb.sloppy.SloppyCounter", "gb.randomkit.random.RNG", "gb.stamps.Timestamps" ]
[((379, 392), 'gb.stamps.Timestamps', 'Timestamps', (['d'], {}), '(d)\n', (389, 392), False, 'from gb.stamps import Timestamps\n'), ((654, 686), 'numpy.array', 'np.array', (['[5, 5]'], {'dtype': '"""uint64"""'}), "([5, 5], dtype='uint64')\n", (662, 686), True, 'import numpy as np\n'), ((704, 738), 'numpy.array', 'np.ar...
import tensorflow as tf """Class for KDD10 percent GAN architecture. Generator and discriminator. """ learning_rate = 0.00001 batch_size = 50 layer = 1 latent_dim = 32 dis_inter_layer_dim = 128 init_kernel = tf.contrib.layers.xavier_initializer() def generator(z_inp, is_training=False, getter=None, reuse=False): ...
[ "tensorflow.variable_scope", "tensorflow.nn.relu", "tensorflow.contrib.layers.xavier_initializer", "tensorflow.layers.dropout", "tensorflow.layers.dense", "tensorflow.squeeze" ]
[((213, 251), 'tensorflow.contrib.layers.xavier_initializer', 'tf.contrib.layers.xavier_initializer', ([], {}), '()\n', (249, 251), True, 'import tensorflow as tf\n'), ((607, 672), 'tensorflow.variable_scope', 'tf.variable_scope', (['"""generator"""'], {'reuse': 'reuse', 'custom_getter': 'getter'}), "('generator', reus...
#!/usr/bin/env python from setuptools import setup, find_packages setup( name='setuptools-wotmod', version='0.2', packages=find_packages(), description='setuptools integration for creating World of Tanks mods', long_description=open('README.md').read(), author='jhakonen', url='https://gith...
[ "setuptools.find_packages" ]
[((137, 152), 'setuptools.find_packages', 'find_packages', ([], {}), '()\n', (150, 152), False, 'from setuptools import setup, find_packages\n')]
from django.http import HttpResponse from django.shortcuts import render, redirect from lab1.views import login def docs(request): if not request.user.is_authenticated: return redirect(login.login_view) return render(request, 'docs.html')
[ "django.shortcuts.render", "django.shortcuts.redirect" ]
[((229, 257), 'django.shortcuts.render', 'render', (['request', '"""docs.html"""'], {}), "(request, 'docs.html')\n", (235, 257), False, 'from django.shortcuts import render, redirect\n'), ((191, 217), 'django.shortcuts.redirect', 'redirect', (['login.login_view'], {}), '(login.login_view)\n', (199, 217), False, 'from d...
import unittest import coverage from flask_script import Manager from project import create_app, db from project.api.models import User COV = coverage.coverage( branch=True, include='project/*', omit=[ 'project/tests/*', 'project/server/config.py', 'project/server/*/__init__.py' ...
[ "project.api.models.User", "project.db.drop_all", "project.db.create_all", "flask_script.Manager", "project.create_app", "coverage.coverage", "unittest.TextTestRunner", "unittest.TestLoader", "project.db.session.commit" ]
[((145, 286), 'coverage.coverage', 'coverage.coverage', ([], {'branch': '(True)', 'include': '"""project/*"""', 'omit': "['project/tests/*', 'project/server/config.py', 'project/server/*/__init__.py']"}), "(branch=True, include='project/*', omit=['project/tests/*',\n 'project/server/config.py', 'project/server/*/__i...
# ----------------------------------------------------------------------------- # Clorm ORM FactBase implementation. FactBase provides a set-like container # specifically for storing facts (Predicate instances). # ------------------------------------------------------------------------------ import abc import io impor...
[ "itertools.chain", "io.StringIO", "typing.cast", "itertools.groupby" ]
[((8102, 8156), 'itertools.groupby', 'itertools.groupby', (['sorted_facts', '(lambda x: x.__class__)'], {}), '(sorted_facts, lambda x: x.__class__)\n', (8119, 8156), False, 'import itertools\n'), ((16691, 16704), 'io.StringIO', 'io.StringIO', ([], {}), '()\n', (16702, 16704), False, 'import io\n'), ((15172, 15193), 'it...
# -*- coding: utf-8 -*- # # ***** BEGIN GPL LICENSE BLOCK ***** # # -------------------------------------------------------------------------- # Blender 2.5 Extensions Framework # -------------------------------------------------------------------------- # # Authors: # <NAME> # # This program is free software; you can ...
[ "bpy.utils.unregister_class", "bpy.ops.ef.msg", "bpy.utils.register_class", "time.strftime" ]
[((1023, 1058), 'bpy.utils.register_class', 'bpy.utils.register_class', (['EF_OT_msg'], {}), '(EF_OT_msg)\n', (1047, 1058), False, 'import bpy\n'), ((1450, 1498), 'bpy.ops.ef.msg', 'bpy.ops.ef.msg', ([], {'msg_type': '"""WARNING"""', 'msg_text': 'str'}), "(msg_type='WARNING', msg_text=str)\n", (1464, 1498), False, 'imp...
# pylint: disable=invalid-name ''' Pytests for the common utilities included in this package. Includes: - conversions.py - specs.py - utils.py To run the tests, type the following in the top level repo directory: python -m pytest --nat-file [path/to/gribfile] --prs-file [path/to/gribfile] ''' from...
[ "adb_graphics.conversions.m_to_mi", "adb_graphics.conversions.pa_to_hpa", "inspect.getfullargspec", "adb_graphics.conversions.m_to_kft", "adb_graphics.datahandler.gribdata.__getattribute__", "adb_graphics.conversions.weasd_to_1hsnw", "adb_graphics.conversions.k_to_f", "adb_graphics.conversions.ms_to_k...
[((830, 845), 'numpy.ones', 'np.ones', (['[3, 2]'], {}), '([3, 2])\n', (837, 845), True, 'import numpy as np\n'), ((928, 949), 'adb_graphics.conversions.k_to_c', 'conversions.k_to_c', (['a'], {}), '(a)\n', (946, 949), True, 'import adb_graphics.conversions as conversions\n'), ((989, 1010), 'adb_graphics.conversions.k_t...
from pyte import tokens, util from pyte.superclasses import _PyteAugmentedValidator, _PyteOp from pyte.util import PY36 class FOR_LOOP(_PyteOp): """ Represents a for loop. """ def __init__(self, iterator: _PyteAugmentedValidator, body: list): """ Represents a for operator. :p...
[ "pyte.util.generate_simple_call", "pyte.util.flatten", "pyte.util.ensure_instruction", "pyte.util.generate_bytecode_from_obb" ]
[((1050, 1106), 'pyte.util.generate_bytecode_from_obb', 'util.generate_bytecode_from_obb', (['self.iterator', 'previous'], {}), '(self.iterator, previous)\n', (1081, 1106), False, 'from pyte import tokens, util\n'), ((1172, 1225), 'pyte.util.generate_bytecode_from_obb', 'util.generate_bytecode_from_obb', (['tokens.GET_...
# This little bit of magic fills the __all__ list # with every plugin name, and means that calling: # from plugins import * # within inmembrane.py will import every plugin import pkgutil __all__ = [] for p in pkgutil.iter_modules(__path__): __all__.append(p[1])
[ "pkgutil.iter_modules" ]
[((211, 241), 'pkgutil.iter_modules', 'pkgutil.iter_modules', (['__path__'], {}), '(__path__)\n', (231, 241), False, 'import pkgutil\n')]
# DomirScire import math import os import random import re import sys import collections if __name__ == '__main__': s = sorted(input().strip()) s_counter = collections.Counter(s).most_common() s_counter = sorted(s_counter, key=lambda x: (x[1] * -1, x[0])) for i in range(0, 3): print(s_counter[i...
[ "collections.Counter" ]
[((165, 187), 'collections.Counter', 'collections.Counter', (['s'], {}), '(s)\n', (184, 187), False, 'import collections\n')]
import math import torch import torch.nn as nn from torch.nn import functional as F class KLRegression(nn.Module): """KL-divergence loss for probabilistic regression. It is computed using Monte Carlo (MC) samples from an arbitrary distribution.""" def __init__(self, eps=0.0): super()._...
[ "torch.log", "torch.mean", "math.log", "torch.sum", "torch.logsumexp" ]
[((725, 761), 'torch.log', 'torch.log', (['(sample_density + self.eps)'], {}), '(sample_density + self.eps)\n', (734, 761), False, 'import torch\n'), ((864, 939), 'torch.mean', 'torch.mean', (['(scores * (gt_density / (sample_density + self.eps)))'], {'dim': 'mc_dim'}), '(scores * (gt_density / (sample_density + self.e...
# coding=utf-8 import arrow from bs4 import BeautifulSoup from digesters.base_digester import BaseDigester TEMPLATE = """<html> <head> <meta content="text/html; charset=utf-8" http-equiv="Content-Type"/> <title>Atlassian HipChat</title> </head> <body style="box-sizing: border-box; height: 100%; width: 100%;...
[ "bs4.BeautifulSoup" ]
[((2916, 2958), 'bs4.BeautifulSoup', 'BeautifulSoup', (['html_message', '"""html.parser"""'], {}), "(html_message, 'html.parser')\n", (2929, 2958), False, 'from bs4 import BeautifulSoup\n')]
import os import shutil import tempfile import numpy as np from yt.loaders import load, load_uniform_grid from yt.testing import ( assert_array_equal, assert_fname, fake_random_ds, requires_file, requires_module, ) from yt.utilities.answer_testing.framework import data_dir_load from yt.visualizati...
[ "yt.loaders.load_uniform_grid", "yt.visualization.plot_window.SlicePlot", "yt.testing.assert_fname", "yt.testing.assert_array_equal", "numpy.arange", "yt.testing.fake_random_ds", "os.path.join", "yt.loaders.load", "os.getcwd", "os.chdir", "yt.utilities.answer_testing.framework.data_dir_load", ...
[((399, 422), 'yt.testing.requires_module', 'requires_module', (['"""h5py"""'], {}), "('h5py')\n", (414, 422), False, 'from yt.testing import assert_array_equal, assert_fname, fake_random_ds, requires_file, requires_module\n'), ((1300, 1323), 'yt.testing.requires_module', 'requires_module', (['"""h5py"""'], {}), "('h5p...
# Generated by Django 3.2.6 on 2021-09-05 19:39 from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('core', '0011_auto_20210905_1619'), ] operations = [ migrations.AlterField( model_name='preco', name='categoria', ...
[ "django.db.models.CharField" ]
[((335, 453), 'django.db.models.CharField', 'models.CharField', ([], {'choices': "[('Premium', 'C'), ('Pro', 'B'), ('Plus', 'A')]", 'max_length': '(15)', 'verbose_name': '"""categoria"""'}), "(choices=[('Premium', 'C'), ('Pro', 'B'), ('Plus', 'A')],\n max_length=15, verbose_name='categoria')\n", (351, 453), False, '...
#!/usr/bin/env python import panflute as pf """ Pandoc filter that causes emphasis to be rendered using the custom macro '\myemph{...}' rather than '\emph{...}' in latex. Other output formats are unaffected. """ def latex(s): return pf.RawInline(s, format='latex') def myemph(e, doc): if type(e)==pf.Emph a...
[ "panflute.RawInline", "panflute.toJSONFilter" ]
[((241, 272), 'panflute.RawInline', 'pf.RawInline', (['s'], {'format': '"""latex"""'}), "(s, format='latex')\n", (253, 272), True, 'import panflute as pf\n'), ((442, 465), 'panflute.toJSONFilter', 'pf.toJSONFilter', (['myemph'], {}), '(myemph)\n', (457, 465), True, 'import panflute as pf\n')]
#! /usr/bin/env python # -*- coding: utf-8 -*- # # Copyright 2008,2009 <NAME> <<EMAIL>> # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at# # # http://www.apache.org/licenses/LICENSE-2.0...
[ "couchdbkit.contrib.WSGIHandler", "json.dumps" ]
[((1108, 1124), 'couchdbkit.contrib.WSGIHandler', 'WSGIHandler', (['app'], {}), '(app)\n', (1119, 1124), False, 'from couchdbkit.contrib import WSGIHandler\n'), ((848, 883), 'json.dumps', 'json.dumps', (["environ['COUCHDB_INFO']"], {}), "(environ['COUCHDB_INFO'])\n", (858, 883), False, 'import json\n')]
import json from .life import BinaryLife class GOL(object): team_names: list = [] columns = 0 rows = 0 def __init__(self, **kwargs): self.load_config(**kwargs) self.create_life() def __repr__(self): s = [] s.append("+" + "-" * (self.columns) + "+") for i i...
[ "json.loads" ]
[((2829, 2849), 'json.loads', 'json.loads', (['self.ic1'], {}), '(self.ic1)\n', (2839, 2849), False, 'import json\n'), ((3045, 3065), 'json.loads', 'json.loads', (['self.ic2'], {}), '(self.ic2)\n', (3055, 3065), False, 'import json\n')]
from django.core.management.base import BaseCommand from django.db import connection class Command(BaseCommand): help = "Add missing id column for hostingstats." def handle(self, *args, **options): with connection.cursor() as cursor: self.cursor = cursor self.cursor.execute( ...
[ "django.db.connection.cursor" ]
[((222, 241), 'django.db.connection.cursor', 'connection.cursor', ([], {}), '()\n', (239, 241), False, 'from django.db import connection\n')]
import numba as nb import numpy as np import scipy.stats @nb.njit(parallel=True) def tiecorrect(rankvals): """ parallelized version of scipy.stats.tiecorrect :param rankvals: p x n array of ranked data (output of rankdata function) """ tc = np.ones(rankvals.shape[1], dtype=np.float64) for j i...
[ "numpy.sqrt", "numpy.ones", "numpy.float64", "numpy.log", "numba.njit", "numpy.asarray", "numpy.diff", "numpy.argsort", "numpy.sum", "numpy.errstate", "numpy.array", "numpy.empty", "numpy.concatenate", "numpy.nonzero", "numpy.maximum", "numba.prange", "numpy.all", "numpy.ravel" ]
[((60, 82), 'numba.njit', 'nb.njit', ([], {'parallel': '(True)'}), '(parallel=True)\n', (67, 82), True, 'import numba as nb\n'), ((718, 740), 'numba.njit', 'nb.njit', ([], {'parallel': '(True)'}), '(parallel=True)\n', (725, 740), True, 'import numba as nb\n'), ((264, 308), 'numpy.ones', 'np.ones', (['rankvals.shape[1]'...
#!/usr/bin/env python3 import codekit.pygithub import github import itertools import pytest @pytest.fixture def git_author(): return github.InputGitAuthor(name='foo', email='<EMAIL>') def test_init(git_author): """Test TargetTag object instantiation""" t_tag = codekit.pygithub.TargetTag( name=...
[ "github.InputGitAuthor", "pytest.raises" ]
[((140, 190), 'github.InputGitAuthor', 'github.InputGitAuthor', ([], {'name': '"""foo"""', 'email': '"""<EMAIL>"""'}), "(name='foo', email='<EMAIL>')\n", (161, 190), False, 'import github\n'), ((1377, 1406), 'pytest.raises', 'pytest.raises', (['AssertionError'], {}), '(AssertionError)\n', (1390, 1406), False, 'import p...
import functools def parse_input() -> list[list[str]]: groups = [[]] with open("input.txt") as file: for line in file: line_ = line.rstrip() if len(line_) > 0: groups[-1].append(line_) else: if not groups[-1] == []: ...
[ "functools.reduce" ]
[((915, 955), 'functools.reduce', 'functools.reduce', (['set.intersection', 'sets'], {}), '(set.intersection, sets)\n', (931, 955), False, 'import functools\n')]
import box from typing import Text import yaml def load_config(config_path: Text) -> box.ConfigBox: """Loads yaml config in instance of box.ConfigBox. Args: config_path {Text}: path to config Returns: box.ConfigBox """ with open(config_path) as config_file: config = yaml....
[ "box.ConfigBox", "yaml.safe_load" ]
[((315, 342), 'yaml.safe_load', 'yaml.safe_load', (['config_file'], {}), '(config_file)\n', (329, 342), False, 'import yaml\n'), ((360, 381), 'box.ConfigBox', 'box.ConfigBox', (['config'], {}), '(config)\n', (373, 381), False, 'import box\n')]
import boto3 from botocore.exceptions import ClientError import datetime import pytest from moto import mock_sagemaker from moto.sts.models import ACCOUNT_ID FAKE_ROLE_ARN = "arn:aws:iam::{}:role/FakeRole".format(ACCOUNT_ID) TEST_REGION_NAME = "us-east-1" class MyProcessingJobModel(object): def __init__( ...
[ "boto3.client", "pytest.raises" ]
[((3482, 3537), 'boto3.client', 'boto3.client', (['"""sagemaker"""'], {'region_name': 'TEST_REGION_NAME'}), "('sagemaker', region_name=TEST_REGION_NAME)\n", (3494, 3537), False, 'import boto3\n'), ((5306, 5356), 'boto3.client', 'boto3.client', (['"""sagemaker"""'], {'region_name': '"""us-east-1"""'}), "('sagemaker', re...
# -*- coding: utf-8 -*- # Form implementation generated from reading ui file 'direction_sprite_widget.ui' # # Created: Wed Jul 30 18:37:40 2014 # by: PyQt4 UI code generator 4.10.4 # # WARNING! All changes made in this file will be lost! from PyQt4 import QtCore, QtGui try: _fromUtf8 = QtCore.QString.fromUt...
[ "PyQt4.QtCore.QSize", "PyQt4.QtGui.QWidget", "PyQt4.QtGui.QScrollArea", "PyQt4.QtGui.QFormLayout", "PyQt4.QtGui.QSpinBox", "PyQt4.QtCore.QMetaObject.connectSlotsByName", "PyQt4.QtGui.QLabel", "PyQt4.QtGui.QPushButton", "PyQt4.QtGui.QComboBox", "PyQt4.QtCore.QRect", "PyQt4.QtGui.QVBoxLayout", "...
[((498, 562), 'PyQt4.QtGui.QApplication.translate', 'QtGui.QApplication.translate', (['context', 'text', 'disambig', '_encoding'], {}), '(context, text, disambig, _encoding)\n', (526, 562), False, 'from PyQt4 import QtCore, QtGui\n'), ((960, 998), 'PyQt4.QtGui.QWidget', 'QtGui.QWidget', (['direction_sprite_widget'], {}...
import nltk import string import argparse parser = argparse.ArgumentParser(description='.') parser.add_argument('-text', help='') parser.add_argument('-meta', help='') parser.add_argument('-output', help='') args = parser.parse_args() # parser.add_argument('-iter', dest='iter', type=int, # defaul...
[ "nltk.corpus.stopwords.words", "argparse.ArgumentParser" ]
[((53, 93), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {'description': '"""."""'}), "(description='.')\n", (76, 93), False, 'import argparse\n'), ((558, 596), 'nltk.corpus.stopwords.words', 'nltk.corpus.stopwords.words', (['"""english"""'], {}), "('english')\n", (585, 596), False, 'import nltk\n')]
from collections import deque import numpy as np import os from abc import ABCMeta, abstractmethod import random random.seed(42) from common import config, VehicleState from helper import Helper INFO = """Average merging time: {} s Traffic flow: {} vehicle/s Average speed: {} km/h Average fuel consumptio...
[ "matplotlib.pyplot.grid", "bokeh.plotting.figure", "helper.Helper.getTmOptimal2", "matplotlib.pyplot.ylabel", "numpy.random.rand", "helper.Helper.getConfigVec", "numpy.random.exponential", "numpy.array", "numpy.arange", "helper.Helper.getTc", "helper.Helper.getTimeMatrix", "collections.deque",...
[((118, 133), 'random.seed', 'random.seed', (['(42)'], {}), '(42)\n', (129, 133), False, 'import random\n'), ((2806, 2813), 'collections.deque', 'deque', ([], {}), '()\n', (2811, 2813), False, 'from collections import deque\n'), ((5912, 5936), 'numpy.arange', 'np.arange', (['(10)', '(30.1)', '(2.0)'], {}), '(10, 30.1, ...
import pytest import datetime import json import functools from urllib.parse import urlencode, parse_qs from descarteslabs.common.graft import client as graft_client from ... import types from .. import tile_url def test_url(): base = "foo" base_q = base + "?" url = functools.partial(tile_url.tile_url...
[ "datetime.datetime", "descarteslabs.common.graft.client.consistent_guid", "json.dumps", "urllib.parse.parse_qs", "functools.partial", "pytest.raises", "urllib.parse.urlencode" ]
[((1587, 1648), 'urllib.parse.parse_qs', 'parse_qs', (['params'], {'strict_parsing': '(True)', 'keep_blank_values': '(True)'}), '(params, strict_parsing=True, keep_blank_values=True)\n', (1595, 1648), False, 'from urllib.parse import urlencode, parse_qs\n'), ((2434, 2482), 'functools.partial', 'functools.partial', (['t...
# Generated by Django 3.0.2 on 2020-01-31 20:33 from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('collections', '0002_auto_20200109_1348'), ] operations = [ migrations.AlterField( model_name='collection', name='com...
[ "django.db.models.DateTimeField", "django.db.models.CharField", "django.db.models.BooleanField" ]
[((345, 442), 'django.db.models.CharField', 'models.CharField', ([], {'help_text': '"""Comment for collection."""', 'max_length': '(256)', 'verbose_name': '"""Comment"""'}), "(help_text='Comment for collection.', max_length=256,\n verbose_name='Comment')\n", (361, 442), False, 'from django.db import migrations, mode...
# -*- coding: utf-8 -*- import os import torch from torch.autograd import Variable import numpy as np import scipy import matplotlib.pyplot as plt import cv2 import scipy.ndimage import shutil import scipy.misc as misc from PIL import Image def mkdirs(folders, erase=False): if type(folders) is not list: ...
[ "torch.LongTensor", "torch.cuda.is_available", "scipy.misc.imresize", "numpy.arange", "matplotlib.pyplot.imshow", "os.path.exists", "numpy.mean", "numpy.reshape", "scipy.misc.imsave", "numpy.max", "numpy.stack", "numpy.min", "numpy.random.permutation", "numpy.tile", "torch.abs", "itert...
[((724, 747), 'os.path.exists', 'os.path.exists', (['imgfile'], {}), '(imgfile)\n', (738, 747), False, 'import os\n'), ((799, 818), 'cv2.imread', 'cv2.imread', (['imgfile'], {}), '(imgfile)\n', (809, 818), False, 'import cv2\n'), ((833, 872), 'cv2.cvtColor', 'cv2.cvtColor', (['srcBGR', 'cv2.COLOR_BGR2RGB'], {}), '(srcB...
from flask import render_template from app import app, html_generator import app.calculator.calculator as calc from app.route_helpers import units_from_form, options_from_form, options_list, flash_errors from app.forms import InputForm from collections import defaultdict @app.route('/', methods=['GET', 'POST']) @app....
[ "flask.render_template", "app.route_helpers.units_from_form", "app.html_generator.make_boxes", "app.forms.InputForm", "app.route_helpers.options_list", "app.calculator.calculator.calculate", "collections.defaultdict", "app.app.route", "app.route_helpers.flash_errors", "app.route_helpers.options_fr...
[((275, 314), 'app.app.route', 'app.route', (['"""/"""'], {'methods': "['GET', 'POST']"}), "('/', methods=['GET', 'POST'])\n", (284, 314), False, 'from app import app, html_generator\n'), ((316, 360), 'app.app.route', 'app.route', (['"""/index"""'], {'methods': "['GET', 'POST']"}), "('/index', methods=['GET', 'POST'])\...
import re def find_indices(): return [m.start(0) for m in re.finditer(reg, content)] def find_content(): return re.findall(reg, content) if __name__ == "__main__": content = 'an example word:cat and word:dog' reg = r'word:\w' print(find_indices()) print(find_content())
[ "re.findall", "re.finditer" ]
[((124, 148), 're.findall', 're.findall', (['reg', 'content'], {}), '(reg, content)\n', (134, 148), False, 'import re\n'), ((64, 89), 're.finditer', 're.finditer', (['reg', 'content'], {}), '(reg, content)\n', (75, 89), False, 'import re\n')]
import smart_imports smart_imports.all() class AccountPrototypeTests(utils_testcase.TestCase, personal_messages_helpers.Mixin): def setUp(self): super(AccountPrototypeTests, self).setUp() self.place_1, self.place_2, self.place_3 = game_logic.create_test_map() self.account = self.account...
[ "smart_imports.all" ]
[((23, 42), 'smart_imports.all', 'smart_imports.all', ([], {}), '()\n', (40, 42), False, 'import smart_imports\n')]
#!/usr/bin/env python3 import json import time from run_common import AWSCli from run_common import print_message from run_create_codebuild_common import create_base_iam_policy from run_create_codebuild_common import create_iam_service_role from run_create_codebuild_common import create_managed_secret_iam_policy from ...
[ "run_common.AWSCli", "run_create_codebuild_common.create_managed_secret_iam_policy", "run_common.print_message", "json.dumps", "run_create_codebuild_common.create_notification_rule", "run_create_codebuild_common.get_notification_rule", "time.sleep", "run_create_codebuild_common.update_notification_rul...
[((632, 662), 'run_common.AWSCli', 'AWSCli', (["settings['AWS_REGION']"], {}), "(settings['AWS_REGION'])\n", (638, 662), False, 'from run_common import AWSCli\n'), ((1035, 1074), 'run_common.print_message', 'print_message', (['"""check previous version"""'], {}), "('check previous version')\n", (1048, 1074), False, 'fr...
import typing as t from pylogview import datefinder from pylogview.record import LogRecord if t.TYPE_CHECKING: from pylogview.window import Window class LogReader: __slots__ = [ "_window", "filename", "_lines", "records", "_record_prefix_length", "_fd", ...
[ "pylogview.datefinder.find_dates" ]
[((4113, 4174), 'pylogview.datefinder.find_dates', 'datefinder.find_dates', (['buffer_string'], {'source': '(True)', 'index': '(True)'}), '(buffer_string, source=True, index=True)\n', (4134, 4174), False, 'from pylogview import datefinder\n'), ((5490, 5551), 'pylogview.datefinder.find_dates', 'datefinder.find_dates', (...
from mailmerge import MailMerge import re import os.path from ElvantoAPIExtensions import Enums, Helpers from modules.__stub__ import ModuleStub class Module(ModuleStub): __VERSION__ = "1.0" __NAME__ = "bulletinGenerator_Kingsgrove" # __executeTime__ = "16:00" # __executeDay__ = "thursday" setting...
[ "re.search", "math.ceil", "re.sub", "ElvantoAPIExtensions.Helpers.NextDate", "ElvantoAPIExtensions.Helpers.ServicesOnDate", "mailmerge.MailMerge" ]
[((903, 938), 'ElvantoAPIExtensions.Helpers.NextDate', 'Helpers.NextDate', (['Enums.Days.SUNDAY'], {}), '(Enums.Days.SUNDAY)\n', (919, 938), False, 'from ElvantoAPIExtensions import Enums, Helpers\n'), ((5204, 5272), 're.search', 're.search', (['"""^Bible Reading (?:- )?(.*)$"""', 'scripturePassageItem.title'], {}), "(...
"""plot.py: Utility builder class for ML plots. Uses scikit-learn code samples and framework """ __author__ = "<NAME>" __license__ = "BSD" __email__ = "<EMAIL>" import numpy as np import pandas as pd import matplotlib.pyplot as plt from mpl_toolkits.mplot3d import Axes3D import randomcolor import math from sklearn.me...
[ "matplotlib.pyplot.ylabel", "sklearn.metrics.auc", "numpy.argsort", "numpy.array", "sklearn.metrics.roc_curve", "randomcolor.RandomColor", "numpy.arange", "matplotlib.pyplot.xlabel", "matplotlib.pyplot.plot", "matplotlib.pyplot.close", "matplotlib.pyplot.ylim", "matplotlib.pyplot.xticks", "m...
[((697, 707), 'matplotlib.pyplot.show', 'plt.show', ([], {}), '()\n', (705, 707), True, 'import matplotlib.pyplot as plt\n'), ((2273, 2298), 'randomcolor.RandomColor', 'randomcolor.RandomColor', ([], {}), '()\n', (2296, 2298), False, 'import randomcolor\n'), ((3572, 3597), 'randomcolor.RandomColor', 'randomcolor.Random...
#! /usr/bin/env python # -*- coding: utf-8 -* import collections from census_data_downloader.core.tables import BaseTableConfig from census_data_downloader.core.decorators import register @register class MedianAgeDownloader(BaseTableConfig): PROCESSED_TABLE_NAME = 'medianage' UNIVERSE = "total population" ...
[ "collections.OrderedDict" ]
[((373, 447), 'collections.OrderedDict', 'collections.OrderedDict', (["{'001': 'median', '002': 'male', '003': 'female'}"], {}), "({'001': 'median', '002': 'male', '003': 'female'})\n", (396, 447), False, 'import collections\n')]
# Copyright 2018 Bloomberg Finance L.P. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to i...
[ "prometheus_client.Counter" ]
[((943, 1056), 'prometheus_client.Counter', 'Counter', (['POD_KILLS_METRIC_NAME', '"""Number of pods killed (including failures)"""', "['status', 'namespace', 'name']"], {}), "(POD_KILLS_METRIC_NAME, 'Number of pods killed (including failures)',\n ['status', 'namespace', 'name'])\n", (950, 1056), False, 'from promet...
import dectate import pytest from kaybee.plugins.widgets.directive import WidgetDirective from kaybee.plugins.widgets.action import WidgetAction class Dummy: pass @pytest.fixture() def widgets_kb_app(): class widgets_kb_app(dectate.App): widget = dectate.directive(WidgetAction) yield widgets_k...
[ "pytest.fixture", "dectate.directive" ]
[((173, 189), 'pytest.fixture', 'pytest.fixture', ([], {}), '()\n', (187, 189), False, 'import pytest\n'), ((329, 345), 'pytest.fixture', 'pytest.fixture', ([], {}), '()\n', (343, 345), False, 'import pytest\n'), ((465, 481), 'pytest.fixture', 'pytest.fixture', ([], {}), '()\n', (479, 481), False, 'import pytest\n'), (...
import os import time import random import scipy.sparse as sp import numpy as np import tensorflow as tf import argparse from models import SpHGAT from utils import process parser = argparse.ArgumentParser() parser.add_argument('--dataset', help='Dataset.', default='imdb', type=str) parser.add_argument('--epochs', he...
[ "tensorflow.local_variables_initializer", "numpy.random.standard_normal", "tensorflow.sparse_placeholder", "utils.process.load_heterogeneous_data", "numpy.array", "utils.process.preprocess_features", "tensorflow.Graph", "argparse.ArgumentParser", "tensorflow.placeholder", "tensorflow.Session", "...
[((184, 209), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {}), '()\n', (207, 209), False, 'import argparse\n'), ((2336, 2377), 'tensorflow.ConfigProto', 'tf.ConfigProto', ([], {'allow_soft_placement': '(True)'}), '(allow_soft_placement=True)\n', (2350, 2377), True, 'import tensorflow as tf\n'), ((3550, 3...
from django import forms from .models import Problem class ProblemForm(forms.ModelForm): options = ( ('A', 'A'), ('B', 'B'), ('C', 'C'), ('D', 'D'), ) choice = forms.ChoiceField(choices=options) class Meta: model = Problem fields = '__all__' # exclude = ['answer'] widgets = { 'answer': forms.Hid...
[ "django.forms.ChoiceField", "django.forms.HiddenInput", "django.forms.Textarea", "django.forms.TextInput" ]
[((172, 206), 'django.forms.ChoiceField', 'forms.ChoiceField', ([], {'choices': 'options'}), '(choices=options)\n', (189, 206), False, 'from django import forms\n'), ((311, 330), 'django.forms.HiddenInput', 'forms.HiddenInput', ([], {}), '()\n', (328, 330), False, 'from django import forms\n'), ((350, 421), 'django.for...
#!/usr/bin/env python """ Create a spreadsheet with two tables, using some named ranges. """ import os from odfdo import Document, Table if __name__ == "__main__": document = Document('spreadsheet') body = document.body table = Table("First Table") body.append(table) # populate the table : for...
[ "os.path.exists", "odfdo.Document", "os.path.join", "os.mkdir", "odfdo.Table" ]
[((181, 204), 'odfdo.Document', 'Document', (['"""spreadsheet"""'], {}), "('spreadsheet')\n", (189, 204), False, 'from odfdo import Document, Table\n'), ((242, 262), 'odfdo.Table', 'Table', (['"""First Table"""'], {}), "('First Table')\n", (247, 262), False, 'from odfdo import Document, Table\n'), ((1017, 1038), 'odfdo...
from nltk.corpus.reader import CategorizedPlaintextCorpusReader from nltk.tokenize.casual import TweetTokenizer from nltk.classify.scikitlearn import SklearnClassifier from sklearn.naive_bayes import BernoulliNB, MultinomialNB from sklearn.svm import SVC, LinearSVC, NuSVC, LinearSVR, NuSVR from sklearn.linear_model im...
[ "nltk.pos_tag", "utils.precision_recall_2step", "nltk.corpus.reader.CategorizedPlaintextCorpusReader", "featureExtractors.unigramsFeatures", "featureExtractors.bigramsFeatures", "pickle.load", "lexicons.mpqa.mpqaDictionary.MpqaDictionaryWrapper", "normalization.normalizeTwitterWordsWithExtraFeatures",...
[((831, 903), 'nltk.tokenize.casual.TweetTokenizer', 'TweetTokenizer', ([], {'reduce_len': '(True)', 'preserve_case': '(True)', 'strip_handles': '(False)'}), '(reduce_len=True, preserve_case=True, strip_handles=False)\n', (845, 903), False, 'from nltk.tokenize.casual import TweetTokenizer\n'), ((917, 1089), 'nltk.corpu...
# # ⚠ Warning # # THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT # LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN # NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIA...
[ "time.time", "decimal.Decimal" ]
[((7684, 7695), 'time.time', 'time.time', ([], {}), '()\n', (7693, 7695), False, 'import time\n'), ((5951, 5965), 'decimal.Decimal', 'Decimal', (['""".01"""'], {}), "('.01')\n", (5958, 5965), False, 'from decimal import Decimal\n'), ((6025, 6039), 'decimal.Decimal', 'Decimal', (['""".01"""'], {}), "('.01')\n", (6032, 6...
#!/usr/bin/env python # -*- coding: utf-8 -*- # @Author : yag8009 # @FileName : md5_hede # @Time : 2020/3/18 import hashlib import time def md5_hede(md5data): md5 = hashlib.md5() # 使用MD5加密模式 md5.update(md5data.encode("utf8")) # 将参数字符串传入 sign = md5.hexdigest() return sign if __name__ == '__mai...
[ "time.time", "hashlib.md5" ]
[((177, 190), 'hashlib.md5', 'hashlib.md5', ([], {}), '()\n', (188, 190), False, 'import hashlib\n'), ((573, 584), 'time.time', 'time.time', ([], {}), '()\n', (582, 584), False, 'import time\n')]
import os import datetime from dateutil.parser import parse as dateutil_parser from jinja2 import Environment, PackageLoader from kinto_http import cli_utils from . import constants from .logger import logger JSON_DATE_FORMAT = "%Y-%m-%dT%H:%M:%SZ" COLLECTION_FORMAT = '/buckets/{bucket_id}/collections/{collection_i...
[ "dateutil.parser.parse", "os.path.exists", "jinja2.Environment", "os.makedirs", "os.path.join", "kinto_http.cli_utils.add_parser_options", "datetime.date.fromtimestamp", "kinto_http.cli_utils.create_client_from_args", "kinto_http.cli_utils.setup_logger", "jinja2.PackageLoader" ]
[((359, 381), 'dateutil.parser.parse', 'dateutil_parser', (['value'], {}), '(value)\n', (374, 381), True, 'from dateutil.parser import parse as dateutil_parser\n'), ((459, 498), 'jinja2.PackageLoader', 'PackageLoader', (['"""amo2kinto"""', '"""templates"""'], {}), "('amo2kinto', 'templates')\n", (472, 498), False, 'fro...
"""Make some fields on Chart and Table nullable We want to copy chart and table data across to these tables but have no way to add a classification for each one, so we'll have to live with some nulls in here. Revision ID: 2019_03_04_make_fields_nullable Revises: 2019_03_04_chart_table_settings Create Date: 2019-03-05...
[ "sqlalchemy.VARCHAR", "sqlalchemy.BOOLEAN" ]
[((657, 679), 'sqlalchemy.VARCHAR', 'sa.VARCHAR', ([], {'length': '(255)'}), '(length=255)\n', (667, 679), True, 'import sqlalchemy as sa\n'), ((765, 777), 'sqlalchemy.BOOLEAN', 'sa.BOOLEAN', ([], {}), '()\n', (775, 777), True, 'import sqlalchemy as sa\n'), ((867, 879), 'sqlalchemy.BOOLEAN', 'sa.BOOLEAN', ([], {}), '()...
from django.contrib import admin from .models import ContactForm class ContactFormAdmin(admin.ModelAdmin): class Meta: model = ContactForm admin.site.register(ContactForm, ContactFormAdmin)
[ "django.contrib.admin.site.register" ]
[((154, 204), 'django.contrib.admin.site.register', 'admin.site.register', (['ContactForm', 'ContactFormAdmin'], {}), '(ContactForm, ContactFormAdmin)\n', (173, 204), False, 'from django.contrib import admin\n')]
import numpy as np import h5py import os import sys from copy import deepcopy #handle .(period) and slash specially since it is part of path #replace with \period or \slash-forward when store, recover later #not using '\forward-slash' is because \f is a special character PERIOD='\period' SLASH='\slash-forward' ''' ...
[ "os.path.join", "h5py.File", "numpy.array", "sys.stdout.flush", "sys.stdout.write" ]
[((892, 915), 'h5py.File', 'h5py.File', (['f_name', 'mode'], {}), '(f_name, mode)\n', (901, 915), False, 'import h5py\n'), ((1759, 1781), 'h5py.File', 'h5py.File', (['f_name', '"""r"""'], {}), "(f_name, 'r')\n", (1768, 1781), False, 'import h5py\n'), ((1923, 1943), 'numpy.array', 'np.array', (['group[key]'], {}), '(gro...
import re import string from libs.dataset.core import SampleEntry, oov_replacement_vocabulary, letter_replacements __word_start_regex = f'[ \t\n]|^|[{string.punctuation}]' __word_end_regex = f'[ \t\n]|$|[{string.punctuation}]' def does_not_have_numbers(s: SampleEntry): return not any(char.isdigit() for char in s...
[ "re.sub", "libs.dataset.core.oov_replacement_vocabulary.items", "libs.dataset.core.letter_replacements.items" ]
[((656, 690), 'libs.dataset.core.oov_replacement_vocabulary.items', 'oov_replacement_vocabulary.items', ([], {}), '()\n', (688, 690), False, 'from libs.dataset.core import SampleEntry, oov_replacement_vocabulary, letter_replacements\n'), ((758, 786), 're.sub', 're.sub', (['"""[ ]{2,}"""', '""" """', 'line'], {}), "('[ ...
import unittest import silasdk from tests.test_config import ( app, eth_private_key, eth_private_key_4, instant_ach_handle, user_handle) class Test007CheckInstantAchTest(unittest.TestCase): def test_check_instant_ach(self): payload = { "user_handle": instant_ach_handle, "accou...
[ "unittest.main", "silasdk.User.check_instant_ach" ]
[((601, 616), 'unittest.main', 'unittest.main', ([], {}), '()\n', (614, 616), False, 'import unittest\n'), ((376, 439), 'silasdk.User.check_instant_ach', 'silasdk.User.check_instant_ach', (['app', 'payload', 'eth_private_key_4'], {}), '(app, payload, eth_private_key_4)\n', (406, 439), False, 'import silasdk\n')]
# Copyright 2021 Google LLC # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, ...
[ "bigbench.api.util.postprocess_output", "numpy.log" ]
[((1171, 1239), 'bigbench.api.util.postprocess_output', 'util.postprocess_output', (['text', 'max_length', 'stop_string', 'output_regex'], {}), '(text, max_length, stop_string, output_regex)\n', (1194, 1239), True, 'import bigbench.api.util as util\n'), ((1417, 1436), 'numpy.log', 'np.log', (['(1 / (i + 1))'], {}), '(1...
"""Click parameter types for osxphotos CLI""" import datetime import os import pathlib import re import bitmath import click import pytimeparse2 from osxphotos.export_db_utils import export_db_get_version from osxphotos.photoinfo import PhotoInfoNone from osxphotos.phototemplate import PhotoTemplate, RenderOptions fr...
[ "osxphotos.photoinfo.PhotoInfoNone", "osxphotos.utils.expand_and_validate_filepath", "osxphotos.utils.load_function", "pytimeparse2.parse", "pathlib.Path", "re.match", "os.getcwd", "click.BadParameter", "bitmath.Byte", "osxphotos.phototemplate.RenderOptions", "osxphotos.timeutils.time_string_to_...
[((2547, 2585), 'osxphotos.utils.expand_and_validate_filepath', 'expand_and_validate_filepath', (['filename'], {}), '(filename)\n', (2575, 2585), False, 'from osxphotos.utils import expand_and_validate_filepath, load_function\n'), ((5048, 5083), 're.match', 're.match', (['"""^[+-]?\\\\s*?\\\\d+$"""', 'value'], {}), "('...
# -*- coding: utf-8 -*- import datetime from south.db import db from south.v2 import SchemaMigration from django.db import models class Migration(SchemaMigration): def forwards(self, orm): # Removing unique constraint on 'Licence', fields ['sn'] db.delete_unique('ralph_assets_licence', ['sn']) ...
[ "south.db.db.delete_column", "south.db.db.send_create_signal", "south.db.db.delete_unique", "django.db.models.ForeignKey", "south.db.db.create_unique", "django.db.models.AutoField", "south.db.db.delete_table" ]
[((269, 317), 'south.db.db.delete_unique', 'db.delete_unique', (['"""ralph_assets_licence"""', "['sn']"], {}), "('ralph_assets_licence', ['sn'])\n", (285, 317), False, 'from south.db import db\n'), ((1526, 1587), 'south.db.db.send_create_signal', 'db.send_create_signal', (['"""ralph_assets"""', "['TransitionsHistory']"...
import connexion import six from openapi_server import query_manager from openapi_server.utils.vars import VARIABLE_TYPE_NAME, VARIABLE_TYPE_URI from openapi_server.models.variable import Variable # noqa: E501 from openapi_server import util def variables_get(username=None, label=None, page=None, per_page=None): # ...
[ "openapi_server.query_manager.delete_resource", "openapi_server.query_manager.post_resource", "openapi_server.query_manager.get_resource", "connexion.request.get_json", "openapi_server.query_manager.put_resource" ]
[((776, 953), 'openapi_server.query_manager.get_resource', 'query_manager.get_resource', ([], {'username': 'username', 'label': 'label', 'page': 'page', 'per_page': 'per_page', 'rdf_type_uri': 'VARIABLE_TYPE_URI', 'rdf_type_name': 'VARIABLE_TYPE_NAME', 'kls': 'Variable'}), '(username=username, label=label, page=page,\n...
import pycxsimulator from pylab import * import copy as cp nr = 500. # carrying capacity of rabbits r_init = 100 # initial rabbit population mr = 0.03 # magnitude of movement of rabbits dr = 1.0 # death rate of rabbits when it faces foxes rr = 0.1 # reproduction rate of rabbits f_init = 30 # initial fox population ...
[ "copy.copy", "pycxsimulator.GUI" ]
[((2420, 2439), 'pycxsimulator.GUI', 'pycxsimulator.GUI', ([], {}), '()\n', (2437, 2439), False, 'import pycxsimulator\n'), ((1988, 1999), 'copy.copy', 'cp.copy', (['ag'], {}), '(ag)\n', (1995, 1999), True, 'import copy as cp\n'), ((2267, 2278), 'copy.copy', 'cp.copy', (['ag'], {}), '(ag)\n', (2274, 2278), True, 'impor...
# Copyright 2021 The ML Collections Authors. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed...
[ "pickle.dumps", "ml_collections.FrozenConfigDict", "ml_collections.ConfigDict", "absl.testing.absltest.main", "copy.deepcopy", "ml_collections.FieldReference" ]
[((1167, 1208), 'ml_collections.FieldReference', 'ml_collections.FieldReference', (["{'int': 0}"], {}), "({'int': 0})\n", (1196, 1208), False, 'import ml_collections\n'), ((1249, 1274), 'copy.deepcopy', 'copy.deepcopy', (['_TEST_DICT'], {}), '(_TEST_DICT)\n', (1262, 1274), False, 'import copy\n'), ((1310, 1347), 'ml_co...
#!/usr/bin/env python # -*- coding: UTF-8 -*- ''' パッケージpyserialをインストールすること pytho2.x系で動作(python3.*系も動作検証済み) Creater:<NAME> ''' import serial import binascii import signal import sys import platform from serial.tools import list_ports #platformの切り替え if platform.system() == 'Windows': #windows用 ports = list_ports.co...
[ "signal.signal", "serial.tools.list_ports.comports", "platform.system", "serial.Serial", "sys.exit" ]
[((252, 269), 'platform.system', 'platform.system', ([], {}), '()\n', (267, 269), False, 'import platform\n'), ((307, 328), 'serial.tools.list_ports.comports', 'list_ports.comports', ([], {}), '()\n', (326, 328), False, 'from serial.tools import list_ports\n'), ((793, 803), 'sys.exit', 'sys.exit', ([], {}), '()\n', (80...
from test.webdnn_test.graph_test.operators_test.util import template_test_unary_operator from webdnn.graph.operators.sigmoid import Sigmoid def template(): template_test_unary_operator(Sigmoid) def test(): template()
[ "test.webdnn_test.graph_test.operators_test.util.template_test_unary_operator" ]
[((162, 199), 'test.webdnn_test.graph_test.operators_test.util.template_test_unary_operator', 'template_test_unary_operator', (['Sigmoid'], {}), '(Sigmoid)\n', (190, 199), False, 'from test.webdnn_test.graph_test.operators_test.util import template_test_unary_operator\n')]
import pybullet_data import pybullet as p import time import numpy as np from src.utils_geom import * from src.utils_depth import * from src.panda import Panda def full_jacob_pb(jac_t, jac_r): return np.vstack((jac_t[0], jac_t[1], jac_t[2], jac_r[0], jac_r[1], jac_r[2])) class pandaEnv(): def __init__(self, ...
[ "pybullet_data.getDataPath", "numpy.linalg.pinv", "numpy.hstack", "pybullet.setTimeStep", "pybullet.setGravity", "numpy.array", "pybullet.setPhysicsEngineParameter", "src.panda.Panda", "pybullet.calculateJacobian", "pybullet.createConstraint", "numpy.vstack", "pybullet.changeConstraint", "py...
[((207, 278), 'numpy.vstack', 'np.vstack', (['(jac_t[0], jac_t[1], jac_t[2], jac_r[0], jac_r[1], jac_r[2])'], {}), '((jac_t[0], jac_t[1], jac_t[2], jac_r[0], jac_r[1], jac_r[2]))\n', (216, 278), True, 'import numpy as np\n'), ((333, 360), 'pybullet_data.getDataPath', 'pybullet_data.getDataPath', ([], {}), '()\n', (358,...
import datetime import decimal import enum import typing as T import uuid import graphene import graphene.types import pydantic import pytest from pydantic import BaseModel, create_model import graphene_pydantic.converters as converters from graphene_pydantic.converters import ConversionError, convert_pydantic_field ...
[ "datetime.datetime", "datetime.time", "uuid.uuid4", "pytest.raises", "datetime.date", "pydantic.create_model", "graphene_pydantic.registry.get_global_registry", "decimal.Decimal" ]
[((558, 589), 'pydantic.create_model', 'create_model', (['"""model"""'], {}), "('model', **kwargs)\n", (570, 589), False, 'from pydantic import BaseModel, create_model\n'), ((781, 820), 'graphene_pydantic.registry.get_global_registry', 'get_global_registry', (['PydanticObjectType'], {}), '(PydanticObjectType)\n', (800,...
import numpy as np import pandas as pd from pathlib import Path import multiprocessing as mp from functools import partial from src.audio import read_as_melspectrogram from src.utils import get_params_hash from src import config NOISE_SOUNDS = [ 'Buzz', 'Car_passing_by', 'Crackle', 'Cricket', 'Hi...
[ "pandas.read_csv", "pathlib.Path", "src.audio.read_as_melspectrogram", "multiprocessing.cpu_count", "functools.partial", "numpy.save" ]
[((906, 953), 'src.audio.read_as_melspectrogram', 'read_as_melspectrogram', (['file_path', 'audio_params'], {}), '(file_path, audio_params)\n', (928, 953), False, 'from src.audio import read_as_melspectrogram\n'), ((1194, 1208), 'pathlib.Path', 'Path', (['dir_path'], {}), '(dir_path)\n', (1198, 1208), False, 'from path...
# Copyright (c) 2015, Frappe Technologies Pvt. Ltd. and Contributors # License: GNU General Public License v3. See license.txt from __future__ import unicode_literals import frappe, unittest from erpnext.accounts.doctype.fiscal_year.fiscal_year import FiscalYearIncorrectDate test_records = frappe.get_test_records('...
[ "frappe.db.exists", "frappe.delete_doc", "frappe.get_test_records", "frappe.get_doc" ]
[((295, 333), 'frappe.get_test_records', 'frappe.get_test_records', (['"""Fiscal Year"""'], {}), "('Fiscal Year')\n", (318, 333), False, 'import frappe, unittest\n'), ((436, 493), 'frappe.db.exists', 'frappe.db.exists', (['"""Fiscal Year"""', '"""_Test Fiscal Year 2000"""'], {}), "('Fiscal Year', '_Test Fiscal Year 200...
from typing import Callable, Generator, Any from unittest import mock import pytest from screenpy import AnActor, pacing, settings from screenpy.abilities import AuthenticateWith2FA, BrowseTheWeb, MakeAPIRequests from screenpy.narration.narrator import Narrator @pytest.fixture(scope="function") def Tester() -> AnAc...
[ "pytest.fixture", "unittest.mock.MagicMock", "unittest.mock.Mock", "screenpy.AnActor.named" ]
[((267, 299), 'pytest.fixture', 'pytest.fixture', ([], {'scope': '"""function"""'}), "(scope='function')\n", (281, 299), False, 'import pytest\n'), ((717, 749), 'pytest.fixture', 'pytest.fixture', ([], {'scope': '"""function"""'}), "(scope='function')\n", (731, 749), False, 'import pytest\n'), ((1021, 1053), 'pytest.fi...
import numpy as np import pandas as pd from sklearn.externals import joblib #from sklearn.ensemble import RandomForestRegressor #from sklearn.multioutput import MultiOutputRegressor #from sklearn.multioutput import MultiOutputRegressor from sklearn.model_selection import train_test_split df = pd.read_csv('https://dr...
[ "numpy.abs", "numpy.mean", "pandas.read_csv", "sklearn.externals.joblib.load", "pandas.get_dummies" ]
[((297, 411), 'pandas.read_csv', 'pd.read_csv', (['"""https://drive.google.com/uc?export=download&id=1XoV8SfvHmzaxRuDRe81OWSQu10dYTbO5"""'], {'sep': '""","""'}), "(\n 'https://drive.google.com/uc?export=download&id=1XoV8SfvHmzaxRuDRe81OWSQu10dYTbO5'\n , sep=',')\n", (308, 411), True, 'import pandas as pd\n'), ((4...
#! /usr/bin/env python # -*- coding: utf-8 -*- import os import argparse import docx from docx.shared import Cm import pylatex from pytablewriter import MarkdownTableWriter def set_column_width(column, width): column.width = width for cell in column.cells: cell.width = width def generate_word_revisio...
[ "os.path.exists", "pylatex.Section", "argparse.ArgumentParser", "pylatex.LongTable", "docx.shared.Cm", "pytablewriter.MarkdownTableWriter", "pylatex.Document", "docx.Document" ]
[((389, 404), 'docx.Document', 'docx.Document', ([], {}), '()\n', (402, 404), False, 'import docx\n'), ((1377, 1428), 'pylatex.Document', 'pylatex.Document', ([], {'geometry_options': 'geometry_options'}), '(geometry_options=geometry_options)\n', (1393, 1428), False, 'import pylatex\n'), ((2242, 2263), 'pytablewriter.M...
import torch import numpy as np import utility from decimal import Decimal from tqdm import tqdm from option import args from torchvision import transforms from PIL import Image import matplotlib matplotlib.use('TkAgg') import matplotlib.pyplot as plt import copy class Trainer(): def __init__(self, opt, loader...
[ "utility.timer", "utility.quantize", "utility.calc_psnr", "matplotlib.use", "utility.make_dual_scheduler", "tqdm.tqdm", "utility.SSIM", "utility.make_optimizer", "utility.make_dual_optimizer", "utility.make_scheduler", "torch.no_grad", "torch.zeros", "decimal.Decimal", "torch.device" ]
[((200, 223), 'matplotlib.use', 'matplotlib.use', (['"""TkAgg"""'], {}), "('TkAgg')\n", (214, 223), False, 'import matplotlib\n'), ((601, 640), 'utility.make_optimizer', 'utility.make_optimizer', (['opt', 'self.model'], {}), '(opt, self.model)\n', (623, 640), False, 'import utility\n'), ((666, 709), 'utility.make_sched...
# Get the database using the method we defined in pymongo_test_insert file from pymongo_test_insert import get_database dbname = get_database() # Create a new collection collection_name = dbname["user_1_items"] item_details = collection_name.find() for item in item_details: # This will give readable output...
[ "pandas.DataFrame", "pymongo_test_insert.get_database" ]
[((132, 146), 'pymongo_test_insert.get_database', 'get_database', ([], {}), '()\n', (144, 146), False, 'from pymongo_test_insert import get_database\n'), ((710, 733), 'pandas.DataFrame', 'DataFrame', (['item_details'], {}), '(item_details)\n', (719, 733), False, 'from pandas import DataFrame\n')]
import json from django.core.serializers.json import DjangoJSONEncoder from django.http import HttpResponse from django.template import Context from django.template import RequestContext from django.template.loader import render_to_string, select_template from django.utils.encoding import force_unicode from ..compat i...
[ "django.http.HttpResponse", "django.utils.encoding.force_unicode", "json.dumps", "django.template.RequestContext" ]
[((730, 780), 'json.dumps', 'json.dumps', (['json_data'], {'cls': 'self.json_encoder_class'}), '(json_data, cls=self.json_encoder_class)\n', (740, 780), False, 'import json\n'), ((879, 980), 'django.http.HttpResponse', 'HttpResponse', (['json_serialized'], {'content_type': '"""application/json; charset=utf-8"""'}), "(j...
#! /usr/bin/env python from setuptools import setup, Extension import importlib import os # copied from kymatio's setup.py: https://github.com/kymatio/kymatio/blob/master/setup.py sfm_version_spec = importlib.util.spec_from_file_location('sfm_version', 'sfm/version.py') sfm_version_module = importlib.util.module_from...
[ "importlib.util.spec_from_file_location", "setuptools.setup", "importlib.util.module_from_spec" ]
[((201, 272), 'importlib.util.spec_from_file_location', 'importlib.util.spec_from_file_location', (['"""sfm_version"""', '"""sfm/version.py"""'], {}), "('sfm_version', 'sfm/version.py')\n", (239, 272), False, 'import importlib\n'), ((294, 343), 'importlib.util.module_from_spec', 'importlib.util.module_from_spec', (['sf...
# from pipet.core.sql.query_interface import * from pypipet.core.operations.inventory import * import pytest from pprint import pprint _supplie_id = 1 def test_update_invs(session, obj_classes, shop_conn): invs = [ {'sku':'s22456', 'supplier_id':_supplie_id, 'qty':20}] update_inventory_bulk(obj_classes, sess...
[ "pprint.pprint" ]
[((502, 513), 'pprint.pprint', 'pprint', (['res'], {}), '(res)\n', (508, 513), False, 'from pprint import pprint\n')]
# Costs # 2019 - Luque, Straub - Risk-based optimal inspection strategies for # structural systems using dynamic Bayesian networks # Table 4, case 1 import numpy as np class InspectionMaintenance: """ Inspection and Maintenance ========================== Cost calculation for inspection and main...
[ "numpy.unique", "numpy.diff", "numpy.array", "numpy.vstack", "numpy.zeros_like" ]
[((1359, 1392), 'numpy.vstack', 'np.vstack', (['system_model.system_pf'], {}), '(system_model.system_pf)\n', (1368, 1392), True, 'import numpy as np\n'), ((1417, 1427), 'numpy.diff', 'np.diff', (['t'], {}), '(t)\n', (1424, 1427), True, 'import numpy as np\n'), ((1452, 1463), 'numpy.diff', 'np.diff', (['pf'], {}), '(pf)...
from __future__ import absolute_import, division, print_function import json import os.path as op import six import numpy as np import scipy.signal as sig import scipy.io.wavfile as sciwav MAXINT16 = 2**15 - 1 FS = 44100 COEFF_DIR = op.join(op.dirname(op.abspath(__file__)), 'coeffs') def normalize(data, maxamp=1): ...
[ "numpy.random.random", "os.path.join", "json.load", "numpy.array", "scipy.signal.lfilter", "scipy.io.wavfile.read", "numpy.around", "numpy.random.seed", "scipy.io.wavfile.write", "os.path.abspath", "numpy.savetxt", "numpy.loadtxt" ]
[((255, 275), 'os.path.abspath', 'op.abspath', (['__file__'], {}), '(__file__)\n', (265, 275), True, 'import os.path as op\n'), ((443, 455), 'json.load', 'json.load', (['f'], {}), '(f)\n', (452, 455), False, 'import json\n'), ((565, 585), 'numpy.random.seed', 'np.random.seed', (['seed'], {}), '(seed)\n', (579, 585), Tr...
# Standard library import atexit import os os.environ["OMP_NUM_THREADS"] = "1" import sys import traceback # Third-party from astropy.utils import iers iers.conf.auto_download = False import astropy.table as at import numpy as np # This project from totoro.config import cache_path from totoro.data import datasets, el...
[ "os.path.exists", "numpy.random.default_rng", "argparse.ArgumentParser", "astropy.table.Table", "astropy.table.vstack", "totoro.data.datasets.items", "numpy.array", "os.unlink", "sys.exit", "totoro.objective.TorusImagingObjective", "traceback.print_exc", "atexit.register", "glob.glob", "as...
[((1200, 1228), 'os.path.exists', 'os.path.exists', (['all_filename'], {}), '(all_filename)\n', (1214, 1228), False, 'import os\n'), ((1430, 1457), 'glob.glob', 'glob.glob', (['cache_glob_pattr'], {}), '(cache_glob_pattr)\n', (1439, 1457), False, 'import glob\n'), ((2009, 2025), 'totoro.data.datasets.items', 'datasets....
""" Problem: You are given a huge list of airline ticket prices between different cities around the world on a given day. These are all direct flights. Each element in the list has the format (source_city, destination, price). Consider a user who is willing to take up to k connections from their origin city A to thei...
[ "DataStructures.Graph.GraphDirectedWeighted", "DataStructures.PriorityQueue.MinPriorityQueue" ]
[((1315, 1333), 'DataStructures.PriorityQueue.MinPriorityQueue', 'MinPriorityQueue', ([], {}), '()\n', (1331, 1333), False, 'from DataStructures.PriorityQueue import MinPriorityQueue\n'), ((2281, 2304), 'DataStructures.Graph.GraphDirectedWeighted', 'GraphDirectedWeighted', ([], {}), '()\n', (2302, 2304), False, 'from D...
"""PyTorch Distributed Data Parallel example from NVIDIA.""" # https://github.com/NVIDIA/DeepLearningExamples import argparse import utils import virtual_machine def main(): parser = argparse.ArgumentParser(description='Optional app description') parser.add_argument('--vm-name', dest='vm_name', type=str, require...
[ "utils.run_threads", "virtual_machine.VirtualMachine", "argparse.ArgumentParser" ]
[((188, 251), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {'description': '"""Optional app description"""'}), "(description='Optional app description')\n", (211, 251), False, 'import argparse\n'), ((2429, 2479), 'utils.run_threads', 'utils.run_threads', (['vms', 'commands'], {'login_shell': '(True)'}), '...
from torch.utils.data import Dataset from skimage import io import os import torch class MnistData(Dataset): def __init__(self, root_dir): self.root_dir = root_dir img_list = [] label_list = os.listdir(self.root_dir) for label in label_list: file_names = os.li...
[ "skimage.io.imread", "os.listdir", "os.path.join" ]
[((229, 254), 'os.listdir', 'os.listdir', (['self.root_dir'], {}), '(self.root_dir)\n', (239, 254), False, 'import os\n'), ((735, 773), 'os.path.join', 'os.path.join', (['self.root_dir', 'img_label'], {}), '(self.root_dir, img_label)\n', (747, 773), False, 'import os\n'), ((793, 825), 'os.path.join', 'os.path.join', ([...
from qunetsim.backends.rw_lock import RWLock from qunetsim.objects.logger import Logger import queue class QuantumStorage(object): """ An object which stores qubits. """ STORAGE_LIMIT_ALL = 1 STORAGE_LIMIT_PER_HOST = 2 STORAGE_LIMIT_INDIVIDUALLY_PER_HOST = 3 def __init__(self): #...
[ "qunetsim.objects.logger.Logger.get_instance", "queue.Queue", "qunetsim.backends.rw_lock.RWLock" ]
[((1044, 1052), 'qunetsim.backends.rw_lock.RWLock', 'RWLock', ([], {}), '()\n', (1050, 1052), False, 'from qunetsim.backends.rw_lock import RWLock\n'), ((1076, 1097), 'qunetsim.objects.logger.Logger.get_instance', 'Logger.get_instance', ([], {}), '()\n', (1095, 1097), False, 'from qunetsim.objects.logger import Logger\...
# -*- coding: utf-8 -*- import datetime from pagseguro.utils import (is_valid_cpf, is_valid_cnpj, is_valid_email, parse_date) from pagseguro.exceptions import PagSeguroValidationError import pytest from dateutil.tz import tzutc def test_is_valid_email(): valid = '<EMAIL>' valid2...
[ "pagseguro.utils.is_valid_cpf", "dateutil.tz.tzutc", "pagseguro.utils.parse_date", "pagseguro.utils.is_valid_cnpj", "pytest.raises", "pagseguro.utils.is_valid_email" ]
[((427, 466), 'pytest.raises', 'pytest.raises', (['PagSeguroValidationError'], {}), '(PagSeguroValidationError)\n', (440, 466), False, 'import pytest\n'), ((476, 501), 'pagseguro.utils.is_valid_email', 'is_valid_email', (['not_valid'], {}), '(not_valid)\n', (490, 501), False, 'from pagseguro.utils import is_valid_cpf, ...
#-*- conding:utf-8 -*- #2018-02-02 11:04:54 import time,random,requests from lxml import etree from fake_useragent import UserAgent import urllib.request as ur from pymongo import MongoClient from multiprocessing.dummy import Pool as ThreadPool client = MongoClient('localhost',27017) db = client['test'] ip_list = db[...
[ "urllib.request.install_opener", "urllib.request.ProxyHandler", "requests.get", "urllib.request.build_opener", "lxml.etree.HTML", "multiprocessing.dummy.Pool", "pymongo.MongoClient", "random.random", "fake_useragent.UserAgent" ]
[((256, 287), 'pymongo.MongoClient', 'MongoClient', (['"""localhost"""', '(27017)'], {}), "('localhost', 27017)\n", (267, 287), False, 'from pymongo import MongoClient\n'), ((336, 347), 'fake_useragent.UserAgent', 'UserAgent', ([], {}), '()\n', (345, 347), False, 'from fake_useragent import UserAgent\n'), ((1119, 1141)...
#!/usr/bin/python3 # -*- coding: utf-8 -*- """ Read & write snapshot of node status data dumped to local file system. This data is used to detect things like sytem changes that may need to be reported. """ import json import time import logging import os import socket import psutil import netifaces as ni logging.ba...
[ "logging.basicConfig", "os.path.exists", "logging.getLogger", "time.ctime", "netifaces.ifaddresses", "psutil.boot_time", "time.time", "json.load", "netifaces.interfaces", "socket.gethostname", "json.dump" ]
[((310, 349), 'logging.basicConfig', 'logging.basicConfig', ([], {'level': 'logging.INFO'}), '(level=logging.INFO)\n', (329, 349), False, 'import logging\n'), ((365, 398), 'logging.getLogger', 'logging.getLogger', (['"""DataSnapshot"""'], {}), "('DataSnapshot')\n", (382, 398), False, 'import logging\n'), ((1889, 1920),...
""" 開発初期段階では、サイトレポートをVIEWに直書きしていた パフォーマンス改善のため、バッチ処理に変更 このVIEWは今は使われていないが、一応残しておく """ from django.conf import settings from django.db.models import Q, Count from django.utils import html from .models import Post, Category, Tag from datetime import datetime from janome.tokenizer import Tokenizer from janome...
[ "janome.analyzer.Analyzer", "django.db.models.Count", "itertools.product", "janome.tokenizer.Tokenizer", "django.db.models.Q" ]
[((3315, 3359), 'itertools.product', 'itertools.product', (['month_list', 'category_list'], {}), '(month_list, category_list)\n', (3332, 3359), False, 'import itertools\n'), ((4727, 4792), 'janome.tokenizer.Tokenizer', 'Tokenizer', ([], {'udic': 'udic_path', 'udic_type': '"""simpledic"""', 'udic_enc': '"""utf8"""'}), "...
import os def check(cmd, mf): m = mf.findNode('matplotlib') if m is None or m.filename is None: return None if cmd.matplotlib_backends: backends = {} for backend in cmd.matplotlib_backends: if backend == '-': pass elif backend == '*': ...
[ "os.path.dirname" ]
[((669, 696), 'os.path.dirname', 'os.path.dirname', (['m.filename'], {}), '(m.filename)\n', (684, 696), False, 'import os\n')]
# Generated by Django 3.1.6 on 2021-02-14 18:13 from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('core', '0008_auto_20210214_2039'), ] operations = [ migrations.AlterField( model_name='device', name='ip_address', ...
[ "django.db.models.GenericIPAddressField", "django.db.models.CharField" ]
[((337, 450), 'django.db.models.GenericIPAddressField', 'models.GenericIPAddressField', ([], {'blank': '(True)', 'help_text': '"""e.g. 192.168.0.17"""', 'null': '(True)', 'verbose_name': '"""IP address"""'}), "(blank=True, help_text='e.g. 192.168.0.17',\n null=True, verbose_name='IP address')\n", (365, 450), False, ...
import requests import re pattern_hashes = "^[a-f0-9]{64}$" pattern_url = "https?:\/\/.*" pattern_ip = "(?:(?:\d|[01]?\d\d|2[0-4]\d|25[0-5])\.){3}(?:25[0-5]|2[0-4]\d|[01]?\d\d|\d)(?:\/\d{1,2})?" pattern_domain = "^[a-zA-Z0-9][a-zA-Z0-9-_]{0,61}[a-zA-Z0-9]{0,1}\.([a-zA-Z]{1,6}|[a-zA-Z0-9-]{1,30}\.[a-zA-Z]{2,3})$" head...
[ "re.findall", "requests.get" ]
[((366, 438), 'requests.get', 'requests.get', (['"""https://labs.inquest.net/api/iocdb/list"""'], {'headers': 'headers'}), "('https://labs.inquest.net/api/iocdb/list', headers=headers)\n", (378, 438), False, 'import requests\n'), ((604, 640), 're.findall', 're.findall', (['pattern', 'data', 're.DOTALL'], {}), '(pattern...
# coding: utf-8 import wx import wx.lib.sized_controls as sc from dataclasses import dataclass from functools import partial from wx.adv import CommandLinkButton from bookworm import app from bookworm import config from bookworm import typehints as t from bookworm.i18n import LocaleInfo from bookworm.concurrency impor...
[ "bookworm.gui.components.ImmutableObjectListView", "bookworm.ocr_engines.tesseract_ocr_engine.TesseractOcrEngine.get_recognition_languages", "bookworm.i18n.LocaleInfo.from_three_letter_code", "wx.StdDialogButtonSizer", "wx.GetApp", "bookworm.logger.logger.getChild", "bookworm.platform_services._win32.te...
[((1281, 1306), 'bookworm.logger.logger.getChild', 'logger.getChild', (['__name__'], {}), '(__name__)\n', (1296, 1306), False, 'from bookworm.logger import logger\n'), ((6838, 6908), 'wx.Choice', 'wx.Choice', (['parent', '(-1)'], {'choices': '[l.description for l in self.languages]'}), '(parent, -1, choices=[l.descript...
# Copyright (c) 2012 <NAME> <<EMAIL>> # # This is free software released under the MIT license. # See COPYING file for details, or visit: # http://www.opensource.org/licenses/mit-license.php # # The file is part of FSMonitor, a file-system monitoring library. # https://github.com/shaurz/fsmonitor import sys, os, time,...
[ "os.listdir", "threading.Lock", "os.path.join", "time.sleep", "os.stat", "time.time" ]
[((730, 741), 'time.time', 'time.time', ([], {}), '()\n', (739, 741), False, 'import sys, os, time, threading, errno\n'), ((1690, 1701), 'time.time', 'time.time', ([], {}), '()\n', (1699, 1701), False, 'import sys, os, time, threading, errno\n'), ((2042, 2055), 'os.stat', 'os.stat', (['path'], {}), '(path)\n', (2049, 2...
import math import random import string from datetime import datetime, timedelta from textwrap import dedent import config import discord from discord.ext import commands import asyncio class etc(commands.Cog): def __init__(self, bot): self.bot = bot @commands.command(name="clear") async def _m...
[ "discord.ext.commands.command" ]
[((273, 303), 'discord.ext.commands.command', 'commands.command', ([], {'name': '"""clear"""'}), "(name='clear')\n", (289, 303), False, 'from discord.ext import commands\n'), ((503, 534), 'discord.ext.commands.command', 'commands.command', ([], {'name': '"""recent"""'}), "(name='recent')\n", (519, 534), False, 'from di...
# Copyright 2019 Xilinx Inc. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, ...
[ "sys.path.insert", "numpy.reshape", "numpy.minimum", "numpy.hstack", "caffe.io.Transformer", "numpy.where", "cv2.cvtColor", "caffe.Net", "numpy.meshgrid", "numpy.maximum", "numpy.arange" ]
[((1052, 1084), 'numpy.maximum', 'np.maximum', (['x1[i]', 'x1[order[1:]]'], {}), '(x1[i], x1[order[1:]])\n', (1062, 1084), True, 'import numpy as np\n'), ((1099, 1131), 'numpy.maximum', 'np.maximum', (['y1[i]', 'y1[order[1:]]'], {}), '(y1[i], y1[order[1:]])\n', (1109, 1131), True, 'import numpy as np\n'), ((1146, 1178)...
#!/usr/bin/env python3 # -*- coding: utf-8 -*- __author__ = 'cnheider' import csv import matplotlib.pyplot as plt import utilities as U # print(plt.style.available) plot_style = 'fivethirtyeight' # plot_style='bmh' # plot_style='ggplot' plt.style.use('seaborn-poster') plt.style.use(plot_style) plt.rcParams['axes.e...
[ "matplotlib.pyplot.get_backend", "matplotlib.pyplot.plot", "matplotlib.pyplot.style.use", "utilities.StatisticAggregator", "tkinter.Tk", "matplotlib.pyplot.ion", "matplotlib.pyplot.title", "csv.reader", "tkinter.filedialog.askopenfilename", "matplotlib.pyplot.show" ]
[((242, 273), 'matplotlib.pyplot.style.use', 'plt.style.use', (['"""seaborn-poster"""'], {}), "('seaborn-poster')\n", (255, 273), True, 'import matplotlib.pyplot as plt\n'), ((274, 299), 'matplotlib.pyplot.style.use', 'plt.style.use', (['plot_style'], {}), '(plot_style)\n', (287, 299), True, 'import matplotlib.pyplot a...
# ============================================================================= # Copyright 2020 NVIDIA. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://ww...
[ "nemo.logging.info", "numpy.all" ]
[((2735, 2798), 'nemo.logging.info', 'logging.info', (['f"""Total errors (multiplied by 2): {total_errors}"""'], {}), "(f'Total errors (multiplied by 2): {total_errors}')\n", (2747, 2798), False, 'from nemo import logging\n'), ((3193, 3262), 'nemo.logging.info', 'logging.info', (['f"""*** Misclassified intent queries (...
from datetime import datetime import pytz from .constants import TIMEZONE timezone = pytz.timezone(TIMEZONE) def get_ist_now(): """Returns Indian Standard Time datetime object. Returns: object -- Datetime object """ return datetime.now(timezone)
[ "pytz.timezone", "datetime.datetime.now" ]
[((88, 111), 'pytz.timezone', 'pytz.timezone', (['TIMEZONE'], {}), '(TIMEZONE)\n', (101, 111), False, 'import pytz\n'), ((253, 275), 'datetime.datetime.now', 'datetime.now', (['timezone'], {}), '(timezone)\n', (265, 275), False, 'from datetime import datetime\n')]
""" LibriParty Dataset creation by using official metadata. Author ------ <NAME>, 2020 <NAME>, 2020 """ import os import sys import speechbrain as sb from hyperpyyaml import load_hyperpyyaml from speechbrain.utils.data_utils import download_file from local.create_mixtures_from_metadata import create_mixture import js...
[ "os.path.exists", "os.makedirs", "speechbrain.core.parse_arguments", "os.path.join", "local.create_mixtures_from_metadata.create_mixture", "json.load", "speechbrain.utils.data_utils.download_file", "hyperpyyaml.load_hyperpyyaml" ]
[((534, 571), 'speechbrain.core.parse_arguments', 'sb.core.parse_arguments', (['sys.argv[1:]'], {}), '(sys.argv[1:])\n', (557, 571), True, 'import speechbrain as sb\n'), ((802, 906), 'speechbrain.utils.data_utils.download_file', 'download_file', (['URL_METADATA', "(metadata_folder + '/meta.zip')"], {'unpack': '(True)',...
import subprocess from pathlib import Path import pytest # from py4gh import __version__ from py4gh.utility import decrypt_files, encrypt_files, get_files # def test_version(): # assert __version__ == "0.1.0" @pytest.fixture(scope="session") def keys(tmpdir_factory): test_pub1 = tmpdir_factory.mktemp("data...
[ "py4gh.utility.encrypt_files", "py4gh.utility.get_files", "pathlib.Path", "subprocess.Popen", "subprocess.run", "py4gh.utility.decrypt_files", "pytest.fixture" ]
[((219, 250), 'pytest.fixture', 'pytest.fixture', ([], {'scope': '"""session"""'}), "(scope='session')\n", (233, 250), False, 'import pytest\n'), ((1113, 1144), 'pytest.fixture', 'pytest.fixture', ([], {'scope': '"""session"""'}), "(scope='session')\n", (1127, 1144), False, 'import pytest\n'), ((641, 756), 'subprocess....
from unittest import TestCase from parameterized import parameterized from pyecsca.ec.context import local from pyecsca.ec.mult import LTRMultiplier, BinaryNAFMultiplier, WindowNAFMultiplier, LadderMultiplier, \ DifferentialLadderMultiplier from pyecsca.ec.params import get_params from pyecsca.sca.re.rpa import M...
[ "parameterized.parameterized.expand", "pyecsca.ec.mult.LadderMultiplier", "pyecsca.ec.params.get_params", "pyecsca.ec.mult.DifferentialLadderMultiplier", "pyecsca.ec.mult.BinaryNAFMultiplier", "pyecsca.sca.re.rpa.MultipleContext", "pyecsca.ec.mult.LTRMultiplier", "pyecsca.ec.mult.WindowNAFMultiplier" ...
[((785, 990), 'parameterized.parameterized.expand', 'parameterized.expand', (["[('10', 10), ('2355498743', 2355498743), (\n '325385790209017329644351321912443757746', \n 325385790209017329644351321912443757746), ('13613624287328732', \n 13613624287328732)]"], {}), "([('10', 10), ('2355498743', 2355498743), (\n...