code stringlengths 22 1.05M | apis listlengths 1 3.31k | extract_api stringlengths 75 3.25M |
|---|---|---|
import tkinter as tk
from tkinter import filedialog
from urllib.request import urlopen
from pathlib import Path
from tkinter import ttk
import numpy as np
import base64
import io
import re
from src.theme import theme
from src.algorithm import blosum
from src.utils import RichText
def qopen(path:str):
'''Opens and... | [
"tkinter.ttk.Button",
"src.theme.theme",
"numpy.hstack",
"pathlib.Path.home",
"src.algorithm.blosum",
"tkinter.ttk.LabelFrame",
"base64.encodestring",
"src.utils.RichText",
"tkinter.Label",
"tkinter.ttk.Entry",
"tkinter.ttk.Frame",
"tkinter.ttk.Label",
"numpy.asarray",
"numpy.ndenumerate",... | [((500, 507), 'tkinter.Tk', 'tk.Tk', ([], {}), '()\n', (505, 507), True, 'import tkinter as tk\n'), ((702, 713), 'tkinter.ttk.Style', 'ttk.Style', ([], {}), '()\n', (711, 713), False, 'from tkinter import ttk\n'), ((984, 1019), 'tkinter.ttk.Notebook', 'ttk.Notebook', (['self.root'], {'padding': '(10)'}), '(self.root, p... |
import numpy as np
import re
def apply_colormaps_based_on_mask(mask, data_for_inside_mask, data_for_outside_mask,
colormap_inside_mask, colormap_outside_mask):
"""
Returns the combination of applying two colormaps to two datasets on two mutually exclusive sets of pixels
a... | [
"re.fullmatch",
"numpy.stack",
"numpy.clip"
] | [((2605, 2638), 'numpy.stack', 'np.stack', (['([frame] * depth)'], {'axis': '(2)'}), '([frame] * depth, axis=2)\n', (2613, 2638), True, 'import numpy as np\n'), ((2915, 2968), 're.fullmatch', 're.fullmatch', (['"""[ra][\\\\-\\\\.0-9]+"""', 'mv_thresholdOnValue'], {}), "('[ra][\\\\-\\\\.0-9]+', mv_thresholdOnValue)\n", ... |
import roslib;roslib.load_manifest('kobuki_dashboard')
import rospy
import diagnostic_msgs
from rqt_robot_dashboard.dashboard import Dashboard
from rqt_robot_dashboard.widgets import ConsoleDashWidget, MenuDashWidget, IconToolButton
from python_qt_binding.QtWidgets import QMessageBox, QAction
from python_qt_binding.Q... | [
"rqt_robot_dashboard.widgets.ConsoleDashWidget",
"rospy.logwarn",
"rospy.get_time",
"roslib.load_manifest",
"python_qt_binding.QtCore.QTimer",
"rospy.Subscriber"
] | [((14, 54), 'roslib.load_manifest', 'roslib.load_manifest', (['"""kobuki_dashboard"""'], {}), "('kobuki_dashboard')\n", (34, 54), False, 'import roslib\n'), ((818, 826), 'python_qt_binding.QtCore.QTimer', 'QTimer', ([], {}), '()\n', (824, 826), False, 'from python_qt_binding.QtCore import QSize, QTimer\n'), ((1171, 127... |
from __future__ import absolute_import, division, print_function
import glob
# Format expected by setup.py and doc/source/conf.py: string of form "X.Y.Z"
_version_major = 0
_version_minor = 1
_version_micro = '' # use '' for first of series, number for 1 and above
_version_extra = 'dev'
# _version_extra = '' # Uncom... | [
"glob.glob"
] | [((2104, 2129), 'glob.glob', 'glob.glob', (['"""scripts/*.py"""'], {}), "('scripts/*.py')\n", (2113, 2129), False, 'import glob\n')] |
# -*- coding: utf8
from gb.randomkit.random import RNG
from gb.samplers import BaseSampler
from gb.samplers import CollapsedGibbsSampler
from gb.stamps import Timestamps
from gb.sloppy import SloppyCounter
from numpy.testing import assert_equal
import numpy as np
def test_get_probability():
d = {}
d[0] = ... | [
"numpy.array",
"gb.sloppy.SloppyCounter",
"gb.randomkit.random.RNG",
"gb.stamps.Timestamps"
] | [((379, 392), 'gb.stamps.Timestamps', 'Timestamps', (['d'], {}), '(d)\n', (389, 392), False, 'from gb.stamps import Timestamps\n'), ((654, 686), 'numpy.array', 'np.array', (['[5, 5]'], {'dtype': '"""uint64"""'}), "([5, 5], dtype='uint64')\n", (662, 686), True, 'import numpy as np\n'), ((704, 738), 'numpy.array', 'np.ar... |
import tensorflow as tf
"""Class for KDD10 percent GAN architecture.
Generator and discriminator.
"""
learning_rate = 0.00001
batch_size = 50
layer = 1
latent_dim = 32
dis_inter_layer_dim = 128
init_kernel = tf.contrib.layers.xavier_initializer()
def generator(z_inp, is_training=False, getter=None, reuse=False):
... | [
"tensorflow.variable_scope",
"tensorflow.nn.relu",
"tensorflow.contrib.layers.xavier_initializer",
"tensorflow.layers.dropout",
"tensorflow.layers.dense",
"tensorflow.squeeze"
] | [((213, 251), 'tensorflow.contrib.layers.xavier_initializer', 'tf.contrib.layers.xavier_initializer', ([], {}), '()\n', (249, 251), True, 'import tensorflow as tf\n'), ((607, 672), 'tensorflow.variable_scope', 'tf.variable_scope', (['"""generator"""'], {'reuse': 'reuse', 'custom_getter': 'getter'}), "('generator', reus... |
#!/usr/bin/env python
from setuptools import setup, find_packages
setup(
name='setuptools-wotmod',
version='0.2',
packages=find_packages(),
description='setuptools integration for creating World of Tanks mods',
long_description=open('README.md').read(),
author='jhakonen',
url='https://gith... | [
"setuptools.find_packages"
] | [((137, 152), 'setuptools.find_packages', 'find_packages', ([], {}), '()\n', (150, 152), False, 'from setuptools import setup, find_packages\n')] |
from django.http import HttpResponse
from django.shortcuts import render, redirect
from lab1.views import login
def docs(request):
if not request.user.is_authenticated:
return redirect(login.login_view)
return render(request, 'docs.html')
| [
"django.shortcuts.render",
"django.shortcuts.redirect"
] | [((229, 257), 'django.shortcuts.render', 'render', (['request', '"""docs.html"""'], {}), "(request, 'docs.html')\n", (235, 257), False, 'from django.shortcuts import render, redirect\n'), ((191, 217), 'django.shortcuts.redirect', 'redirect', (['login.login_view'], {}), '(login.login_view)\n', (199, 217), False, 'from d... |
import unittest
import coverage
from flask_script import Manager
from project import create_app, db
from project.api.models import User
COV = coverage.coverage(
branch=True,
include='project/*',
omit=[
'project/tests/*',
'project/server/config.py',
'project/server/*/__init__.py'
... | [
"project.api.models.User",
"project.db.drop_all",
"project.db.create_all",
"flask_script.Manager",
"project.create_app",
"coverage.coverage",
"unittest.TextTestRunner",
"unittest.TestLoader",
"project.db.session.commit"
] | [((145, 286), 'coverage.coverage', 'coverage.coverage', ([], {'branch': '(True)', 'include': '"""project/*"""', 'omit': "['project/tests/*', 'project/server/config.py', 'project/server/*/__init__.py']"}), "(branch=True, include='project/*', omit=['project/tests/*',\n 'project/server/config.py', 'project/server/*/__i... |
# -----------------------------------------------------------------------------
# Clorm ORM FactBase implementation. FactBase provides a set-like container
# specifically for storing facts (Predicate instances).
# ------------------------------------------------------------------------------
import abc
import io
impor... | [
"itertools.chain",
"io.StringIO",
"typing.cast",
"itertools.groupby"
] | [((8102, 8156), 'itertools.groupby', 'itertools.groupby', (['sorted_facts', '(lambda x: x.__class__)'], {}), '(sorted_facts, lambda x: x.__class__)\n', (8119, 8156), False, 'import itertools\n'), ((16691, 16704), 'io.StringIO', 'io.StringIO', ([], {}), '()\n', (16702, 16704), False, 'import io\n'), ((15172, 15193), 'it... |
# -*- coding: utf-8 -*-
#
# ***** BEGIN GPL LICENSE BLOCK *****
#
# --------------------------------------------------------------------------
# Blender 2.5 Extensions Framework
# --------------------------------------------------------------------------
#
# Authors:
# <NAME>
#
# This program is free software; you can ... | [
"bpy.utils.unregister_class",
"bpy.ops.ef.msg",
"bpy.utils.register_class",
"time.strftime"
] | [((1023, 1058), 'bpy.utils.register_class', 'bpy.utils.register_class', (['EF_OT_msg'], {}), '(EF_OT_msg)\n', (1047, 1058), False, 'import bpy\n'), ((1450, 1498), 'bpy.ops.ef.msg', 'bpy.ops.ef.msg', ([], {'msg_type': '"""WARNING"""', 'msg_text': 'str'}), "(msg_type='WARNING', msg_text=str)\n", (1464, 1498), False, 'imp... |
# pylint: disable=invalid-name
'''
Pytests for the common utilities included in this package. Includes:
- conversions.py
- specs.py
- utils.py
To run the tests, type the following in the top level repo directory:
python -m pytest --nat-file [path/to/gribfile] --prs-file [path/to/gribfile]
'''
from... | [
"adb_graphics.conversions.m_to_mi",
"adb_graphics.conversions.pa_to_hpa",
"inspect.getfullargspec",
"adb_graphics.conversions.m_to_kft",
"adb_graphics.datahandler.gribdata.__getattribute__",
"adb_graphics.conversions.weasd_to_1hsnw",
"adb_graphics.conversions.k_to_f",
"adb_graphics.conversions.ms_to_k... | [((830, 845), 'numpy.ones', 'np.ones', (['[3, 2]'], {}), '([3, 2])\n', (837, 845), True, 'import numpy as np\n'), ((928, 949), 'adb_graphics.conversions.k_to_c', 'conversions.k_to_c', (['a'], {}), '(a)\n', (946, 949), True, 'import adb_graphics.conversions as conversions\n'), ((989, 1010), 'adb_graphics.conversions.k_t... |
from pyte import tokens, util
from pyte.superclasses import _PyteAugmentedValidator, _PyteOp
from pyte.util import PY36
class FOR_LOOP(_PyteOp):
"""
Represents a for loop.
"""
def __init__(self, iterator: _PyteAugmentedValidator, body: list):
"""
Represents a for operator.
:p... | [
"pyte.util.generate_simple_call",
"pyte.util.flatten",
"pyte.util.ensure_instruction",
"pyte.util.generate_bytecode_from_obb"
] | [((1050, 1106), 'pyte.util.generate_bytecode_from_obb', 'util.generate_bytecode_from_obb', (['self.iterator', 'previous'], {}), '(self.iterator, previous)\n', (1081, 1106), False, 'from pyte import tokens, util\n'), ((1172, 1225), 'pyte.util.generate_bytecode_from_obb', 'util.generate_bytecode_from_obb', (['tokens.GET_... |
# This little bit of magic fills the __all__ list
# with every plugin name, and means that calling:
# from plugins import *
# within inmembrane.py will import every plugin
import pkgutil
__all__ = []
for p in pkgutil.iter_modules(__path__):
__all__.append(p[1])
| [
"pkgutil.iter_modules"
] | [((211, 241), 'pkgutil.iter_modules', 'pkgutil.iter_modules', (['__path__'], {}), '(__path__)\n', (231, 241), False, 'import pkgutil\n')] |
# DomirScire
import math
import os
import random
import re
import sys
import collections
if __name__ == '__main__':
s = sorted(input().strip())
s_counter = collections.Counter(s).most_common()
s_counter = sorted(s_counter, key=lambda x: (x[1] * -1, x[0]))
for i in range(0, 3):
print(s_counter[i... | [
"collections.Counter"
] | [((165, 187), 'collections.Counter', 'collections.Counter', (['s'], {}), '(s)\n', (184, 187), False, 'import collections\n')] |
import math
import torch
import torch.nn as nn
from torch.nn import functional as F
class KLRegression(nn.Module):
"""KL-divergence loss for probabilistic regression.
It is computed using Monte Carlo (MC) samples from an arbitrary distribution."""
def __init__(self, eps=0.0):
super()._... | [
"torch.log",
"torch.mean",
"math.log",
"torch.sum",
"torch.logsumexp"
] | [((725, 761), 'torch.log', 'torch.log', (['(sample_density + self.eps)'], {}), '(sample_density + self.eps)\n', (734, 761), False, 'import torch\n'), ((864, 939), 'torch.mean', 'torch.mean', (['(scores * (gt_density / (sample_density + self.eps)))'], {'dim': 'mc_dim'}), '(scores * (gt_density / (sample_density + self.e... |
# coding=utf-8
import arrow
from bs4 import BeautifulSoup
from digesters.base_digester import BaseDigester
TEMPLATE = """<html>
<head>
<meta content="text/html; charset=utf-8" http-equiv="Content-Type"/>
<title>Atlassian HipChat</title>
</head>
<body style="box-sizing: border-box; height: 100%; width: 100%;... | [
"bs4.BeautifulSoup"
] | [((2916, 2958), 'bs4.BeautifulSoup', 'BeautifulSoup', (['html_message', '"""html.parser"""'], {}), "(html_message, 'html.parser')\n", (2929, 2958), False, 'from bs4 import BeautifulSoup\n')] |
import os
import shutil
import tempfile
import numpy as np
from yt.loaders import load, load_uniform_grid
from yt.testing import (
assert_array_equal,
assert_fname,
fake_random_ds,
requires_file,
requires_module,
)
from yt.utilities.answer_testing.framework import data_dir_load
from yt.visualizati... | [
"yt.loaders.load_uniform_grid",
"yt.visualization.plot_window.SlicePlot",
"yt.testing.assert_fname",
"yt.testing.assert_array_equal",
"numpy.arange",
"yt.testing.fake_random_ds",
"os.path.join",
"yt.loaders.load",
"os.getcwd",
"os.chdir",
"yt.utilities.answer_testing.framework.data_dir_load",
... | [((399, 422), 'yt.testing.requires_module', 'requires_module', (['"""h5py"""'], {}), "('h5py')\n", (414, 422), False, 'from yt.testing import assert_array_equal, assert_fname, fake_random_ds, requires_file, requires_module\n'), ((1300, 1323), 'yt.testing.requires_module', 'requires_module', (['"""h5py"""'], {}), "('h5p... |
# Generated by Django 3.2.6 on 2021-09-05 19:39
from django.db import migrations, models
class Migration(migrations.Migration):
dependencies = [
('core', '0011_auto_20210905_1619'),
]
operations = [
migrations.AlterField(
model_name='preco',
name='categoria',
... | [
"django.db.models.CharField"
] | [((335, 453), 'django.db.models.CharField', 'models.CharField', ([], {'choices': "[('Premium', 'C'), ('Pro', 'B'), ('Plus', 'A')]", 'max_length': '(15)', 'verbose_name': '"""categoria"""'}), "(choices=[('Premium', 'C'), ('Pro', 'B'), ('Plus', 'A')],\n max_length=15, verbose_name='categoria')\n", (351, 453), False, '... |
#!/usr/bin/env python
import panflute as pf
"""
Pandoc filter that causes emphasis to be rendered using
the custom macro '\myemph{...}' rather than '\emph{...}'
in latex. Other output formats are unaffected.
"""
def latex(s):
return pf.RawInline(s, format='latex')
def myemph(e, doc):
if type(e)==pf.Emph a... | [
"panflute.RawInline",
"panflute.toJSONFilter"
] | [((241, 272), 'panflute.RawInline', 'pf.RawInline', (['s'], {'format': '"""latex"""'}), "(s, format='latex')\n", (253, 272), True, 'import panflute as pf\n'), ((442, 465), 'panflute.toJSONFilter', 'pf.toJSONFilter', (['myemph'], {}), '(myemph)\n', (457, 465), True, 'import panflute as pf\n')] |
#! /usr/bin/env python
# -*- coding: utf-8 -*-
#
# Copyright 2008,2009 <NAME> <<EMAIL>>
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at#
#
# http://www.apache.org/licenses/LICENSE-2.0... | [
"couchdbkit.contrib.WSGIHandler",
"json.dumps"
] | [((1108, 1124), 'couchdbkit.contrib.WSGIHandler', 'WSGIHandler', (['app'], {}), '(app)\n', (1119, 1124), False, 'from couchdbkit.contrib import WSGIHandler\n'), ((848, 883), 'json.dumps', 'json.dumps', (["environ['COUCHDB_INFO']"], {}), "(environ['COUCHDB_INFO'])\n", (858, 883), False, 'import json\n')] |
import json
from .life import BinaryLife
class GOL(object):
team_names: list = []
columns = 0
rows = 0
def __init__(self, **kwargs):
self.load_config(**kwargs)
self.create_life()
def __repr__(self):
s = []
s.append("+" + "-" * (self.columns) + "+")
for i i... | [
"json.loads"
] | [((2829, 2849), 'json.loads', 'json.loads', (['self.ic1'], {}), '(self.ic1)\n', (2839, 2849), False, 'import json\n'), ((3045, 3065), 'json.loads', 'json.loads', (['self.ic2'], {}), '(self.ic2)\n', (3055, 3065), False, 'import json\n')] |
from django.core.management.base import BaseCommand
from django.db import connection
class Command(BaseCommand):
help = "Add missing id column for hostingstats."
def handle(self, *args, **options):
with connection.cursor() as cursor:
self.cursor = cursor
self.cursor.execute(
... | [
"django.db.connection.cursor"
] | [((222, 241), 'django.db.connection.cursor', 'connection.cursor', ([], {}), '()\n', (239, 241), False, 'from django.db import connection\n')] |
import numba as nb
import numpy as np
import scipy.stats
@nb.njit(parallel=True)
def tiecorrect(rankvals):
"""
parallelized version of scipy.stats.tiecorrect
:param rankvals: p x n array of ranked data (output of rankdata function)
"""
tc = np.ones(rankvals.shape[1], dtype=np.float64)
for j i... | [
"numpy.sqrt",
"numpy.ones",
"numpy.float64",
"numpy.log",
"numba.njit",
"numpy.asarray",
"numpy.diff",
"numpy.argsort",
"numpy.sum",
"numpy.errstate",
"numpy.array",
"numpy.empty",
"numpy.concatenate",
"numpy.nonzero",
"numpy.maximum",
"numba.prange",
"numpy.all",
"numpy.ravel"
] | [((60, 82), 'numba.njit', 'nb.njit', ([], {'parallel': '(True)'}), '(parallel=True)\n', (67, 82), True, 'import numba as nb\n'), ((718, 740), 'numba.njit', 'nb.njit', ([], {'parallel': '(True)'}), '(parallel=True)\n', (725, 740), True, 'import numba as nb\n'), ((264, 308), 'numpy.ones', 'np.ones', (['rankvals.shape[1]'... |
#!/usr/bin/env python3
import codekit.pygithub
import github
import itertools
import pytest
@pytest.fixture
def git_author():
return github.InputGitAuthor(name='foo', email='<EMAIL>')
def test_init(git_author):
"""Test TargetTag object instantiation"""
t_tag = codekit.pygithub.TargetTag(
name=... | [
"github.InputGitAuthor",
"pytest.raises"
] | [((140, 190), 'github.InputGitAuthor', 'github.InputGitAuthor', ([], {'name': '"""foo"""', 'email': '"""<EMAIL>"""'}), "(name='foo', email='<EMAIL>')\n", (161, 190), False, 'import github\n'), ((1377, 1406), 'pytest.raises', 'pytest.raises', (['AssertionError'], {}), '(AssertionError)\n', (1390, 1406), False, 'import p... |
import functools
def parse_input() -> list[list[str]]:
groups = [[]]
with open("input.txt") as file:
for line in file:
line_ = line.rstrip()
if len(line_) > 0:
groups[-1].append(line_)
else:
if not groups[-1] == []:
... | [
"functools.reduce"
] | [((915, 955), 'functools.reduce', 'functools.reduce', (['set.intersection', 'sets'], {}), '(set.intersection, sets)\n', (931, 955), False, 'import functools\n')] |
import box
from typing import Text
import yaml
def load_config(config_path: Text) -> box.ConfigBox:
"""Loads yaml config in instance of box.ConfigBox.
Args:
config_path {Text}: path to config
Returns:
box.ConfigBox
"""
with open(config_path) as config_file:
config = yaml.... | [
"box.ConfigBox",
"yaml.safe_load"
] | [((315, 342), 'yaml.safe_load', 'yaml.safe_load', (['config_file'], {}), '(config_file)\n', (329, 342), False, 'import yaml\n'), ((360, 381), 'box.ConfigBox', 'box.ConfigBox', (['config'], {}), '(config)\n', (373, 381), False, 'import box\n')] |
import boto3
from botocore.exceptions import ClientError
import datetime
import pytest
from moto import mock_sagemaker
from moto.sts.models import ACCOUNT_ID
FAKE_ROLE_ARN = "arn:aws:iam::{}:role/FakeRole".format(ACCOUNT_ID)
TEST_REGION_NAME = "us-east-1"
class MyProcessingJobModel(object):
def __init__(
... | [
"boto3.client",
"pytest.raises"
] | [((3482, 3537), 'boto3.client', 'boto3.client', (['"""sagemaker"""'], {'region_name': 'TEST_REGION_NAME'}), "('sagemaker', region_name=TEST_REGION_NAME)\n", (3494, 3537), False, 'import boto3\n'), ((5306, 5356), 'boto3.client', 'boto3.client', (['"""sagemaker"""'], {'region_name': '"""us-east-1"""'}), "('sagemaker', re... |
# -*- coding: utf-8 -*-
# Form implementation generated from reading ui file 'direction_sprite_widget.ui'
#
# Created: Wed Jul 30 18:37:40 2014
# by: PyQt4 UI code generator 4.10.4
#
# WARNING! All changes made in this file will be lost!
from PyQt4 import QtCore, QtGui
try:
_fromUtf8 = QtCore.QString.fromUt... | [
"PyQt4.QtCore.QSize",
"PyQt4.QtGui.QWidget",
"PyQt4.QtGui.QScrollArea",
"PyQt4.QtGui.QFormLayout",
"PyQt4.QtGui.QSpinBox",
"PyQt4.QtCore.QMetaObject.connectSlotsByName",
"PyQt4.QtGui.QLabel",
"PyQt4.QtGui.QPushButton",
"PyQt4.QtGui.QComboBox",
"PyQt4.QtCore.QRect",
"PyQt4.QtGui.QVBoxLayout",
"... | [((498, 562), 'PyQt4.QtGui.QApplication.translate', 'QtGui.QApplication.translate', (['context', 'text', 'disambig', '_encoding'], {}), '(context, text, disambig, _encoding)\n', (526, 562), False, 'from PyQt4 import QtCore, QtGui\n'), ((960, 998), 'PyQt4.QtGui.QWidget', 'QtGui.QWidget', (['direction_sprite_widget'], {}... |
import nltk
import string
import argparse
parser = argparse.ArgumentParser(description='.')
parser.add_argument('-text', help='')
parser.add_argument('-meta', help='')
parser.add_argument('-output', help='')
args = parser.parse_args()
# parser.add_argument('-iter', dest='iter', type=int,
# defaul... | [
"nltk.corpus.stopwords.words",
"argparse.ArgumentParser"
] | [((53, 93), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {'description': '"""."""'}), "(description='.')\n", (76, 93), False, 'import argparse\n'), ((558, 596), 'nltk.corpus.stopwords.words', 'nltk.corpus.stopwords.words', (['"""english"""'], {}), "('english')\n", (585, 596), False, 'import nltk\n')] |
from collections import deque
import numpy as np
import os
from abc import ABCMeta, abstractmethod
import random
random.seed(42)
from common import config, VehicleState
from helper import Helper
INFO = """Average merging time: {} s
Traffic flow: {} vehicle/s
Average speed: {} km/h
Average fuel consumptio... | [
"matplotlib.pyplot.grid",
"bokeh.plotting.figure",
"helper.Helper.getTmOptimal2",
"matplotlib.pyplot.ylabel",
"numpy.random.rand",
"helper.Helper.getConfigVec",
"numpy.random.exponential",
"numpy.array",
"numpy.arange",
"helper.Helper.getTc",
"helper.Helper.getTimeMatrix",
"collections.deque",... | [((118, 133), 'random.seed', 'random.seed', (['(42)'], {}), '(42)\n', (129, 133), False, 'import random\n'), ((2806, 2813), 'collections.deque', 'deque', ([], {}), '()\n', (2811, 2813), False, 'from collections import deque\n'), ((5912, 5936), 'numpy.arange', 'np.arange', (['(10)', '(30.1)', '(2.0)'], {}), '(10, 30.1, ... |
import pytest
import datetime
import json
import functools
from urllib.parse import urlencode, parse_qs
from descarteslabs.common.graft import client as graft_client
from ... import types
from .. import tile_url
def test_url():
base = "foo"
base_q = base + "?"
url = functools.partial(tile_url.tile_url... | [
"datetime.datetime",
"descarteslabs.common.graft.client.consistent_guid",
"json.dumps",
"urllib.parse.parse_qs",
"functools.partial",
"pytest.raises",
"urllib.parse.urlencode"
] | [((1587, 1648), 'urllib.parse.parse_qs', 'parse_qs', (['params'], {'strict_parsing': '(True)', 'keep_blank_values': '(True)'}), '(params, strict_parsing=True, keep_blank_values=True)\n', (1595, 1648), False, 'from urllib.parse import urlencode, parse_qs\n'), ((2434, 2482), 'functools.partial', 'functools.partial', (['t... |
# Generated by Django 3.0.2 on 2020-01-31 20:33
from django.db import migrations, models
class Migration(migrations.Migration):
dependencies = [
('collections', '0002_auto_20200109_1348'),
]
operations = [
migrations.AlterField(
model_name='collection',
name='com... | [
"django.db.models.DateTimeField",
"django.db.models.CharField",
"django.db.models.BooleanField"
] | [((345, 442), 'django.db.models.CharField', 'models.CharField', ([], {'help_text': '"""Comment for collection."""', 'max_length': '(256)', 'verbose_name': '"""Comment"""'}), "(help_text='Comment for collection.', max_length=256,\n verbose_name='Comment')\n", (361, 442), False, 'from django.db import migrations, mode... |
# -*- coding: utf-8 -*-
import os
import torch
from torch.autograd import Variable
import numpy as np
import scipy
import matplotlib.pyplot as plt
import cv2
import scipy.ndimage
import shutil
import scipy.misc as misc
from PIL import Image
def mkdirs(folders, erase=False):
if type(folders) is not list:
... | [
"torch.LongTensor",
"torch.cuda.is_available",
"scipy.misc.imresize",
"numpy.arange",
"matplotlib.pyplot.imshow",
"os.path.exists",
"numpy.mean",
"numpy.reshape",
"scipy.misc.imsave",
"numpy.max",
"numpy.stack",
"numpy.min",
"numpy.random.permutation",
"numpy.tile",
"torch.abs",
"itert... | [((724, 747), 'os.path.exists', 'os.path.exists', (['imgfile'], {}), '(imgfile)\n', (738, 747), False, 'import os\n'), ((799, 818), 'cv2.imread', 'cv2.imread', (['imgfile'], {}), '(imgfile)\n', (809, 818), False, 'import cv2\n'), ((833, 872), 'cv2.cvtColor', 'cv2.cvtColor', (['srcBGR', 'cv2.COLOR_BGR2RGB'], {}), '(srcB... |
from flask import render_template
from app import app, html_generator
import app.calculator.calculator as calc
from app.route_helpers import units_from_form, options_from_form, options_list, flash_errors
from app.forms import InputForm
from collections import defaultdict
@app.route('/', methods=['GET', 'POST'])
@app.... | [
"flask.render_template",
"app.route_helpers.units_from_form",
"app.html_generator.make_boxes",
"app.forms.InputForm",
"app.route_helpers.options_list",
"app.calculator.calculator.calculate",
"collections.defaultdict",
"app.app.route",
"app.route_helpers.flash_errors",
"app.route_helpers.options_fr... | [((275, 314), 'app.app.route', 'app.route', (['"""/"""'], {'methods': "['GET', 'POST']"}), "('/', methods=['GET', 'POST'])\n", (284, 314), False, 'from app import app, html_generator\n'), ((316, 360), 'app.app.route', 'app.route', (['"""/index"""'], {'methods': "['GET', 'POST']"}), "('/index', methods=['GET', 'POST'])\... |
import re
def find_indices():
return [m.start(0) for m in re.finditer(reg, content)]
def find_content():
return re.findall(reg, content)
if __name__ == "__main__":
content = 'an example word:cat and word:dog'
reg = r'word:\w'
print(find_indices())
print(find_content())
| [
"re.findall",
"re.finditer"
] | [((124, 148), 're.findall', 're.findall', (['reg', 'content'], {}), '(reg, content)\n', (134, 148), False, 'import re\n'), ((64, 89), 're.finditer', 're.finditer', (['reg', 'content'], {}), '(reg, content)\n', (75, 89), False, 'import re\n')] |
import smart_imports
smart_imports.all()
class AccountPrototypeTests(utils_testcase.TestCase, personal_messages_helpers.Mixin):
def setUp(self):
super(AccountPrototypeTests, self).setUp()
self.place_1, self.place_2, self.place_3 = game_logic.create_test_map()
self.account = self.account... | [
"smart_imports.all"
] | [((23, 42), 'smart_imports.all', 'smart_imports.all', ([], {}), '()\n', (40, 42), False, 'import smart_imports\n')] |
#!/usr/bin/env python3
import json
import time
from run_common import AWSCli
from run_common import print_message
from run_create_codebuild_common import create_base_iam_policy
from run_create_codebuild_common import create_iam_service_role
from run_create_codebuild_common import create_managed_secret_iam_policy
from ... | [
"run_common.AWSCli",
"run_create_codebuild_common.create_managed_secret_iam_policy",
"run_common.print_message",
"json.dumps",
"run_create_codebuild_common.create_notification_rule",
"run_create_codebuild_common.get_notification_rule",
"time.sleep",
"run_create_codebuild_common.update_notification_rul... | [((632, 662), 'run_common.AWSCli', 'AWSCli', (["settings['AWS_REGION']"], {}), "(settings['AWS_REGION'])\n", (638, 662), False, 'from run_common import AWSCli\n'), ((1035, 1074), 'run_common.print_message', 'print_message', (['"""check previous version"""'], {}), "('check previous version')\n", (1048, 1074), False, 'fr... |
import typing as t
from pylogview import datefinder
from pylogview.record import LogRecord
if t.TYPE_CHECKING:
from pylogview.window import Window
class LogReader:
__slots__ = [
"_window",
"filename",
"_lines",
"records",
"_record_prefix_length",
"_fd",
... | [
"pylogview.datefinder.find_dates"
] | [((4113, 4174), 'pylogview.datefinder.find_dates', 'datefinder.find_dates', (['buffer_string'], {'source': '(True)', 'index': '(True)'}), '(buffer_string, source=True, index=True)\n', (4134, 4174), False, 'from pylogview import datefinder\n'), ((5490, 5551), 'pylogview.datefinder.find_dates', 'datefinder.find_dates', (... |
from mailmerge import MailMerge
import re
import os.path
from ElvantoAPIExtensions import Enums, Helpers
from modules.__stub__ import ModuleStub
class Module(ModuleStub):
__VERSION__ = "1.0"
__NAME__ = "bulletinGenerator_Kingsgrove"
# __executeTime__ = "16:00"
# __executeDay__ = "thursday"
setting... | [
"re.search",
"math.ceil",
"re.sub",
"ElvantoAPIExtensions.Helpers.NextDate",
"ElvantoAPIExtensions.Helpers.ServicesOnDate",
"mailmerge.MailMerge"
] | [((903, 938), 'ElvantoAPIExtensions.Helpers.NextDate', 'Helpers.NextDate', (['Enums.Days.SUNDAY'], {}), '(Enums.Days.SUNDAY)\n', (919, 938), False, 'from ElvantoAPIExtensions import Enums, Helpers\n'), ((5204, 5272), 're.search', 're.search', (['"""^Bible Reading (?:- )?(.*)$"""', 'scripturePassageItem.title'], {}), "(... |
"""plot.py: Utility builder class for ML plots.
Uses scikit-learn code samples and framework
"""
__author__ = "<NAME>"
__license__ = "BSD"
__email__ = "<EMAIL>"
import numpy as np
import pandas as pd
import matplotlib.pyplot as plt
from mpl_toolkits.mplot3d import Axes3D
import randomcolor
import math
from sklearn.me... | [
"matplotlib.pyplot.ylabel",
"sklearn.metrics.auc",
"numpy.argsort",
"numpy.array",
"sklearn.metrics.roc_curve",
"randomcolor.RandomColor",
"numpy.arange",
"matplotlib.pyplot.xlabel",
"matplotlib.pyplot.plot",
"matplotlib.pyplot.close",
"matplotlib.pyplot.ylim",
"matplotlib.pyplot.xticks",
"m... | [((697, 707), 'matplotlib.pyplot.show', 'plt.show', ([], {}), '()\n', (705, 707), True, 'import matplotlib.pyplot as plt\n'), ((2273, 2298), 'randomcolor.RandomColor', 'randomcolor.RandomColor', ([], {}), '()\n', (2296, 2298), False, 'import randomcolor\n'), ((3572, 3597), 'randomcolor.RandomColor', 'randomcolor.Random... |
#! /usr/bin/env python
# -*- coding: utf-8 -*
import collections
from census_data_downloader.core.tables import BaseTableConfig
from census_data_downloader.core.decorators import register
@register
class MedianAgeDownloader(BaseTableConfig):
PROCESSED_TABLE_NAME = 'medianage'
UNIVERSE = "total population"
... | [
"collections.OrderedDict"
] | [((373, 447), 'collections.OrderedDict', 'collections.OrderedDict', (["{'001': 'median', '002': 'male', '003': 'female'}"], {}), "({'001': 'median', '002': 'male', '003': 'female'})\n", (396, 447), False, 'import collections\n')] |
# Copyright 2018 Bloomberg Finance L.P.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to i... | [
"prometheus_client.Counter"
] | [((943, 1056), 'prometheus_client.Counter', 'Counter', (['POD_KILLS_METRIC_NAME', '"""Number of pods killed (including failures)"""', "['status', 'namespace', 'name']"], {}), "(POD_KILLS_METRIC_NAME, 'Number of pods killed (including failures)',\n ['status', 'namespace', 'name'])\n", (950, 1056), False, 'from promet... |
import dectate
import pytest
from kaybee.plugins.widgets.directive import WidgetDirective
from kaybee.plugins.widgets.action import WidgetAction
class Dummy:
pass
@pytest.fixture()
def widgets_kb_app():
class widgets_kb_app(dectate.App):
widget = dectate.directive(WidgetAction)
yield widgets_k... | [
"pytest.fixture",
"dectate.directive"
] | [((173, 189), 'pytest.fixture', 'pytest.fixture', ([], {}), '()\n', (187, 189), False, 'import pytest\n'), ((329, 345), 'pytest.fixture', 'pytest.fixture', ([], {}), '()\n', (343, 345), False, 'import pytest\n'), ((465, 481), 'pytest.fixture', 'pytest.fixture', ([], {}), '()\n', (479, 481), False, 'import pytest\n'), (... |
import os
import time
import random
import scipy.sparse as sp
import numpy as np
import tensorflow as tf
import argparse
from models import SpHGAT
from utils import process
parser = argparse.ArgumentParser()
parser.add_argument('--dataset', help='Dataset.', default='imdb', type=str)
parser.add_argument('--epochs', he... | [
"tensorflow.local_variables_initializer",
"numpy.random.standard_normal",
"tensorflow.sparse_placeholder",
"utils.process.load_heterogeneous_data",
"numpy.array",
"utils.process.preprocess_features",
"tensorflow.Graph",
"argparse.ArgumentParser",
"tensorflow.placeholder",
"tensorflow.Session",
"... | [((184, 209), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {}), '()\n', (207, 209), False, 'import argparse\n'), ((2336, 2377), 'tensorflow.ConfigProto', 'tf.ConfigProto', ([], {'allow_soft_placement': '(True)'}), '(allow_soft_placement=True)\n', (2350, 2377), True, 'import tensorflow as tf\n'), ((3550, 3... |
from django import forms
from .models import Problem
class ProblemForm(forms.ModelForm):
options = (
('A', 'A'),
('B', 'B'),
('C', 'C'),
('D', 'D'),
)
choice = forms.ChoiceField(choices=options)
class Meta:
model = Problem
fields = '__all__'
# exclude = ['answer']
widgets = {
'answer': forms.Hid... | [
"django.forms.ChoiceField",
"django.forms.HiddenInput",
"django.forms.Textarea",
"django.forms.TextInput"
] | [((172, 206), 'django.forms.ChoiceField', 'forms.ChoiceField', ([], {'choices': 'options'}), '(choices=options)\n', (189, 206), False, 'from django import forms\n'), ((311, 330), 'django.forms.HiddenInput', 'forms.HiddenInput', ([], {}), '()\n', (328, 330), False, 'from django import forms\n'), ((350, 421), 'django.for... |
#!/usr/bin/env python
"""
Create a spreadsheet with two tables, using some named ranges.
"""
import os
from odfdo import Document, Table
if __name__ == "__main__":
document = Document('spreadsheet')
body = document.body
table = Table("First Table")
body.append(table)
# populate the table :
for... | [
"os.path.exists",
"odfdo.Document",
"os.path.join",
"os.mkdir",
"odfdo.Table"
] | [((181, 204), 'odfdo.Document', 'Document', (['"""spreadsheet"""'], {}), "('spreadsheet')\n", (189, 204), False, 'from odfdo import Document, Table\n'), ((242, 262), 'odfdo.Table', 'Table', (['"""First Table"""'], {}), "('First Table')\n", (247, 262), False, 'from odfdo import Document, Table\n'), ((1017, 1038), 'odfdo... |
from nltk.corpus.reader import CategorizedPlaintextCorpusReader
from nltk.tokenize.casual import TweetTokenizer
from nltk.classify.scikitlearn import SklearnClassifier
from sklearn.naive_bayes import BernoulliNB, MultinomialNB
from sklearn.svm import SVC, LinearSVC, NuSVC, LinearSVR, NuSVR
from sklearn.linear_model im... | [
"nltk.pos_tag",
"utils.precision_recall_2step",
"nltk.corpus.reader.CategorizedPlaintextCorpusReader",
"featureExtractors.unigramsFeatures",
"featureExtractors.bigramsFeatures",
"pickle.load",
"lexicons.mpqa.mpqaDictionary.MpqaDictionaryWrapper",
"normalization.normalizeTwitterWordsWithExtraFeatures",... | [((831, 903), 'nltk.tokenize.casual.TweetTokenizer', 'TweetTokenizer', ([], {'reduce_len': '(True)', 'preserve_case': '(True)', 'strip_handles': '(False)'}), '(reduce_len=True, preserve_case=True, strip_handles=False)\n', (845, 903), False, 'from nltk.tokenize.casual import TweetTokenizer\n'), ((917, 1089), 'nltk.corpu... |
# # ⚠ Warning
#
# THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT
# LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN
# NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIA... | [
"time.time",
"decimal.Decimal"
] | [((7684, 7695), 'time.time', 'time.time', ([], {}), '()\n', (7693, 7695), False, 'import time\n'), ((5951, 5965), 'decimal.Decimal', 'Decimal', (['""".01"""'], {}), "('.01')\n", (5958, 5965), False, 'from decimal import Decimal\n'), ((6025, 6039), 'decimal.Decimal', 'Decimal', (['""".01"""'], {}), "('.01')\n", (6032, 6... |
#!/usr/bin/env python
# -*- coding: utf-8 -*-
# @Author : yag8009
# @FileName : md5_hede
# @Time : 2020/3/18
import hashlib
import time
def md5_hede(md5data):
md5 = hashlib.md5() # 使用MD5加密模式
md5.update(md5data.encode("utf8")) # 将参数字符串传入
sign = md5.hexdigest()
return sign
if __name__ == '__mai... | [
"time.time",
"hashlib.md5"
] | [((177, 190), 'hashlib.md5', 'hashlib.md5', ([], {}), '()\n', (188, 190), False, 'import hashlib\n'), ((573, 584), 'time.time', 'time.time', ([], {}), '()\n', (582, 584), False, 'import time\n')] |
import os
import datetime
from dateutil.parser import parse as dateutil_parser
from jinja2 import Environment, PackageLoader
from kinto_http import cli_utils
from . import constants
from .logger import logger
JSON_DATE_FORMAT = "%Y-%m-%dT%H:%M:%SZ"
COLLECTION_FORMAT = '/buckets/{bucket_id}/collections/{collection_i... | [
"dateutil.parser.parse",
"os.path.exists",
"jinja2.Environment",
"os.makedirs",
"os.path.join",
"kinto_http.cli_utils.add_parser_options",
"datetime.date.fromtimestamp",
"kinto_http.cli_utils.create_client_from_args",
"kinto_http.cli_utils.setup_logger",
"jinja2.PackageLoader"
] | [((359, 381), 'dateutil.parser.parse', 'dateutil_parser', (['value'], {}), '(value)\n', (374, 381), True, 'from dateutil.parser import parse as dateutil_parser\n'), ((459, 498), 'jinja2.PackageLoader', 'PackageLoader', (['"""amo2kinto"""', '"""templates"""'], {}), "('amo2kinto', 'templates')\n", (472, 498), False, 'fro... |
"""Make some fields on Chart and Table nullable
We want to copy chart and table data across to these tables but have no way to add a
classification for each one, so we'll have to live with some nulls in here.
Revision ID: 2019_03_04_make_fields_nullable
Revises: 2019_03_04_chart_table_settings
Create Date: 2019-03-05... | [
"sqlalchemy.VARCHAR",
"sqlalchemy.BOOLEAN"
] | [((657, 679), 'sqlalchemy.VARCHAR', 'sa.VARCHAR', ([], {'length': '(255)'}), '(length=255)\n', (667, 679), True, 'import sqlalchemy as sa\n'), ((765, 777), 'sqlalchemy.BOOLEAN', 'sa.BOOLEAN', ([], {}), '()\n', (775, 777), True, 'import sqlalchemy as sa\n'), ((867, 879), 'sqlalchemy.BOOLEAN', 'sa.BOOLEAN', ([], {}), '()... |
from django.contrib import admin
from .models import ContactForm
class ContactFormAdmin(admin.ModelAdmin):
class Meta:
model = ContactForm
admin.site.register(ContactForm, ContactFormAdmin)
| [
"django.contrib.admin.site.register"
] | [((154, 204), 'django.contrib.admin.site.register', 'admin.site.register', (['ContactForm', 'ContactFormAdmin'], {}), '(ContactForm, ContactFormAdmin)\n', (173, 204), False, 'from django.contrib import admin\n')] |
import numpy as np
import h5py
import os
import sys
from copy import deepcopy
#handle .(period) and slash specially since it is part of path
#replace with \period or \slash-forward when store, recover later
#not using '\forward-slash' is because \f is a special character
PERIOD='\period'
SLASH='\slash-forward'
'''
... | [
"os.path.join",
"h5py.File",
"numpy.array",
"sys.stdout.flush",
"sys.stdout.write"
] | [((892, 915), 'h5py.File', 'h5py.File', (['f_name', 'mode'], {}), '(f_name, mode)\n', (901, 915), False, 'import h5py\n'), ((1759, 1781), 'h5py.File', 'h5py.File', (['f_name', '"""r"""'], {}), "(f_name, 'r')\n", (1768, 1781), False, 'import h5py\n'), ((1923, 1943), 'numpy.array', 'np.array', (['group[key]'], {}), '(gro... |
import re
import string
from libs.dataset.core import SampleEntry, oov_replacement_vocabulary, letter_replacements
__word_start_regex = f'[ \t\n]|^|[{string.punctuation}]'
__word_end_regex = f'[ \t\n]|$|[{string.punctuation}]'
def does_not_have_numbers(s: SampleEntry):
return not any(char.isdigit() for char in s... | [
"re.sub",
"libs.dataset.core.oov_replacement_vocabulary.items",
"libs.dataset.core.letter_replacements.items"
] | [((656, 690), 'libs.dataset.core.oov_replacement_vocabulary.items', 'oov_replacement_vocabulary.items', ([], {}), '()\n', (688, 690), False, 'from libs.dataset.core import SampleEntry, oov_replacement_vocabulary, letter_replacements\n'), ((758, 786), 're.sub', 're.sub', (['"""[ ]{2,}"""', '""" """', 'line'], {}), "('[ ... |
import unittest
import silasdk
from tests.test_config import (
app, eth_private_key, eth_private_key_4, instant_ach_handle, user_handle)
class Test007CheckInstantAchTest(unittest.TestCase):
def test_check_instant_ach(self):
payload = {
"user_handle": instant_ach_handle,
"accou... | [
"unittest.main",
"silasdk.User.check_instant_ach"
] | [((601, 616), 'unittest.main', 'unittest.main', ([], {}), '()\n', (614, 616), False, 'import unittest\n'), ((376, 439), 'silasdk.User.check_instant_ach', 'silasdk.User.check_instant_ach', (['app', 'payload', 'eth_private_key_4'], {}), '(app, payload, eth_private_key_4)\n', (406, 439), False, 'import silasdk\n')] |
# Copyright 2021 Google LLC
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, ... | [
"bigbench.api.util.postprocess_output",
"numpy.log"
] | [((1171, 1239), 'bigbench.api.util.postprocess_output', 'util.postprocess_output', (['text', 'max_length', 'stop_string', 'output_regex'], {}), '(text, max_length, stop_string, output_regex)\n', (1194, 1239), True, 'import bigbench.api.util as util\n'), ((1417, 1436), 'numpy.log', 'np.log', (['(1 / (i + 1))'], {}), '(1... |
"""Click parameter types for osxphotos CLI"""
import datetime
import os
import pathlib
import re
import bitmath
import click
import pytimeparse2
from osxphotos.export_db_utils import export_db_get_version
from osxphotos.photoinfo import PhotoInfoNone
from osxphotos.phototemplate import PhotoTemplate, RenderOptions
fr... | [
"osxphotos.photoinfo.PhotoInfoNone",
"osxphotos.utils.expand_and_validate_filepath",
"osxphotos.utils.load_function",
"pytimeparse2.parse",
"pathlib.Path",
"re.match",
"os.getcwd",
"click.BadParameter",
"bitmath.Byte",
"osxphotos.phototemplate.RenderOptions",
"osxphotos.timeutils.time_string_to_... | [((2547, 2585), 'osxphotos.utils.expand_and_validate_filepath', 'expand_and_validate_filepath', (['filename'], {}), '(filename)\n', (2575, 2585), False, 'from osxphotos.utils import expand_and_validate_filepath, load_function\n'), ((5048, 5083), 're.match', 're.match', (['"""^[+-]?\\\\s*?\\\\d+$"""', 'value'], {}), "('... |
# -*- coding: utf-8 -*-
import datetime
from south.db import db
from south.v2 import SchemaMigration
from django.db import models
class Migration(SchemaMigration):
def forwards(self, orm):
# Removing unique constraint on 'Licence', fields ['sn']
db.delete_unique('ralph_assets_licence', ['sn'])
... | [
"south.db.db.delete_column",
"south.db.db.send_create_signal",
"south.db.db.delete_unique",
"django.db.models.ForeignKey",
"south.db.db.create_unique",
"django.db.models.AutoField",
"south.db.db.delete_table"
] | [((269, 317), 'south.db.db.delete_unique', 'db.delete_unique', (['"""ralph_assets_licence"""', "['sn']"], {}), "('ralph_assets_licence', ['sn'])\n", (285, 317), False, 'from south.db import db\n'), ((1526, 1587), 'south.db.db.send_create_signal', 'db.send_create_signal', (['"""ralph_assets"""', "['TransitionsHistory']"... |
import connexion
import six
from openapi_server import query_manager
from openapi_server.utils.vars import VARIABLE_TYPE_NAME, VARIABLE_TYPE_URI
from openapi_server.models.variable import Variable # noqa: E501
from openapi_server import util
def variables_get(username=None, label=None, page=None, per_page=None): # ... | [
"openapi_server.query_manager.delete_resource",
"openapi_server.query_manager.post_resource",
"openapi_server.query_manager.get_resource",
"connexion.request.get_json",
"openapi_server.query_manager.put_resource"
] | [((776, 953), 'openapi_server.query_manager.get_resource', 'query_manager.get_resource', ([], {'username': 'username', 'label': 'label', 'page': 'page', 'per_page': 'per_page', 'rdf_type_uri': 'VARIABLE_TYPE_URI', 'rdf_type_name': 'VARIABLE_TYPE_NAME', 'kls': 'Variable'}), '(username=username, label=label, page=page,\n... |
import pycxsimulator
from pylab import *
import copy as cp
nr = 500. # carrying capacity of rabbits
r_init = 100 # initial rabbit population
mr = 0.03 # magnitude of movement of rabbits
dr = 1.0 # death rate of rabbits when it faces foxes
rr = 0.1 # reproduction rate of rabbits
f_init = 30 # initial fox population
... | [
"copy.copy",
"pycxsimulator.GUI"
] | [((2420, 2439), 'pycxsimulator.GUI', 'pycxsimulator.GUI', ([], {}), '()\n', (2437, 2439), False, 'import pycxsimulator\n'), ((1988, 1999), 'copy.copy', 'cp.copy', (['ag'], {}), '(ag)\n', (1995, 1999), True, 'import copy as cp\n'), ((2267, 2278), 'copy.copy', 'cp.copy', (['ag'], {}), '(ag)\n', (2274, 2278), True, 'impor... |
# Copyright 2021 The ML Collections Authors.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed... | [
"pickle.dumps",
"ml_collections.FrozenConfigDict",
"ml_collections.ConfigDict",
"absl.testing.absltest.main",
"copy.deepcopy",
"ml_collections.FieldReference"
] | [((1167, 1208), 'ml_collections.FieldReference', 'ml_collections.FieldReference', (["{'int': 0}"], {}), "({'int': 0})\n", (1196, 1208), False, 'import ml_collections\n'), ((1249, 1274), 'copy.deepcopy', 'copy.deepcopy', (['_TEST_DICT'], {}), '(_TEST_DICT)\n', (1262, 1274), False, 'import copy\n'), ((1310, 1347), 'ml_co... |
#!/usr/bin/env python
# -*- coding: UTF-8 -*-
'''
パッケージpyserialをインストールすること
pytho2.x系で動作(python3.*系も動作検証済み)
Creater:<NAME>
'''
import serial
import binascii
import signal
import sys
import platform
from serial.tools import list_ports
#platformの切り替え
if platform.system() == 'Windows': #windows用
ports = list_ports.co... | [
"signal.signal",
"serial.tools.list_ports.comports",
"platform.system",
"serial.Serial",
"sys.exit"
] | [((252, 269), 'platform.system', 'platform.system', ([], {}), '()\n', (267, 269), False, 'import platform\n'), ((307, 328), 'serial.tools.list_ports.comports', 'list_ports.comports', ([], {}), '()\n', (326, 328), False, 'from serial.tools import list_ports\n'), ((793, 803), 'sys.exit', 'sys.exit', ([], {}), '()\n', (80... |
from test.webdnn_test.graph_test.operators_test.util import template_test_unary_operator
from webdnn.graph.operators.sigmoid import Sigmoid
def template():
template_test_unary_operator(Sigmoid)
def test():
template()
| [
"test.webdnn_test.graph_test.operators_test.util.template_test_unary_operator"
] | [((162, 199), 'test.webdnn_test.graph_test.operators_test.util.template_test_unary_operator', 'template_test_unary_operator', (['Sigmoid'], {}), '(Sigmoid)\n', (190, 199), False, 'from test.webdnn_test.graph_test.operators_test.util import template_test_unary_operator\n')] |
import pybullet_data
import pybullet as p
import time
import numpy as np
from src.utils_geom import *
from src.utils_depth import *
from src.panda import Panda
def full_jacob_pb(jac_t, jac_r):
return np.vstack((jac_t[0], jac_t[1], jac_t[2], jac_r[0], jac_r[1], jac_r[2]))
class pandaEnv():
def __init__(self,
... | [
"pybullet_data.getDataPath",
"numpy.linalg.pinv",
"numpy.hstack",
"pybullet.setTimeStep",
"pybullet.setGravity",
"numpy.array",
"pybullet.setPhysicsEngineParameter",
"src.panda.Panda",
"pybullet.calculateJacobian",
"pybullet.createConstraint",
"numpy.vstack",
"pybullet.changeConstraint",
"py... | [((207, 278), 'numpy.vstack', 'np.vstack', (['(jac_t[0], jac_t[1], jac_t[2], jac_r[0], jac_r[1], jac_r[2])'], {}), '((jac_t[0], jac_t[1], jac_t[2], jac_r[0], jac_r[1], jac_r[2]))\n', (216, 278), True, 'import numpy as np\n'), ((333, 360), 'pybullet_data.getDataPath', 'pybullet_data.getDataPath', ([], {}), '()\n', (358,... |
import datetime
import decimal
import enum
import typing as T
import uuid
import graphene
import graphene.types
import pydantic
import pytest
from pydantic import BaseModel, create_model
import graphene_pydantic.converters as converters
from graphene_pydantic.converters import ConversionError, convert_pydantic_field
... | [
"datetime.datetime",
"datetime.time",
"uuid.uuid4",
"pytest.raises",
"datetime.date",
"pydantic.create_model",
"graphene_pydantic.registry.get_global_registry",
"decimal.Decimal"
] | [((558, 589), 'pydantic.create_model', 'create_model', (['"""model"""'], {}), "('model', **kwargs)\n", (570, 589), False, 'from pydantic import BaseModel, create_model\n'), ((781, 820), 'graphene_pydantic.registry.get_global_registry', 'get_global_registry', (['PydanticObjectType'], {}), '(PydanticObjectType)\n', (800,... |
import numpy as np
import pandas as pd
from pathlib import Path
import multiprocessing as mp
from functools import partial
from src.audio import read_as_melspectrogram
from src.utils import get_params_hash
from src import config
NOISE_SOUNDS = [
'Buzz',
'Car_passing_by',
'Crackle',
'Cricket',
'Hi... | [
"pandas.read_csv",
"pathlib.Path",
"src.audio.read_as_melspectrogram",
"multiprocessing.cpu_count",
"functools.partial",
"numpy.save"
] | [((906, 953), 'src.audio.read_as_melspectrogram', 'read_as_melspectrogram', (['file_path', 'audio_params'], {}), '(file_path, audio_params)\n', (928, 953), False, 'from src.audio import read_as_melspectrogram\n'), ((1194, 1208), 'pathlib.Path', 'Path', (['dir_path'], {}), '(dir_path)\n', (1198, 1208), False, 'from path... |
# Copyright (c) 2015, Frappe Technologies Pvt. Ltd. and Contributors
# License: GNU General Public License v3. See license.txt
from __future__ import unicode_literals
import frappe, unittest
from erpnext.accounts.doctype.fiscal_year.fiscal_year import FiscalYearIncorrectDate
test_records = frappe.get_test_records('... | [
"frappe.db.exists",
"frappe.delete_doc",
"frappe.get_test_records",
"frappe.get_doc"
] | [((295, 333), 'frappe.get_test_records', 'frappe.get_test_records', (['"""Fiscal Year"""'], {}), "('Fiscal Year')\n", (318, 333), False, 'import frappe, unittest\n'), ((436, 493), 'frappe.db.exists', 'frappe.db.exists', (['"""Fiscal Year"""', '"""_Test Fiscal Year 2000"""'], {}), "('Fiscal Year', '_Test Fiscal Year 200... |
from typing import Callable, Generator, Any
from unittest import mock
import pytest
from screenpy import AnActor, pacing, settings
from screenpy.abilities import AuthenticateWith2FA, BrowseTheWeb, MakeAPIRequests
from screenpy.narration.narrator import Narrator
@pytest.fixture(scope="function")
def Tester() -> AnAc... | [
"pytest.fixture",
"unittest.mock.MagicMock",
"unittest.mock.Mock",
"screenpy.AnActor.named"
] | [((267, 299), 'pytest.fixture', 'pytest.fixture', ([], {'scope': '"""function"""'}), "(scope='function')\n", (281, 299), False, 'import pytest\n'), ((717, 749), 'pytest.fixture', 'pytest.fixture', ([], {'scope': '"""function"""'}), "(scope='function')\n", (731, 749), False, 'import pytest\n'), ((1021, 1053), 'pytest.fi... |
import numpy as np
import pandas as pd
from sklearn.externals import joblib
#from sklearn.ensemble import RandomForestRegressor
#from sklearn.multioutput import MultiOutputRegressor
#from sklearn.multioutput import MultiOutputRegressor
from sklearn.model_selection import train_test_split
df = pd.read_csv('https://dr... | [
"numpy.abs",
"numpy.mean",
"pandas.read_csv",
"sklearn.externals.joblib.load",
"pandas.get_dummies"
] | [((297, 411), 'pandas.read_csv', 'pd.read_csv', (['"""https://drive.google.com/uc?export=download&id=1XoV8SfvHmzaxRuDRe81OWSQu10dYTbO5"""'], {'sep': '""","""'}), "(\n 'https://drive.google.com/uc?export=download&id=1XoV8SfvHmzaxRuDRe81OWSQu10dYTbO5'\n , sep=',')\n", (308, 411), True, 'import pandas as pd\n'), ((4... |
#! /usr/bin/env python
# -*- coding: utf-8 -*-
import os
import argparse
import docx
from docx.shared import Cm
import pylatex
from pytablewriter import MarkdownTableWriter
def set_column_width(column, width):
column.width = width
for cell in column.cells:
cell.width = width
def generate_word_revisio... | [
"os.path.exists",
"pylatex.Section",
"argparse.ArgumentParser",
"pylatex.LongTable",
"docx.shared.Cm",
"pytablewriter.MarkdownTableWriter",
"pylatex.Document",
"docx.Document"
] | [((389, 404), 'docx.Document', 'docx.Document', ([], {}), '()\n', (402, 404), False, 'import docx\n'), ((1377, 1428), 'pylatex.Document', 'pylatex.Document', ([], {'geometry_options': 'geometry_options'}), '(geometry_options=geometry_options)\n', (1393, 1428), False, 'import pylatex\n'), ((2242, 2263), 'pytablewriter.M... |
import torch
import numpy as np
import utility
from decimal import Decimal
from tqdm import tqdm
from option import args
from torchvision import transforms
from PIL import Image
import matplotlib
matplotlib.use('TkAgg')
import matplotlib.pyplot as plt
import copy
class Trainer():
def __init__(self, opt, loader... | [
"utility.timer",
"utility.quantize",
"utility.calc_psnr",
"matplotlib.use",
"utility.make_dual_scheduler",
"tqdm.tqdm",
"utility.SSIM",
"utility.make_optimizer",
"utility.make_dual_optimizer",
"utility.make_scheduler",
"torch.no_grad",
"torch.zeros",
"decimal.Decimal",
"torch.device"
] | [((200, 223), 'matplotlib.use', 'matplotlib.use', (['"""TkAgg"""'], {}), "('TkAgg')\n", (214, 223), False, 'import matplotlib\n'), ((601, 640), 'utility.make_optimizer', 'utility.make_optimizer', (['opt', 'self.model'], {}), '(opt, self.model)\n', (623, 640), False, 'import utility\n'), ((666, 709), 'utility.make_sched... |
# Get the database using the method we defined in pymongo_test_insert file
from pymongo_test_insert import get_database
dbname = get_database()
# Create a new collection
collection_name = dbname["user_1_items"]
item_details = collection_name.find()
for item in item_details:
# This will give readable output... | [
"pandas.DataFrame",
"pymongo_test_insert.get_database"
] | [((132, 146), 'pymongo_test_insert.get_database', 'get_database', ([], {}), '()\n', (144, 146), False, 'from pymongo_test_insert import get_database\n'), ((710, 733), 'pandas.DataFrame', 'DataFrame', (['item_details'], {}), '(item_details)\n', (719, 733), False, 'from pandas import DataFrame\n')] |
import json
from django.core.serializers.json import DjangoJSONEncoder
from django.http import HttpResponse
from django.template import Context
from django.template import RequestContext
from django.template.loader import render_to_string, select_template
from django.utils.encoding import force_unicode
from ..compat i... | [
"django.http.HttpResponse",
"django.utils.encoding.force_unicode",
"json.dumps",
"django.template.RequestContext"
] | [((730, 780), 'json.dumps', 'json.dumps', (['json_data'], {'cls': 'self.json_encoder_class'}), '(json_data, cls=self.json_encoder_class)\n', (740, 780), False, 'import json\n'), ((879, 980), 'django.http.HttpResponse', 'HttpResponse', (['json_serialized'], {'content_type': '"""application/json; charset=utf-8"""'}), "(j... |
#! /usr/bin/env python
from setuptools import setup, Extension
import importlib
import os
# copied from kymatio's setup.py: https://github.com/kymatio/kymatio/blob/master/setup.py
sfm_version_spec = importlib.util.spec_from_file_location('sfm_version', 'sfm/version.py')
sfm_version_module = importlib.util.module_from... | [
"importlib.util.spec_from_file_location",
"setuptools.setup",
"importlib.util.module_from_spec"
] | [((201, 272), 'importlib.util.spec_from_file_location', 'importlib.util.spec_from_file_location', (['"""sfm_version"""', '"""sfm/version.py"""'], {}), "('sfm_version', 'sfm/version.py')\n", (239, 272), False, 'import importlib\n'), ((294, 343), 'importlib.util.module_from_spec', 'importlib.util.module_from_spec', (['sf... |
# from pipet.core.sql.query_interface import *
from pypipet.core.operations.inventory import *
import pytest
from pprint import pprint
_supplie_id = 1
def test_update_invs(session, obj_classes, shop_conn):
invs = [ {'sku':'s22456', 'supplier_id':_supplie_id, 'qty':20}]
update_inventory_bulk(obj_classes, sess... | [
"pprint.pprint"
] | [((502, 513), 'pprint.pprint', 'pprint', (['res'], {}), '(res)\n', (508, 513), False, 'from pprint import pprint\n')] |
# Costs
# 2019 - Luque, Straub - Risk-based optimal inspection strategies for
# structural systems using dynamic Bayesian networks
# Table 4, case 1
import numpy as np
class InspectionMaintenance:
"""
Inspection and Maintenance
==========================
Cost calculation for inspection and main... | [
"numpy.unique",
"numpy.diff",
"numpy.array",
"numpy.vstack",
"numpy.zeros_like"
] | [((1359, 1392), 'numpy.vstack', 'np.vstack', (['system_model.system_pf'], {}), '(system_model.system_pf)\n', (1368, 1392), True, 'import numpy as np\n'), ((1417, 1427), 'numpy.diff', 'np.diff', (['t'], {}), '(t)\n', (1424, 1427), True, 'import numpy as np\n'), ((1452, 1463), 'numpy.diff', 'np.diff', (['pf'], {}), '(pf)... |
from __future__ import absolute_import, division, print_function
import json
import os.path as op
import six
import numpy as np
import scipy.signal as sig
import scipy.io.wavfile as sciwav
MAXINT16 = 2**15 - 1
FS = 44100
COEFF_DIR = op.join(op.dirname(op.abspath(__file__)), 'coeffs')
def normalize(data, maxamp=1):
... | [
"numpy.random.random",
"os.path.join",
"json.load",
"numpy.array",
"scipy.signal.lfilter",
"scipy.io.wavfile.read",
"numpy.around",
"numpy.random.seed",
"scipy.io.wavfile.write",
"os.path.abspath",
"numpy.savetxt",
"numpy.loadtxt"
] | [((255, 275), 'os.path.abspath', 'op.abspath', (['__file__'], {}), '(__file__)\n', (265, 275), True, 'import os.path as op\n'), ((443, 455), 'json.load', 'json.load', (['f'], {}), '(f)\n', (452, 455), False, 'import json\n'), ((565, 585), 'numpy.random.seed', 'np.random.seed', (['seed'], {}), '(seed)\n', (579, 585), Tr... |
# Standard library
import atexit
import os
os.environ["OMP_NUM_THREADS"] = "1"
import sys
import traceback
# Third-party
from astropy.utils import iers
iers.conf.auto_download = False
import astropy.table as at
import numpy as np
# This project
from totoro.config import cache_path
from totoro.data import datasets, el... | [
"os.path.exists",
"numpy.random.default_rng",
"argparse.ArgumentParser",
"astropy.table.Table",
"astropy.table.vstack",
"totoro.data.datasets.items",
"numpy.array",
"os.unlink",
"sys.exit",
"totoro.objective.TorusImagingObjective",
"traceback.print_exc",
"atexit.register",
"glob.glob",
"as... | [((1200, 1228), 'os.path.exists', 'os.path.exists', (['all_filename'], {}), '(all_filename)\n', (1214, 1228), False, 'import os\n'), ((1430, 1457), 'glob.glob', 'glob.glob', (['cache_glob_pattr'], {}), '(cache_glob_pattr)\n', (1439, 1457), False, 'import glob\n'), ((2009, 2025), 'totoro.data.datasets.items', 'datasets.... |
"""
Problem:
You are given a huge list of airline ticket prices between different cities around the
world on a given day. These are all direct flights. Each element in the list has the
format (source_city, destination, price).
Consider a user who is willing to take up to k connections from their origin city A to
thei... | [
"DataStructures.Graph.GraphDirectedWeighted",
"DataStructures.PriorityQueue.MinPriorityQueue"
] | [((1315, 1333), 'DataStructures.PriorityQueue.MinPriorityQueue', 'MinPriorityQueue', ([], {}), '()\n', (1331, 1333), False, 'from DataStructures.PriorityQueue import MinPriorityQueue\n'), ((2281, 2304), 'DataStructures.Graph.GraphDirectedWeighted', 'GraphDirectedWeighted', ([], {}), '()\n', (2302, 2304), False, 'from D... |
"""PyTorch Distributed Data Parallel example from NVIDIA."""
# https://github.com/NVIDIA/DeepLearningExamples
import argparse
import utils
import virtual_machine
def main():
parser = argparse.ArgumentParser(description='Optional app description')
parser.add_argument('--vm-name', dest='vm_name', type=str, require... | [
"utils.run_threads",
"virtual_machine.VirtualMachine",
"argparse.ArgumentParser"
] | [((188, 251), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {'description': '"""Optional app description"""'}), "(description='Optional app description')\n", (211, 251), False, 'import argparse\n'), ((2429, 2479), 'utils.run_threads', 'utils.run_threads', (['vms', 'commands'], {'login_shell': '(True)'}), '... |
from torch.utils.data import Dataset
from skimage import io
import os
import torch
class MnistData(Dataset):
def __init__(self, root_dir):
self.root_dir = root_dir
img_list = []
label_list = os.listdir(self.root_dir)
for label in label_list:
file_names = os.li... | [
"skimage.io.imread",
"os.listdir",
"os.path.join"
] | [((229, 254), 'os.listdir', 'os.listdir', (['self.root_dir'], {}), '(self.root_dir)\n', (239, 254), False, 'import os\n'), ((735, 773), 'os.path.join', 'os.path.join', (['self.root_dir', 'img_label'], {}), '(self.root_dir, img_label)\n', (747, 773), False, 'import os\n'), ((793, 825), 'os.path.join', 'os.path.join', ([... |
from qunetsim.backends.rw_lock import RWLock
from qunetsim.objects.logger import Logger
import queue
class QuantumStorage(object):
"""
An object which stores qubits.
"""
STORAGE_LIMIT_ALL = 1
STORAGE_LIMIT_PER_HOST = 2
STORAGE_LIMIT_INDIVIDUALLY_PER_HOST = 3
def __init__(self):
#... | [
"qunetsim.objects.logger.Logger.get_instance",
"queue.Queue",
"qunetsim.backends.rw_lock.RWLock"
] | [((1044, 1052), 'qunetsim.backends.rw_lock.RWLock', 'RWLock', ([], {}), '()\n', (1050, 1052), False, 'from qunetsim.backends.rw_lock import RWLock\n'), ((1076, 1097), 'qunetsim.objects.logger.Logger.get_instance', 'Logger.get_instance', ([], {}), '()\n', (1095, 1097), False, 'from qunetsim.objects.logger import Logger\... |
# -*- coding: utf-8 -*-
import datetime
from pagseguro.utils import (is_valid_cpf, is_valid_cnpj, is_valid_email,
parse_date)
from pagseguro.exceptions import PagSeguroValidationError
import pytest
from dateutil.tz import tzutc
def test_is_valid_email():
valid = '<EMAIL>'
valid2... | [
"pagseguro.utils.is_valid_cpf",
"dateutil.tz.tzutc",
"pagseguro.utils.parse_date",
"pagseguro.utils.is_valid_cnpj",
"pytest.raises",
"pagseguro.utils.is_valid_email"
] | [((427, 466), 'pytest.raises', 'pytest.raises', (['PagSeguroValidationError'], {}), '(PagSeguroValidationError)\n', (440, 466), False, 'import pytest\n'), ((476, 501), 'pagseguro.utils.is_valid_email', 'is_valid_email', (['not_valid'], {}), '(not_valid)\n', (490, 501), False, 'from pagseguro.utils import is_valid_cpf, ... |
#-*- conding:utf-8 -*-
#2018-02-02 11:04:54
import time,random,requests
from lxml import etree
from fake_useragent import UserAgent
import urllib.request as ur
from pymongo import MongoClient
from multiprocessing.dummy import Pool as ThreadPool
client = MongoClient('localhost',27017)
db = client['test']
ip_list = db[... | [
"urllib.request.install_opener",
"urllib.request.ProxyHandler",
"requests.get",
"urllib.request.build_opener",
"lxml.etree.HTML",
"multiprocessing.dummy.Pool",
"pymongo.MongoClient",
"random.random",
"fake_useragent.UserAgent"
] | [((256, 287), 'pymongo.MongoClient', 'MongoClient', (['"""localhost"""', '(27017)'], {}), "('localhost', 27017)\n", (267, 287), False, 'from pymongo import MongoClient\n'), ((336, 347), 'fake_useragent.UserAgent', 'UserAgent', ([], {}), '()\n', (345, 347), False, 'from fake_useragent import UserAgent\n'), ((1119, 1141)... |
#!/usr/bin/python3
# -*- coding: utf-8 -*-
"""
Read & write snapshot of node status data dumped to local file system.
This data is used to detect things like sytem changes that may need to be reported.
"""
import json
import time
import logging
import os
import socket
import psutil
import netifaces as ni
logging.ba... | [
"logging.basicConfig",
"os.path.exists",
"logging.getLogger",
"time.ctime",
"netifaces.ifaddresses",
"psutil.boot_time",
"time.time",
"json.load",
"netifaces.interfaces",
"socket.gethostname",
"json.dump"
] | [((310, 349), 'logging.basicConfig', 'logging.basicConfig', ([], {'level': 'logging.INFO'}), '(level=logging.INFO)\n', (329, 349), False, 'import logging\n'), ((365, 398), 'logging.getLogger', 'logging.getLogger', (['"""DataSnapshot"""'], {}), "('DataSnapshot')\n", (382, 398), False, 'import logging\n'), ((1889, 1920),... |
"""
開発初期段階では、サイトレポートをVIEWに直書きしていた
パフォーマンス改善のため、バッチ処理に変更
このVIEWは今は使われていないが、一応残しておく
"""
from django.conf import settings
from django.db.models import Q, Count
from django.utils import html
from .models import Post, Category, Tag
from datetime import datetime
from janome.tokenizer import Tokenizer
from janome... | [
"janome.analyzer.Analyzer",
"django.db.models.Count",
"itertools.product",
"janome.tokenizer.Tokenizer",
"django.db.models.Q"
] | [((3315, 3359), 'itertools.product', 'itertools.product', (['month_list', 'category_list'], {}), '(month_list, category_list)\n', (3332, 3359), False, 'import itertools\n'), ((4727, 4792), 'janome.tokenizer.Tokenizer', 'Tokenizer', ([], {'udic': 'udic_path', 'udic_type': '"""simpledic"""', 'udic_enc': '"""utf8"""'}), "... |
import os
def check(cmd, mf):
m = mf.findNode('matplotlib')
if m is None or m.filename is None:
return None
if cmd.matplotlib_backends:
backends = {}
for backend in cmd.matplotlib_backends:
if backend == '-':
pass
elif backend == '*':
... | [
"os.path.dirname"
] | [((669, 696), 'os.path.dirname', 'os.path.dirname', (['m.filename'], {}), '(m.filename)\n', (684, 696), False, 'import os\n')] |
# Generated by Django 3.1.6 on 2021-02-14 18:13
from django.db import migrations, models
class Migration(migrations.Migration):
dependencies = [
('core', '0008_auto_20210214_2039'),
]
operations = [
migrations.AlterField(
model_name='device',
name='ip_address',
... | [
"django.db.models.GenericIPAddressField",
"django.db.models.CharField"
] | [((337, 450), 'django.db.models.GenericIPAddressField', 'models.GenericIPAddressField', ([], {'blank': '(True)', 'help_text': '"""e.g. 192.168.0.17"""', 'null': '(True)', 'verbose_name': '"""IP address"""'}), "(blank=True, help_text='e.g. 192.168.0.17',\n null=True, verbose_name='IP address')\n", (365, 450), False, ... |
import requests
import re
pattern_hashes = "^[a-f0-9]{64}$"
pattern_url = "https?:\/\/.*"
pattern_ip = "(?:(?:\d|[01]?\d\d|2[0-4]\d|25[0-5])\.){3}(?:25[0-5]|2[0-4]\d|[01]?\d\d|\d)(?:\/\d{1,2})?"
pattern_domain = "^[a-zA-Z0-9][a-zA-Z0-9-_]{0,61}[a-zA-Z0-9]{0,1}\.([a-zA-Z]{1,6}|[a-zA-Z0-9-]{1,30}\.[a-zA-Z]{2,3})$"
head... | [
"re.findall",
"requests.get"
] | [((366, 438), 'requests.get', 'requests.get', (['"""https://labs.inquest.net/api/iocdb/list"""'], {'headers': 'headers'}), "('https://labs.inquest.net/api/iocdb/list', headers=headers)\n", (378, 438), False, 'import requests\n'), ((604, 640), 're.findall', 're.findall', (['pattern', 'data', 're.DOTALL'], {}), '(pattern... |
# coding: utf-8
import wx
import wx.lib.sized_controls as sc
from dataclasses import dataclass
from functools import partial
from wx.adv import CommandLinkButton
from bookworm import app
from bookworm import config
from bookworm import typehints as t
from bookworm.i18n import LocaleInfo
from bookworm.concurrency impor... | [
"bookworm.gui.components.ImmutableObjectListView",
"bookworm.ocr_engines.tesseract_ocr_engine.TesseractOcrEngine.get_recognition_languages",
"bookworm.i18n.LocaleInfo.from_three_letter_code",
"wx.StdDialogButtonSizer",
"wx.GetApp",
"bookworm.logger.logger.getChild",
"bookworm.platform_services._win32.te... | [((1281, 1306), 'bookworm.logger.logger.getChild', 'logger.getChild', (['__name__'], {}), '(__name__)\n', (1296, 1306), False, 'from bookworm.logger import logger\n'), ((6838, 6908), 'wx.Choice', 'wx.Choice', (['parent', '(-1)'], {'choices': '[l.description for l in self.languages]'}), '(parent, -1, choices=[l.descript... |
# Copyright (c) 2012 <NAME> <<EMAIL>>
#
# This is free software released under the MIT license.
# See COPYING file for details, or visit:
# http://www.opensource.org/licenses/mit-license.php
#
# The file is part of FSMonitor, a file-system monitoring library.
# https://github.com/shaurz/fsmonitor
import sys, os, time,... | [
"os.listdir",
"threading.Lock",
"os.path.join",
"time.sleep",
"os.stat",
"time.time"
] | [((730, 741), 'time.time', 'time.time', ([], {}), '()\n', (739, 741), False, 'import sys, os, time, threading, errno\n'), ((1690, 1701), 'time.time', 'time.time', ([], {}), '()\n', (1699, 1701), False, 'import sys, os, time, threading, errno\n'), ((2042, 2055), 'os.stat', 'os.stat', (['path'], {}), '(path)\n', (2049, 2... |
import math
import random
import string
from datetime import datetime, timedelta
from textwrap import dedent
import config
import discord
from discord.ext import commands
import asyncio
class etc(commands.Cog):
def __init__(self, bot):
self.bot = bot
@commands.command(name="clear")
async def _m... | [
"discord.ext.commands.command"
] | [((273, 303), 'discord.ext.commands.command', 'commands.command', ([], {'name': '"""clear"""'}), "(name='clear')\n", (289, 303), False, 'from discord.ext import commands\n'), ((503, 534), 'discord.ext.commands.command', 'commands.command', ([], {'name': '"""recent"""'}), "(name='recent')\n", (519, 534), False, 'from di... |
# Copyright 2019 Xilinx Inc.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, ... | [
"sys.path.insert",
"numpy.reshape",
"numpy.minimum",
"numpy.hstack",
"caffe.io.Transformer",
"numpy.where",
"cv2.cvtColor",
"caffe.Net",
"numpy.meshgrid",
"numpy.maximum",
"numpy.arange"
] | [((1052, 1084), 'numpy.maximum', 'np.maximum', (['x1[i]', 'x1[order[1:]]'], {}), '(x1[i], x1[order[1:]])\n', (1062, 1084), True, 'import numpy as np\n'), ((1099, 1131), 'numpy.maximum', 'np.maximum', (['y1[i]', 'y1[order[1:]]'], {}), '(y1[i], y1[order[1:]])\n', (1109, 1131), True, 'import numpy as np\n'), ((1146, 1178)... |
#!/usr/bin/env python3
# -*- coding: utf-8 -*-
__author__ = 'cnheider'
import csv
import matplotlib.pyplot as plt
import utilities as U
# print(plt.style.available)
plot_style = 'fivethirtyeight'
# plot_style='bmh'
# plot_style='ggplot'
plt.style.use('seaborn-poster')
plt.style.use(plot_style)
plt.rcParams['axes.e... | [
"matplotlib.pyplot.get_backend",
"matplotlib.pyplot.plot",
"matplotlib.pyplot.style.use",
"utilities.StatisticAggregator",
"tkinter.Tk",
"matplotlib.pyplot.ion",
"matplotlib.pyplot.title",
"csv.reader",
"tkinter.filedialog.askopenfilename",
"matplotlib.pyplot.show"
] | [((242, 273), 'matplotlib.pyplot.style.use', 'plt.style.use', (['"""seaborn-poster"""'], {}), "('seaborn-poster')\n", (255, 273), True, 'import matplotlib.pyplot as plt\n'), ((274, 299), 'matplotlib.pyplot.style.use', 'plt.style.use', (['plot_style'], {}), '(plot_style)\n', (287, 299), True, 'import matplotlib.pyplot a... |
# =============================================================================
# Copyright 2020 NVIDIA. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://ww... | [
"nemo.logging.info",
"numpy.all"
] | [((2735, 2798), 'nemo.logging.info', 'logging.info', (['f"""Total errors (multiplied by 2): {total_errors}"""'], {}), "(f'Total errors (multiplied by 2): {total_errors}')\n", (2747, 2798), False, 'from nemo import logging\n'), ((3193, 3262), 'nemo.logging.info', 'logging.info', (['f"""*** Misclassified intent queries (... |
from datetime import datetime
import pytz
from .constants import TIMEZONE
timezone = pytz.timezone(TIMEZONE)
def get_ist_now():
"""Returns Indian Standard Time datetime object.
Returns:
object -- Datetime object
"""
return datetime.now(timezone)
| [
"pytz.timezone",
"datetime.datetime.now"
] | [((88, 111), 'pytz.timezone', 'pytz.timezone', (['TIMEZONE'], {}), '(TIMEZONE)\n', (101, 111), False, 'import pytz\n'), ((253, 275), 'datetime.datetime.now', 'datetime.now', (['timezone'], {}), '(timezone)\n', (265, 275), False, 'from datetime import datetime\n')] |
"""
LibriParty Dataset creation by using official metadata.
Author
------
<NAME>, 2020
<NAME>, 2020
"""
import os
import sys
import speechbrain as sb
from hyperpyyaml import load_hyperpyyaml
from speechbrain.utils.data_utils import download_file
from local.create_mixtures_from_metadata import create_mixture
import js... | [
"os.path.exists",
"os.makedirs",
"speechbrain.core.parse_arguments",
"os.path.join",
"local.create_mixtures_from_metadata.create_mixture",
"json.load",
"speechbrain.utils.data_utils.download_file",
"hyperpyyaml.load_hyperpyyaml"
] | [((534, 571), 'speechbrain.core.parse_arguments', 'sb.core.parse_arguments', (['sys.argv[1:]'], {}), '(sys.argv[1:])\n', (557, 571), True, 'import speechbrain as sb\n'), ((802, 906), 'speechbrain.utils.data_utils.download_file', 'download_file', (['URL_METADATA', "(metadata_folder + '/meta.zip')"], {'unpack': '(True)',... |
import subprocess
from pathlib import Path
import pytest
# from py4gh import __version__
from py4gh.utility import decrypt_files, encrypt_files, get_files
# def test_version():
# assert __version__ == "0.1.0"
@pytest.fixture(scope="session")
def keys(tmpdir_factory):
test_pub1 = tmpdir_factory.mktemp("data... | [
"py4gh.utility.encrypt_files",
"py4gh.utility.get_files",
"pathlib.Path",
"subprocess.Popen",
"subprocess.run",
"py4gh.utility.decrypt_files",
"pytest.fixture"
] | [((219, 250), 'pytest.fixture', 'pytest.fixture', ([], {'scope': '"""session"""'}), "(scope='session')\n", (233, 250), False, 'import pytest\n'), ((1113, 1144), 'pytest.fixture', 'pytest.fixture', ([], {'scope': '"""session"""'}), "(scope='session')\n", (1127, 1144), False, 'import pytest\n'), ((641, 756), 'subprocess.... |
from unittest import TestCase
from parameterized import parameterized
from pyecsca.ec.context import local
from pyecsca.ec.mult import LTRMultiplier, BinaryNAFMultiplier, WindowNAFMultiplier, LadderMultiplier, \
DifferentialLadderMultiplier
from pyecsca.ec.params import get_params
from pyecsca.sca.re.rpa import M... | [
"parameterized.parameterized.expand",
"pyecsca.ec.mult.LadderMultiplier",
"pyecsca.ec.params.get_params",
"pyecsca.ec.mult.DifferentialLadderMultiplier",
"pyecsca.ec.mult.BinaryNAFMultiplier",
"pyecsca.sca.re.rpa.MultipleContext",
"pyecsca.ec.mult.LTRMultiplier",
"pyecsca.ec.mult.WindowNAFMultiplier"
... | [((785, 990), 'parameterized.parameterized.expand', 'parameterized.expand', (["[('10', 10), ('2355498743', 2355498743), (\n '325385790209017329644351321912443757746', \n 325385790209017329644351321912443757746), ('13613624287328732', \n 13613624287328732)]"], {}), "([('10', 10), ('2355498743', 2355498743), (\n... |