code stringlengths 22 1.05M | apis listlengths 1 3.31k | extract_api stringlengths 75 3.25M |
|---|---|---|
from collections import OrderedDict as odict
import time
import numpy as np
from .json_encoder import JsonNumEncoder
import os
class Profiler:
"""This class provides a very simple yet light implementation of function profiling.
It is very easy to use:
>>> profiler.reset()
>>> profiler.start("... | [
"os.path.join",
"collections.OrderedDict",
"time.time"
] | [((957, 964), 'collections.OrderedDict', 'odict', ([], {}), '()\n', (962, 964), True, 'from collections import OrderedDict as odict\n'), ((1279, 1290), 'time.time', 'time.time', ([], {}), '()\n', (1288, 1290), False, 'import time\n'), ((2513, 2520), 'collections.OrderedDict', 'odict', ([], {}), '()\n', (2518, 2520), Tr... |
import math
import hashlib
import zlib
from bitarray import bitarray
from globals import G
class BFsignature():
def __init__(self, total_chunks):
self.total_chunks = total_chunks
if self.total_chunks > 0:
self.cal_m()
#print("bf size = ",self.m)
else:
#lo... | [
"zlib.crc32",
"hashlib.sha1",
"bitarray.bitarray",
"math.log"
] | [((635, 651), 'bitarray.bitarray', 'bitarray', (['self.m'], {}), '(self.m)\n', (643, 651), False, 'from bitarray import bitarray\n'), ((999, 1013), 'hashlib.sha1', 'hashlib.sha1', ([], {}), '()\n', (1011, 1013), False, 'import hashlib\n'), ((467, 478), 'math.log', 'math.log', (['(2)'], {}), '(2)\n', (475, 478), False, ... |
"""
The main window for the GUI.
"""
from msl.qt import (
QtWidgets,
QtGui,
Qt,
Signal,
convert,
prompt,
application,
)
from msl.qt.utils import (
screen_geometry,
drag_drop_paths,
)
from msl.io import read
from .line_edit import LineEdit
from .. import logger
from .equipment import... | [
"msl.qt.QtGui.QAction",
"msl.qt.application",
"msl.qt.prompt.critical",
"msl.qt.convert.to_qicon",
"msl.io.read",
"msl.qt.Signal",
"msl.qt.QtGui.QColor",
"msl.qt.prompt.yes_no",
"msl.qt.QtWidgets.QProgressBar",
"msl.qt.utils.screen_geometry",
"msl.qt.QtGui.QPalette",
"msl.qt.QtWidgets.QDockWid... | [((561, 574), 'msl.qt.Signal', 'Signal', (['float'], {}), '(float)\n', (567, 574), False, 'from msl.qt import QtWidgets, QtGui, Qt, Signal, convert, prompt, application\n'), ((646, 654), 'msl.qt.Signal', 'Signal', ([], {}), '()\n', (652, 654), False, 'from msl.qt import QtWidgets, QtGui, Qt, Signal, convert, prompt, ap... |
# -*- coding: utf-8 -*-
"""
Copyright (c) 2020 Ephreal under the MIT License.
To view the license and requirements when distributing this software, please
view the license at https://github.com/ephreal/catapi/LICENSE.
"""
from catapi.analysis import Analysis
from tests import async_capable
class Testanaylsis(async... | [
"catapi.analysis.Analysis"
] | [((596, 623), 'catapi.analysis.Analysis', 'Analysis', ([], {}), '(**analysis_params)\n', (604, 623), False, 'from catapi.analysis import Analysis\n')] |
# -*- coding: utf-8 -*-
# Authors: <NAME> <<EMAIL>>
#
# License: BSD (3-clause)
import numpy as np
from os import path as op
from .utils import _check_pytables
from .externals.six import string_types, text_type
##############################################################################
# WRITE
def write_hdf5(fn... | [
"os.path.isfile",
"numpy.array",
"numpy.atleast_1d"
] | [((780, 796), 'os.path.isfile', 'op.isfile', (['fname'], {}), '(fname)\n', (789, 796), True, 'from os import path as op\n'), ((4036, 4052), 'os.path.isfile', 'op.isfile', (['fname'], {}), '(fname)\n', (4045, 4052), True, 'from os import path as op\n'), ((5339, 5353), 'numpy.array', 'np.array', (['node'], {}), '(node)\n... |
import requests
import os
import datetime
from bs4 import BeautifulSoup
from dotenv import load_dotenv
from twilio.rest import Client
from sendgrid import SendGridAPIClient
from sendgrid.helpers.mail import Mail
load_dotenv()
TWILIO_ACCOUNT_SID = os.environ.get("TWILIO_ACCOUNT_SID", "OOPS, please specify env var call... | [
"sendgrid.helpers.mail.Mail",
"os.environ.get",
"sendgrid.SendGridAPIClient",
"requests.get",
"dotenv.load_dotenv",
"bs4.BeautifulSoup",
"datetime.datetime.now",
"twilio.rest.Client"
] | [((213, 226), 'dotenv.load_dotenv', 'load_dotenv', ([], {}), '()\n', (224, 226), False, 'from dotenv import load_dotenv\n'), ((249, 349), 'os.environ.get', 'os.environ.get', (['"""TWILIO_ACCOUNT_SID"""', '"""OOPS, please specify env var called \'TWILIO_ACCOUNT_SID\'"""'], {}), '(\'TWILIO_ACCOUNT_SID\',\n "OOPS, plea... |
#!/usr/bin/python3
import sys
import numpy as np
import numpysane as nps
import os
testdir = os.path.dirname(os.path.realpath(__file__))
# I import the LOCAL mrcal since that's what I'm testing
sys.path[:0] = f"{testdir}/..",
import mrcal
import testutils
import cv2
image = cv2.imread(f'{testdir}/data/figueroa-ove... | [
"testutils.finish",
"os.path.realpath",
"numpy.array",
"testutils.confirm",
"numpy.linalg.inv",
"numpysane.dummy",
"mrcal.match_feature",
"testutils.confirm_equal",
"cv2.imread",
"mrcal.apply_homography",
"numpy.arange"
] | [((280, 381), 'cv2.imread', 'cv2.imread', (['f"""{testdir}/data/figueroa-overpass-looking-S.0.downsampled.jpg"""', 'cv2.IMREAD_GRAYSCALE'], {}), "(f'{testdir}/data/figueroa-overpass-looking-S.0.downsampled.jpg',\n cv2.IMREAD_GRAYSCALE)\n", (290, 381), False, 'import cv2\n'), ((478, 568), 'numpy.array', 'np.array', (... |
#!/usr/bin/env python
"""Solve the range minimum query problem."""
def read_numbers(number_file, query_file):
"""
Parameters
----------
number_file : str
query_file : str
Returns
-------
tuple
(numbers, queries) - both are lists
"""
with open(number_file) as f:
... | [
"argparse.ArgumentParser",
"sys.exit"
] | [((1613, 1700), 'argparse.ArgumentParser', 'ArgumentParser', ([], {'description': '__doc__', 'formatter_class': 'ArgumentDefaultsHelpFormatter'}), '(description=__doc__, formatter_class=\n ArgumentDefaultsHelpFormatter)\n', (1627, 1700), False, 'from argparse import ArgumentDefaultsHelpFormatter, ArgumentParser\n'),... |
from urllib.parse import urlparse
class Card:
AUTO_GENERATED_TEXT = 'Auto-created by TrelloNextActions'
def __init__(self, trello, json):
self._trello = trello
self.id = json['id']
self.name = json['name']
self.board_id = json['idBoard']
self.description = json['desc'... | [
"urllib.parse.urlparse"
] | [((633, 659), 'urllib.parse.urlparse', 'urlparse', (['self.description'], {}), '(self.description)\n', (641, 659), False, 'from urllib.parse import urlparse\n')] |
# -*- coding: utf-8 -*-
from __future__ import unicode_literals
from django.db import models, migrations
class Migration(migrations.Migration):
dependencies = [
('forum', '0063_auto_20150611_1052'),
]
operations = [
migrations.AddField(
model_name='poll',
name='a... | [
"django.db.models.DateTimeField"
] | [((342, 436), 'django.db.models.DateTimeField', 'models.DateTimeField', ([], {'default': "b'2000-01-01 00:00:00'", 'verbose_name': '"""aa"""', 'auto_now_add': '(True)'}), "(default=b'2000-01-01 00:00:00', verbose_name='aa',\n auto_now_add=True)\n", (362, 436), False, 'from django.db import models, migrations\n')] |
# Copyright 2017 FUJITSU LIMITED
#
# Licensed under the Apache License, Version 2.0 (the "License"); you may
# not use this file except in compliance with the License. You may obtain
# a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writ... | [
"monasca_notification.config.parse_args",
"oslo_config.cfg.CONF.set_default",
"monasca_notification.conf.register_opts",
"oslo_config.cfg.CONF.set_override",
"unittest.mock.patch"
] | [((960, 998), 'unittest.mock.patch', 'mock.patch', (['"""monascastatsd.Connection"""'], {}), "('monascastatsd.Connection')\n", (970, 998), False, 'from unittest import mock\n'), ((1352, 1372), 'monasca_notification.conf.register_opts', 'conf.register_opts', ([], {}), '()\n', (1370, 1372), False, 'from monasca_notificat... |
#!/usr/bin/python
# -*- coding: utf-8 -*-
#
# Copyright 2017 Falseen
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by a... | [
"sys.path.pop",
"sys.path.insert"
] | [((847, 862), 'sys.path.pop', 'sys.path.pop', (['(0)'], {}), '(0)\n', (859, 862), False, 'import sys\n'), ((930, 954), 'sys.path.insert', 'sys.path.insert', (['(0)', 'path'], {}), '(0, path)\n', (945, 954), False, 'import sys\n')] |
# Copyright (c) 2019 <NAME>
#
# Permission is hereby granted, free of charge, to any person obtaining a copy
# of this software and associated documentation files (the "Software"), to deal
# in the Software without restriction, including without limitation the rights
# to use, copy, modify, merge, publish, distribute, ... | [
"fla.ndfa.instance.Instance",
"copy.copy"
] | [((3970, 4008), 'copy.copy', 'copy.copy', (['self.current_configurations'], {}), '(self.current_configurations)\n', (3979, 4008), False, 'import copy\n'), ((3798, 3847), 'fla.ndfa.instance.Instance', 'Instance', (['self', 'state', 'configuration.current_word'], {}), '(self, state, configuration.current_word)\n', (3806,... |
# -*- coding: utf-8 -*-
# python-holidays
# ---------------
# A fast, efficient Python library for generating country, province and state
# specific sets of holidays on the fly. It aims to make determining whether a
# specific date is a holiday as fast and flexible as possible.
#
# Author: ryanss <<EMAIL>> (c) ... | [
"dateutil.relativedelta.MO",
"dateutil.relativedelta.relativedelta",
"datetime.date",
"holidays.holiday_base.HolidayBase.__init__"
] | [((813, 849), 'holidays.holiday_base.HolidayBase.__init__', 'HolidayBase.__init__', (['self'], {}), '(self, **kwargs)\n', (833, 849), False, 'from holidays.holiday_base import HolidayBase\n'), ((964, 982), 'datetime.date', 'date', (['year', 'JAN', '(1)'], {}), '(year, JAN, 1)\n', (968, 982), False, 'from datetime impor... |
import sys
import os
sys.path.append(os.path.join(os.path.dirname(os.path.abspath(__file__)), "../../"))
from api.fact.Fact_Market_Stock_Data import Fact_Market_Stock_Data
from api.common.Saver import Saver
from datetime import datetime,timedelta
from api.common.Logs import Logger_process, Logger_process_error
dda = F... | [
"datetime.datetime.now",
"api.common.Logs.Logger_process.log",
"os.path.abspath",
"datetime.timedelta",
"api.fact.Fact_Market_Stock_Data.Fact_Market_Stock_Data"
] | [((319, 343), 'api.fact.Fact_Market_Stock_Data.Fact_Market_Stock_Data', 'Fact_Market_Stock_Data', ([], {}), '()\n', (341, 343), False, 'from api.fact.Fact_Market_Stock_Data import Fact_Market_Stock_Data\n'), ((383, 448), 'api.common.Logs.Logger_process.log', 'Logger_process.log', (["('call Fact_Market_Stock_History,' +... |
# Copyright 2020 Huawei Technologies Co., Ltd
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to... | [
"copy.deepcopy",
"mindspore.ops.operations.Mul",
"numpy.ones",
"mindspore.nn.LogSoftmax",
"mindspore.ops.operations.MatMul",
"mindspore.ops.operations.Shape",
"mindspore.ops.operations.OnesLike",
"mindspore.ops.operations.Reshape",
"mindspore.common.tensor.Tensor",
"numpy.tril",
"mindspore.ops.o... | [((1983, 1994), 'mindspore.ops.operations.Reshape', 'P.Reshape', ([], {}), '()\n', (1992, 1994), True, 'from mindspore.ops import operations as P\n'), ((2017, 2043), 'mindspore.ops.operations.MatMul', 'P.MatMul', ([], {'transpose_b': '(True)'}), '(transpose_b=True)\n', (2025, 2043), True, 'from mindspore.ops import ope... |
#_*_coding=UTF-8_*_
from PIL import Image
#编码
#内置 bin() 的替代,返回固定长度的二进制字符串
def constLenBin(int):
binary = "0" * (8 - (len(bin(int)) - 2)) + bin(int).replace('0b','') #去掉 bin() 返回的二进制字符串中的 '0b',并在左边补足'0'直到字符串长度为8
return binary
#取得一个 PIL 图像并且更改所有值为偶数(使最低有效位为0)
def makeImageEven(image):
pixels = list(image.g... | [
"PIL.Image.new",
"PIL.Image.open"
] | [((483, 516), 'PIL.Image.new', 'Image.new', (['image.mode', 'image.size'], {}), '(image.mode, image.size)\n', (492, 516), False, 'from PIL import Image\n'), ((1279, 1320), 'PIL.Image.new', 'Image.new', (['evenImage.mode', 'evenImage.size'], {}), '(evenImage.mode, evenImage.size)\n', (1288, 1320), False, 'from PIL impor... |
import lambda_function
channel_id = 'CNCRFR7KP'
team_id = 'TE4VDPM2L'
user_id = 'UE6P69HFH'
response_url = "no"
channel_name = "local"
command_list = ['init',
'init <https://charasheet.vampire-blood.net/2796211>',
'1D100',
'result',
's',
... | [
"lambda_function.bootstrap"
] | [((1835, 1871), 'lambda_function.bootstrap', 'lambda_function.bootstrap', (['event', '{}'], {}), '(event, {})\n', (1860, 1871), False, 'import lambda_function\n')] |
import django_filters
from django.contrib.auth.models import User
from .constants import *
from .models import ObjectChange
class ObjectChangeFilter(django_filters.FilterSet):
q = django_filters.CharFilter(method="search", label="Search")
time = django_filters.DateTimeFromToRangeFilter()
action = django... | [
"django.contrib.auth.models.User.objects.all",
"django_filters.DateTimeFromToRangeFilter",
"django_filters.MultipleChoiceFilter",
"django_filters.CharFilter"
] | [((188, 246), 'django_filters.CharFilter', 'django_filters.CharFilter', ([], {'method': '"""search"""', 'label': '"""Search"""'}), "(method='search', label='Search')\n", (213, 246), False, 'import django_filters\n'), ((258, 300), 'django_filters.DateTimeFromToRangeFilter', 'django_filters.DateTimeFromToRangeFilter', ([... |
__copyright__ = "Copyright (c) 2020-2021 Jina AI Limited. All rights reserved."
__license__ = "Apache-2.0"
import pytest
from jina import Document, DocumentArray
from ...laser_encoder import LaserEncoder
@pytest.fixture()
def docs_generator():
return DocumentArray((Document(text='random text') for _ in range(30)... | [
"pytest.fixture",
"jina.Document"
] | [((208, 224), 'pytest.fixture', 'pytest.fixture', ([], {}), '()\n', (222, 224), False, 'import pytest\n'), ((731, 764), 'jina.Document', 'Document', ([], {'id': '"""chunk11"""', 'text': 'text'}), "(id='chunk11', text=text)\n", (739, 764), False, 'from jina import Document, DocumentArray\n'), ((788, 821), 'jina.Document... |
from typing import Tuple
from ground.hints import (Point,
Scalar,
Segment)
from hypothesis import given
from locus.segmental import Tree
from tests.utils import (is_segment,
to_segment_point_distance)
from . import strategies
@given(strate... | [
"hypothesis.given",
"tests.utils.to_segment_point_distance",
"tests.utils.is_segment"
] | [((308, 343), 'hypothesis.given', 'given', (['strategies.trees_with_points'], {}), '(strategies.trees_with_points)\n', (313, 343), False, 'from hypothesis import given\n'), ((524, 559), 'hypothesis.given', 'given', (['strategies.trees_with_points'], {}), '(strategies.trees_with_points)\n', (529, 559), False, 'from hypo... |
import gym
import gym_pigchase_topdown
env = gym.make('PigChase-v0')
_ = env.reset()
env.render()
done = False
while True:
action = env.read_action()
if action != -1:
_, reward, done, _ = env.step(action)
print(reward)
env.render()
if done:
break | [
"gym.make"
] | [((46, 69), 'gym.make', 'gym.make', (['"""PigChase-v0"""'], {}), "('PigChase-v0')\n", (54, 69), False, 'import gym\n')] |
from dataclasses import field, dataclass
from typing import List
import marshmallow_dataclass
from flask_marshmallow import Schema
@dataclass
class UserResponse:
UserName: str
UserResponseSchema = marshmallow_dataclass.class_schema(UserResponse, base_schema=Schema)
@dataclass
class AllUsersResponse:
User... | [
"dataclasses.field",
"marshmallow_dataclass.class_schema"
] | [((206, 274), 'marshmallow_dataclass.class_schema', 'marshmallow_dataclass.class_schema', (['UserResponse'], {'base_schema': 'Schema'}), '(UserResponse, base_schema=Schema)\n', (240, 274), False, 'import marshmallow_dataclass\n'), ((399, 471), 'marshmallow_dataclass.class_schema', 'marshmallow_dataclass.class_schema', ... |
#!/usr/bin/env python
"""
@package mi.dataset.parser
@file marine-integrations/mi/dataset/parser/cspp_base.py
@author <NAME>
@brief Base Parser for a cspp dataset driver
Release notes:
initial release
"""
import copy
import re
import string
from mi.core.log import get_logger
from mi.core.common i... | [
"re.split",
"mi.core.exceptions.ConfigurationException",
"re.compile",
"mi.core.exceptions.DatasetParserException",
"mi.core.log.get_logger",
"string.rstrip",
"mi.core.exceptions.RecoverableSampleException",
"copy.copy"
] | [((745, 757), 'mi.core.log.get_logger', 'get_logger', ([], {}), '()\n', (755, 757), False, 'from mi.core.log import get_logger\n'), ((893, 922), 're.compile', 're.compile', (['HEADER_PART_REGEX'], {}), '(HEADER_PART_REGEX)\n', (903, 922), False, 'import re\n'), ((1010, 1042), 're.compile', 're.compile', (['TIMESTAMP_LI... |
import collections
from typing import Counter
class Solution:
def getHint(self, secret: str, guess: str) -> str:
bulls = cows = 0
seen = collections.Counter() # type: Counter[str]
for s, g in zip(secret, guess):
if s == g:
bulls += 1
else:
... | [
"collections.Counter"
] | [((160, 181), 'collections.Counter', 'collections.Counter', ([], {}), '()\n', (179, 181), False, 'import collections\n')] |
# Licensed to the Apache Software Foundation (ASF) under one
# or more contributor license agreements. See the NOTICE file
# distributed with this work for additional information
# regarding copyright ownership. The ASF licenses this file
# to you under the Apache License, Version 2.0 (the
# "License"); you may not u... | [
"pydolphinscheduler.core.process_definition.ProcessDefinition",
"pydolphinscheduler.tasks.shell.Shell",
"pydolphinscheduler.tasks.dependent.DependentItem"
] | [((1580, 1653), 'pydolphinscheduler.core.process_definition.ProcessDefinition', 'ProcessDefinition', ([], {'name': '"""task_dependent_external"""', 'tenant': '"""tenant_exists"""'}), "(name='task_dependent_external', tenant='tenant_exists')\n", (1597, 1653), False, 'from pydolphinscheduler.core.process_definition impor... |
from subprocess import call
from sniffer.api import runnable, file_validator
import os
@file_validator
def py_files(filename):
is_match = (filename.endswith('.py')
and not filename.endswith('_flymake.py')
and not os.path.basename(filename).startswith('.'))
return is_match
@ru... | [
"os.path.basename",
"subprocess.call"
] | [((425, 433), 'subprocess.call', 'call', (['fn'], {}), '(fn)\n', (429, 433), False, 'from subprocess import call\n'), ((251, 277), 'os.path.basename', 'os.path.basename', (['filename'], {}), '(filename)\n', (267, 277), False, 'import os\n')] |
# import_export_ballotpedia/views_admin.py
# Brought to you by We Vote. Be good.
# -*- coding: UTF-8 -*-
from .controllers import attach_ballotpedia_election_by_district_from_api, \
retrieve_ballot_items_from_polling_location, \
retrieve_ballotpedia_candidates_by_district_from_api, retrieve_ballotpedia_measure... | [
"django.shortcuts.render",
"config.base.get_environment_variable",
"admin_tools.views.redirect_to_sign_in_page",
"datetime.date.today",
"election.models.Election.objects.get",
"django.core.urlresolvers.reverse",
"wevote_functions.functions.positive_value_exists",
"django.db.models.Q",
"polling_locat... | [((1420, 1476), 'config.base.get_environment_variable', 'get_environment_variable', (['"""BALLOTPEDIA_API_CONTAINS_URL"""'], {}), "('BALLOTPEDIA_API_CONTAINS_URL')\n", (1444, 1476), False, 'from config.base import get_environment_variable\n'), ((3695, 3733), 'wevote_functions.functions.positive_value_exists', 'positive... |
from flask import current_app as app
from flask_migrate import Migrate, migrate, upgrade, stamp, current
from alembic.migration import MigrationContext
from sqlalchemy import create_engine
from sqlalchemy.engine.url import make_url
from sqlalchemy_utils import (
database_exists as database_exists_util,
create_d... | [
"sqlalchemy_utils.drop_database",
"sqlalchemy_utils.database_exists",
"sqlalchemy.engine.url.make_url",
"sqlalchemy_utils.create_database",
"flask_migrate.Migrate",
"alembic.migration.MigrationContext.configure",
"flask.current_app.config.get"
] | [((434, 443), 'flask_migrate.Migrate', 'Migrate', ([], {}), '()\n', (441, 443), False, 'from flask_migrate import Migrate, migrate, upgrade, stamp, current\n'), ((479, 526), 'sqlalchemy.engine.url.make_url', 'make_url', (["app.config['SQLALCHEMY_DATABASE_URI']"], {}), "(app.config['SQLALCHEMY_DATABASE_URI'])\n", (487, ... |
import os
import inspect
import pandas as pd
class ICDConversionHelper:
ICD9_TO_ICD10_CONVERSION_MAPPING = 'icd9_to_icd10.csv'
ICD10_6_TO_ICD10_8_CONVERSION_MAPPING = 'icd10_6_to_icd10_8.csv'
def __init__(self):
curr_dir = os.path.dirname(os.path.abspath(inspect.stack()[0][1]))
self.ICD9_... | [
"inspect.stack",
"pandas.read_csv"
] | [((331, 398), 'pandas.read_csv', 'pd.read_csv', (["(curr_dir + '/' + self.ICD9_TO_ICD10_CONVERSION_MAPPING)"], {}), "(curr_dir + '/' + self.ICD9_TO_ICD10_CONVERSION_MAPPING)\n", (342, 398), True, 'import pandas as pd\n'), ((433, 505), 'pandas.read_csv', 'pd.read_csv', (["(curr_dir + '/' + self.ICD10_6_TO_ICD10_8_CONVER... |
import socket
import sys
import threading
import elgamal
numBits = 32
keyOne = elgamal.generate_keys(iNumBits=numBits, iConfidence=1000)
privKeyB = keyOne["privateKey"]
pubKeyB = keyOne["publicKey"]
print(f"Public Key: {pubKeyB.p} {pubKeyB.g} {pubKeyB.h}")
# ----------------------------------------------------------... | [
"elgamal.generate_keys",
"socket.socket",
"elgamal.PublicKey",
"elgamal.encrypt",
"threading.Thread",
"elgamal.decrypt"
] | [((81, 138), 'elgamal.generate_keys', 'elgamal.generate_keys', ([], {'iNumBits': 'numBits', 'iConfidence': '(1000)'}), '(iNumBits=numBits, iConfidence=1000)\n', (102, 138), False, 'import elgamal\n'), ((454, 502), 'socket.socket', 'socket.socket', (['socket.AF_INET', 'socket.SOCK_DGRAM'], {}), '(socket.AF_INET, socket.... |
from django import forms
from django.contrib.auth.forms import UserCreationForm
from django.contrib.auth import get_user_model
from sistema.mail import send_mail_template
from sistema.utils import generate_hash_key
from usuario.models import PasswordReset
User = get_user_model()
class PasswordResetForm(forms.Form):
... | [
"django.contrib.auth.get_user_model",
"sistema.utils.generate_hash_key",
"django.forms.ValidationError",
"usuario.models.PasswordReset",
"django.forms.EmailField",
"sistema.mail.send_mail_template"
] | [((264, 280), 'django.contrib.auth.get_user_model', 'get_user_model', ([], {}), '()\n', (278, 280), False, 'from django.contrib.auth import get_user_model\n'), ((329, 361), 'django.forms.EmailField', 'forms.EmailField', ([], {'label': '"""E-mail"""'}), "(label='E-mail')\n", (345, 361), False, 'from django import forms\... |
import os
import sys
sys.path.append('.')
import cv2
import numpy as np
import time
from Utilities import drawRegion, drawBox, col_rgb, CVConstants
try:
import pyMTF
mtf_available = 1
except ImportError as e:
print('MTF unavailable: {}'.format(e))
mtf_available = 0
from siamfc.SiamFC import SiamFC... | [
"cv2.TrackerGOTURN_create",
"siamfc.SiamFC.SiamFC",
"time.clock",
"cv2.TrackerKCF_create",
"cv2.imshow",
"numpy.array",
"siamfc.SiamFC.SiamFCParams",
"sys.path.append",
"cv2.TrackerMedianFlow_create",
"cv2.__version__.split",
"os.path.exists",
"pyMTF.getRegion",
"DaSiamRPN.DaSiamRPN.DaSiamRP... | [((22, 42), 'sys.path.append', 'sys.path.append', (['"""."""'], {}), "('.')\n", (37, 42), False, 'import sys\n'), ((1618, 1632), 'siamfc.SiamFC.SiamFCParams', 'SiamFCParams', ([], {}), '()\n', (1630, 1632), False, 'from siamfc.SiamFC import SiamFC, SiamFCParams\n'), ((1658, 1674), 'SiamMask.SiamMask.SiamMaskParams', 'S... |
from abc import abstractmethod
from typing import Optional, Tuple
from pylidar_slam.common.modules import _with_cv2
if _with_cv2:
import cv2
import numpy as np
from omegaconf import DictConfig
from pylidar_slam.common.utils import check_sizes, assert_debug
class ImageBased2DRegistration:
... | [
"numpy.clip",
"cv2.BFMatcher",
"numpy.eye",
"numpy.ones",
"cv2.findHomography",
"numpy.round",
"pylidar_slam.common.utils.assert_debug",
"cv2.AKAZE_create",
"numpy.argsort",
"numpy.array",
"cv2.ORB_create",
"pylidar_slam.common.utils.check_sizes",
"numpy.linalg.svd",
"matplotlib.cm.get_cma... | [((619, 661), 'pylidar_slam.common.utils.assert_debug', 'assert_debug', (["(features in ['orb', 'akaze'])"], {}), "(features in ['orb', 'akaze'])\n", (631, 661), False, 'from pylidar_slam.common.utils import check_sizes, assert_debug\n'), ((863, 906), 'cv2.BFMatcher', 'cv2.BFMatcher', (['cv2.NORM_L2'], {'crossCheck': '... |
# Generated by Django 3.1 on 2021-01-26 08:00
from django.db import migrations, models
class Migration(migrations.Migration):
dependencies = [
('tms', '0001_initial'),
]
operations = [
migrations.AlterField(
model_name='ts2pstransfer',
name='sub_type',
... | [
"django.db.models.IntegerField"
] | [((328, 468), 'django.db.models.IntegerField', 'models.IntegerField', ([], {'choices': "[(4, 'TS to PS (Single Degree)'), (5, 'PS-TS to PS-PS (Dual Degree)'), (6,\n 'TS-TS to TS-PS (Dual Degree)')]"}), "(choices=[(4, 'TS to PS (Single Degree)'), (5,\n 'PS-TS to PS-PS (Dual Degree)'), (6, 'TS-TS to TS-PS (Dual Deg... |
import ray
import torch
from collections import namedtuple, defaultdict
from itertools import count, chain
from flatland.envs.agent_utils import RailAgentStatus
from agent.PPO.PPORollout import PPORollout, PPOTransition
class PPORunner():
def _select_actions(self, state, done):
valid_handles = list()
... | [
"torch.tensor",
"torch.stack",
"collections.defaultdict",
"torch.zeros"
] | [((2958, 2974), 'collections.defaultdict', 'defaultdict', (['int'], {}), '(int)\n', (2969, 2974), False, 'from collections import namedtuple, defaultdict\n'), ((3076, 3099), 'collections.defaultdict', 'defaultdict', (['PPORollout'], {}), '(PPORollout)\n', (3087, 3099), False, 'from collections import namedtuple, defaul... |
import unittest
from app.models import User
class userModelTest(unittest.TestCase):
def setUp(self):
'''
function that creates instance of User class
'''
self.new_user = User(password='<PASSWORD>')
def test_set_password(self):
'''
test case that ascertains tha... | [
"app.models.User"
] | [((209, 236), 'app.models.User', 'User', ([], {'password': '"""<PASSWORD>"""'}), "(password='<PASSWORD>')\n", (213, 236), False, 'from app.models import User\n')] |
#!/usr/bin/env python3
# Copyright 2021 iRobot Corporation. All Rights Reserved.
# @author <NAME> (<EMAIL>)
#
# Launch Create(R) 3 nodes
from ament_index_python.packages import get_package_share_directory
from launch import LaunchDescription
from launch.actions import DeclareLaunchArgument, IncludeLaunchDescription
fr... | [
"launch.substitutions.LaunchConfiguration",
"ament_index_python.packages.get_package_share_directory",
"launch.LaunchDescription",
"launch.actions.DeclareLaunchArgument",
"launch_ros.actions.Node",
"launch.launch_description_sources.PythonLaunchDescriptionSource",
"launch.substitutions.PathJoinSubstitut... | [((580, 718), 'launch.actions.DeclareLaunchArgument', 'DeclareLaunchArgument', (['"""gazebo"""'], {'default_value': '"""classic"""', 'choices': "['classic', 'ignition']", 'description': '"""Which gazebo simulator to use"""'}), "('gazebo', default_value='classic', choices=['classic',\n 'ignition'], description='Which... |
import json
from topojson.core.extract import Extract
from shapely import geometry
import geopandas
import geojson
# extract copies coordinates sequentially into a buffer
def test_extract_linestring():
data = {
"foo": {"type": "LineString", "coordinates": [[0, 0], [1, 0], [2, 0]]},
"bar": {"type":... | [
"shapefile.Reader",
"geopandas.read_file",
"shapely.geometry.MultiLineString",
"topojson.core.extract.Extract",
"shapely.geometry.Polygon",
"shapely.geometry.LineString",
"geojson.load"
] | [((8484, 8542), 'shapefile.Reader', 'shapefile.Reader', (['"""tests/files_shapefile/southamerica.shp"""'], {}), "('tests/files_shapefile/southamerica.shp')\n", (8500, 8542), False, 'import shapefile\n'), ((9103, 9161), 'shapely.geometry.Polygon', 'geometry.Polygon', (['[[0, 0], [1, 0], [1, 1], [0, 1], [0, 0]]'], {}), '... |
"""
tricks:
1.torch-optimizer:实现了最新的一些优化器.
2.numba:import numba as nb,纯python或numpy加速,加@nb.njit或@nb.jit(nopython=True)
3.swifter:df.apply()→·df.swifter.apply(),加速pandas
4.cupy:1000万以上数据更快
5.modin:import modin.pandas as mdpd,用mdpd代替pd即可,加速pandas,加载数据和查询数据更快,统计方法pandas更快
"""
import os
import sys
import argparse
import ti... | [
"torch.cuda.manual_seed_all",
"torch.manual_seed",
"torch.utils.tensorboard.SummaryWriter",
"numpy.mean",
"torch.nn.CrossEntropyLoss",
"argparse.ArgumentParser",
"data.custom_dataset.MyDataset",
"torch.load",
"wandb.init",
"wandb.watch",
"torch.cuda.is_available",
"numpy.random.seed",
"torch... | [((729, 740), 'time.time', 'time.time', ([], {}), '()\n', (738, 740), False, 'import time\n'), ((745, 769), 'torch.cuda.empty_cache', 'torch.cuda.empty_cache', ([], {}), '()\n', (767, 769), False, 'import torch\n'), ((1817, 1879), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {'description': '"""Train a ne... |
# -*- coding:utf-8 -*-
# Create by 27
# @Time : 2020/2/21 22:45
__author__ = '27'
'''
使用python发送http请求的两个方法:
1、urllib
2、requests (推荐)
小原则:
对于外部获取的数据,如果出现异常,最好不要抛出来,而是应该默认值处理。
'''
import requests
class HTTP: # python3中写与不写(object)没有区别
@staticmethod
def get(url, return_json=True):
r = requests.get(url... | [
"requests.get"
] | [((304, 321), 'requests.get', 'requests.get', (['url'], {}), '(url)\n', (316, 321), False, 'import requests\n')] |
import Crypto.Util.number as cun
import Crypto.Random.random as crr
import Crypto.Util.Padding as cup
from Crypto.Cipher import AES
import os
import hashlib
class DiffieHellman:
def __init__(self, p: int):
self.p = p
self.g = 8
self.private_key = crr.getrandbits(128)
def public_key(se... | [
"os.urandom",
"Crypto.Util.number.size",
"Crypto.Util.number.long_to_bytes",
"Crypto.Cipher.AES.new",
"Crypto.Random.random.getrandbits",
"Crypto.Util.number.isPrime"
] | [((1401, 1431), 'Crypto.Cipher.AES.new', 'AES.new', (['aes_key', 'AES.MODE_ECB'], {}), '(aes_key, AES.MODE_ECB)\n', (1408, 1431), False, 'from Crypto.Cipher import AES\n'), ((277, 297), 'Crypto.Random.random.getrandbits', 'crr.getrandbits', (['(128)'], {}), '(128)\n', (292, 297), True, 'import Crypto.Random.random as c... |
import os
import json
import argparse
from tqdm import tqdm
def extract_ingredients(result_dict, all_ingredient_dict):
for method, result in result_dict.items():
for recipe_id, annotation in result.items():
ingredient_list = annotation["ingredients"]
sentences = annotation["sentence... | [
"json.load",
"os.path.join",
"argparse.ArgumentParser"
] | [((3001, 3026), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {}), '()\n', (3024, 3026), False, 'import argparse\n'), ((3453, 3465), 'json.load', 'json.load', (['f'], {}), '(f)\n', (3462, 3465), False, 'import json\n'), ((2746, 2758), 'json.load', 'json.load', (['f'], {}), '(f)\n', (2755, 2758), False, 'im... |
"""
This module defines custom converters for routing. These converters streamline
resource code by performing the process of getting the model instance from the
database before the view method is called. The converter converts a resource's
slug to the corresponding model instance, which is then placed into the view
me... | [
"app.models.Observer.query.filter_by",
"app.models.Map.query.filter_by"
] | [((1159, 1193), 'app.models.Map.query.filter_by', 'Map.query.filter_by', ([], {'slug': 'map_slug'}), '(slug=map_slug)\n', (1178, 1193), False, 'from app.models import Map, Observer, Obstacle\n'), ((1650, 1689), 'app.models.Observer.query.filter_by', 'Observer.query.filter_by', ([], {'slug': 'obs_slug'}), '(slug=obs_slu... |
# Licensed under the GPL: https://www.gnu.org/licenses/old-licenses/gpl-2.0.html
# For details: https://github.com/PyCQA/pylint/blob/master/LICENSE
from glob import glob
from os.path import basename, join, splitext
from pylint.testutils.constants import SYS_VERS_STR
def _get_tests_info(input_dir, msg_dir, prefix, s... | [
"os.path.splitext",
"os.path.join",
"os.path.basename"
] | [((820, 858), 'os.path.join', 'join', (['input_dir', "(prefix + '*' + suffix)"], {}), "(input_dir, prefix + '*' + suffix)\n", (824, 858), False, 'from os.path import basename, join, splitext\n'), ((878, 893), 'os.path.basename', 'basename', (['fname'], {}), '(fname)\n', (886, 893), False, 'from os.path import basename,... |
from aiogram import types
from .dataset import PHOTO
photo = types.PhotoSize(**PHOTO)
def test_export():
exported = photo.to_python()
assert isinstance(exported, dict)
assert exported == PHOTO
def test_file_id():
assert isinstance(photo.file_id, str)
assert photo.file_id == PHOTO['file_id']
d... | [
"aiogram.types.PhotoSize"
] | [((62, 86), 'aiogram.types.PhotoSize', 'types.PhotoSize', ([], {}), '(**PHOTO)\n', (77, 86), False, 'from aiogram import types\n')] |
from Models.EAE_models.ensemblance_interface import EAE_interface
from Models.EAE_models.EncoderCNN import EncoderCNN
from Models.EAE_models.DecoderCNN import DecoderCNN
import tensorflow as tf
class Vanilla_CNN_ensemblance(EAE_interface):
def __init__(self, filters, dim, p_m, **kwargs):
super(Vanilla_CN... | [
"Models.EAE_models.DecoderCNN.DecoderCNN",
"Models.EAE_models.EncoderCNN.EncoderCNN",
"tensorflow.keras.layers.Dense"
] | [((384, 417), 'tensorflow.keras.layers.Dense', 'tf.keras.layers.Dense', (['self.z_dim'], {}), '(self.z_dim)\n', (405, 417), True, 'import tensorflow as tf\n'), ((437, 470), 'tensorflow.keras.layers.Dense', 'tf.keras.layers.Dense', (['(4 * 4 * 32)'], {}), '(4 * 4 * 32)\n', (458, 470), True, 'import tensorflow as tf\n'),... |
from collections import Counter
from analysis.computation import utils
def frequency(ambitus_list):
freq = Counter(ambitus_list)
r = [['Ambitus', 'Pieces']]
for k, v in sorted(freq.items()):
r.append([k, v])
return r
def frequency_pie(ambitus_list):
r = utils.aux_pie_chart(Counter(ambit... | [
"analysis.computation.utils.boxplot",
"collections.Counter",
"analysis.computation.utils.aux_basic_stats",
"analysis.computation.utils.histogram",
"analysis.computation.utils.distribution",
"analysis.computation.utils.get_music_data_attrib"
] | [((113, 134), 'collections.Counter', 'Counter', (['ambitus_list'], {}), '(ambitus_list)\n', (120, 134), False, 'from collections import Counter\n'), ((431, 493), 'analysis.computation.utils.get_music_data_attrib', 'utils.get_music_data_attrib', (['compositions', '"""ambitus"""', '"""append"""'], {}), "(compositions, 'a... |
"""Test harness for Image Sequence References."""
import unittest
import sys
import opentimelineio as otio
import opentimelineio.test_utils as otio_test_utils
IS_PYTHON_2 = (sys.version_info < (3, 0))
class ImageSequenceReferenceTests(
unittest.TestCase, otio_test_utils.OTIOAssertions
):
def test_create(se... | [
"opentimelineio.adapters.otio_json.read_from_string",
"opentimelineio.opentime.RationalTime",
"opentimelineio.schema.ImageSequenceReference",
"opentimelineio.schema.V2d",
"opentimelineio.opentime.from_timecode",
"unittest.skipIf",
"opentimelineio.adapters.otio_json.write_to_string",
"copy.deepcopy",
... | [((1688, 1762), 'unittest.skipIf', 'unittest.skipIf', (['IS_PYTHON_2', '"""unicode strings do funny things in python2"""'], {}), "(IS_PYTHON_2, 'unicode strings do funny things in python2')\n", (1703, 1762), False, 'import unittest\n'), ((2914, 2988), 'unittest.skipIf', 'unittest.skipIf', (['IS_PYTHON_2', '"""unicode s... |
#
# Copyright 2013 Red Hat, Inc
#
# Licensed under the Apache License, Version 2.0 (the "License"); you may
# not use this file except in compliance with the License. You may obtain
# a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agr... | [
"warnings.warn",
"options.build_url"
] | [((7653, 7679), 'options.build_url', 'options.build_url', (['path', 'q'], {}), '(path, q)\n', (7670, 7679), False, 'import options\n'), ((8002, 8036), 'options.build_url', 'options.build_url', (['path', 'q', 'params'], {}), '(path, q, params)\n', (8019, 8036), False, 'import options\n'), ((3360, 3445), 'warnings.warn',... |
from yo_fluq_ds__tests.common import *
import numpy as np
class MiscMethodsTests(TestCase):
def test_pairwise(self):
result = Query.args(1,2,3).feed(fluq.pairwise()).to_list()
self.assertListEqual([(1,2),(2,3)],result)
def test_strjoin(self):
result = Query.args(1,2,3).feed(fluq.strjoi... | [
"numpy.random.RandomState"
] | [((819, 843), 'numpy.random.RandomState', 'np.random.RandomState', (['(1)'], {}), '(1)\n', (840, 843), True, 'import numpy as np\n')] |
from direct.directnotify import DirectNotifyGlobal
from direct.distributed.DistributedObjectUD import DistributedObjectUD
class AwardManagerUD(DistributedObjectUD):
notify = DirectNotifyGlobal.directNotify.newCategory('AwardManagerUD')
| [
"direct.directnotify.DirectNotifyGlobal.directNotify.newCategory"
] | [((179, 240), 'direct.directnotify.DirectNotifyGlobal.directNotify.newCategory', 'DirectNotifyGlobal.directNotify.newCategory', (['"""AwardManagerUD"""'], {}), "('AwardManagerUD')\n", (222, 240), False, 'from direct.directnotify import DirectNotifyGlobal\n')] |
import numpy as np
import torch
import torch.nn as nn
import torch.nn.functional as F
import math, copy, time
from torch.autograd import Variable
from utils import outputActivation
import pdb
# Customizations
# - DONE Embeddings: linear transform d_feats -> d_model features
# - DONE Generator
# - DONE Batching
# D... | [
"torch.nn.Dropout",
"torch.sin",
"math.sqrt",
"torch.from_numpy",
"utils.outputActivation",
"math.log",
"numpy.array",
"torch.cos",
"torch.cuda.is_available",
"torch.squeeze",
"copy.deepcopy",
"torch.nn.init.xavier_uniform",
"copy.copy",
"torch.repeat_interleave",
"torch.nn.functional.so... | [((4461, 4486), 'torch.nn.functional.softmax', 'F.softmax', (['scores'], {'dim': '(-1)'}), '(scores, dim=-1)\n', (4470, 4486), True, 'import torch.nn.functional as F\n'), ((735, 753), 'copy.copy', 'copy.copy', (['d_model'], {}), '(d_model)\n', (744, 753), False, 'import math, copy, time\n'), ((2080, 2115), 'torch.nn.Li... |
import pytest
from multi_bracket_validation import multi_bracket_validation
def test_mbv_true_case_simple():
"""test function on balanced str"""
assert multi_bracket_validation('[{()}]') == True
def test_mbv_true_case_empty_str():
"""test function with empty string"""
assert multi_bracket_validation('... | [
"multi_bracket_validation.multi_bracket_validation"
] | [((161, 195), 'multi_bracket_validation.multi_bracket_validation', 'multi_bracket_validation', (['"""[{()}]"""'], {}), "('[{()}]')\n", (185, 195), False, 'from multi_bracket_validation import multi_bracket_validation\n'), ((294, 322), 'multi_bracket_validation.multi_bracket_validation', 'multi_bracket_validation', (['"... |
# -*- coding: utf-8 -*-
# Part of Odoo. See LICENSE file for full copyright and licensing details.
import ir
import workflow
import module
import res
import report
import tests
def post_init(cr, registry):
"""Rewrite ICP's to force groups"""
from odoo import api, SUPERUSER_ID
from odoo.addons.base.ir.ir_c... | [
"odoo.api.Environment",
"odoo.addons.base.ir.ir_config_parameter._default_parameters.iteritems"
] | [((374, 411), 'odoo.api.Environment', 'api.Environment', (['cr', 'SUPERUSER_ID', '{}'], {}), '(cr, SUPERUSER_ID, {})\n', (389, 411), False, 'from odoo import api, SUPERUSER_ID\n'), ((470, 501), 'odoo.addons.base.ir.ir_config_parameter._default_parameters.iteritems', '_default_parameters.iteritems', ([], {}), '()\n', (4... |
"""
===================================
Toy Glucose Phosphorylation Process
===================================
This is a toy example referenced in the documentation.
"""
from vivarium.core.process import Process
from vivarium.core.composition import simulate_process
from vivarium.plots.simulation_output import plot_... | [
"vivarium.core.composition.simulate_process",
"vivarium.plots.simulation_output.plot_simulation_output"
] | [((3674, 3712), 'vivarium.core.composition.simulate_process', 'simulate_process', (['my_process', 'settings'], {}), '(my_process, settings)\n', (3690, 3712), False, 'from vivarium.core.composition import simulate_process\n'), ((3717, 3761), 'vivarium.plots.simulation_output.plot_simulation_output', 'plot_simulation_out... |
# -*- coding: utf-8 -*-
from __future__ import print_function
import datetime
import collections
from midas.compat import unittest
class PigSchemaToPyStructTests(unittest.TestCase):
def _get_target(self):
from midas.pig_schema import pig_schema_to_py_struct
return pig_schema_to_py_struct
... | [
"midas.pig_schema.make_serializer",
"collections.namedtuple",
"midas.pig_schema.make_parser",
"midas.pig_schema.pig_schema_to_py_struct",
"midas.compat.unittest.main",
"datetime.date",
"midas.pig_schema.make_parser_from_schema"
] | [((7811, 7826), 'midas.compat.unittest.main', 'unittest.main', ([], {}), '()\n', (7824, 7826), False, 'from midas.compat import unittest\n'), ((2594, 2618), 'midas.pig_schema.make_parser', 'make_parser', (['"""chararray"""'], {}), "('chararray')\n", (2605, 2618), False, 'from midas.pig_schema import make_parser\n'), ((... |
#!/usr/bin/env python
# -*- coding: utf-8 -*-
"""General utilities (like verbose output decroator)."""
import base64
import os
import sys
from termcolor import colored
def verbose(in_progress: str, success: str):
"""Print message to stdout/stderr using decorator syntax.
Parameters
----------
in_pro... | [
"termcolor.colored",
"sys.exit"
] | [((1399, 1438), 'termcolor.colored', 'colored', (['message', '"""red"""'], {'attrs': "['bold']"}), "(message, 'red', attrs=['bold'])\n", (1406, 1438), False, 'from termcolor import colored\n'), ((1820, 1859), 'termcolor.colored', 'colored', (['message', 'color'], {'attrs': "['bold']"}), "(message, color, attrs=['bold']... |
#!/usr/bin python3
# Write and send emails from the Command Line.
import sys
import time
from selenium import webdriver
from selenium.webdriver.common.keys import Keys
recipient = sys.argv[1]
subject = sys.argv[2]
message = sys.argv[3]
# Username Login
browser = webdriver.Firefox()
browser.get('https://accounts.goo... | [
"selenium.webdriver.Firefox",
"time.sleep"
] | [((267, 286), 'selenium.webdriver.Firefox', 'webdriver.Firefox', ([], {}), '()\n', (284, 286), False, 'from selenium import webdriver\n'), ((476, 489), 'time.sleep', 'time.sleep', (['(3)'], {}), '(3)\n', (486, 489), False, 'import time\n'), ((625, 638), 'time.sleep', 'time.sleep', (['(5)'], {}), '(5)\n', (635, 638), Fa... |
from fluiddb.testing.basic import FluidinfoTestCase
from fluiddb.util.constant import Constant, ConstantEnum
class ConstantTest(FluidinfoTestCase):
def testInstantiate(self):
"""A L{Constant} includes an ID and a name."""
constant = Constant(42, 'VALUE')
self.assertEqual(42, constant.id)
... | [
"fluiddb.util.constant.ConstantEnum",
"fluiddb.util.constant.Constant"
] | [((256, 277), 'fluiddb.util.constant.Constant', 'Constant', (['(42)', '"""VALUE"""'], {}), "(42, 'VALUE')\n", (264, 277), False, 'from fluiddb.util.constant import Constant, ConstantEnum\n'), ((515, 536), 'fluiddb.util.constant.Constant', 'Constant', (['(42)', '"""VALUE"""'], {}), "(42, 'VALUE')\n", (523, 536), False, ... |
import sqlite3
class Database:
def __init__(self):
self.connection = sqlite3.connect('./database.db')
self.cursor = self.connection.cursor()
def insert_item(self, item):
cur = self.cursor.execute(
"insert into posts (title, file, link) values (?, ?, ?)",
(it... | [
"sqlite3.connect"
] | [((82, 114), 'sqlite3.connect', 'sqlite3.connect', (['"""./database.db"""'], {}), "('./database.db')\n", (97, 114), False, 'import sqlite3\n')] |
#!/usr/bin/env python
# -*- coding: utf-8 -*-
###############################################################################
# Copyright 2015 Kitware Inc.
#
# Licensed under the Apache License, Version 2.0 ( the 'License' );
# you may not use this file except in compliance with the License.
# You may obtain a cop... | [
"cumulus.common.girder._get_profile",
"girder.api.rest.getCurrentUser",
"cumulus.constants.VolumeType.is_valid_type",
"girder.api.rest.loadmodel",
"bson.objectid.ObjectId",
"cumulus.common.girder.get_task_token",
"girder.api.describe.Description",
"cumulus.common.girder.send_status_notification",
"g... | [((2908, 2975), 'girder.api.rest.loadmodel', 'loadmodel', ([], {'model': '"""volume"""', 'plugin': '"""cumulus"""', 'level': 'AccessType.WRITE'}), "(model='volume', plugin='cumulus', level=AccessType.WRITE)\n", (2917, 2975), False, 'from girder.api.rest import loadmodel\n'), ((4799, 5363), 'girder.api.docs.addModel', '... |
import numpy as np
import time
import ray
import ray.autoscaler.sdk
from ray._private.test_utils import Semaphore
import json
import os
from time import perf_counter
from tqdm import trange, tqdm
MAX_ARGS = 10000
MAX_RETURNS = 3000
MAX_RAY_GET_ARGS = 10000
MAX_QUEUED_TASKS = 1_000_000
MAX_RAY_GET_SIZE = 100 * 2**30
... | [
"ray.init",
"ray.cluster_resources",
"ray.get",
"numpy.arange",
"tqdm.tqdm",
"time.perf_counter",
"time.sleep",
"ray._private.test_utils.Semaphore.remote",
"numpy.zeros",
"ray.put",
"ray.remote",
"ray.available_resources",
"tqdm.trange",
"json.dump"
] | [((3347, 3371), 'ray.init', 'ray.init', ([], {'address': '"""auto"""'}), "(address='auto')\n", (3355, 3371), False, 'import ray\n'), ((3386, 3400), 'time.perf_counter', 'perf_counter', ([], {}), '()\n', (3398, 3400), False, 'from time import perf_counter\n'), ((3429, 3443), 'time.perf_counter', 'perf_counter', ([], {})... |
import cv2
import numpy as np
import matplotlib.pyplot as plt
IMAGE = "b&w2.jpg"
prototxt = "./Models/colorization_deploy_v2.prototxt.txt"
model = "./Models/colorization_release_v2.caffemodel"
points = "./Models/pts_in_hull.npy"
image = "./input_images/"+IMAGE
net = cv2.dnn.readNetFromCaffe(prototxt, mod... | [
"numpy.clip",
"matplotlib.pyplot.imshow",
"cv2.dnn.blobFromImage",
"numpy.full",
"cv2.dnn.readNetFromCaffe",
"cv2.cvtColor",
"numpy.concatenate",
"cv2.split",
"matplotlib.pyplot.axis",
"cv2.resize",
"numpy.load",
"cv2.imread",
"matplotlib.pyplot.show"
] | [((282, 323), 'cv2.dnn.readNetFromCaffe', 'cv2.dnn.readNetFromCaffe', (['prototxt', 'model'], {}), '(prototxt, model)\n', (306, 323), False, 'import cv2\n'), ((331, 346), 'numpy.load', 'np.load', (['points'], {}), '(points)\n', (338, 346), True, 'import numpy as np\n'), ((610, 627), 'cv2.imread', 'cv2.imread', (['image... |
import pprint
import sys
import youtube_dl
def main():
for url in sys.argv[1:]:
with youtube_dl.YoutubeDL() as ydl:
info_dict = ydl.extract_info(url, download=False)
pprint.pprint(info_dict)
if __name__ == '__main__':
main()
| [
"youtube_dl.YoutubeDL",
"pprint.pprint"
] | [((100, 122), 'youtube_dl.YoutubeDL', 'youtube_dl.YoutubeDL', ([], {}), '()\n', (120, 122), False, 'import youtube_dl\n'), ((205, 229), 'pprint.pprint', 'pprint.pprint', (['info_dict'], {}), '(info_dict)\n', (218, 229), False, 'import pprint\n')] |
import tensorflow as tf
x = tf.placeholder(tf.float32, [None,3])
x_data = [[1,2,3],[4,5,6]]
w = tf.Variable(tf.random_normal([3,2]))
b = tf.Variable(tf.random_normal([2,1]))
expr = tf.matmul(x,w) + b
sess = tf.Session()
sess.run(tf.global_variables_initializer())
print(x_data)
print(sess.run(w))
print(sess.run(b... | [
"tensorflow.random_normal",
"tensorflow.placeholder",
"tensorflow.Session",
"tensorflow.global_variables_initializer",
"tensorflow.matmul"
] | [((29, 66), 'tensorflow.placeholder', 'tf.placeholder', (['tf.float32', '[None, 3]'], {}), '(tf.float32, [None, 3])\n', (43, 66), True, 'import tensorflow as tf\n'), ((212, 224), 'tensorflow.Session', 'tf.Session', ([], {}), '()\n', (222, 224), True, 'import tensorflow as tf\n'), ((111, 135), 'tensorflow.random_normal'... |
#!/bin/env python
import StringIO
import unittest
import naivetrack
import naivetrack.utils
class TestNaivetrackSummaries(unittest.TestCase):
def setUp(self):
# we will use a global chromosome list
self.chromosome_list = []
self.file = StringIO.StringIO("""#
#
chr1\t2\tT\t2\t10\t50
\tG\t... | [
"StringIO.StringIO",
"naivetrack.utils.aggregate_allele_information",
"naivetrack.SummarizeEntries",
"naivetrack.TrackParser"
] | [((268, 479), 'StringIO.StringIO', 'StringIO.StringIO', (['"""#\n#\nchr1\t2\tT\t2\t10\t50\n\tG\t4\t20\tsample_1, forward\n\tT\t6\t30\tsample_1, forward\nchr1\t20000\tT\t3\t40\t130\n\tA\t20\t60\tsample_1, forward\n\tA\t10\t30\tsample_1, reverse\n\tG\t10\t40\tsample_1, forward\n"""'], {}), '(\n """#\n#\nchr1\t2\tT\t2\... |
import numpy as np
from netCDF4 import Dataset
from datetime import datetime
from datetime import timedelta
import os
import sys
import matplotlib.pyplot as plt
from matplotlib import gridspec
import matplotlib.colors as mcolors
import matplotlib.patches as patches
from matplotlib.colors import BoundaryNorm
from to... | [
"tools_LT.setup_12p",
"numpy.array",
"datetime.timedelta",
"numpy.arange",
"datetime.datetime",
"numpy.where",
"netCDF4.Dataset",
"matplotlib.colors.ListedColormap",
"matplotlib.pyplot.close",
"numpy.meshgrid",
"matplotlib.pyplot.cla",
"matplotlib.pyplot.cm.get_cmap",
"matplotlib.pyplot.show... | [((17338, 17372), 'numpy.arange', 'np.arange', (['(DX * 0.5)', '(DX * XDIM)', 'DX'], {}), '(DX * 0.5, DX * XDIM, DX)\n', (17347, 17372), True, 'import numpy as np\n'), ((17375, 17409), 'numpy.arange', 'np.arange', (['(DY * 0.5)', '(DY * YDIM)', 'DY'], {}), '(DY * 0.5, DY * YDIM, DY)\n', (17384, 17409), True, 'import nu... |
from django import forms
from filament.models import *
class FilamentForm(forms.Form):
material = forms.ChoiceField(
required=True,
choices=FILAMENT_MATERIAL_CHOICES,
widget=forms.Select(attrs={
'class': 'form-control',
'form': 'filament_form',
}),
... | [
"django.forms.ClearableFileInput",
"django.forms.Select",
"django.forms.NumberInput",
"django.forms.URLInput",
"django.forms.TextInput"
] | [((210, 280), 'django.forms.Select', 'forms.Select', ([], {'attrs': "{'class': 'form-control', 'form': 'filament_form'}"}), "(attrs={'class': 'form-control', 'form': 'filament_form'})\n", (222, 280), False, 'from django import forms\n'), ((399, 510), 'django.forms.NumberInput', 'forms.NumberInput', ([], {'attrs': "{'cl... |
from django.shortcuts import render
from django.http import HttpResponse
# Create your views here.
def hello_world(request):
return HttpResponse("Hello world!!!")
| [
"django.http.HttpResponse"
] | [((138, 168), 'django.http.HttpResponse', 'HttpResponse', (['"""Hello world!!!"""'], {}), "('Hello world!!!')\n", (150, 168), False, 'from django.http import HttpResponse\n')] |
# Generated by Django 3.1 on 2020-08-13 02:38
from django.db import migrations, models
import django.utils.timezone
class Migration(migrations.Migration):
dependencies = [
("core", "0011_auto_20200214_1939"),
]
operations = [
migrations.AlterField(
model_name="note",
... | [
"django.db.models.DateTimeField"
] | [((356, 432), 'django.db.models.DateTimeField', 'models.DateTimeField', ([], {'default': 'django.utils.timezone.now', 'verbose_name': '"""Time"""'}), "(default=django.utils.timezone.now, verbose_name='Time')\n", (376, 432), False, 'from django.db import migrations, models\n')] |
"""Module that holds tickets endpoints"""
from datetime import date
from flask_jwt_extended import jwt_required, get_jwt_identity
from flask_restplus import Resource
from sqlalchemy import func
from api.utilities.helpers.return_value import return_value
from api.utilities.messages.error import error_messages
from api... | [
"flask_jwt_extended.get_jwt_identity",
"datetime.date.today",
"sqlalchemy.func.DATE",
"main.api.route"
] | [((656, 701), 'main.api.route', 'api.route', (['"""/flights/<string:flight_id>/book"""'], {}), "('/flights/<string:flight_id>/book')\n", (665, 701), False, 'from main import api\n'), ((1895, 1916), 'main.api.route', 'api.route', (['"""/tickets"""'], {}), "('/tickets')\n", (1904, 1916), False, 'from main import api\n'),... |
#! /usr/bin/python
import sys, struct
import xml.dom.minidom
from lmcp import LMCPObject
## ===============================================================================
## Authors: AFRL/RQQA
## Organization: Air Force Research Laboratory, Aerospace Systems Directorate, Power and Control Division
##
## Copyright (... | [
"lmcp.LMCPObject.LMCPObject.unpackFromXMLNode",
"lmcp.LMCPFactory.LMCPFactory",
"lmcp.LMCPObject.LMCPObject.unpack",
"lmcp.LMCPObject.LMCPObject.toString",
"lmcp.LMCPObject.LMCPObject.toDictMembers",
"struct.pack",
"lmcp.LMCPObject.LMCPObject.pack",
"lmcp.LMCPObject.LMCPObject.unpackFromDict",
"lmcp... | [((2372, 2420), 'lmcp.LMCPObject.LMCPObject.unpack', 'LMCPObject.LMCPObject.unpack', (['self', 'buffer', '_pos'], {}), '(self, buffer, _pos)\n', (2400, 2420), False, 'from lmcp import LMCPObject\n'), ((3517, 3581), 'lmcp.LMCPObject.LMCPObject.unpackFromXMLNode', 'LMCPObject.LMCPObject.unpackFromXMLNode', (['self', 'el'... |
from abc import abstractmethod
import numpy as np
from pymoo.core.population import Population
# ---------------------------------------------------------------------------------------------------------
# Survival
# ----------------------------------------------------------------------------------------------------... | [
"numpy.median",
"pymoo.core.population.Population",
"numpy.where",
"numpy.argsort",
"pymoo.core.population.Population.merge"
] | [((2429, 2442), 'numpy.median', 'np.median', (['cv'], {}), '(cv)\n', (2438, 2442), True, 'import numpy as np\n'), ((2183, 2194), 'numpy.where', 'np.where', (['b'], {}), '(b)\n', (2191, 2194), True, 'import numpy as np\n'), ((2215, 2227), 'numpy.where', 'np.where', (['(~b)'], {}), '(~b)\n', (2223, 2227), True, 'import n... |
from os import mkdir
from baelfire.dependencies import AlwaysTrue
from baelfire.dependencies import FileChanged
from baelfire.dependencies import FileDoesNotExists
from baelfire.dependencies import PidIsNotRunning
from baelfire.dependencies import TaskRebuilded
from baelfire.task import FileTask
from baelfire.task imp... | [
"baelfire.dependencies.PidIsNotRunning",
"baelfire.dependencies.FileDoesNotExists",
"bdjango.dependency.MigrationsChanged",
"baelfire.dependencies.AlwaysTrue",
"os.mkdir",
"baelfire.dependencies.FileChanged"
] | [((608, 626), 'os.mkdir', 'mkdir', (['self.output'], {}), '(self.output)\n', (613, 626), False, 'from os import mkdir\n'), ((1089, 1111), 'baelfire.dependencies.FileChanged', 'FileChanged', (['"""setuppy"""'], {}), "('setuppy')\n", (1100, 1111), False, 'from baelfire.dependencies import FileChanged\n'), ((1561, 1600), ... |
"""
Copyright (c) 2016, <NAME> .All rights reserved.
Redistribution and use in source and binary forms, with or without modification,
are permitted provided that the following conditions are met:
1. Redistributions of source code must retain the above copyright notice,
this list of conditions and the foll... | [
"numpy.ones",
"nibabel.load",
"scipy.io.loadmat",
"numpy.issubdtype",
"ImgOperations.imgOp.applyPadding",
"numpy.empty",
"numpy.rint",
"numpy.random.shuffle"
] | [((1511, 1534), 'nibabel.load', 'nib.load', (['imageFileName'], {}), '(imageFileName)\n', (1519, 1534), True, 'import nibabel as nib\n'), ((1987, 2013), 'scipy.io.loadmat', 'sio.loadmat', (['imageFileName'], {}), '(imageFileName)\n', (1998, 2013), True, 'import scipy.io as sio\n'), ((5928, 5951), 'numpy.random.shuffle'... |
#!/usr/bin/env python3
import argparse
import pandas as pd
import os
from tree import *
import copy
def main():
parser = argparse.ArgumentParser()
parser.add_argument('--json',
help='auspice json')
parser.add_argument('--ids',
help='Drop all ids in the list... | [
"argparse.ArgumentParser"
] | [((128, 153), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {}), '()\n', (151, 153), False, 'import argparse\n')] |
# -*- coding: utf-8 -*-
from __future__ import unicode_literals
import flask_restplus as restplus
class Foo(restplus.Resource):
def get(self):
return "data"
class ErrorsTest(object):
def test_accept_default_application_json(self, app, client):
api = restplus.Api(app)
api.add_resourc... | [
"flask_restplus.Api"
] | [((279, 296), 'flask_restplus.Api', 'restplus.Api', (['app'], {}), '(app)\n', (291, 296), True, 'import flask_restplus as restplus\n'), ((574, 591), 'flask_restplus.Api', 'restplus.Api', (['app'], {}), '(app)\n', (586, 591), True, 'import flask_restplus as restplus\n'), ((885, 926), 'flask_restplus.Api', 'restplus.Api'... |
"""Fuzzy matching to a library of songs in RTTTL format.
Takes an input string and matches against song title in a dictionary of available songs.
Returns song title, RTTTL string, and match confidence metric. Fuzzy matching applied so
a 'match' is guaranteed, however wayward the confidence.
Dependencies:
pip3 in... | [
"fuzzywuzzy.fuzz.token_set_ratio"
] | [((897, 936), 'fuzzywuzzy.fuzz.token_set_ratio', 'fuzz.token_set_ratio', (['input_string', 'key'], {}), '(input_string, key)\n', (917, 936), False, 'from fuzzywuzzy import fuzz\n')] |
"""
Test nqueens algorithm
"""
from modules.queens.algorithms.nqueens import NQueens
from . import app
# n solutions
# 1 1
# 2 0
# 3 0
# 4 2
# 5 10
# 6 4
# 7 40
# 8 92
# 9 352
# 10 724
# 11 2,680
# 12 14,200
# 13 73,712
# 14 365,596
# 15 2,279,184
"""Set the N size board"""
n = 10
solutions = 724
def test_n_queens(... | [
"modules.queens.algorithms.nqueens.NQueens"
] | [((378, 395), 'modules.queens.algorithms.nqueens.NQueens', 'NQueens', (['n', '(False)'], {}), '(n, False)\n', (385, 395), False, 'from modules.queens.algorithms.nqueens import NQueens\n'), ((643, 659), 'modules.queens.algorithms.nqueens.NQueens', 'NQueens', (['n', '(True)'], {}), '(n, True)\n', (650, 659), False, 'from... |
import pygame as pg
from Data.images import *
pg.init()
def draw_minimap(surface, player):
pg.draw.rect(surface, pg.Color("black"), (595, 0, 205, 205), 0)
surface.blit(pg.transform.scale(player.level.background, (200, 200)), (600,0))
surface.blit(IMAGES['minimap_player'], \
((600 + ((pla... | [
"pygame.Color",
"pygame.init",
"pygame.transform.scale"
] | [((46, 55), 'pygame.init', 'pg.init', ([], {}), '()\n', (53, 55), True, 'import pygame as pg\n'), ((119, 136), 'pygame.Color', 'pg.Color', (['"""black"""'], {}), "('black')\n", (127, 136), True, 'import pygame as pg\n'), ((178, 233), 'pygame.transform.scale', 'pg.transform.scale', (['player.level.background', '(200, 20... |
from GUI import *
import Backend as core
app = None
def view_command():
rows = core.view()
app.listClientes.delete(0, END)
for r in rows:
app.listClientes.insert(END, r)
def search_command():
app.listClientes.delete(0, END)
rows = core.search(app.txtNome.get(), app.txtSobreno... | [
"Backend.view",
"Backend.delete"
] | [((91, 102), 'Backend.view', 'core.view', ([], {}), '()\n', (100, 102), True, 'import Backend as core\n'), ((774, 789), 'Backend.delete', 'core.delete', (['id'], {}), '(id)\n', (785, 789), True, 'import Backend as core\n')] |
import sys, os
sys.path.append(os.pardir)
from fluctana import *
# HOW TO RUN
# ./python3 check_hurst.py 10186 [15.9,16] ECEI_L1303-1305
shot = int(sys.argv[1])
trange = eval(sys.argv[2])
clist = sys.argv[3].split(',')
# call fluctana
A = FluctAna()
# add data
A.add_data(KstarEcei(shot=shot, clist=clist), trange=t... | [
"sys.path.append"
] | [((15, 41), 'sys.path.append', 'sys.path.append', (['os.pardir'], {}), '(os.pardir)\n', (30, 41), False, 'import sys, os\n')] |
from typing import List
from dataclasses import dataclass, field
@dataclass()
class DownloadParams:
paths: List[str]
output_folder: str
s3_bucket: str = field(default="for-dvc")
| [
"dataclasses.dataclass",
"dataclasses.field"
] | [((69, 80), 'dataclasses.dataclass', 'dataclass', ([], {}), '()\n', (78, 80), False, 'from dataclasses import dataclass, field\n'), ((168, 192), 'dataclasses.field', 'field', ([], {'default': '"""for-dvc"""'}), "(default='for-dvc')\n", (173, 192), False, 'from dataclasses import dataclass, field\n')] |
import datetime
from asgiref.sync import sync_to_async
from django.urls import reverse_lazy
from django.utils.timezone import now
from django.shortcuts import render, reverse, get_object_or_404, get_list_or_404, HttpResponse
from django.views.generic import ListView, DetailView, TemplateView, CreateView
from django.co... | [
"django.shortcuts.render",
"Laelia.pdf.engine.PdfCreator",
"Laelia.apps.care.models.Visit.objects.get",
"django.contrib.contenttypes.models.ContentType.objects.get",
"Laelia.apps.care.forms.VisitModelForm",
"django.shortcuts.HttpResponse",
"django.shortcuts.get_object_or_404",
"django.shortcuts.get_li... | [((944, 1007), 'django.contrib.contenttypes.models.ContentType.objects.get', 'ContentType.objects.get', ([], {'app_label': '"""base"""', 'model': '"""professional"""'}), "(app_label='base', model='professional')\n", (967, 1007), False, 'from django.contrib.contenttypes.models import ContentType\n'), ((17686, 17730), 'd... |
#!/usr/bin/env python3
# -*- coding: utf-8 -*-
"""Collection of helper methods for rdm module
@author: baihan
"""
import numpy as np
from scipy.spatial.distance import squareform
def batch_to_vectors(x):
"""converts a *stack* of RDMs in vector or matrix form into vector form
Args:
x: stack of RDMs
... | [
"scipy.spatial.distance.squareform",
"numpy.sqrt",
"numpy.unique",
"numpy.array",
"numpy.ndarray",
"numpy.arange"
] | [((2313, 2338), 'numpy.unique', 'np.unique', (['pattern_select'], {}), '(pattern_select)\n', (2322, 2338), True, 'import numpy as np\n'), ((1457, 1492), 'numpy.ndarray', 'np.ndarray', (['(n_rdm, n_cond, n_cond)'], {}), '((n_rdm, n_cond, n_cond))\n', (1467, 1492), True, 'import numpy as np\n'), ((1512, 1528), 'numpy.ara... |
from passlib.hash import ldap_salted_sha1
from passlib.hash import ldap_pbkdf2_sha256
from sshpubkeys import SSHKey
import datetime
class Usermgmt(object):
def attrs(self):
raise NotImplementedError
def get_dict(self):
return dict((key, value) for key, value in self.__dict__.items()
... | [
"passlib.hash.ldap_pbkdf2_sha256.hash",
"passlib.hash.ldap_pbkdf2_sha256.verify",
"passlib.hash.ldap_salted_sha1.identify",
"datetime.datetime.now",
"sshpubkeys.SSHKey",
"passlib.hash.ldap_pbkdf2_sha256.identify",
"passlib.hash.ldap_salted_sha1.verify"
] | [((4687, 4698), 'sshpubkeys.SSHKey', 'SSHKey', (['key'], {}), '(key)\n', (4693, 4698), False, 'from sshpubkeys import SSHKey\n'), ((4798, 4809), 'sshpubkeys.SSHKey', 'SSHKey', (['key'], {}), '(key)\n', (4804, 4809), False, 'from sshpubkeys import SSHKey\n'), ((4129, 4162), 'passlib.hash.ldap_pbkdf2_sha256.hash', 'ldap_... |
#!/usr/bin/env python2
# -*- coding: utf-8 -*-
"""
Created on Fri Apr 28 17:57:57 2017
@author: pipolose
"""
from polyML.plot_overlay import plot_n_save_3plane
import matplotlib.pyplot as plt
from matplotlib import cm
in_dict = {}
in_dict['do_save'] = True # True #
in_dict['formats_used'] = ['pdf', 'png']
in_dict[... | [
"polyML.plot_overlay.plot_n_save_3plane"
] | [((761, 788), 'polyML.plot_overlay.plot_n_save_3plane', 'plot_n_save_3plane', (['in_dict'], {}), '(in_dict)\n', (779, 788), False, 'from polyML.plot_overlay import plot_n_save_3plane\n'), ((1132, 1159), 'polyML.plot_overlay.plot_n_save_3plane', 'plot_n_save_3plane', (['in_dict'], {}), '(in_dict)\n', (1150, 1159), False... |
from app import db
ROLE_USER = 0
ROLE_ADMIN = 1
class User(db.Model):
id = db.Column(db.Integer, primary_key = True)
nickname = db.Column(db.String(64), index = True, unique = True)
email = db.Column(db.String(120), index = True, unique = True)
role = db.Column(db.SmallInteger, default = ROLE_USER)
... | [
"app.db.String",
"app.db.Column"
] | [((81, 120), 'app.db.Column', 'db.Column', (['db.Integer'], {'primary_key': '(True)'}), '(db.Integer, primary_key=True)\n', (90, 120), False, 'from app import db\n'), ((270, 315), 'app.db.Column', 'db.Column', (['db.SmallInteger'], {'default': 'ROLE_USER'}), '(db.SmallInteger, default=ROLE_USER)\n', (279, 315), False, ... |
from flask import Blueprint, render_template, request
from wtforms import Form, IntegerField, validators
from flask_table import Table, Col
from app.helpers.requestid import requestid
from app.modules.books.models import Book
# pylint: disable=invalid-name
frontend_blueprint = Blueprint('frontend', __name__, url_prefi... | [
"flask.render_template",
"app.modules.books.models.Book.query.filter",
"wtforms.validators.Optional",
"flask_table.Col",
"flask.Blueprint",
"app.modules.books.models.Book.query.all"
] | [((279, 326), 'flask.Blueprint', 'Blueprint', (['"""frontend"""', '__name__'], {'url_prefix': '"""/"""'}), "('frontend', __name__, url_prefix='/')\n", (288, 326), False, 'from flask import Blueprint, render_template, request\n'), ((418, 427), 'flask_table.Col', 'Col', (['"""id"""'], {}), "('id')\n", (421, 427), False, ... |
# Generated by Django 2.2.13 on 2020-06-12 19:50
from django.db import migrations, models
class Migration(migrations.Migration):
dependencies = [
('ebooks', '0106_auto_20200524_1227'),
('tooltips', '0006_auto_20200612_2140'),
]
operations = [
migrations.RemoveField(
... | [
"django.db.migrations.RemoveField",
"django.db.models.ManyToManyField"
] | [((284, 342), 'django.db.migrations.RemoveField', 'migrations.RemoveField', ([], {'model_name': '"""tooltip"""', 'name': '"""books"""'}), "(model_name='tooltip', name='books')\n", (306, 342), False, 'from django.db import migrations, models\n'), ((491, 552), 'django.db.models.ManyToManyField', 'models.ManyToManyField',... |
# Copyright (c) Facebook, Inc. and its affiliates.
#
# This source code is licensed under the MIT license found in the
# LICENSE file in the root directory of this source tree.
# flake8: noqa
from builtins import __test_sink, __test_source
def sink(json):
__test_sink(json)
def test():
query = {"json": __te... | [
"builtins.__test_sink",
"builtins.__test_source"
] | [((263, 280), 'builtins.__test_sink', '__test_sink', (['json'], {}), '(json)\n', (274, 280), False, 'from builtins import __test_sink, __test_source\n'), ((316, 331), 'builtins.__test_source', '__test_source', ([], {}), '()\n', (329, 331), False, 'from builtins import __test_sink, __test_source\n')] |
import json
import logging
from pprint import pformat
from typing import Any, Dict, Set
import pydantic
from models_library.projects_nodes import NodeID
from models_library.utils.nodes import compute_node_hash
from packaging import version
from ..node_ports_common.dbmanager import DBManager
from ..node_ports_common.e... | [
"logging.getLogger",
"json.loads",
"models_library.projects_nodes.NodeID",
"json.dumps",
"pprint.pformat",
"packaging.version.parse"
] | [((608, 635), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (625, 635), False, 'import logging\n'), ((545, 576), 'packaging.version.parse', 'version.parse', (['pydantic.VERSION'], {}), '(pydantic.VERSION)\n', (558, 576), False, 'from packaging import version\n'), ((579, 599), 'packaging.... |
# Author: <NAME>, <NAME>
# Date: 05/28/2020
# Version: 1.2
import json
import pygame
import GameField
import random
class Dice:
image_list: []
gamefield: GameField
image: pygame.image
coordinates = (5, 5)
surface_dice: pygame.Surface
"""
desc:
- init
param:
- gamefie... | [
"pygame.Surface",
"pygame.transform.smoothscale",
"pygame.display.get_surface",
"json.load",
"pygame.display.update",
"random.randint"
] | [((577, 693), 'pygame.Surface', 'pygame.Surface', (['(self.gamefield.screen_size_multiplier, self.gamefield.screen_size_multiplier)', 'pygame.SRCALPHA'], {}), '((self.gamefield.screen_size_multiplier, self.gamefield.\n screen_size_multiplier), pygame.SRCALPHA)\n', (591, 693), False, 'import pygame\n'), ((1321, 1341)... |
from random import randrange
import allure
from model.contact import Contact
def test_emails_on_home_page(app, db):
if len(db.get_contact_list()) == 0:
with allure.step("If contact's list is empty, I'll create a new contact"):
app.contact.create(Contact(firstname="NAME"))
with allure.step(... | [
"allure.step",
"model.contact.Contact"
] | [((308, 352), 'allure.step', 'allure.step', (['"""Given an index for compering"""'], {}), "('Given an index for compering')\n", (319, 352), False, 'import allure\n'), ((426, 492), 'allure.step', 'allure.step', (['"""Given contact\'s by index information from home page"""'], {}), '("Given contact\'s by index informati... |
import functools
import logging
import typing
from contextlib import contextmanager
from sqlalchemy.exc import SQLAlchemyError
from .exceptions import CorporaException
from ..corpora_orm import Base, DBSessionMaker
logger = logging.getLogger(__name__)
class DbUtils:
"""DbUtils as a singleton to avoid creating ... | [
"logging.getLogger",
"functools.wraps"
] | [((227, 254), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (244, 254), False, 'import logging\n'), ((2741, 2762), 'functools.wraps', 'functools.wraps', (['func'], {}), '(func)\n', (2756, 2762), False, 'import functools\n')] |
import sys
import os
import random
import matplotlib as mpl
import matplotlib.pyplot as plt
from dlimage.mnist import MNISTLoader
def load_data(data_path):
mndata = MNISTLoader(data_path)
images, labels = mndata.load_testing()
return images, labels
def show_image(image, label):
plt.title(label, {'f... | [
"matplotlib.pyplot.title",
"dlimage.mnist.MNISTLoader",
"random.randint",
"matplotlib.pyplot.show"
] | [((172, 194), 'dlimage.mnist.MNISTLoader', 'MNISTLoader', (['data_path'], {}), '(data_path)\n', (183, 194), False, 'from dlimage.mnist import MNISTLoader\n'), ((300, 334), 'matplotlib.pyplot.title', 'plt.title', (['label', "{'fontsize': 36}"], {}), "(label, {'fontsize': 36})\n", (309, 334), True, 'import matplotlib.pyp... |
import discord
import asyncio
import discord
import asyncio
from discord.ext import commands
import random
from discord.ext.commands.cooldowns import BucketType
import datetime
import psycopg2
from Principais.principais import bot, mydb, cursor
bot = bot
mydb = mydb
cursor = cursor
async def On_command_error(ctx,... | [
"Principais.principais.bot.get_guild",
"Principais.principais.cursor.execute",
"discord.utils.find",
"Principais.principais.bot.user.mentioned_in",
"Principais.principais.bot.process_commands",
"Principais.principais.mydb.commit",
"discord.Permissions",
"Principais.principais.cursor.fetchone",
"asyn... | [((794, 824), 'Principais.principais.bot.user.mentioned_in', 'bot.user.mentioned_in', (['message'], {}), '(message)\n', (815, 824), False, 'from Principais.principais import bot, mydb, cursor\n'), ((2928, 2951), 'Principais.principais.cursor.execute', 'cursor.execute', (['inserir'], {}), '(inserir)\n', (2942, 2951), Fa... |
# -*- coding: utf-8 -*-
from __future__ import unicode_literals
import os
from pyarrow import HadoopFileSystem, HdfsFile
from pyarrow.hdfs import connect
from plaso.tarzan.lib.hdfs import Hdfs
class PyArrowHdfs(Hdfs):
"""HDFS driver utilizing PyArrow library."""
fs = None # type: HadoopFileSystem
de... | [
"pyarrow.hdfs.connect"
] | [((836, 914), 'pyarrow.hdfs.connect', 'connect', ([], {'host': "(uri.hostname or 'default')", 'port': '(uri.port or 0)', 'user': 'uri.username'}), "(host=uri.hostname or 'default', port=uri.port or 0, user=uri.username)\n", (843, 914), False, 'from pyarrow.hdfs import connect\n')] |
import tensorflow as tf
from models import Classifier
class Model(Classifier):
name = 'cnnmnist'
def __init__(self, *args, **kwargs):
super().__init__(*args, **kwargs)
def build(self, x, y, *args, **kwargs):
self.nclasses = y._shape_as_list()[1]
logits = self.get_logits(x, self... | [
"tensorflow.layers.dense",
"tensorflow.variable_scope",
"tensorflow.reshape",
"tensorflow.layers.max_pooling2d",
"tensorflow.contrib.layers.l2_regularizer",
"tensorflow.nn.softmax_cross_entropy_with_logits_v2",
"tensorflow.layers.conv2d",
"tensorflow.constant",
"tensorflow.trainable_variables",
"t... | [((789, 811), 'tensorflow.reduce_mean', 'tf.reduce_mean', (['losses'], {}), '(losses)\n', (803, 811), True, 'import tensorflow as tf\n'), ((863, 888), 'tensorflow.identity', 'tf.identity', (['loss', '"""loss"""'], {}), "(loss, 'loss')\n", (874, 888), True, 'import tensorflow as tf\n'), ((999, 1047), 'tensorflow.constan... |