code
stringlengths
22
1.05M
apis
listlengths
1
3.31k
extract_api
stringlengths
75
3.25M
from collections import OrderedDict as odict import time import numpy as np from .json_encoder import JsonNumEncoder import os class Profiler: """This class provides a very simple yet light implementation of function profiling. It is very easy to use: >>> profiler.reset() >>> profiler.start("...
[ "os.path.join", "collections.OrderedDict", "time.time" ]
[((957, 964), 'collections.OrderedDict', 'odict', ([], {}), '()\n', (962, 964), True, 'from collections import OrderedDict as odict\n'), ((1279, 1290), 'time.time', 'time.time', ([], {}), '()\n', (1288, 1290), False, 'import time\n'), ((2513, 2520), 'collections.OrderedDict', 'odict', ([], {}), '()\n', (2518, 2520), Tr...
import math import hashlib import zlib from bitarray import bitarray from globals import G class BFsignature(): def __init__(self, total_chunks): self.total_chunks = total_chunks if self.total_chunks > 0: self.cal_m() #print("bf size = ",self.m) else: #lo...
[ "zlib.crc32", "hashlib.sha1", "bitarray.bitarray", "math.log" ]
[((635, 651), 'bitarray.bitarray', 'bitarray', (['self.m'], {}), '(self.m)\n', (643, 651), False, 'from bitarray import bitarray\n'), ((999, 1013), 'hashlib.sha1', 'hashlib.sha1', ([], {}), '()\n', (1011, 1013), False, 'import hashlib\n'), ((467, 478), 'math.log', 'math.log', (['(2)'], {}), '(2)\n', (475, 478), False, ...
""" The main window for the GUI. """ from msl.qt import ( QtWidgets, QtGui, Qt, Signal, convert, prompt, application, ) from msl.qt.utils import ( screen_geometry, drag_drop_paths, ) from msl.io import read from .line_edit import LineEdit from .. import logger from .equipment import...
[ "msl.qt.QtGui.QAction", "msl.qt.application", "msl.qt.prompt.critical", "msl.qt.convert.to_qicon", "msl.io.read", "msl.qt.Signal", "msl.qt.QtGui.QColor", "msl.qt.prompt.yes_no", "msl.qt.QtWidgets.QProgressBar", "msl.qt.utils.screen_geometry", "msl.qt.QtGui.QPalette", "msl.qt.QtWidgets.QDockWid...
[((561, 574), 'msl.qt.Signal', 'Signal', (['float'], {}), '(float)\n', (567, 574), False, 'from msl.qt import QtWidgets, QtGui, Qt, Signal, convert, prompt, application\n'), ((646, 654), 'msl.qt.Signal', 'Signal', ([], {}), '()\n', (652, 654), False, 'from msl.qt import QtWidgets, QtGui, Qt, Signal, convert, prompt, ap...
# -*- coding: utf-8 -*- """ Copyright (c) 2020 Ephreal under the MIT License. To view the license and requirements when distributing this software, please view the license at https://github.com/ephreal/catapi/LICENSE. """ from catapi.analysis import Analysis from tests import async_capable class Testanaylsis(async...
[ "catapi.analysis.Analysis" ]
[((596, 623), 'catapi.analysis.Analysis', 'Analysis', ([], {}), '(**analysis_params)\n', (604, 623), False, 'from catapi.analysis import Analysis\n')]
# -*- coding: utf-8 -*- # Authors: <NAME> <<EMAIL>> # # License: BSD (3-clause) import numpy as np from os import path as op from .utils import _check_pytables from .externals.six import string_types, text_type ############################################################################## # WRITE def write_hdf5(fn...
[ "os.path.isfile", "numpy.array", "numpy.atleast_1d" ]
[((780, 796), 'os.path.isfile', 'op.isfile', (['fname'], {}), '(fname)\n', (789, 796), True, 'from os import path as op\n'), ((4036, 4052), 'os.path.isfile', 'op.isfile', (['fname'], {}), '(fname)\n', (4045, 4052), True, 'from os import path as op\n'), ((5339, 5353), 'numpy.array', 'np.array', (['node'], {}), '(node)\n...
import requests import os import datetime from bs4 import BeautifulSoup from dotenv import load_dotenv from twilio.rest import Client from sendgrid import SendGridAPIClient from sendgrid.helpers.mail import Mail load_dotenv() TWILIO_ACCOUNT_SID = os.environ.get("TWILIO_ACCOUNT_SID", "OOPS, please specify env var call...
[ "sendgrid.helpers.mail.Mail", "os.environ.get", "sendgrid.SendGridAPIClient", "requests.get", "dotenv.load_dotenv", "bs4.BeautifulSoup", "datetime.datetime.now", "twilio.rest.Client" ]
[((213, 226), 'dotenv.load_dotenv', 'load_dotenv', ([], {}), '()\n', (224, 226), False, 'from dotenv import load_dotenv\n'), ((249, 349), 'os.environ.get', 'os.environ.get', (['"""TWILIO_ACCOUNT_SID"""', '"""OOPS, please specify env var called \'TWILIO_ACCOUNT_SID\'"""'], {}), '(\'TWILIO_ACCOUNT_SID\',\n "OOPS, plea...
#!/usr/bin/python3 import sys import numpy as np import numpysane as nps import os testdir = os.path.dirname(os.path.realpath(__file__)) # I import the LOCAL mrcal since that's what I'm testing sys.path[:0] = f"{testdir}/..", import mrcal import testutils import cv2 image = cv2.imread(f'{testdir}/data/figueroa-ove...
[ "testutils.finish", "os.path.realpath", "numpy.array", "testutils.confirm", "numpy.linalg.inv", "numpysane.dummy", "mrcal.match_feature", "testutils.confirm_equal", "cv2.imread", "mrcal.apply_homography", "numpy.arange" ]
[((280, 381), 'cv2.imread', 'cv2.imread', (['f"""{testdir}/data/figueroa-overpass-looking-S.0.downsampled.jpg"""', 'cv2.IMREAD_GRAYSCALE'], {}), "(f'{testdir}/data/figueroa-overpass-looking-S.0.downsampled.jpg',\n cv2.IMREAD_GRAYSCALE)\n", (290, 381), False, 'import cv2\n'), ((478, 568), 'numpy.array', 'np.array', (...
#!/usr/bin/env python """Solve the range minimum query problem.""" def read_numbers(number_file, query_file): """ Parameters ---------- number_file : str query_file : str Returns ------- tuple (numbers, queries) - both are lists """ with open(number_file) as f: ...
[ "argparse.ArgumentParser", "sys.exit" ]
[((1613, 1700), 'argparse.ArgumentParser', 'ArgumentParser', ([], {'description': '__doc__', 'formatter_class': 'ArgumentDefaultsHelpFormatter'}), '(description=__doc__, formatter_class=\n ArgumentDefaultsHelpFormatter)\n', (1627, 1700), False, 'from argparse import ArgumentDefaultsHelpFormatter, ArgumentParser\n'),...
from urllib.parse import urlparse class Card: AUTO_GENERATED_TEXT = 'Auto-created by TrelloNextActions' def __init__(self, trello, json): self._trello = trello self.id = json['id'] self.name = json['name'] self.board_id = json['idBoard'] self.description = json['desc'...
[ "urllib.parse.urlparse" ]
[((633, 659), 'urllib.parse.urlparse', 'urlparse', (['self.description'], {}), '(self.description)\n', (641, 659), False, 'from urllib.parse import urlparse\n')]
# -*- coding: utf-8 -*- from __future__ import unicode_literals from django.db import models, migrations class Migration(migrations.Migration): dependencies = [ ('forum', '0063_auto_20150611_1052'), ] operations = [ migrations.AddField( model_name='poll', name='a...
[ "django.db.models.DateTimeField" ]
[((342, 436), 'django.db.models.DateTimeField', 'models.DateTimeField', ([], {'default': "b'2000-01-01 00:00:00'", 'verbose_name': '"""aa"""', 'auto_now_add': '(True)'}), "(default=b'2000-01-01 00:00:00', verbose_name='aa',\n auto_now_add=True)\n", (362, 436), False, 'from django.db import models, migrations\n')]
# Copyright 2017 FUJITSU LIMITED # # Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writ...
[ "monasca_notification.config.parse_args", "oslo_config.cfg.CONF.set_default", "monasca_notification.conf.register_opts", "oslo_config.cfg.CONF.set_override", "unittest.mock.patch" ]
[((960, 998), 'unittest.mock.patch', 'mock.patch', (['"""monascastatsd.Connection"""'], {}), "('monascastatsd.Connection')\n", (970, 998), False, 'from unittest import mock\n'), ((1352, 1372), 'monasca_notification.conf.register_opts', 'conf.register_opts', ([], {}), '()\n', (1370, 1372), False, 'from monasca_notificat...
#!/usr/bin/python # -*- coding: utf-8 -*- # # Copyright 2017 Falseen # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by a...
[ "sys.path.pop", "sys.path.insert" ]
[((847, 862), 'sys.path.pop', 'sys.path.pop', (['(0)'], {}), '(0)\n', (859, 862), False, 'import sys\n'), ((930, 954), 'sys.path.insert', 'sys.path.insert', (['(0)', 'path'], {}), '(0, path)\n', (945, 954), False, 'import sys\n')]
# Copyright (c) 2019 <NAME> # # Permission is hereby granted, free of charge, to any person obtaining a copy # of this software and associated documentation files (the "Software"), to deal # in the Software without restriction, including without limitation the rights # to use, copy, modify, merge, publish, distribute, ...
[ "fla.ndfa.instance.Instance", "copy.copy" ]
[((3970, 4008), 'copy.copy', 'copy.copy', (['self.current_configurations'], {}), '(self.current_configurations)\n', (3979, 4008), False, 'import copy\n'), ((3798, 3847), 'fla.ndfa.instance.Instance', 'Instance', (['self', 'state', 'configuration.current_word'], {}), '(self, state, configuration.current_word)\n', (3806,...
# -*- coding: utf-8 -*- # python-holidays # --------------- # A fast, efficient Python library for generating country, province and state # specific sets of holidays on the fly. It aims to make determining whether a # specific date is a holiday as fast and flexible as possible. # # Author: ryanss <<EMAIL>> (c) ...
[ "dateutil.relativedelta.MO", "dateutil.relativedelta.relativedelta", "datetime.date", "holidays.holiday_base.HolidayBase.__init__" ]
[((813, 849), 'holidays.holiday_base.HolidayBase.__init__', 'HolidayBase.__init__', (['self'], {}), '(self, **kwargs)\n', (833, 849), False, 'from holidays.holiday_base import HolidayBase\n'), ((964, 982), 'datetime.date', 'date', (['year', 'JAN', '(1)'], {}), '(year, JAN, 1)\n', (968, 982), False, 'from datetime impor...
import sys import os sys.path.append(os.path.join(os.path.dirname(os.path.abspath(__file__)), "../../")) from api.fact.Fact_Market_Stock_Data import Fact_Market_Stock_Data from api.common.Saver import Saver from datetime import datetime,timedelta from api.common.Logs import Logger_process, Logger_process_error dda = F...
[ "datetime.datetime.now", "api.common.Logs.Logger_process.log", "os.path.abspath", "datetime.timedelta", "api.fact.Fact_Market_Stock_Data.Fact_Market_Stock_Data" ]
[((319, 343), 'api.fact.Fact_Market_Stock_Data.Fact_Market_Stock_Data', 'Fact_Market_Stock_Data', ([], {}), '()\n', (341, 343), False, 'from api.fact.Fact_Market_Stock_Data import Fact_Market_Stock_Data\n'), ((383, 448), 'api.common.Logs.Logger_process.log', 'Logger_process.log', (["('call Fact_Market_Stock_History,' +...
# Copyright 2020 Huawei Technologies Co., Ltd # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to...
[ "copy.deepcopy", "mindspore.ops.operations.Mul", "numpy.ones", "mindspore.nn.LogSoftmax", "mindspore.ops.operations.MatMul", "mindspore.ops.operations.Shape", "mindspore.ops.operations.OnesLike", "mindspore.ops.operations.Reshape", "mindspore.common.tensor.Tensor", "numpy.tril", "mindspore.ops.o...
[((1983, 1994), 'mindspore.ops.operations.Reshape', 'P.Reshape', ([], {}), '()\n', (1992, 1994), True, 'from mindspore.ops import operations as P\n'), ((2017, 2043), 'mindspore.ops.operations.MatMul', 'P.MatMul', ([], {'transpose_b': '(True)'}), '(transpose_b=True)\n', (2025, 2043), True, 'from mindspore.ops import ope...
#_*_coding=UTF-8_*_ from PIL import Image #编码 #内置 bin() 的替代,返回固定长度的二进制字符串 def constLenBin(int): binary = "0" * (8 - (len(bin(int)) - 2)) + bin(int).replace('0b','') #去掉 bin() 返回的二进制字符串中的 '0b',并在左边补足'0'直到字符串长度为8 return binary #取得一个 PIL 图像并且更改所有值为偶数(使最低有效位为0) def makeImageEven(image): pixels = list(image.g...
[ "PIL.Image.new", "PIL.Image.open" ]
[((483, 516), 'PIL.Image.new', 'Image.new', (['image.mode', 'image.size'], {}), '(image.mode, image.size)\n', (492, 516), False, 'from PIL import Image\n'), ((1279, 1320), 'PIL.Image.new', 'Image.new', (['evenImage.mode', 'evenImage.size'], {}), '(evenImage.mode, evenImage.size)\n', (1288, 1320), False, 'from PIL impor...
import lambda_function channel_id = 'CNCRFR7KP' team_id = 'TE4VDPM2L' user_id = 'UE6P69HFH' response_url = "no" channel_name = "local" command_list = ['init', 'init <https://charasheet.vampire-blood.net/2796211>', '1D100', 'result', 's', ...
[ "lambda_function.bootstrap" ]
[((1835, 1871), 'lambda_function.bootstrap', 'lambda_function.bootstrap', (['event', '{}'], {}), '(event, {})\n', (1860, 1871), False, 'import lambda_function\n')]
import django_filters from django.contrib.auth.models import User from .constants import * from .models import ObjectChange class ObjectChangeFilter(django_filters.FilterSet): q = django_filters.CharFilter(method="search", label="Search") time = django_filters.DateTimeFromToRangeFilter() action = django...
[ "django.contrib.auth.models.User.objects.all", "django_filters.DateTimeFromToRangeFilter", "django_filters.MultipleChoiceFilter", "django_filters.CharFilter" ]
[((188, 246), 'django_filters.CharFilter', 'django_filters.CharFilter', ([], {'method': '"""search"""', 'label': '"""Search"""'}), "(method='search', label='Search')\n", (213, 246), False, 'import django_filters\n'), ((258, 300), 'django_filters.DateTimeFromToRangeFilter', 'django_filters.DateTimeFromToRangeFilter', ([...
__copyright__ = "Copyright (c) 2020-2021 Jina AI Limited. All rights reserved." __license__ = "Apache-2.0" import pytest from jina import Document, DocumentArray from ...laser_encoder import LaserEncoder @pytest.fixture() def docs_generator(): return DocumentArray((Document(text='random text') for _ in range(30)...
[ "pytest.fixture", "jina.Document" ]
[((208, 224), 'pytest.fixture', 'pytest.fixture', ([], {}), '()\n', (222, 224), False, 'import pytest\n'), ((731, 764), 'jina.Document', 'Document', ([], {'id': '"""chunk11"""', 'text': 'text'}), "(id='chunk11', text=text)\n", (739, 764), False, 'from jina import Document, DocumentArray\n'), ((788, 821), 'jina.Document...
from typing import Tuple from ground.hints import (Point, Scalar, Segment) from hypothesis import given from locus.segmental import Tree from tests.utils import (is_segment, to_segment_point_distance) from . import strategies @given(strate...
[ "hypothesis.given", "tests.utils.to_segment_point_distance", "tests.utils.is_segment" ]
[((308, 343), 'hypothesis.given', 'given', (['strategies.trees_with_points'], {}), '(strategies.trees_with_points)\n', (313, 343), False, 'from hypothesis import given\n'), ((524, 559), 'hypothesis.given', 'given', (['strategies.trees_with_points'], {}), '(strategies.trees_with_points)\n', (529, 559), False, 'from hypo...
import gym import gym_pigchase_topdown env = gym.make('PigChase-v0') _ = env.reset() env.render() done = False while True: action = env.read_action() if action != -1: _, reward, done, _ = env.step(action) print(reward) env.render() if done: break
[ "gym.make" ]
[((46, 69), 'gym.make', 'gym.make', (['"""PigChase-v0"""'], {}), "('PigChase-v0')\n", (54, 69), False, 'import gym\n')]
from dataclasses import field, dataclass from typing import List import marshmallow_dataclass from flask_marshmallow import Schema @dataclass class UserResponse: UserName: str UserResponseSchema = marshmallow_dataclass.class_schema(UserResponse, base_schema=Schema) @dataclass class AllUsersResponse: User...
[ "dataclasses.field", "marshmallow_dataclass.class_schema" ]
[((206, 274), 'marshmallow_dataclass.class_schema', 'marshmallow_dataclass.class_schema', (['UserResponse'], {'base_schema': 'Schema'}), '(UserResponse, base_schema=Schema)\n', (240, 274), False, 'import marshmallow_dataclass\n'), ((399, 471), 'marshmallow_dataclass.class_schema', 'marshmallow_dataclass.class_schema', ...
#!/usr/bin/env python """ @package mi.dataset.parser @file marine-integrations/mi/dataset/parser/cspp_base.py @author <NAME> @brief Base Parser for a cspp dataset driver Release notes: initial release """ import copy import re import string from mi.core.log import get_logger from mi.core.common i...
[ "re.split", "mi.core.exceptions.ConfigurationException", "re.compile", "mi.core.exceptions.DatasetParserException", "mi.core.log.get_logger", "string.rstrip", "mi.core.exceptions.RecoverableSampleException", "copy.copy" ]
[((745, 757), 'mi.core.log.get_logger', 'get_logger', ([], {}), '()\n', (755, 757), False, 'from mi.core.log import get_logger\n'), ((893, 922), 're.compile', 're.compile', (['HEADER_PART_REGEX'], {}), '(HEADER_PART_REGEX)\n', (903, 922), False, 'import re\n'), ((1010, 1042), 're.compile', 're.compile', (['TIMESTAMP_LI...
import collections from typing import Counter class Solution: def getHint(self, secret: str, guess: str) -> str: bulls = cows = 0 seen = collections.Counter() # type: Counter[str] for s, g in zip(secret, guess): if s == g: bulls += 1 else: ...
[ "collections.Counter" ]
[((160, 181), 'collections.Counter', 'collections.Counter', ([], {}), '()\n', (179, 181), False, 'import collections\n')]
# Licensed to the Apache Software Foundation (ASF) under one # or more contributor license agreements. See the NOTICE file # distributed with this work for additional information # regarding copyright ownership. The ASF licenses this file # to you under the Apache License, Version 2.0 (the # "License"); you may not u...
[ "pydolphinscheduler.core.process_definition.ProcessDefinition", "pydolphinscheduler.tasks.shell.Shell", "pydolphinscheduler.tasks.dependent.DependentItem" ]
[((1580, 1653), 'pydolphinscheduler.core.process_definition.ProcessDefinition', 'ProcessDefinition', ([], {'name': '"""task_dependent_external"""', 'tenant': '"""tenant_exists"""'}), "(name='task_dependent_external', tenant='tenant_exists')\n", (1597, 1653), False, 'from pydolphinscheduler.core.process_definition impor...
from subprocess import call from sniffer.api import runnable, file_validator import os @file_validator def py_files(filename): is_match = (filename.endswith('.py') and not filename.endswith('_flymake.py') and not os.path.basename(filename).startswith('.')) return is_match @ru...
[ "os.path.basename", "subprocess.call" ]
[((425, 433), 'subprocess.call', 'call', (['fn'], {}), '(fn)\n', (429, 433), False, 'from subprocess import call\n'), ((251, 277), 'os.path.basename', 'os.path.basename', (['filename'], {}), '(filename)\n', (267, 277), False, 'import os\n')]
# import_export_ballotpedia/views_admin.py # Brought to you by We Vote. Be good. # -*- coding: UTF-8 -*- from .controllers import attach_ballotpedia_election_by_district_from_api, \ retrieve_ballot_items_from_polling_location, \ retrieve_ballotpedia_candidates_by_district_from_api, retrieve_ballotpedia_measure...
[ "django.shortcuts.render", "config.base.get_environment_variable", "admin_tools.views.redirect_to_sign_in_page", "datetime.date.today", "election.models.Election.objects.get", "django.core.urlresolvers.reverse", "wevote_functions.functions.positive_value_exists", "django.db.models.Q", "polling_locat...
[((1420, 1476), 'config.base.get_environment_variable', 'get_environment_variable', (['"""BALLOTPEDIA_API_CONTAINS_URL"""'], {}), "('BALLOTPEDIA_API_CONTAINS_URL')\n", (1444, 1476), False, 'from config.base import get_environment_variable\n'), ((3695, 3733), 'wevote_functions.functions.positive_value_exists', 'positive...
from flask import current_app as app from flask_migrate import Migrate, migrate, upgrade, stamp, current from alembic.migration import MigrationContext from sqlalchemy import create_engine from sqlalchemy.engine.url import make_url from sqlalchemy_utils import ( database_exists as database_exists_util, create_d...
[ "sqlalchemy_utils.drop_database", "sqlalchemy_utils.database_exists", "sqlalchemy.engine.url.make_url", "sqlalchemy_utils.create_database", "flask_migrate.Migrate", "alembic.migration.MigrationContext.configure", "flask.current_app.config.get" ]
[((434, 443), 'flask_migrate.Migrate', 'Migrate', ([], {}), '()\n', (441, 443), False, 'from flask_migrate import Migrate, migrate, upgrade, stamp, current\n'), ((479, 526), 'sqlalchemy.engine.url.make_url', 'make_url', (["app.config['SQLALCHEMY_DATABASE_URI']"], {}), "(app.config['SQLALCHEMY_DATABASE_URI'])\n", (487, ...
import os import inspect import pandas as pd class ICDConversionHelper: ICD9_TO_ICD10_CONVERSION_MAPPING = 'icd9_to_icd10.csv' ICD10_6_TO_ICD10_8_CONVERSION_MAPPING = 'icd10_6_to_icd10_8.csv' def __init__(self): curr_dir = os.path.dirname(os.path.abspath(inspect.stack()[0][1])) self.ICD9_...
[ "inspect.stack", "pandas.read_csv" ]
[((331, 398), 'pandas.read_csv', 'pd.read_csv', (["(curr_dir + '/' + self.ICD9_TO_ICD10_CONVERSION_MAPPING)"], {}), "(curr_dir + '/' + self.ICD9_TO_ICD10_CONVERSION_MAPPING)\n", (342, 398), True, 'import pandas as pd\n'), ((433, 505), 'pandas.read_csv', 'pd.read_csv', (["(curr_dir + '/' + self.ICD10_6_TO_ICD10_8_CONVER...
import socket import sys import threading import elgamal numBits = 32 keyOne = elgamal.generate_keys(iNumBits=numBits, iConfidence=1000) privKeyB = keyOne["privateKey"] pubKeyB = keyOne["publicKey"] print(f"Public Key: {pubKeyB.p} {pubKeyB.g} {pubKeyB.h}") # ----------------------------------------------------------...
[ "elgamal.generate_keys", "socket.socket", "elgamal.PublicKey", "elgamal.encrypt", "threading.Thread", "elgamal.decrypt" ]
[((81, 138), 'elgamal.generate_keys', 'elgamal.generate_keys', ([], {'iNumBits': 'numBits', 'iConfidence': '(1000)'}), '(iNumBits=numBits, iConfidence=1000)\n', (102, 138), False, 'import elgamal\n'), ((454, 502), 'socket.socket', 'socket.socket', (['socket.AF_INET', 'socket.SOCK_DGRAM'], {}), '(socket.AF_INET, socket....
from django import forms from django.contrib.auth.forms import UserCreationForm from django.contrib.auth import get_user_model from sistema.mail import send_mail_template from sistema.utils import generate_hash_key from usuario.models import PasswordReset User = get_user_model() class PasswordResetForm(forms.Form): ...
[ "django.contrib.auth.get_user_model", "sistema.utils.generate_hash_key", "django.forms.ValidationError", "usuario.models.PasswordReset", "django.forms.EmailField", "sistema.mail.send_mail_template" ]
[((264, 280), 'django.contrib.auth.get_user_model', 'get_user_model', ([], {}), '()\n', (278, 280), False, 'from django.contrib.auth import get_user_model\n'), ((329, 361), 'django.forms.EmailField', 'forms.EmailField', ([], {'label': '"""E-mail"""'}), "(label='E-mail')\n", (345, 361), False, 'from django import forms\...
import os import sys sys.path.append('.') import cv2 import numpy as np import time from Utilities import drawRegion, drawBox, col_rgb, CVConstants try: import pyMTF mtf_available = 1 except ImportError as e: print('MTF unavailable: {}'.format(e)) mtf_available = 0 from siamfc.SiamFC import SiamFC...
[ "cv2.TrackerGOTURN_create", "siamfc.SiamFC.SiamFC", "time.clock", "cv2.TrackerKCF_create", "cv2.imshow", "numpy.array", "siamfc.SiamFC.SiamFCParams", "sys.path.append", "cv2.TrackerMedianFlow_create", "cv2.__version__.split", "os.path.exists", "pyMTF.getRegion", "DaSiamRPN.DaSiamRPN.DaSiamRP...
[((22, 42), 'sys.path.append', 'sys.path.append', (['"""."""'], {}), "('.')\n", (37, 42), False, 'import sys\n'), ((1618, 1632), 'siamfc.SiamFC.SiamFCParams', 'SiamFCParams', ([], {}), '()\n', (1630, 1632), False, 'from siamfc.SiamFC import SiamFC, SiamFCParams\n'), ((1658, 1674), 'SiamMask.SiamMask.SiamMaskParams', 'S...
from abc import abstractmethod from typing import Optional, Tuple from pylidar_slam.common.modules import _with_cv2 if _with_cv2: import cv2 import numpy as np from omegaconf import DictConfig from pylidar_slam.common.utils import check_sizes, assert_debug class ImageBased2DRegistration: ...
[ "numpy.clip", "cv2.BFMatcher", "numpy.eye", "numpy.ones", "cv2.findHomography", "numpy.round", "pylidar_slam.common.utils.assert_debug", "cv2.AKAZE_create", "numpy.argsort", "numpy.array", "cv2.ORB_create", "pylidar_slam.common.utils.check_sizes", "numpy.linalg.svd", "matplotlib.cm.get_cma...
[((619, 661), 'pylidar_slam.common.utils.assert_debug', 'assert_debug', (["(features in ['orb', 'akaze'])"], {}), "(features in ['orb', 'akaze'])\n", (631, 661), False, 'from pylidar_slam.common.utils import check_sizes, assert_debug\n'), ((863, 906), 'cv2.BFMatcher', 'cv2.BFMatcher', (['cv2.NORM_L2'], {'crossCheck': '...
# Generated by Django 3.1 on 2021-01-26 08:00 from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('tms', '0001_initial'), ] operations = [ migrations.AlterField( model_name='ts2pstransfer', name='sub_type', ...
[ "django.db.models.IntegerField" ]
[((328, 468), 'django.db.models.IntegerField', 'models.IntegerField', ([], {'choices': "[(4, 'TS to PS (Single Degree)'), (5, 'PS-TS to PS-PS (Dual Degree)'), (6,\n 'TS-TS to TS-PS (Dual Degree)')]"}), "(choices=[(4, 'TS to PS (Single Degree)'), (5,\n 'PS-TS to PS-PS (Dual Degree)'), (6, 'TS-TS to TS-PS (Dual Deg...
import ray import torch from collections import namedtuple, defaultdict from itertools import count, chain from flatland.envs.agent_utils import RailAgentStatus from agent.PPO.PPORollout import PPORollout, PPOTransition class PPORunner(): def _select_actions(self, state, done): valid_handles = list() ...
[ "torch.tensor", "torch.stack", "collections.defaultdict", "torch.zeros" ]
[((2958, 2974), 'collections.defaultdict', 'defaultdict', (['int'], {}), '(int)\n', (2969, 2974), False, 'from collections import namedtuple, defaultdict\n'), ((3076, 3099), 'collections.defaultdict', 'defaultdict', (['PPORollout'], {}), '(PPORollout)\n', (3087, 3099), False, 'from collections import namedtuple, defaul...
import unittest from app.models import User class userModelTest(unittest.TestCase): def setUp(self): ''' function that creates instance of User class ''' self.new_user = User(password='<PASSWORD>') def test_set_password(self): ''' test case that ascertains tha...
[ "app.models.User" ]
[((209, 236), 'app.models.User', 'User', ([], {'password': '"""<PASSWORD>"""'}), "(password='<PASSWORD>')\n", (213, 236), False, 'from app.models import User\n')]
#!/usr/bin/env python3 # Copyright 2021 iRobot Corporation. All Rights Reserved. # @author <NAME> (<EMAIL>) # # Launch Create(R) 3 nodes from ament_index_python.packages import get_package_share_directory from launch import LaunchDescription from launch.actions import DeclareLaunchArgument, IncludeLaunchDescription fr...
[ "launch.substitutions.LaunchConfiguration", "ament_index_python.packages.get_package_share_directory", "launch.LaunchDescription", "launch.actions.DeclareLaunchArgument", "launch_ros.actions.Node", "launch.launch_description_sources.PythonLaunchDescriptionSource", "launch.substitutions.PathJoinSubstitut...
[((580, 718), 'launch.actions.DeclareLaunchArgument', 'DeclareLaunchArgument', (['"""gazebo"""'], {'default_value': '"""classic"""', 'choices': "['classic', 'ignition']", 'description': '"""Which gazebo simulator to use"""'}), "('gazebo', default_value='classic', choices=['classic',\n 'ignition'], description='Which...
import json from topojson.core.extract import Extract from shapely import geometry import geopandas import geojson # extract copies coordinates sequentially into a buffer def test_extract_linestring(): data = { "foo": {"type": "LineString", "coordinates": [[0, 0], [1, 0], [2, 0]]}, "bar": {"type":...
[ "shapefile.Reader", "geopandas.read_file", "shapely.geometry.MultiLineString", "topojson.core.extract.Extract", "shapely.geometry.Polygon", "shapely.geometry.LineString", "geojson.load" ]
[((8484, 8542), 'shapefile.Reader', 'shapefile.Reader', (['"""tests/files_shapefile/southamerica.shp"""'], {}), "('tests/files_shapefile/southamerica.shp')\n", (8500, 8542), False, 'import shapefile\n'), ((9103, 9161), 'shapely.geometry.Polygon', 'geometry.Polygon', (['[[0, 0], [1, 0], [1, 1], [0, 1], [0, 0]]'], {}), '...
""" tricks: 1.torch-optimizer:实现了最新的一些优化器. 2.numba:import numba as nb,纯python或numpy加速,加@nb.njit或@nb.jit(nopython=True) 3.swifter:df.apply()→·df.swifter.apply(),加速pandas 4.cupy:1000万以上数据更快 5.modin:import modin.pandas as mdpd,用mdpd代替pd即可,加速pandas,加载数据和查询数据更快,统计方法pandas更快 """ import os import sys import argparse import ti...
[ "torch.cuda.manual_seed_all", "torch.manual_seed", "torch.utils.tensorboard.SummaryWriter", "numpy.mean", "torch.nn.CrossEntropyLoss", "argparse.ArgumentParser", "data.custom_dataset.MyDataset", "torch.load", "wandb.init", "wandb.watch", "torch.cuda.is_available", "numpy.random.seed", "torch...
[((729, 740), 'time.time', 'time.time', ([], {}), '()\n', (738, 740), False, 'import time\n'), ((745, 769), 'torch.cuda.empty_cache', 'torch.cuda.empty_cache', ([], {}), '()\n', (767, 769), False, 'import torch\n'), ((1817, 1879), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {'description': '"""Train a ne...
# -*- coding:utf-8 -*- # Create by 27 # @Time : 2020/2/21 22:45 __author__ = '27' ''' 使用python发送http请求的两个方法: 1、urllib 2、requests (推荐) 小原则: 对于外部获取的数据,如果出现异常,最好不要抛出来,而是应该默认值处理。 ''' import requests class HTTP: # python3中写与不写(object)没有区别 @staticmethod def get(url, return_json=True): r = requests.get(url...
[ "requests.get" ]
[((304, 321), 'requests.get', 'requests.get', (['url'], {}), '(url)\n', (316, 321), False, 'import requests\n')]
import Crypto.Util.number as cun import Crypto.Random.random as crr import Crypto.Util.Padding as cup from Crypto.Cipher import AES import os import hashlib class DiffieHellman: def __init__(self, p: int): self.p = p self.g = 8 self.private_key = crr.getrandbits(128) def public_key(se...
[ "os.urandom", "Crypto.Util.number.size", "Crypto.Util.number.long_to_bytes", "Crypto.Cipher.AES.new", "Crypto.Random.random.getrandbits", "Crypto.Util.number.isPrime" ]
[((1401, 1431), 'Crypto.Cipher.AES.new', 'AES.new', (['aes_key', 'AES.MODE_ECB'], {}), '(aes_key, AES.MODE_ECB)\n', (1408, 1431), False, 'from Crypto.Cipher import AES\n'), ((277, 297), 'Crypto.Random.random.getrandbits', 'crr.getrandbits', (['(128)'], {}), '(128)\n', (292, 297), True, 'import Crypto.Random.random as c...
import os import json import argparse from tqdm import tqdm def extract_ingredients(result_dict, all_ingredient_dict): for method, result in result_dict.items(): for recipe_id, annotation in result.items(): ingredient_list = annotation["ingredients"] sentences = annotation["sentence...
[ "json.load", "os.path.join", "argparse.ArgumentParser" ]
[((3001, 3026), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {}), '()\n', (3024, 3026), False, 'import argparse\n'), ((3453, 3465), 'json.load', 'json.load', (['f'], {}), '(f)\n', (3462, 3465), False, 'import json\n'), ((2746, 2758), 'json.load', 'json.load', (['f'], {}), '(f)\n', (2755, 2758), False, 'im...
""" This module defines custom converters for routing. These converters streamline resource code by performing the process of getting the model instance from the database before the view method is called. The converter converts a resource's slug to the corresponding model instance, which is then placed into the view me...
[ "app.models.Observer.query.filter_by", "app.models.Map.query.filter_by" ]
[((1159, 1193), 'app.models.Map.query.filter_by', 'Map.query.filter_by', ([], {'slug': 'map_slug'}), '(slug=map_slug)\n', (1178, 1193), False, 'from app.models import Map, Observer, Obstacle\n'), ((1650, 1689), 'app.models.Observer.query.filter_by', 'Observer.query.filter_by', ([], {'slug': 'obs_slug'}), '(slug=obs_slu...
# Licensed under the GPL: https://www.gnu.org/licenses/old-licenses/gpl-2.0.html # For details: https://github.com/PyCQA/pylint/blob/master/LICENSE from glob import glob from os.path import basename, join, splitext from pylint.testutils.constants import SYS_VERS_STR def _get_tests_info(input_dir, msg_dir, prefix, s...
[ "os.path.splitext", "os.path.join", "os.path.basename" ]
[((820, 858), 'os.path.join', 'join', (['input_dir', "(prefix + '*' + suffix)"], {}), "(input_dir, prefix + '*' + suffix)\n", (824, 858), False, 'from os.path import basename, join, splitext\n'), ((878, 893), 'os.path.basename', 'basename', (['fname'], {}), '(fname)\n', (886, 893), False, 'from os.path import basename,...
from aiogram import types from .dataset import PHOTO photo = types.PhotoSize(**PHOTO) def test_export(): exported = photo.to_python() assert isinstance(exported, dict) assert exported == PHOTO def test_file_id(): assert isinstance(photo.file_id, str) assert photo.file_id == PHOTO['file_id'] d...
[ "aiogram.types.PhotoSize" ]
[((62, 86), 'aiogram.types.PhotoSize', 'types.PhotoSize', ([], {}), '(**PHOTO)\n', (77, 86), False, 'from aiogram import types\n')]
from Models.EAE_models.ensemblance_interface import EAE_interface from Models.EAE_models.EncoderCNN import EncoderCNN from Models.EAE_models.DecoderCNN import DecoderCNN import tensorflow as tf class Vanilla_CNN_ensemblance(EAE_interface): def __init__(self, filters, dim, p_m, **kwargs): super(Vanilla_CN...
[ "Models.EAE_models.DecoderCNN.DecoderCNN", "Models.EAE_models.EncoderCNN.EncoderCNN", "tensorflow.keras.layers.Dense" ]
[((384, 417), 'tensorflow.keras.layers.Dense', 'tf.keras.layers.Dense', (['self.z_dim'], {}), '(self.z_dim)\n', (405, 417), True, 'import tensorflow as tf\n'), ((437, 470), 'tensorflow.keras.layers.Dense', 'tf.keras.layers.Dense', (['(4 * 4 * 32)'], {}), '(4 * 4 * 32)\n', (458, 470), True, 'import tensorflow as tf\n'),...
from collections import Counter from analysis.computation import utils def frequency(ambitus_list): freq = Counter(ambitus_list) r = [['Ambitus', 'Pieces']] for k, v in sorted(freq.items()): r.append([k, v]) return r def frequency_pie(ambitus_list): r = utils.aux_pie_chart(Counter(ambit...
[ "analysis.computation.utils.boxplot", "collections.Counter", "analysis.computation.utils.aux_basic_stats", "analysis.computation.utils.histogram", "analysis.computation.utils.distribution", "analysis.computation.utils.get_music_data_attrib" ]
[((113, 134), 'collections.Counter', 'Counter', (['ambitus_list'], {}), '(ambitus_list)\n', (120, 134), False, 'from collections import Counter\n'), ((431, 493), 'analysis.computation.utils.get_music_data_attrib', 'utils.get_music_data_attrib', (['compositions', '"""ambitus"""', '"""append"""'], {}), "(compositions, 'a...
"""Test harness for Image Sequence References.""" import unittest import sys import opentimelineio as otio import opentimelineio.test_utils as otio_test_utils IS_PYTHON_2 = (sys.version_info < (3, 0)) class ImageSequenceReferenceTests( unittest.TestCase, otio_test_utils.OTIOAssertions ): def test_create(se...
[ "opentimelineio.adapters.otio_json.read_from_string", "opentimelineio.opentime.RationalTime", "opentimelineio.schema.ImageSequenceReference", "opentimelineio.schema.V2d", "opentimelineio.opentime.from_timecode", "unittest.skipIf", "opentimelineio.adapters.otio_json.write_to_string", "copy.deepcopy", ...
[((1688, 1762), 'unittest.skipIf', 'unittest.skipIf', (['IS_PYTHON_2', '"""unicode strings do funny things in python2"""'], {}), "(IS_PYTHON_2, 'unicode strings do funny things in python2')\n", (1703, 1762), False, 'import unittest\n'), ((2914, 2988), 'unittest.skipIf', 'unittest.skipIf', (['IS_PYTHON_2', '"""unicode s...
# # Copyright 2013 Red Hat, Inc # # Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agr...
[ "warnings.warn", "options.build_url" ]
[((7653, 7679), 'options.build_url', 'options.build_url', (['path', 'q'], {}), '(path, q)\n', (7670, 7679), False, 'import options\n'), ((8002, 8036), 'options.build_url', 'options.build_url', (['path', 'q', 'params'], {}), '(path, q, params)\n', (8019, 8036), False, 'import options\n'), ((3360, 3445), 'warnings.warn',...
from yo_fluq_ds__tests.common import * import numpy as np class MiscMethodsTests(TestCase): def test_pairwise(self): result = Query.args(1,2,3).feed(fluq.pairwise()).to_list() self.assertListEqual([(1,2),(2,3)],result) def test_strjoin(self): result = Query.args(1,2,3).feed(fluq.strjoi...
[ "numpy.random.RandomState" ]
[((819, 843), 'numpy.random.RandomState', 'np.random.RandomState', (['(1)'], {}), '(1)\n', (840, 843), True, 'import numpy as np\n')]
from direct.directnotify import DirectNotifyGlobal from direct.distributed.DistributedObjectUD import DistributedObjectUD class AwardManagerUD(DistributedObjectUD): notify = DirectNotifyGlobal.directNotify.newCategory('AwardManagerUD')
[ "direct.directnotify.DirectNotifyGlobal.directNotify.newCategory" ]
[((179, 240), 'direct.directnotify.DirectNotifyGlobal.directNotify.newCategory', 'DirectNotifyGlobal.directNotify.newCategory', (['"""AwardManagerUD"""'], {}), "('AwardManagerUD')\n", (222, 240), False, 'from direct.directnotify import DirectNotifyGlobal\n')]
import numpy as np import torch import torch.nn as nn import torch.nn.functional as F import math, copy, time from torch.autograd import Variable from utils import outputActivation import pdb # Customizations # - DONE Embeddings: linear transform d_feats -> d_model features # - DONE Generator # - DONE Batching # D...
[ "torch.nn.Dropout", "torch.sin", "math.sqrt", "torch.from_numpy", "utils.outputActivation", "math.log", "numpy.array", "torch.cos", "torch.cuda.is_available", "torch.squeeze", "copy.deepcopy", "torch.nn.init.xavier_uniform", "copy.copy", "torch.repeat_interleave", "torch.nn.functional.so...
[((4461, 4486), 'torch.nn.functional.softmax', 'F.softmax', (['scores'], {'dim': '(-1)'}), '(scores, dim=-1)\n', (4470, 4486), True, 'import torch.nn.functional as F\n'), ((735, 753), 'copy.copy', 'copy.copy', (['d_model'], {}), '(d_model)\n', (744, 753), False, 'import math, copy, time\n'), ((2080, 2115), 'torch.nn.Li...
import pytest from multi_bracket_validation import multi_bracket_validation def test_mbv_true_case_simple(): """test function on balanced str""" assert multi_bracket_validation('[{()}]') == True def test_mbv_true_case_empty_str(): """test function with empty string""" assert multi_bracket_validation('...
[ "multi_bracket_validation.multi_bracket_validation" ]
[((161, 195), 'multi_bracket_validation.multi_bracket_validation', 'multi_bracket_validation', (['"""[{()}]"""'], {}), "('[{()}]')\n", (185, 195), False, 'from multi_bracket_validation import multi_bracket_validation\n'), ((294, 322), 'multi_bracket_validation.multi_bracket_validation', 'multi_bracket_validation', (['"...
# -*- coding: utf-8 -*- # Part of Odoo. See LICENSE file for full copyright and licensing details. import ir import workflow import module import res import report import tests def post_init(cr, registry): """Rewrite ICP's to force groups""" from odoo import api, SUPERUSER_ID from odoo.addons.base.ir.ir_c...
[ "odoo.api.Environment", "odoo.addons.base.ir.ir_config_parameter._default_parameters.iteritems" ]
[((374, 411), 'odoo.api.Environment', 'api.Environment', (['cr', 'SUPERUSER_ID', '{}'], {}), '(cr, SUPERUSER_ID, {})\n', (389, 411), False, 'from odoo import api, SUPERUSER_ID\n'), ((470, 501), 'odoo.addons.base.ir.ir_config_parameter._default_parameters.iteritems', '_default_parameters.iteritems', ([], {}), '()\n', (4...
""" =================================== Toy Glucose Phosphorylation Process =================================== This is a toy example referenced in the documentation. """ from vivarium.core.process import Process from vivarium.core.composition import simulate_process from vivarium.plots.simulation_output import plot_...
[ "vivarium.core.composition.simulate_process", "vivarium.plots.simulation_output.plot_simulation_output" ]
[((3674, 3712), 'vivarium.core.composition.simulate_process', 'simulate_process', (['my_process', 'settings'], {}), '(my_process, settings)\n', (3690, 3712), False, 'from vivarium.core.composition import simulate_process\n'), ((3717, 3761), 'vivarium.plots.simulation_output.plot_simulation_output', 'plot_simulation_out...
# -*- coding: utf-8 -*- from __future__ import print_function import datetime import collections from midas.compat import unittest class PigSchemaToPyStructTests(unittest.TestCase): def _get_target(self): from midas.pig_schema import pig_schema_to_py_struct return pig_schema_to_py_struct ...
[ "midas.pig_schema.make_serializer", "collections.namedtuple", "midas.pig_schema.make_parser", "midas.pig_schema.pig_schema_to_py_struct", "midas.compat.unittest.main", "datetime.date", "midas.pig_schema.make_parser_from_schema" ]
[((7811, 7826), 'midas.compat.unittest.main', 'unittest.main', ([], {}), '()\n', (7824, 7826), False, 'from midas.compat import unittest\n'), ((2594, 2618), 'midas.pig_schema.make_parser', 'make_parser', (['"""chararray"""'], {}), "('chararray')\n", (2605, 2618), False, 'from midas.pig_schema import make_parser\n'), ((...
#!/usr/bin/env python # -*- coding: utf-8 -*- """General utilities (like verbose output decroator).""" import base64 import os import sys from termcolor import colored def verbose(in_progress: str, success: str): """Print message to stdout/stderr using decorator syntax. Parameters ---------- in_pro...
[ "termcolor.colored", "sys.exit" ]
[((1399, 1438), 'termcolor.colored', 'colored', (['message', '"""red"""'], {'attrs': "['bold']"}), "(message, 'red', attrs=['bold'])\n", (1406, 1438), False, 'from termcolor import colored\n'), ((1820, 1859), 'termcolor.colored', 'colored', (['message', 'color'], {'attrs': "['bold']"}), "(message, color, attrs=['bold']...
#!/usr/bin python3 # Write and send emails from the Command Line. import sys import time from selenium import webdriver from selenium.webdriver.common.keys import Keys recipient = sys.argv[1] subject = sys.argv[2] message = sys.argv[3] # Username Login browser = webdriver.Firefox() browser.get('https://accounts.goo...
[ "selenium.webdriver.Firefox", "time.sleep" ]
[((267, 286), 'selenium.webdriver.Firefox', 'webdriver.Firefox', ([], {}), '()\n', (284, 286), False, 'from selenium import webdriver\n'), ((476, 489), 'time.sleep', 'time.sleep', (['(3)'], {}), '(3)\n', (486, 489), False, 'import time\n'), ((625, 638), 'time.sleep', 'time.sleep', (['(5)'], {}), '(5)\n', (635, 638), Fa...
from fluiddb.testing.basic import FluidinfoTestCase from fluiddb.util.constant import Constant, ConstantEnum class ConstantTest(FluidinfoTestCase): def testInstantiate(self): """A L{Constant} includes an ID and a name.""" constant = Constant(42, 'VALUE') self.assertEqual(42, constant.id) ...
[ "fluiddb.util.constant.ConstantEnum", "fluiddb.util.constant.Constant" ]
[((256, 277), 'fluiddb.util.constant.Constant', 'Constant', (['(42)', '"""VALUE"""'], {}), "(42, 'VALUE')\n", (264, 277), False, 'from fluiddb.util.constant import Constant, ConstantEnum\n'), ((515, 536), 'fluiddb.util.constant.Constant', 'Constant', (['(42)', '"""VALUE"""'], {}), "(42, 'VALUE')\n", (523, 536), False, ...
import sqlite3 class Database: def __init__(self): self.connection = sqlite3.connect('./database.db') self.cursor = self.connection.cursor() def insert_item(self, item): cur = self.cursor.execute( "insert into posts (title, file, link) values (?, ?, ?)", (it...
[ "sqlite3.connect" ]
[((82, 114), 'sqlite3.connect', 'sqlite3.connect', (['"""./database.db"""'], {}), "('./database.db')\n", (97, 114), False, 'import sqlite3\n')]
#!/usr/bin/env python # -*- coding: utf-8 -*- ############################################################################### # Copyright 2015 Kitware Inc. # # Licensed under the Apache License, Version 2.0 ( the 'License' ); # you may not use this file except in compliance with the License. # You may obtain a cop...
[ "cumulus.common.girder._get_profile", "girder.api.rest.getCurrentUser", "cumulus.constants.VolumeType.is_valid_type", "girder.api.rest.loadmodel", "bson.objectid.ObjectId", "cumulus.common.girder.get_task_token", "girder.api.describe.Description", "cumulus.common.girder.send_status_notification", "g...
[((2908, 2975), 'girder.api.rest.loadmodel', 'loadmodel', ([], {'model': '"""volume"""', 'plugin': '"""cumulus"""', 'level': 'AccessType.WRITE'}), "(model='volume', plugin='cumulus', level=AccessType.WRITE)\n", (2917, 2975), False, 'from girder.api.rest import loadmodel\n'), ((4799, 5363), 'girder.api.docs.addModel', '...
import numpy as np import time import ray import ray.autoscaler.sdk from ray._private.test_utils import Semaphore import json import os from time import perf_counter from tqdm import trange, tqdm MAX_ARGS = 10000 MAX_RETURNS = 3000 MAX_RAY_GET_ARGS = 10000 MAX_QUEUED_TASKS = 1_000_000 MAX_RAY_GET_SIZE = 100 * 2**30 ...
[ "ray.init", "ray.cluster_resources", "ray.get", "numpy.arange", "tqdm.tqdm", "time.perf_counter", "time.sleep", "ray._private.test_utils.Semaphore.remote", "numpy.zeros", "ray.put", "ray.remote", "ray.available_resources", "tqdm.trange", "json.dump" ]
[((3347, 3371), 'ray.init', 'ray.init', ([], {'address': '"""auto"""'}), "(address='auto')\n", (3355, 3371), False, 'import ray\n'), ((3386, 3400), 'time.perf_counter', 'perf_counter', ([], {}), '()\n', (3398, 3400), False, 'from time import perf_counter\n'), ((3429, 3443), 'time.perf_counter', 'perf_counter', ([], {})...
import cv2 import numpy as np import matplotlib.pyplot as plt IMAGE = "b&w2.jpg" prototxt = "./Models/colorization_deploy_v2.prototxt.txt" model = "./Models/colorization_release_v2.caffemodel" points = "./Models/pts_in_hull.npy" image = "./input_images/"+IMAGE net = cv2.dnn.readNetFromCaffe(prototxt, mod...
[ "numpy.clip", "matplotlib.pyplot.imshow", "cv2.dnn.blobFromImage", "numpy.full", "cv2.dnn.readNetFromCaffe", "cv2.cvtColor", "numpy.concatenate", "cv2.split", "matplotlib.pyplot.axis", "cv2.resize", "numpy.load", "cv2.imread", "matplotlib.pyplot.show" ]
[((282, 323), 'cv2.dnn.readNetFromCaffe', 'cv2.dnn.readNetFromCaffe', (['prototxt', 'model'], {}), '(prototxt, model)\n', (306, 323), False, 'import cv2\n'), ((331, 346), 'numpy.load', 'np.load', (['points'], {}), '(points)\n', (338, 346), True, 'import numpy as np\n'), ((610, 627), 'cv2.imread', 'cv2.imread', (['image...
import pprint import sys import youtube_dl def main(): for url in sys.argv[1:]: with youtube_dl.YoutubeDL() as ydl: info_dict = ydl.extract_info(url, download=False) pprint.pprint(info_dict) if __name__ == '__main__': main()
[ "youtube_dl.YoutubeDL", "pprint.pprint" ]
[((100, 122), 'youtube_dl.YoutubeDL', 'youtube_dl.YoutubeDL', ([], {}), '()\n', (120, 122), False, 'import youtube_dl\n'), ((205, 229), 'pprint.pprint', 'pprint.pprint', (['info_dict'], {}), '(info_dict)\n', (218, 229), False, 'import pprint\n')]
import tensorflow as tf x = tf.placeholder(tf.float32, [None,3]) x_data = [[1,2,3],[4,5,6]] w = tf.Variable(tf.random_normal([3,2])) b = tf.Variable(tf.random_normal([2,1])) expr = tf.matmul(x,w) + b sess = tf.Session() sess.run(tf.global_variables_initializer()) print(x_data) print(sess.run(w)) print(sess.run(b...
[ "tensorflow.random_normal", "tensorflow.placeholder", "tensorflow.Session", "tensorflow.global_variables_initializer", "tensorflow.matmul" ]
[((29, 66), 'tensorflow.placeholder', 'tf.placeholder', (['tf.float32', '[None, 3]'], {}), '(tf.float32, [None, 3])\n', (43, 66), True, 'import tensorflow as tf\n'), ((212, 224), 'tensorflow.Session', 'tf.Session', ([], {}), '()\n', (222, 224), True, 'import tensorflow as tf\n'), ((111, 135), 'tensorflow.random_normal'...
#!/bin/env python import StringIO import unittest import naivetrack import naivetrack.utils class TestNaivetrackSummaries(unittest.TestCase): def setUp(self): # we will use a global chromosome list self.chromosome_list = [] self.file = StringIO.StringIO("""# # chr1\t2\tT\t2\t10\t50 \tG\t...
[ "StringIO.StringIO", "naivetrack.utils.aggregate_allele_information", "naivetrack.SummarizeEntries", "naivetrack.TrackParser" ]
[((268, 479), 'StringIO.StringIO', 'StringIO.StringIO', (['"""#\n#\nchr1\t2\tT\t2\t10\t50\n\tG\t4\t20\tsample_1, forward\n\tT\t6\t30\tsample_1, forward\nchr1\t20000\tT\t3\t40\t130\n\tA\t20\t60\tsample_1, forward\n\tA\t10\t30\tsample_1, reverse\n\tG\t10\t40\tsample_1, forward\n"""'], {}), '(\n """#\n#\nchr1\t2\tT\t2\...
import numpy as np from netCDF4 import Dataset from datetime import datetime from datetime import timedelta import os import sys import matplotlib.pyplot as plt from matplotlib import gridspec import matplotlib.colors as mcolors import matplotlib.patches as patches from matplotlib.colors import BoundaryNorm from to...
[ "tools_LT.setup_12p", "numpy.array", "datetime.timedelta", "numpy.arange", "datetime.datetime", "numpy.where", "netCDF4.Dataset", "matplotlib.colors.ListedColormap", "matplotlib.pyplot.close", "numpy.meshgrid", "matplotlib.pyplot.cla", "matplotlib.pyplot.cm.get_cmap", "matplotlib.pyplot.show...
[((17338, 17372), 'numpy.arange', 'np.arange', (['(DX * 0.5)', '(DX * XDIM)', 'DX'], {}), '(DX * 0.5, DX * XDIM, DX)\n', (17347, 17372), True, 'import numpy as np\n'), ((17375, 17409), 'numpy.arange', 'np.arange', (['(DY * 0.5)', '(DY * YDIM)', 'DY'], {}), '(DY * 0.5, DY * YDIM, DY)\n', (17384, 17409), True, 'import nu...
from django import forms from filament.models import * class FilamentForm(forms.Form): material = forms.ChoiceField( required=True, choices=FILAMENT_MATERIAL_CHOICES, widget=forms.Select(attrs={ 'class': 'form-control', 'form': 'filament_form', }), ...
[ "django.forms.ClearableFileInput", "django.forms.Select", "django.forms.NumberInput", "django.forms.URLInput", "django.forms.TextInput" ]
[((210, 280), 'django.forms.Select', 'forms.Select', ([], {'attrs': "{'class': 'form-control', 'form': 'filament_form'}"}), "(attrs={'class': 'form-control', 'form': 'filament_form'})\n", (222, 280), False, 'from django import forms\n'), ((399, 510), 'django.forms.NumberInput', 'forms.NumberInput', ([], {'attrs': "{'cl...
from django.shortcuts import render from django.http import HttpResponse # Create your views here. def hello_world(request): return HttpResponse("Hello world!!!")
[ "django.http.HttpResponse" ]
[((138, 168), 'django.http.HttpResponse', 'HttpResponse', (['"""Hello world!!!"""'], {}), "('Hello world!!!')\n", (150, 168), False, 'from django.http import HttpResponse\n')]
# Generated by Django 3.1 on 2020-08-13 02:38 from django.db import migrations, models import django.utils.timezone class Migration(migrations.Migration): dependencies = [ ("core", "0011_auto_20200214_1939"), ] operations = [ migrations.AlterField( model_name="note", ...
[ "django.db.models.DateTimeField" ]
[((356, 432), 'django.db.models.DateTimeField', 'models.DateTimeField', ([], {'default': 'django.utils.timezone.now', 'verbose_name': '"""Time"""'}), "(default=django.utils.timezone.now, verbose_name='Time')\n", (376, 432), False, 'from django.db import migrations, models\n')]
"""Module that holds tickets endpoints""" from datetime import date from flask_jwt_extended import jwt_required, get_jwt_identity from flask_restplus import Resource from sqlalchemy import func from api.utilities.helpers.return_value import return_value from api.utilities.messages.error import error_messages from api...
[ "flask_jwt_extended.get_jwt_identity", "datetime.date.today", "sqlalchemy.func.DATE", "main.api.route" ]
[((656, 701), 'main.api.route', 'api.route', (['"""/flights/<string:flight_id>/book"""'], {}), "('/flights/<string:flight_id>/book')\n", (665, 701), False, 'from main import api\n'), ((1895, 1916), 'main.api.route', 'api.route', (['"""/tickets"""'], {}), "('/tickets')\n", (1904, 1916), False, 'from main import api\n'),...
#! /usr/bin/python import sys, struct import xml.dom.minidom from lmcp import LMCPObject ## =============================================================================== ## Authors: AFRL/RQQA ## Organization: Air Force Research Laboratory, Aerospace Systems Directorate, Power and Control Division ## ## Copyright (...
[ "lmcp.LMCPObject.LMCPObject.unpackFromXMLNode", "lmcp.LMCPFactory.LMCPFactory", "lmcp.LMCPObject.LMCPObject.unpack", "lmcp.LMCPObject.LMCPObject.toString", "lmcp.LMCPObject.LMCPObject.toDictMembers", "struct.pack", "lmcp.LMCPObject.LMCPObject.pack", "lmcp.LMCPObject.LMCPObject.unpackFromDict", "lmcp...
[((2372, 2420), 'lmcp.LMCPObject.LMCPObject.unpack', 'LMCPObject.LMCPObject.unpack', (['self', 'buffer', '_pos'], {}), '(self, buffer, _pos)\n', (2400, 2420), False, 'from lmcp import LMCPObject\n'), ((3517, 3581), 'lmcp.LMCPObject.LMCPObject.unpackFromXMLNode', 'LMCPObject.LMCPObject.unpackFromXMLNode', (['self', 'el'...
from abc import abstractmethod import numpy as np from pymoo.core.population import Population # --------------------------------------------------------------------------------------------------------- # Survival # ----------------------------------------------------------------------------------------------------...
[ "numpy.median", "pymoo.core.population.Population", "numpy.where", "numpy.argsort", "pymoo.core.population.Population.merge" ]
[((2429, 2442), 'numpy.median', 'np.median', (['cv'], {}), '(cv)\n', (2438, 2442), True, 'import numpy as np\n'), ((2183, 2194), 'numpy.where', 'np.where', (['b'], {}), '(b)\n', (2191, 2194), True, 'import numpy as np\n'), ((2215, 2227), 'numpy.where', 'np.where', (['(~b)'], {}), '(~b)\n', (2223, 2227), True, 'import n...
from os import mkdir from baelfire.dependencies import AlwaysTrue from baelfire.dependencies import FileChanged from baelfire.dependencies import FileDoesNotExists from baelfire.dependencies import PidIsNotRunning from baelfire.dependencies import TaskRebuilded from baelfire.task import FileTask from baelfire.task imp...
[ "baelfire.dependencies.PidIsNotRunning", "baelfire.dependencies.FileDoesNotExists", "bdjango.dependency.MigrationsChanged", "baelfire.dependencies.AlwaysTrue", "os.mkdir", "baelfire.dependencies.FileChanged" ]
[((608, 626), 'os.mkdir', 'mkdir', (['self.output'], {}), '(self.output)\n', (613, 626), False, 'from os import mkdir\n'), ((1089, 1111), 'baelfire.dependencies.FileChanged', 'FileChanged', (['"""setuppy"""'], {}), "('setuppy')\n", (1100, 1111), False, 'from baelfire.dependencies import FileChanged\n'), ((1561, 1600), ...
""" Copyright (c) 2016, <NAME> .All rights reserved. Redistribution and use in source and binary forms, with or without modification, are permitted provided that the following conditions are met: 1. Redistributions of source code must retain the above copyright notice, this list of conditions and the foll...
[ "numpy.ones", "nibabel.load", "scipy.io.loadmat", "numpy.issubdtype", "ImgOperations.imgOp.applyPadding", "numpy.empty", "numpy.rint", "numpy.random.shuffle" ]
[((1511, 1534), 'nibabel.load', 'nib.load', (['imageFileName'], {}), '(imageFileName)\n', (1519, 1534), True, 'import nibabel as nib\n'), ((1987, 2013), 'scipy.io.loadmat', 'sio.loadmat', (['imageFileName'], {}), '(imageFileName)\n', (1998, 2013), True, 'import scipy.io as sio\n'), ((5928, 5951), 'numpy.random.shuffle'...
#!/usr/bin/env python3 import argparse import pandas as pd import os from tree import * import copy def main(): parser = argparse.ArgumentParser() parser.add_argument('--json', help='auspice json') parser.add_argument('--ids', help='Drop all ids in the list...
[ "argparse.ArgumentParser" ]
[((128, 153), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {}), '()\n', (151, 153), False, 'import argparse\n')]
# -*- coding: utf-8 -*- from __future__ import unicode_literals import flask_restplus as restplus class Foo(restplus.Resource): def get(self): return "data" class ErrorsTest(object): def test_accept_default_application_json(self, app, client): api = restplus.Api(app) api.add_resourc...
[ "flask_restplus.Api" ]
[((279, 296), 'flask_restplus.Api', 'restplus.Api', (['app'], {}), '(app)\n', (291, 296), True, 'import flask_restplus as restplus\n'), ((574, 591), 'flask_restplus.Api', 'restplus.Api', (['app'], {}), '(app)\n', (586, 591), True, 'import flask_restplus as restplus\n'), ((885, 926), 'flask_restplus.Api', 'restplus.Api'...
"""Fuzzy matching to a library of songs in RTTTL format. Takes an input string and matches against song title in a dictionary of available songs. Returns song title, RTTTL string, and match confidence metric. Fuzzy matching applied so a 'match' is guaranteed, however wayward the confidence. Dependencies: pip3 in...
[ "fuzzywuzzy.fuzz.token_set_ratio" ]
[((897, 936), 'fuzzywuzzy.fuzz.token_set_ratio', 'fuzz.token_set_ratio', (['input_string', 'key'], {}), '(input_string, key)\n', (917, 936), False, 'from fuzzywuzzy import fuzz\n')]
""" Test nqueens algorithm """ from modules.queens.algorithms.nqueens import NQueens from . import app # n solutions # 1 1 # 2 0 # 3 0 # 4 2 # 5 10 # 6 4 # 7 40 # 8 92 # 9 352 # 10 724 # 11 2,680 # 12 14,200 # 13 73,712 # 14 365,596 # 15 2,279,184 """Set the N size board""" n = 10 solutions = 724 def test_n_queens(...
[ "modules.queens.algorithms.nqueens.NQueens" ]
[((378, 395), 'modules.queens.algorithms.nqueens.NQueens', 'NQueens', (['n', '(False)'], {}), '(n, False)\n', (385, 395), False, 'from modules.queens.algorithms.nqueens import NQueens\n'), ((643, 659), 'modules.queens.algorithms.nqueens.NQueens', 'NQueens', (['n', '(True)'], {}), '(n, True)\n', (650, 659), False, 'from...
import pygame as pg from Data.images import * pg.init() def draw_minimap(surface, player): pg.draw.rect(surface, pg.Color("black"), (595, 0, 205, 205), 0) surface.blit(pg.transform.scale(player.level.background, (200, 200)), (600,0)) surface.blit(IMAGES['minimap_player'], \ ((600 + ((pla...
[ "pygame.Color", "pygame.init", "pygame.transform.scale" ]
[((46, 55), 'pygame.init', 'pg.init', ([], {}), '()\n', (53, 55), True, 'import pygame as pg\n'), ((119, 136), 'pygame.Color', 'pg.Color', (['"""black"""'], {}), "('black')\n", (127, 136), True, 'import pygame as pg\n'), ((178, 233), 'pygame.transform.scale', 'pg.transform.scale', (['player.level.background', '(200, 20...
from GUI import * import Backend as core app = None def view_command(): rows = core.view() app.listClientes.delete(0, END) for r in rows: app.listClientes.insert(END, r) def search_command(): app.listClientes.delete(0, END) rows = core.search(app.txtNome.get(), app.txtSobreno...
[ "Backend.view", "Backend.delete" ]
[((91, 102), 'Backend.view', 'core.view', ([], {}), '()\n', (100, 102), True, 'import Backend as core\n'), ((774, 789), 'Backend.delete', 'core.delete', (['id'], {}), '(id)\n', (785, 789), True, 'import Backend as core\n')]
import sys, os sys.path.append(os.pardir) from fluctana import * # HOW TO RUN # ./python3 check_hurst.py 10186 [15.9,16] ECEI_L1303-1305 shot = int(sys.argv[1]) trange = eval(sys.argv[2]) clist = sys.argv[3].split(',') # call fluctana A = FluctAna() # add data A.add_data(KstarEcei(shot=shot, clist=clist), trange=t...
[ "sys.path.append" ]
[((15, 41), 'sys.path.append', 'sys.path.append', (['os.pardir'], {}), '(os.pardir)\n', (30, 41), False, 'import sys, os\n')]
from typing import List from dataclasses import dataclass, field @dataclass() class DownloadParams: paths: List[str] output_folder: str s3_bucket: str = field(default="for-dvc")
[ "dataclasses.dataclass", "dataclasses.field" ]
[((69, 80), 'dataclasses.dataclass', 'dataclass', ([], {}), '()\n', (78, 80), False, 'from dataclasses import dataclass, field\n'), ((168, 192), 'dataclasses.field', 'field', ([], {'default': '"""for-dvc"""'}), "(default='for-dvc')\n", (173, 192), False, 'from dataclasses import dataclass, field\n')]
import datetime from asgiref.sync import sync_to_async from django.urls import reverse_lazy from django.utils.timezone import now from django.shortcuts import render, reverse, get_object_or_404, get_list_or_404, HttpResponse from django.views.generic import ListView, DetailView, TemplateView, CreateView from django.co...
[ "django.shortcuts.render", "Laelia.pdf.engine.PdfCreator", "Laelia.apps.care.models.Visit.objects.get", "django.contrib.contenttypes.models.ContentType.objects.get", "Laelia.apps.care.forms.VisitModelForm", "django.shortcuts.HttpResponse", "django.shortcuts.get_object_or_404", "django.shortcuts.get_li...
[((944, 1007), 'django.contrib.contenttypes.models.ContentType.objects.get', 'ContentType.objects.get', ([], {'app_label': '"""base"""', 'model': '"""professional"""'}), "(app_label='base', model='professional')\n", (967, 1007), False, 'from django.contrib.contenttypes.models import ContentType\n'), ((17686, 17730), 'd...
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """Collection of helper methods for rdm module @author: baihan """ import numpy as np from scipy.spatial.distance import squareform def batch_to_vectors(x): """converts a *stack* of RDMs in vector or matrix form into vector form Args: x: stack of RDMs ...
[ "scipy.spatial.distance.squareform", "numpy.sqrt", "numpy.unique", "numpy.array", "numpy.ndarray", "numpy.arange" ]
[((2313, 2338), 'numpy.unique', 'np.unique', (['pattern_select'], {}), '(pattern_select)\n', (2322, 2338), True, 'import numpy as np\n'), ((1457, 1492), 'numpy.ndarray', 'np.ndarray', (['(n_rdm, n_cond, n_cond)'], {}), '((n_rdm, n_cond, n_cond))\n', (1467, 1492), True, 'import numpy as np\n'), ((1512, 1528), 'numpy.ara...
from passlib.hash import ldap_salted_sha1 from passlib.hash import ldap_pbkdf2_sha256 from sshpubkeys import SSHKey import datetime class Usermgmt(object): def attrs(self): raise NotImplementedError def get_dict(self): return dict((key, value) for key, value in self.__dict__.items() ...
[ "passlib.hash.ldap_pbkdf2_sha256.hash", "passlib.hash.ldap_pbkdf2_sha256.verify", "passlib.hash.ldap_salted_sha1.identify", "datetime.datetime.now", "sshpubkeys.SSHKey", "passlib.hash.ldap_pbkdf2_sha256.identify", "passlib.hash.ldap_salted_sha1.verify" ]
[((4687, 4698), 'sshpubkeys.SSHKey', 'SSHKey', (['key'], {}), '(key)\n', (4693, 4698), False, 'from sshpubkeys import SSHKey\n'), ((4798, 4809), 'sshpubkeys.SSHKey', 'SSHKey', (['key'], {}), '(key)\n', (4804, 4809), False, 'from sshpubkeys import SSHKey\n'), ((4129, 4162), 'passlib.hash.ldap_pbkdf2_sha256.hash', 'ldap_...
#!/usr/bin/env python2 # -*- coding: utf-8 -*- """ Created on Fri Apr 28 17:57:57 2017 @author: pipolose """ from polyML.plot_overlay import plot_n_save_3plane import matplotlib.pyplot as plt from matplotlib import cm in_dict = {} in_dict['do_save'] = True # True # in_dict['formats_used'] = ['pdf', 'png'] in_dict[...
[ "polyML.plot_overlay.plot_n_save_3plane" ]
[((761, 788), 'polyML.plot_overlay.plot_n_save_3plane', 'plot_n_save_3plane', (['in_dict'], {}), '(in_dict)\n', (779, 788), False, 'from polyML.plot_overlay import plot_n_save_3plane\n'), ((1132, 1159), 'polyML.plot_overlay.plot_n_save_3plane', 'plot_n_save_3plane', (['in_dict'], {}), '(in_dict)\n', (1150, 1159), False...
from app import db ROLE_USER = 0 ROLE_ADMIN = 1 class User(db.Model): id = db.Column(db.Integer, primary_key = True) nickname = db.Column(db.String(64), index = True, unique = True) email = db.Column(db.String(120), index = True, unique = True) role = db.Column(db.SmallInteger, default = ROLE_USER) ...
[ "app.db.String", "app.db.Column" ]
[((81, 120), 'app.db.Column', 'db.Column', (['db.Integer'], {'primary_key': '(True)'}), '(db.Integer, primary_key=True)\n', (90, 120), False, 'from app import db\n'), ((270, 315), 'app.db.Column', 'db.Column', (['db.SmallInteger'], {'default': 'ROLE_USER'}), '(db.SmallInteger, default=ROLE_USER)\n', (279, 315), False, ...
from flask import Blueprint, render_template, request from wtforms import Form, IntegerField, validators from flask_table import Table, Col from app.helpers.requestid import requestid from app.modules.books.models import Book # pylint: disable=invalid-name frontend_blueprint = Blueprint('frontend', __name__, url_prefi...
[ "flask.render_template", "app.modules.books.models.Book.query.filter", "wtforms.validators.Optional", "flask_table.Col", "flask.Blueprint", "app.modules.books.models.Book.query.all" ]
[((279, 326), 'flask.Blueprint', 'Blueprint', (['"""frontend"""', '__name__'], {'url_prefix': '"""/"""'}), "('frontend', __name__, url_prefix='/')\n", (288, 326), False, 'from flask import Blueprint, render_template, request\n'), ((418, 427), 'flask_table.Col', 'Col', (['"""id"""'], {}), "('id')\n", (421, 427), False, ...
# Generated by Django 2.2.13 on 2020-06-12 19:50 from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('ebooks', '0106_auto_20200524_1227'), ('tooltips', '0006_auto_20200612_2140'), ] operations = [ migrations.RemoveField( ...
[ "django.db.migrations.RemoveField", "django.db.models.ManyToManyField" ]
[((284, 342), 'django.db.migrations.RemoveField', 'migrations.RemoveField', ([], {'model_name': '"""tooltip"""', 'name': '"""books"""'}), "(model_name='tooltip', name='books')\n", (306, 342), False, 'from django.db import migrations, models\n'), ((491, 552), 'django.db.models.ManyToManyField', 'models.ManyToManyField',...
# Copyright (c) Facebook, Inc. and its affiliates. # # This source code is licensed under the MIT license found in the # LICENSE file in the root directory of this source tree. # flake8: noqa from builtins import __test_sink, __test_source def sink(json): __test_sink(json) def test(): query = {"json": __te...
[ "builtins.__test_sink", "builtins.__test_source" ]
[((263, 280), 'builtins.__test_sink', '__test_sink', (['json'], {}), '(json)\n', (274, 280), False, 'from builtins import __test_sink, __test_source\n'), ((316, 331), 'builtins.__test_source', '__test_source', ([], {}), '()\n', (329, 331), False, 'from builtins import __test_sink, __test_source\n')]
import json import logging from pprint import pformat from typing import Any, Dict, Set import pydantic from models_library.projects_nodes import NodeID from models_library.utils.nodes import compute_node_hash from packaging import version from ..node_ports_common.dbmanager import DBManager from ..node_ports_common.e...
[ "logging.getLogger", "json.loads", "models_library.projects_nodes.NodeID", "json.dumps", "pprint.pformat", "packaging.version.parse" ]
[((608, 635), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (625, 635), False, 'import logging\n'), ((545, 576), 'packaging.version.parse', 'version.parse', (['pydantic.VERSION'], {}), '(pydantic.VERSION)\n', (558, 576), False, 'from packaging import version\n'), ((579, 599), 'packaging....
# Author: <NAME>, <NAME> # Date: 05/28/2020 # Version: 1.2 import json import pygame import GameField import random class Dice: image_list: [] gamefield: GameField image: pygame.image coordinates = (5, 5) surface_dice: pygame.Surface """ desc: - init param: - gamefie...
[ "pygame.Surface", "pygame.transform.smoothscale", "pygame.display.get_surface", "json.load", "pygame.display.update", "random.randint" ]
[((577, 693), 'pygame.Surface', 'pygame.Surface', (['(self.gamefield.screen_size_multiplier, self.gamefield.screen_size_multiplier)', 'pygame.SRCALPHA'], {}), '((self.gamefield.screen_size_multiplier, self.gamefield.\n screen_size_multiplier), pygame.SRCALPHA)\n', (591, 693), False, 'import pygame\n'), ((1321, 1341)...
from random import randrange import allure from model.contact import Contact def test_emails_on_home_page(app, db): if len(db.get_contact_list()) == 0: with allure.step("If contact's list is empty, I'll create a new contact"): app.contact.create(Contact(firstname="NAME")) with allure.step(...
[ "allure.step", "model.contact.Contact" ]
[((308, 352), 'allure.step', 'allure.step', (['"""Given an index for compering"""'], {}), "('Given an index for compering')\n", (319, 352), False, 'import allure\n'), ((426, 492), 'allure.step', 'allure.step', (['"""Given contact\'s by index information from home page"""'], {}), '("Given contact\'s by index informati...
import functools import logging import typing from contextlib import contextmanager from sqlalchemy.exc import SQLAlchemyError from .exceptions import CorporaException from ..corpora_orm import Base, DBSessionMaker logger = logging.getLogger(__name__) class DbUtils: """DbUtils as a singleton to avoid creating ...
[ "logging.getLogger", "functools.wraps" ]
[((227, 254), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (244, 254), False, 'import logging\n'), ((2741, 2762), 'functools.wraps', 'functools.wraps', (['func'], {}), '(func)\n', (2756, 2762), False, 'import functools\n')]
import sys import os import random import matplotlib as mpl import matplotlib.pyplot as plt from dlimage.mnist import MNISTLoader def load_data(data_path): mndata = MNISTLoader(data_path) images, labels = mndata.load_testing() return images, labels def show_image(image, label): plt.title(label, {'f...
[ "matplotlib.pyplot.title", "dlimage.mnist.MNISTLoader", "random.randint", "matplotlib.pyplot.show" ]
[((172, 194), 'dlimage.mnist.MNISTLoader', 'MNISTLoader', (['data_path'], {}), '(data_path)\n', (183, 194), False, 'from dlimage.mnist import MNISTLoader\n'), ((300, 334), 'matplotlib.pyplot.title', 'plt.title', (['label', "{'fontsize': 36}"], {}), "(label, {'fontsize': 36})\n", (309, 334), True, 'import matplotlib.pyp...
import discord import asyncio import discord import asyncio from discord.ext import commands import random from discord.ext.commands.cooldowns import BucketType import datetime import psycopg2 from Principais.principais import bot, mydb, cursor bot = bot mydb = mydb cursor = cursor async def On_command_error(ctx,...
[ "Principais.principais.bot.get_guild", "Principais.principais.cursor.execute", "discord.utils.find", "Principais.principais.bot.user.mentioned_in", "Principais.principais.bot.process_commands", "Principais.principais.mydb.commit", "discord.Permissions", "Principais.principais.cursor.fetchone", "asyn...
[((794, 824), 'Principais.principais.bot.user.mentioned_in', 'bot.user.mentioned_in', (['message'], {}), '(message)\n', (815, 824), False, 'from Principais.principais import bot, mydb, cursor\n'), ((2928, 2951), 'Principais.principais.cursor.execute', 'cursor.execute', (['inserir'], {}), '(inserir)\n', (2942, 2951), Fa...
# -*- coding: utf-8 -*- from __future__ import unicode_literals import os from pyarrow import HadoopFileSystem, HdfsFile from pyarrow.hdfs import connect from plaso.tarzan.lib.hdfs import Hdfs class PyArrowHdfs(Hdfs): """HDFS driver utilizing PyArrow library.""" fs = None # type: HadoopFileSystem de...
[ "pyarrow.hdfs.connect" ]
[((836, 914), 'pyarrow.hdfs.connect', 'connect', ([], {'host': "(uri.hostname or 'default')", 'port': '(uri.port or 0)', 'user': 'uri.username'}), "(host=uri.hostname or 'default', port=uri.port or 0, user=uri.username)\n", (843, 914), False, 'from pyarrow.hdfs import connect\n')]
import tensorflow as tf from models import Classifier class Model(Classifier): name = 'cnnmnist' def __init__(self, *args, **kwargs): super().__init__(*args, **kwargs) def build(self, x, y, *args, **kwargs): self.nclasses = y._shape_as_list()[1] logits = self.get_logits(x, self...
[ "tensorflow.layers.dense", "tensorflow.variable_scope", "tensorflow.reshape", "tensorflow.layers.max_pooling2d", "tensorflow.contrib.layers.l2_regularizer", "tensorflow.nn.softmax_cross_entropy_with_logits_v2", "tensorflow.layers.conv2d", "tensorflow.constant", "tensorflow.trainable_variables", "t...
[((789, 811), 'tensorflow.reduce_mean', 'tf.reduce_mean', (['losses'], {}), '(losses)\n', (803, 811), True, 'import tensorflow as tf\n'), ((863, 888), 'tensorflow.identity', 'tf.identity', (['loss', '"""loss"""'], {}), "(loss, 'loss')\n", (874, 888), True, 'import tensorflow as tf\n'), ((999, 1047), 'tensorflow.constan...