code
stringlengths
22
1.05M
apis
listlengths
1
3.31k
extract_api
stringlengths
75
3.25M
# -*- coding: utf-8 -*- """ LemonSoap - headers scent. Deals with column headers. """ import pandas as pd import inflection import re import logging from ..lemon_bar import LemonBar from .scent_template import ScentTemplate class ColumnsScent(ScentTemplate): """ Manages headers issue identification and fixi...
[ "re.sub", "inflection.underscore" ]
[((2078, 2106), 'inflection.underscore', 'inflection.underscore', (['fixed'], {}), '(fixed)\n', (2099, 2106), False, 'import inflection\n'), ((2123, 2148), 're.sub', 're.sub', (['"""\\\\W+"""', '""""""', 'fixed'], {}), "('\\\\W+', '', fixed)\n", (2129, 2148), False, 'import re\n')]
# Generated by Django 3.2.3 on 2021-10-19 18:54 from django.conf import settings from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('user', '0001_initial'), ] operations = [ migrations.AddField( model_name='userprofile', ...
[ "django.db.models.ManyToManyField" ]
[((361, 463), 'django.db.models.ManyToManyField', 'models.ManyToManyField', ([], {'related_name': '"""_user_userprofile_relations_+"""', 'to': 'settings.AUTH_USER_MODEL'}), "(related_name='_user_userprofile_relations_+', to=\n settings.AUTH_USER_MODEL)\n", (383, 463), False, 'from django.db import migrations, models...
#python -m marbles test_semantic_columns.py import unittest from marbles.mixins import mixins import pandas as pd import requests from pyspark.sql import SparkSession import psycopg2 as pg import pandas as pd import marbles from pyspark.sql.types import StructType, StructField, StringType import psycopg2 as pg #from s...
[ "psycopg2.connect", "pandas.read_sql_query", "pyspark.sql.functions.col", "pyspark.sql.SparkSession.builder.config", "pyspark.sql.types.StringType" ]
[((5059, 5083), 'psycopg2.connect', 'pg.connect', (['config_psyco'], {}), '(config_psyco)\n', (5069, 5083), True, 'import psycopg2 as pg\n'), ((5094, 5164), 'pandas.read_sql_query', 'pd.read_sql_query', (['"""select * from clean.rita limit 1;"""'], {'con': 'connection'}), "('select * from clean.rita limit 1;', con=conn...
import math class polygon: def __init__(self, arr): self.original_arr = arr self.size = len(self.original_arr) self.__set_min_max_by_original__() self.__refactor_original_seq__() self.sorted_arr.append(self.sorted_arr[0]) self.size += 1 def __set_min_max_by_ori...
[ "math.ceil", "math.sqrt" ]
[((4603, 4734), 'math.sqrt', 'math.sqrt', (['((self.sorted_arr[i][0] - self.sorted_arr[i + 1][0]) ** 2 + (self.\n sorted_arr[i][1] - self.sorted_arr[i + 1][1]) ** 2)'], {}), '((self.sorted_arr[i][0] - self.sorted_arr[i + 1][0]) ** 2 + (self.\n sorted_arr[i][1] - self.sorted_arr[i + 1][1]) ** 2)\n', (4612, 4734), ...
# coding: utf-8 # In[20]: import numpy as np import pydensecrf.densecrf as dcrf import os import cv2 import random from tqdm import tqdm # In[21]: from skimage.color import gray2rgb from skimage.color import rgb2gray import matplotlib.pyplot as plt from sklearn.metrics import f1_score, accuracy_score from pyden...
[ "numpy.ascontiguousarray", "numpy.array", "pydensecrf.utils.unary_from_softmax", "matplotlib.pyplot.imshow", "pydensecrf.densecrf.DenseCRF2D", "os.path.exists", "os.listdir", "numpy.mean", "matplotlib.pyplot.close", "numpy.exp", "os.mkdir", "matplotlib.pyplot.axis", "numpy.maximum", "numpy...
[((1290, 1309), 'numpy.zeros', 'np.zeros', (['(h, w, 2)'], {}), '((h, w, 2))\n', (1298, 1309), True, 'import numpy as np\n'), ((2223, 2239), 'numpy.zeros', 'np.zeros', (['(h, w)'], {}), '((h, w))\n', (2231, 2239), True, 'import numpy as np\n'), ((2253, 2272), 'numpy.zeros', 'np.zeros', (['(h, w, 2)'], {}), '((h, w, 2))...
import random from django.core.management.base import BaseCommand from django.contrib.admin.utils import flatten from django_seed import Seed from lists import models as list_models from users import models as user_models from rooms import models as room_models NAME = "lists" class Command(BaseCommand): help =...
[ "users.models.User.objects.all", "rooms.models.Room.objects.all", "lists.models.List.objects.create", "random.randint" ]
[((409, 439), 'users.models.User.objects.all', 'user_models.User.objects.all', ([], {}), '()\n', (437, 439), True, 'from users import models as user_models\n'), ((456, 486), 'rooms.models.Room.objects.all', 'room_models.Room.objects.all', ([], {}), '()\n', (484, 486), True, 'from rooms import models as room_models\n'),...
# -*- coding: utf-8 -*- from __future__ import unicode_literals from django.db import models, migrations class Migration(migrations.Migration): dependencies = [ ('classroom', '0002_assignment_description'), ] operations = [ migrations.AddField( model_name='assignment', ...
[ "django.db.models.DateTimeField", "django.db.models.BooleanField" ]
[((369, 448), 'django.db.models.DateTimeField', 'models.DateTimeField', ([], {'null': '(True)', 'verbose_name': '"""Fecha de evaluación"""', 'blank': '(True)'}), "(null=True, verbose_name='Fecha de evaluación', blank=True)\n", (389, 448), False, 'from django.db import models, migrations\n'), ((616, 768), 'django.db.mod...
import numpy as np import matplotlib.pyplot as plt import os from pyburst.grids import grid_analyser, grid_strings, grid_tools # resolution tests y_factors = {'dt': 3600, 'fluence': 1e39, 'peak': 1e38, } y_labels = {'dt': '$\Delta t$', 'rate': 'Burst rate', ...
[ "matplotlib.pyplot.savefig", "pyburst.grids.grid_tools.enumerate_params", "numpy.where", "os.path.join", "matplotlib.pyplot.close", "pyburst.grids.grid_strings.plots_path", "matplotlib.pyplot.tight_layout", "pyburst.grids.grid_analyser.Kgrid", "numpy.full", "matplotlib.pyplot.subplots", "matplot...
[((1303, 1345), 'pyburst.grids.grid_tools.enumerate_params', 'grid_tools.enumerate_params', (['unique_subset'], {}), '(unique_subset)\n', (1330, 1345), False, 'from pyburst.grids import grid_analyser, grid_strings, grid_tools\n'), ((2387, 2436), 'matplotlib.pyplot.subplots', 'plt.subplots', (['n', '(2)'], {'sharex': '(...
# Copyright (c) 2020-present, Assistive Robotics Lab # All rights reserved. # # This source code is licensed under the license found in the # LICENSE file in the root directory of this source tree. # from transformers.training_utils import fit from transformers.transformers import ( InferenceTransformerEncoder, ...
[ "torch.manual_seed", "torch.optim.lr_scheduler.MultiStepLR", "argparse.ArgumentParser", "common.logging.logger.info", "torch.nn.L1Loss", "torch.cuda.device_count", "transformers.training_utils.fit", "torch.nn.DataParallel", "torch.cuda.is_available", "common.losses.QuatDistance", "numpy.random.s...
[((541, 562), 'torch.manual_seed', 'torch.manual_seed', (['(42)'], {}), '(42)\n', (558, 562), False, 'import torch\n'), ((563, 581), 'numpy.random.seed', 'np.random.seed', (['(42)'], {}), '(42)\n', (577, 581), True, 'import numpy as np\n'), ((833, 858), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {}), '(...
from typing import Tuple import torch import torch.nn as nn from pyro.distributions.util import broadcast_shape from pyro_util.modules.weight_scaling import GammaReLU, WSLinear T = torch.Tensor def make_ws_fc(*dims: int) -> nn.Module: """Helper function for creating a fully connected neural network. This v...
[ "torch.nn.ReLU", "pyro_util.modules.weight_scaling.WSLinear", "torch.nn.Sequential", "torch.nn.BatchNorm1d", "pyro_util.modules.weight_scaling.GammaReLU", "pyro.distributions.util.broadcast_shape", "torch.nn.Linear" ]
[((704, 726), 'torch.nn.Sequential', 'nn.Sequential', (['*layers'], {}), '(*layers)\n', (717, 726), True, 'import torch.nn as nn\n'), ((1263, 1285), 'torch.nn.Sequential', 'nn.Sequential', (['*layers'], {}), '(*layers)\n', (1276, 1285), True, 'import torch.nn as nn\n'), ((527, 553), 'pyro_util.modules.weight_scaling.WS...
#!/usr/bin/env python ''' catalog_harvesting/util.py General utilities for the project ''' import random def unique_id(): ''' Return a random 17-character string that works well for mongo IDs ''' charmap = "abcdefghijklmnopqrstuvwxyzABCDEFGHIJKLMNOPQRSTUVWXYZ0123456789" return ''.join([random.cho...
[ "random.choice" ]
[((310, 332), 'random.choice', 'random.choice', (['charmap'], {}), '(charmap)\n', (323, 332), False, 'import random\n')]
from WeatherScreens.RingScreen import RingScreen from WeatherScreens.QuadrantScreen import QuadrantScreen from WeatherScreens.ImageScreen import ImageScreen from WeatherScreens.ScreenBase import ScreenBase from datetime import datetime, timedelta from suntime import Sun, SunTimeException from dateutil import tz import ...
[ "WeatherScreens.ScreenBase.ScreenBase", "argparse.ArgumentParser", "datetime.datetime.strptime", "pyowm.OWM", "WeatherScreens.ImageScreen.ImageScreen", "datetime.datetime.now", "WeatherScreens.QuadrantScreen.QuadrantScreen", "datetime.timedelta", "WeatherScreens.RingScreen.RingScreen" ]
[((385, 448), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {'description': '"""WeatherFrame CLI Utility"""'}), "(description='WeatherFrame CLI Utility')\n", (408, 448), False, 'import argparse\n'), ((9014, 9028), 'datetime.datetime.now', 'datetime.now', ([], {}), '()\n', (9026, 9028), False, 'from datetim...
from django.conf.urls import url, include from django.conf import settings from . import views # Wire up our API using automatic URL routing. # Additionally, we include login URLs for the browsable API. urlpatterns = [ url(r'manage/', views.index), ]
[ "django.conf.urls.url" ]
[((222, 249), 'django.conf.urls.url', 'url', (['"""manage/"""', 'views.index'], {}), "('manage/', views.index)\n", (225, 249), False, 'from django.conf.urls import url, include\n')]
# This code is generated automatically by ClointFusion BOT Builder Tool. import ClointFusion as cf import time cf.window_show_desktop() cf.mouse_click(int(cf.pg.size()[0]/2),int(cf.pg.size()[1]/2)) try: cf.mouse_click(*cf.mouse_search_snip_return_coordinates_x_y(r'C:\Users\mrmay\AppData\Local\Temp\cf_log_5fa2...
[ "ClointFusion.key_write_enter", "ClointFusion.window_show_desktop", "ClointFusion.key_press", "time.sleep", "ClointFusion.mouse_search_snip_return_coordinates_x_y", "ClointFusion.pg.size", "ClointFusion.mouse_click" ]
[((114, 138), 'ClointFusion.window_show_desktop', 'cf.window_show_desktop', ([], {}), '()\n', (136, 138), True, 'import ClointFusion as cf\n'), ((530, 543), 'time.sleep', 'time.sleep', (['(2)'], {}), '(2)\n', (540, 543), False, 'import time\n'), ((869, 882), 'time.sleep', 'time.sleep', (['(0)'], {}), '(0)\n', (879, 882...
#!/usr/bin/env nemesis """ This script creates a spatial database for the initial stress and state variables for a Maxwell plane strain material. """ sim = "gravity_vardensity" materials = ["crust","mantle"] import numpy import h5py from spatialdata.spatialdb.SimpleIOAscii import SimpleIOAscii from spatialdata.geoco...
[ "spatialdata.spatialdb.SimpleIOAscii.SimpleIOAscii", "h5py.File", "numpy.array", "numpy.zeros", "spatialdata.geocoords.CSCart.CSCart" ]
[((351, 359), 'spatialdata.geocoords.CSCart.CSCart', 'CSCart', ([], {}), '()\n', (357, 359), False, 'from spatialdata.geocoords.CSCart import CSCart\n'), ((659, 903), 'numpy.array', 'numpy.array', (['[[0.62200847, 0.16666667, 0.0446582, 0.16666667], [0.16666667, 0.62200847, \n 0.16666667, 0.0446582], [0.16666667, 0....
import datetime import typing from . import enums, tools class CatalogueAPIWrapper: """Methods for listing objects""" def __init__( self, username: str, password: str, language: enums.Language = enums.Language.GERMAN ): """Create a new Wrapper containing functions for listing different o...
[ "datetime.date.today", "datetime.timedelta" ]
[((15395, 15416), 'datetime.date.today', 'datetime.date.today', ([], {}), '()\n', (15414, 15416), False, 'import datetime\n'), ((15419, 15446), 'datetime.timedelta', 'datetime.timedelta', ([], {'days': '(-7)'}), '(days=-7)\n', (15437, 15446), False, 'import datetime\n'), ((17879, 17900), 'datetime.date.today', 'datetim...
# coding: utf-8 """加密算法:公钥(私钥)加密,私钥解密""" from Crypto.PublicKey import RSA from Crypto import Random DATA = 'Hello, word!' PRIVATE_KEY_PEM = """-----<KEY>""" PUBLIC_KEY_PEM = """-----<KEY>""" def _encrypt_by_public(): random_func = Random.new().read public_key = RSA.importKey(PUBLIC_KEY_PEM) encrypted =...
[ "Crypto.Random.new", "Crypto.PublicKey.RSA.importKey" ]
[((275, 304), 'Crypto.PublicKey.RSA.importKey', 'RSA.importKey', (['PUBLIC_KEY_PEM'], {}), '(PUBLIC_KEY_PEM)\n', (288, 304), False, 'from Crypto.PublicKey import RSA\n'), ((463, 493), 'Crypto.PublicKey.RSA.importKey', 'RSA.importKey', (['PRIVATE_KEY_PEM'], {}), '(PRIVATE_KEY_PEM)\n', (476, 493), False, 'from Crypto.Pub...
from django.contrib.auth import get_user_model from rest_framework import mixins from rest_framework.viewsets import GenericViewSet from users.serializers import UserSerializer class UserViewSet(mixins.RetrieveModelMixin, mixins.ListModelMixin, GenericViewSet): queryset = get_user_model().objects.all() seria...
[ "django.contrib.auth.get_user_model" ]
[((280, 296), 'django.contrib.auth.get_user_model', 'get_user_model', ([], {}), '()\n', (294, 296), False, 'from django.contrib.auth import get_user_model\n')]
import numpy import chainer from chainer.backends import cuda from chainer.functions.activation import sigmoid from chainer.functions.activation import tanh from chainer.functions.array import concat from chainer.functions.math import linear_interpolate from chainer import link from chainer.links.connection import lin...
[ "chainer.functions.array.concat.concat", "chainer.Variable", "chainer.links.connection.linear.Linear", "chainer.get_device", "chainer.functions.math.linear_interpolate.linear_interpolate", "chainer.get_dtype" ]
[((773, 823), 'chainer.functions.math.linear_interpolate.linear_interpolate', 'linear_interpolate.linear_interpolate', (['f', 'h_bar', 'h'], {}), '(f, h_bar, h)\n', (810, 823), False, 'from chainer.functions.math import linear_interpolate\n'), ((1305, 1331), 'chainer.get_device', 'chainer.get_device', (['device'], {}),...
from django.contrib.auth import get_user_model from django.urls import reverse from django.test import TestCase from rest_framework import status from rest_framework.test import APIClient from core.models import Member, Band from rockband.serializers import MemberSerializer MEMBERS_URL = reverse('rockband:member-l...
[ "django.contrib.auth.get_user_model", "django.urls.reverse", "core.models.Member.objects.all", "rest_framework.test.APIClient", "core.models.Member.objects.filter", "core.models.Band.objects.create", "core.models.Member.objects.create", "rockband.serializers.MemberSerializer" ]
[((294, 325), 'django.urls.reverse', 'reverse', (['"""rockband:member-list"""'], {}), "('rockband:member-list')\n", (301, 325), False, 'from django.urls import reverse\n'), ((475, 486), 'rest_framework.test.APIClient', 'APIClient', ([], {}), '()\n', (484, 486), False, 'from rest_framework.test import APIClient\n'), ((8...
from slave.playground.bots import BotInformation from slave.lib.bots import BotBasic, BotV2 config = { 'host': 'chat.freenode.net', 'port': 6667, 'channel': "#slavebotpool666", 'boss_name': 'boss666', 'bot_prefix': "SLAVEBOT" } BotInformation.read_config_from_dict(config) BotInformation.use_other...
[ "slave.playground.bots.BotInformation.start", "slave.playground.bots.BotInformation.read_config_from_dict", "slave.playground.bots.BotInformation.use_other_bot_commands" ]
[((249, 293), 'slave.playground.bots.BotInformation.read_config_from_dict', 'BotInformation.read_config_from_dict', (['config'], {}), '(config)\n', (285, 293), False, 'from slave.playground.bots import BotInformation\n'), ((296, 340), 'slave.playground.bots.BotInformation.use_other_bot_commands', 'BotInformation.use_ot...
############################################################################## # Copyright (c) 2017 <NAME> <<EMAIL>>, Red Hat # # All rights reserved. This program and the accompanying materials # are made available under the terms of the Apache License, Version 2.0 # which accompanies this distribution, and is availab...
[ "hashlib.sha256", "argparse.ArgumentParser", "os.path.join", "binaryornot.check.is_binary", "os.walk" ]
[((670, 686), 'hashlib.sha256', 'hashlib.sha256', ([], {}), '()\n', (684, 686), False, 'import hashlib\n'), ((696, 721), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {}), '()\n', (719, 721), False, 'import argparse\n'), ((960, 981), 'os.walk', 'os.walk', (['args.project'], {}), '(args.project)\n', (967, 9...
import inspect from collections import namedtuple import numpy as np from sklearn.ensemble import GradientBoostingRegressor from sklearn.model_selection import train_test_split from sklearn.exceptions import NotFittedError from uq360.algorithms.posthocuq import PostHocUQ class MetamodelRegression(PostHocUQ): """...
[ "numpy.abs", "collections.namedtuple", "numpy.hstack", "sklearn.model_selection.train_test_split", "numpy.asarray", "numpy.expand_dims", "inspect.isclass", "sklearn.ensemble.GradientBoostingRegressor" ]
[((6132, 6145), 'numpy.asarray', 'np.asarray', (['X'], {}), '(X)\n', (6142, 6145), True, 'import numpy as np\n'), ((6158, 6171), 'numpy.asarray', 'np.asarray', (['y'], {}), '(y)\n', (6168, 6171), True, 'import numpy as np\n'), ((7884, 7916), 'numpy.hstack', 'np.hstack', (['[X, y_hat_meta_prime]'], {}), '([X, y_hat_meta...
import logging import subprocess from threading import Thread from ulauncher.api.client.Extension import Extension from ulauncher.api.shared.event import KeywordQueryEvent, ItemEnterEvent, \ PreferencesEvent, PreferencesUpdateEvent from ulauncher.api.shared.action.ExtensionCustomAction import \ ExtensionCustomA...
[ "logging.getLogger", "ulauncher.api.shared.item.ExtensionResultItem.ExtensionResultItem", "dendron.query_listener.KeywordQueryEventListener", "subprocess.run", "ulauncher.api.shared.action.RenderResultListAction.RenderResultListAction", "ulauncher.api.shared.action.ExtensionCustomAction.ExtensionCustomAct...
[((713, 740), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (730, 740), False, 'import logging\n'), ((1323, 1361), 'threading.Thread', 'Thread', ([], {'target': 'self.dendron.load_notes'}), '(target=self.dendron.load_notes)\n', (1329, 1361), False, 'from threading import Thread\n'), ((23...
# pylint: disable=redefined-outer-name # pylint: disable=too-many-lines import itertools import pytest from buzzard.test.tools import assert_tiles_eq from buzzard.test import make_tile_set ANY = 42 PARAMS1 = { 'extend', 'overlap', 'exclude', 'exception', 'shrink', } PARAMS2 = {'br', 'tr', 'tl',...
[ "itertools.product", "pytest.mark.parametrize", "pytest.raises", "pytest.fixture", "buzzard.test.tools.assert_tiles_eq", "buzzard.test.make_tile_set.make_tile_set" ]
[((1320, 1350), 'pytest.fixture', 'pytest.fixture', ([], {'scope': '"""module"""'}), "(scope='module')\n", (1334, 1350), False, 'import pytest\n'), ((2402, 2432), 'pytest.fixture', 'pytest.fixture', ([], {'params': 'PARAMS2'}), '(params=PARAMS2)\n', (2416, 2432), False, 'import pytest\n'), ((2497, 2527), 'pytest.fixtur...
import unittest from collections import defaultdict import numpy as np import pandas as pd from ife.io.io import ImageReader class TestMomentFeatures(unittest.TestCase): def test_moment_output_type(self) -> None: features = ImageReader.read_from_single_file("ife/data/small_rgb.jpg") moment = fe...
[ "ife.io.io.ImageReader.read_from_single_file", "numpy.zeros" ]
[((240, 299), 'ife.io.io.ImageReader.read_from_single_file', 'ImageReader.read_from_single_file', (['"""ife/data/small_rgb.jpg"""'], {}), "('ife/data/small_rgb.jpg')\n", (273, 299), False, 'from ife.io.io import ImageReader\n'), ((945, 1004), 'ife.io.io.ImageReader.read_from_single_file', 'ImageReader.read_from_single_...
import numpy as np from math import pi import torch from pykeops.torch import LazyTensor from plyfile import PlyData, PlyElement from helper import * import torch.nn as nn import torch.nn.functional as F # from matplotlib import pyplot as plt from pykeops.torch.cluster import grid_cluster, cluster_ranges_centroids, fr...
[ "torch.nn.ReLU", "numpy.sqrt", "torch.max", "math.sqrt", "torch.sum", "torch.nn.GroupNorm", "pyvtk.PointData", "pyvtk.VtkData", "pyvtk.CellData", "torch.zeros_like", "torch.nn.init.uniform_", "pykeops.torch.LazyTensor", "torch.randn", "torch.ones_like", "torch.nn.LeakyReLU", "torch.Ten...
[((2722, 2736), 'pyvtk.VtkData', 'VtkData', (['*data'], {}), '(*data)\n', (2729, 2736), False, 'from pyvtk import PolyData, PointData, CellData, Scalars, Vectors, VtkData, PointData\n'), ((5780, 5805), 'pykeops.torch.LazyTensor', 'LazyTensor', (['x[:, None, :]'], {}), '(x[:, None, :])\n', (5790, 5805), False, 'from pyk...
from random import randint, seed import numpy as np from os import path, mkdir from maze_utils import generate_grid seed_number = 69 training_folder = "training" testing_folder = "testing" tot_elem_training = 100 # numero di matrici da generare tot_elem_testing = 20 # numero di matrici da generare max_w = 10 ...
[ "os.path.exists", "maze_utils.generate_grid", "random.seed", "os.mkdir", "numpy.random.seed", "numpy.savetxt", "random.randint" ]
[((707, 734), 'numpy.random.seed', 'np.random.seed', (['seed_number'], {}), '(seed_number)\n', (721, 734), True, 'import numpy as np\n'), ((739, 756), 'random.seed', 'seed', (['seed_number'], {}), '(seed_number)\n', (743, 756), False, 'from random import randint, seed\n'), ((1101, 1128), 'os.path.exists', 'path.exists'...
"""create tables Revision ID: 6fb351569d30 Revises: 4<PASSWORD>1ff38b Create Date: 2019-05-06 21:59:43.998735 """ from alembic import op import sqlalchemy as sa # revision identifiers, used by Alembic. revision = '<KEY>' down_revision = '4<PASSWORD>' branch_labels = None depends_on = None def upgrade(): # ###...
[ "sqlalchemy.ForeignKeyConstraint", "sqlalchemy.DateTime", "alembic.op.drop_table", "sqlalchemy.Boolean", "sqlalchemy.PrimaryKeyConstraint", "sqlalchemy.Date", "sqlalchemy.Integer", "sqlalchemy.Numeric", "sqlalchemy.String", "sqlalchemy.Enum" ]
[((13130, 13179), 'alembic.op.drop_table', 'op.drop_table', (['"""transaction_check"""'], {'schema': '"""glod"""'}), "('transaction_check', schema='glod')\n", (13143, 13179), False, 'from alembic import op\n'), ((13184, 13227), 'alembic.op.drop_table', 'op.drop_table', (['"""transaction"""'], {'schema': '"""glod"""'}),...
"""Hparams""" import argparse as ap import tensorflow as tf from pathlib import Path HOME = str(Path.home()) HPARAM_CHOICES= { "model": ["cpdb", "copy", "bdrnn", "cpdb2", "cpdb2_prot"], "optimizer": ["adam", "sgd", "adadelta"], "unit_type": ["lstm", "lstmblock", "nlstm", "gru"], "train...
[ "pathlib.Path.home", "tensorflow.contrib.training.HParams", "argparse.ArgumentParser" ]
[((97, 108), 'pathlib.Path.home', 'Path.home', ([], {}), '()\n', (106, 108), False, 'from pathlib import Path\n'), ((1415, 1513), 'argparse.ArgumentParser', 'ap.ArgumentParser', ([], {'description': '"""Hyperparameters"""', 'add_help': '(False)', 'argument_default': 'ap.SUPPRESS'}), "(description='Hyperparameters', add...
from kivy.lang import Builder from kivy.metrics import dp from kivy import properties as p from kivy.animation import Animation from kivymd.app import MDApp as App from kivymd.uix.screen import MDScreen class HomeMainScreen(MDScreen): bg_pos = p.NumericProperty(0) def toggle_bg_pos(self): bg_pos...
[ "kivy.animation.Animation", "kivy.metrics.dp", "kivy.properties.NumericProperty" ]
[((251, 271), 'kivy.properties.NumericProperty', 'p.NumericProperty', (['(0)'], {}), '(0)\n', (268, 271), True, 'from kivy import properties as p\n'), ((349, 368), 'kivy.metrics.dp', 'dp', (['(self.height / 2)'], {}), '(self.height / 2)\n', (351, 368), False, 'from kivy.metrics import dp\n'), ((375, 399), 'kivy.animati...
# def draw_nx(g, labels=None): # import matplotlib.pyplot as plt # if labels is not None: # g = nx.relabel_nodes(g, labels) # pos = nx.kamada_kawai_layout(g) # nx.draw(g, pos, with_labels=True) # plt.show() # # def draw_nx_attributes_as_labels(g, attribute): # # import pylab # impo...
[ "torch.tensor", "networkx.draw", "torch_geometric.data.Data", "torch_geometric.utils.convert.to_networkx" ]
[((2444, 2504), 'torch.tensor', 'torch.tensor', (['[[0, 1, 1, 2], [1, 0, 2, 1]]'], {'dtype': 'torch.long'}), '([[0, 1, 1, 2], [1, 0, 2, 1]], dtype=torch.long)\n', (2456, 2504), False, 'import torch\n'), ((2589, 2625), 'torch.tensor', 'torch.tensor', (['[[0.0], [-1.0], [1.0]]'], {}), '([[0.0], [-1.0], [1.0]])\n', (2601,...
# Copyright 2016 Open Source Robotics Foundation, Inc. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law...
[ "rclpy.ok", "ros_mstar.srv.MStarSrv.Request", "rclpy.spin_once", "rclpy.init", "rclpy.shutdown" ]
[((1467, 1488), 'rclpy.init', 'rclpy.init', ([], {'args': 'args'}), '(args=args)\n', (1477, 1488), False, 'import rclpy\n'), ((1914, 1924), 'rclpy.ok', 'rclpy.ok', ([], {}), '()\n', (1922, 1924), False, 'import rclpy\n'), ((2546, 2562), 'rclpy.shutdown', 'rclpy.shutdown', ([], {}), '()\n', (2560, 2562), False, 'import ...
import atexit import contextlib import time from typing import Any, List, Type from unittest import mock import pytest import region_profiler.global_instance import region_profiler.profiler from region_profiler import RegionProfiler, func from region_profiler import install as install_profiler from region_profiler imp...
[ "region_profiler.reporters.SilentReporter", "unittest.mock.Mock", "region_profiler.install", "region_profiler.func", "time.sleep", "pytest.mark.parametrize", "region_profiler.region", "region_profiler.iter_proxy" ]
[((1040, 1091), 'pytest.mark.parametrize', 'pytest.mark.parametrize', (['"""multiple_runs"""', '[0, 1, 2]'], {}), "('multiple_runs', [0, 1, 2])\n", (1063, 1091), False, 'import pytest\n'), ((1589, 1635), 'pytest.mark.parametrize', 'pytest.mark.parametrize', (['"""use_cython"""', '[False]'], {}), "('use_cython', [False]...
#!/usr/bin/env python import datetime import logging import time from threading import Thread import requests from requests.auth import HTTPBasicAuth import settings def update_notiwire(data=None, relative_url=''): URL = settings.API_URL + settings.NAME + '/' if not data: data = {} data['api_key...
[ "logging.basicConfig", "requests.post", "requests.auth.HTTPBasicAuth", "time.sleep", "requests.get", "datetime.datetime.now", "time.time", "threading.Thread", "logging.info", "logging.error" ]
[((463, 507), 'requests.post', 'requests.post', (['(URL + relative_url)'], {'data': 'data'}), '(URL + relative_url, data=data)\n', (476, 507), False, 'import requests\n'), ((4057, 4127), 'logging.basicConfig', 'logging.basicConfig', ([], {'format': '"""%(asctime)s %(message)s"""', 'level': 'log_level'}), "(format='%(as...
from django.conf import settings from django.conf.urls import url from django.conf.urls.static import static from . import views urlpatterns=[ url('^$',views.index,name = 'index'), url(r'^profile/(\d+)',views.profile,name = "profile"), url(r'^create/post',views.new_post, name = "new-post"), url(r'^foll...
[ "django.conf.urls.static.static", "django.conf.urls.url" ]
[((148, 184), 'django.conf.urls.url', 'url', (['"""^$"""', 'views.index'], {'name': '"""index"""'}), "('^$', views.index, name='index')\n", (151, 184), False, 'from django.conf.urls import url\n'), ((190, 243), 'django.conf.urls.url', 'url', (['"""^profile/(\\\\d+)"""', 'views.profile'], {'name': '"""profile"""'}), "('...
from unittest.mock import MagicMock from unittest.mock import patch import aiofiles from aiofiles import threadpool async def test_unit_get_current_version_both_files_dont_exist(mock_hub, hub, tmp_path): """ SCENARIO #1 - override_version_file DOES NOT EXIST - main_version_file DOES NOT EXIST """...
[ "aiofiles.threadpool.wrap.register", "unittest.mock.MagicMock", "aiofiles.threadpool.AsyncBufferedIOBase", "unittest.mock.patch" ]
[((622, 663), 'unittest.mock.patch', 'patch', (['"""os.getcwd"""'], {'return_value': 'tmp_path'}), "('os.getcwd', return_value=tmp_path)\n", (627, 663), False, 'from unittest.mock import patch\n'), ((1657, 1698), 'unittest.mock.patch', 'patch', (['"""os.getcwd"""'], {'return_value': 'tmp_path'}), "('os.getcwd', return_...
import setuptools setuptools.setup( name="fakeokpy", version='0.1', url="https://github.com/yuvipanda/fakeokpy", author="<NAME>", author_email="<EMAIL>", license="BSD-3-Clause", packages=setuptools.find_packages(), )
[ "setuptools.find_packages" ]
[((216, 242), 'setuptools.find_packages', 'setuptools.find_packages', ([], {}), '()\n', (240, 242), False, 'import setuptools\n')]
# coding: utf-8 # $ \newcommand{\cat}[2][\phantom{i}]{\ket{C^{#2}_{#1\alpha}}} $ # $ \newcommand{\ket}[1]{|#1\rangle} $ # $ \newcommand{\bra}[1]{\langle#1|} $ # $ \newcommand{\braket}[2]{\langle#1|#2\rangle} $ # $\newcommand{\au}{\hat{a}^\dagger}$ # $\newcommand{\ad}{\hat{a}}$ # $\newcommand{\bu}{\hat{b}^\dagger}$ # ...
[ "numpy.sqrt" ]
[((2879, 2890), 'numpy.sqrt', 'sqrt', (['(Δ * χ)'], {}), '(Δ * χ)\n', (2883, 2890), False, 'from numpy import sqrt\n')]
def coding_problem_41(flights_db, starting_airport): """ Given an unordered list of flights taken by someone, each represented as (origin, destination) pairs, and a starting airport, compute the person's itinerary. If no such itinerary exists, return null. If there are multiple possible itineraries, ret...
[ "doctest.testmod" ]
[((935, 964), 'doctest.testmod', 'doctest.testmod', ([], {'verbose': '(True)'}), '(verbose=True)\n', (950, 964), False, 'import doctest\n')]
import itertools import pytest from iterators.invalid_iter import InvalidIter def _grouper_to_keys(grouper): return [g[0] for g in grouper] def _grouper_to_groups(grouper): return [list(g[1]) for g in grouper] @pytest.mark.parametrize("keyfunc, data, expected_keys", [ (lambda x: x, [], []), (lambd...
[ "pytest.mark.parametrize", "iterators.invalid_iter.InvalidIter", "pytest.raises", "itertools.groupby" ]
[((225, 504), 'pytest.mark.parametrize', 'pytest.mark.parametrize', (['"""keyfunc, data, expected_keys"""', "[(lambda x: x, [], []), (lambda x: x, [1, 2, 3], [1, 2, 3]), (lambda x: x,\n [1, 2, 2, 2, 3, 3], [1, 2, 3]), (lambda x: x, '', []), (lambda x: x,\n 'ABC', ['A', 'B', 'C']), (lambda x: x, 'ABBBCC', ['A', 'B...
import serialio class Serial(object): def __init__(self, port, baudrate, timeout): self.port = port self.baudrate = baudrate self.timeout = timeout self._openPort() def _openPort(self): self.hComm = serialio.Serial(self.port, self.baudrate) # Opening the port def read(self): data = seria...
[ "serialio.read", "serialio.Serial" ]
[((223, 264), 'serialio.Serial', 'serialio.Serial', (['self.port', 'self.baudrate'], {}), '(self.port, self.baudrate)\n', (238, 264), False, 'import serialio\n'), ((315, 340), 'serialio.read', 'serialio.read', (['self.hComm'], {}), '(self.hComm)\n', (328, 340), False, 'import serialio\n')]
import numpy as np class NumpyDynamic: def __init__(self, dtype, array_size=(100,)): self.data = np.zeros(array_size, dtype) self.array_size = list(array_size) self.size = 0 def add(self, x): if self.size == self.array_size[0]: self.array_size[0] *= 2 ...
[ "numpy.zeros" ]
[((112, 139), 'numpy.zeros', 'np.zeros', (['array_size', 'dtype'], {}), '(array_size, dtype)\n', (120, 139), True, 'import numpy as np\n'), ((330, 372), 'numpy.zeros', 'np.zeros', (['self.array_size', 'self.data.dtype'], {}), '(self.array_size, self.data.dtype)\n', (338, 372), True, 'import numpy as np\n')]
from rest_framework import serializers from api.models import RouteModel class RouteDistanceSerializer(serializers.ModelSerializer): km = serializers.FloatField(source='distance', read_only=True) class Meta: model = RouteModel fields = ('route_id', 'km')
[ "rest_framework.serializers.FloatField" ]
[((144, 201), 'rest_framework.serializers.FloatField', 'serializers.FloatField', ([], {'source': '"""distance"""', 'read_only': '(True)'}), "(source='distance', read_only=True)\n", (166, 201), False, 'from rest_framework import serializers\n')]
import numpy as np from abc import ABCMeta, abstractmethod class Node(object): """Represents state in MCTS search tree. Args: state (object): The environment state corresponding to this node in the search tree. Note: Node object is immutable. Node is left without exit edges (empty dict)...
[ "numpy.sqrt" ]
[((2113, 2134), 'numpy.sqrt', 'np.sqrt', (['state_visits'], {}), '(state_visits)\n', (2120, 2134), True, 'import numpy as np\n')]
import os import random from sklearn.metrics import mean_squared_error as mse from core.composer.chain import Chain from core.composer.composer import ComposerRequirements, DummyChainTypeEnum, DummyComposer from core.models.data import OutputData from core.models.model import * from core.repository.dataset_t...
[ "core.composer.composer.DummyComposer", "core.repository.model_types_repository.ModelTypesRepository", "core.utils.project_root", "matplotlib.pyplot.plot", "random.seed", "sklearn.metrics.mean_squared_error", "core.repository.model_types_repository.ModelMetaInfoTemplate", "core.composer.composer.Compo...
[((694, 708), 'random.seed', 'random.seed', (['(1)'], {}), '(1)\n', (705, 708), False, 'import random\n'), ((2380, 2402), 'core.repository.model_types_repository.ModelTypesRepository', 'ModelTypesRepository', ([], {}), '()\n', (2400, 2402), False, 'from core.repository.model_types_repository import ModelMetaInfoTemplat...
import os import re import shutil def svnLockFiles(files): fileStr = ' '.join(files) print('Locking files: ', fileStr) os.system('svn lock ' + fileStr) def svnUnlockFiles(files): fileStr = ' '.join(files) print('Unlocking files: ', fileStr) os.system('svn unlock ' + fileStr) ...
[ "os.system", "shutil.move" ]
[((143, 175), 'os.system', 'os.system', (["('svn lock ' + fileStr)"], {}), "('svn lock ' + fileStr)\n", (152, 175), False, 'import os\n'), ((283, 317), 'os.system', 'os.system', (["('svn unlock ' + fileStr)"], {}), "('svn unlock ' + fileStr)\n", (292, 317), False, 'import os\n'), ((1009, 1039), 'shutil.move', 'shutil.m...
import numpy as np import time import pytest import jax.numpy as jnp import jax.config as config import torch import tensorflow as tf from tensornetwork.linalg import linalg from tensornetwork import backends from tensornetwork.backends.numpy import numpy_backend from tensornetwork.backends.jax import jax_backend #pyli...
[ "jax.config.update", "tensornetwork.linalg.linalg.zeros", "tensornetwork.linalg.linalg.randn", "numpy.testing.assert_allclose", "tensornetwork.backends.backend_factory.get_backend", "tensornetwork.linalg.linalg.eye", "numpy.random.seed", "tensornetwork.linalg.linalg.ones", "time.time", "tensornetw...
[((342, 379), 'jax.config.update', 'config.update', (['"""jax_enable_x64"""', '(True)'], {}), "('jax_enable_x64', True)\n", (355, 379), True, 'import jax.config as config\n'), ((1806, 1851), 'tensornetwork.backends.backend_factory.get_backend', 'backends.backend_factory.get_backend', (['backend'], {}), '(backend)\n', (...
"""main.py file representingcomparison statistics for Pyrunc module""" # Python module(s) from timeit import timeit # Project module(s) from Pyrunc import Pyrunc def main(): """Main Method""" pr_c = Pyrunc() # -------------------------------------------------------------------------------- # ----...
[ "timeit.timeit", "Pyrunc.Pyrunc" ]
[((214, 222), 'Pyrunc.Pyrunc', 'Pyrunc', ([], {}), '()\n', (220, 222), False, 'from Pyrunc import Pyrunc\n'), ((2360, 2414), 'timeit.timeit', 'timeit', ([], {'stmt': '"""csummer(30)"""', 'setup': 'c_setup', 'number': '(1000)'}), "(stmt='csummer(30)', setup=c_setup, number=1000)\n", (2366, 2414), False, 'from timeit imp...
from django.urls import reverse_lazy, reverse from django.views.generic import TemplateView from exporter.applications.services import post_applications, post_open_general_licences_applications from exporter.apply_for_a_licence.forms.open_general_licences import ( open_general_licence_forms, open_general_licen...
[ "exporter.apply_for_a_licence.forms.open_general_licences.open_general_licence_forms", "exporter.apply_for_a_licence.forms.triage_questions.transhipment_questions", "exporter.apply_for_a_licence.forms.triage_questions.MOD_questions", "exporter.apply_for_a_licence.forms.triage_questions.trade_control_licence_q...
[((1008, 1026), 'exporter.apply_for_a_licence.forms.triage_questions.opening_question', 'opening_question', ([], {}), '()\n', (1024, 1026), False, 'from exporter.apply_for_a_licence.forms.triage_questions import opening_question, export_licence_questions, MOD_questions, transhipment_questions, trade_control_licence_que...
#!/usr/bin/env python3 import numpy as np import pandas as pd import librosa import os import sys import time from datetime import datetime from pathlib import Path from src.python.audio_transforms import * from src.python.model_predict import * from src.python.graphics import plot_graph # Hardcoding a few variable...
[ "pathlib.Path", "src.python.graphics.plot_graph", "numpy.array", "datetime.datetime.now", "librosa.effects.trim", "librosa.load" ]
[((1090, 1129), 'pathlib.Path', 'Path', (['DIR_ROOT', '"""demo_files"""', '"""results"""'], {}), "(DIR_ROOT, 'demo_files', 'results')\n", (1094, 1129), False, 'from pathlib import Path\n'), ((3100, 3117), 'numpy.array', 'np.array', (['results'], {}), '(results)\n', (3108, 3117), True, 'import numpy as np\n'), ((481, 48...
import socket import struct import json import time import os import platform from optparse import OptionParser import sys import xml.etree.ElementTree as ET import config from device_config import BASE_CONST MCAST_GRP = '192.168.3.11' MCAST_PORT = 8427 DEFAULT_DCID_XML = '/Applications/Shure Update Utility.app/C...
[ "xml.etree.ElementTree.parse", "socket.socket", "device_config.BASE_CONST.items", "optparse.OptionParser", "os.path.isfile", "platform.system", "config.app_dir", "socket.inet_aton", "sys.exit", "json.load", "time.time", "json.dump" ]
[((535, 603), 'socket.socket', 'socket.socket', (['socket.AF_INET', 'socket.SOCK_DGRAM', 'socket.IPPROTO_UDP'], {}), '(socket.AF_INET, socket.SOCK_DGRAM, socket.IPPROTO_UDP)\n', (548, 603), False, 'import socket\n'), ((1916, 1934), 'device_config.BASE_CONST.items', 'BASE_CONST.items', ([], {}), '()\n', (1932, 1934), Fa...
import fastai from neptune.new.integrations.fastai import NeptuneCallback from fastai.vision.all import * import neptune.new as neptune run = neptune.init( project="common/fastai-integration", api_token="<PASSWORD>", tags="basic" ) path = untar_data(URLs.MNIST_TINY) dls = ImageDataLoaders.from_csv(path) # Log al...
[ "neptune.new.integrations.fastai.NeptuneCallback", "neptune.new.init" ]
[((143, 234), 'neptune.new.init', 'neptune.init', ([], {'project': '"""common/fastai-integration"""', 'api_token': '"""<PASSWORD>"""', 'tags': '"""basic"""'}), "(project='common/fastai-integration', api_token='<PASSWORD>',\n tags='basic')\n", (155, 234), True, 'import neptune.new as neptune\n'), ((393, 446), 'neptun...
import unittest import numpy as np from xcube.webapi.controllers.time_series import get_time_series_info, get_time_series_for_point, \ get_time_series_for_geometry, get_time_series_for_geometry_collection from ..helpers import new_test_service_context class TimeSeriesControllerTest(unittest.TestCase): def ...
[ "xcube.webapi.controllers.time_series.get_time_series_for_point", "xcube.webapi.controllers.time_series.get_time_series_info", "numpy.datetime64" ]
[((441, 514), 'xcube.webapi.controllers.time_series.get_time_series_for_point', 'get_time_series_for_point', (['ctx', '"""demo"""', '"""conc_tsm"""'], {'lon': '(-150.0)', 'lat': '(-30.0)'}), "(ctx, 'demo', 'conc_tsm', lon=-150.0, lat=-30.0)\n", (466, 514), False, 'from xcube.webapi.controllers.time_series import get_ti...
from math import sqrt from math import atan2 from math import asin beta = 0.1 sampleFreq = 10.0 #Fastest implementation in python for invsqrt def invsqrt(number): return number ** -0.5 def update_IMU( gx, gy, gz, ax, ay, az, q0, q1, q2, q3): gx = gx * 0.0174533 gy = gy * 0.0174533 gz = gz * 0.017453...
[ "math.asin", "math.sqrt", "math.atan2" ]
[((6068, 6117), 'math.atan2', 'atan2', (['(q0 * q1 + q2 * q3)', '(0.5 - q1 * q1 - q2 * q2)'], {}), '(q0 * q1 + q2 * q3, 0.5 - q1 * q1 - q2 * q2)\n', (6073, 6117), False, 'from math import atan2\n'), ((6120, 6152), 'math.asin', 'asin', (['(-2.0 * (q1 * q3 - q0 * q2))'], {}), '(-2.0 * (q1 * q3 - q0 * q2))\n', (6124, 6152...
import setuptools if __name__ == '__main__': setuptools.setup( name='Name', version='0.1', # this automatically detects the packages in the specified # (or current directory if no directory is given). packages=setuptools.find_packages(exclude=['tests', 'docs']), # ...
[ "setuptools.find_packages" ]
[((256, 307), 'setuptools.find_packages', 'setuptools.find_packages', ([], {'exclude': "['tests', 'docs']"}), "(exclude=['tests', 'docs'])\n", (280, 307), False, 'import setuptools\n')]
""" # Interaction Tracker # @license http://www.apache.org/licenses/LICENSE-2.0 # Author @ <NAME>, Zaki """ from analytics.models import (Log, ActionLog) from rest_framework import serializers class LogSerializer(serializers.ModelSerializer): class Meta: model = Log fields = ('app','appuser',...
[ "analytics.models.ActionLog.objects.create", "rest_framework.serializers.HiddenField", "analytics.models.Log.objects.create" ]
[((658, 700), 'rest_framework.serializers.HiddenField', 'serializers.HiddenField', ([], {'default': '"""Request"""'}), "(default='Request')\n", (681, 700), False, 'from rest_framework import serializers\n'), ((891, 933), 'analytics.models.ActionLog.objects.create', 'ActionLog.objects.create', ([], {}), '(**validated_da...
""" Description: Requirements: pySerial, wxPython Phoenix glossary and of other descriptions: DMM - digital multimeter PSU - power supply SBC - single board computer INS - general instrument commands GEN - general sequence instructions """ import json import logging import serial import serialfunctions as sf imp...
[ "wx.Button", "wx.BoxSizer", "time.sleep", "wx.StaticBoxSizer", "wx.StaticText", "wx.TextCtrl", "serial.Serial", "wx.Panel.__init__" ]
[((624, 655), 'wx.Panel.__init__', 'wx.Panel.__init__', (['self', 'parent'], {}), '(self, parent)\n', (641, 655), False, 'import wx\n'), ((932, 956), 'wx.BoxSizer', 'wx.BoxSizer', (['wx.VERTICAL'], {}), '(wx.VERTICAL)\n', (943, 956), False, 'import wx\n'), ((983, 1009), 'wx.BoxSizer', 'wx.BoxSizer', (['wx.HORIZONTAL'],...
# -*- coding: utf-8 -*- """\ Coroutine utilities ------------------- Some code snippets inspired by http://www.dabeaz.com/coroutines/ """ import re import functools def coroutine(func): """Prime a coroutine for send commands. Args: func (coroutine): A function that takes values via yield Retur...
[ "functools.wraps", "re.compile" ]
[((381, 402), 'functools.wraps', 'functools.wraps', (['func'], {}), '(func)\n', (396, 402), False, 'import functools\n'), ((1371, 1403), 're.compile', 're.compile', (['pattern'], {'flags': 'flags'}), '(pattern, flags=flags)\n', (1381, 1403), False, 'import re\n')]
from sqlalchemy import Boolean, Column, ForeignKey, Integer, String, DateTime from sqlalchemy.orm import relationship import datetime from database import Base class Org(Base): __tablename__ = "orgs" id = Column(Integer, primary_key=True, index=True) name = Column(String, unique=True, index=True) cr...
[ "sqlalchemy.orm.relationship", "sqlalchemy.ForeignKey", "sqlalchemy.Column" ]
[((217, 262), 'sqlalchemy.Column', 'Column', (['Integer'], {'primary_key': '(True)', 'index': '(True)'}), '(Integer, primary_key=True, index=True)\n', (223, 262), False, 'from sqlalchemy import Boolean, Column, ForeignKey, Integer, String, DateTime\n'), ((274, 313), 'sqlalchemy.Column', 'Column', (['String'], {'unique'...
#!/usr/bin/python3 import sys import time import array import numpy as np import pandas as pd import statistics import matplotlib.pyplot as plt import seaborn as sns # sns.set_theme(style="darkgrid") x_b = [1, 10, 100, 1000, 10000, 100000, 1000000] cyc_pi2 = [8379072, 8379072, 3675200, 372864, 37312, 3728, 368] cyc...
[ "matplotlib.pyplot.grid", "matplotlib.pyplot.ylabel", "matplotlib.pyplot.gca", "matplotlib.pyplot.xlabel", "matplotlib.pyplot.bar", "matplotlib.pyplot.title", "matplotlib.pyplot.yscale", "matplotlib.pyplot.legend", "matplotlib.pyplot.show" ]
[((1063, 1072), 'matplotlib.pyplot.gca', 'plt.gca', ([], {}), '()\n', (1070, 1072), True, 'import matplotlib.pyplot as plt\n'), ((1200, 1252), 'matplotlib.pyplot.bar', 'plt.bar', (['r1', 'cyc_lap'], {'width': 'barWidth', 'label': '"""Laptop"""'}), "(r1, cyc_lap, width=barWidth, label='Laptop')\n", (1207, 1252), True, '...
from flask_bcrypt import Bcrypt from flask_caching import Cache from flask_debugtoolbar import DebugToolbarExtension from flask_login import LoginManager from flask_migrate import Migrate from flask_sqlalchemy import SQLAlchemy import logging bcrypt = Bcrypt() login_manager = LoginManager() db = SQLAlchemy() migrate =...
[ "logging.getLogger", "flask_login.LoginManager", "flask_debugtoolbar.DebugToolbarExtension", "flask_bcrypt.Bcrypt", "flask_caching.Cache", "flask_migrate.Migrate", "flask_sqlalchemy.SQLAlchemy" ]
[((253, 261), 'flask_bcrypt.Bcrypt', 'Bcrypt', ([], {}), '()\n', (259, 261), False, 'from flask_bcrypt import Bcrypt\n'), ((278, 292), 'flask_login.LoginManager', 'LoginManager', ([], {}), '()\n', (290, 292), False, 'from flask_login import LoginManager\n'), ((298, 310), 'flask_sqlalchemy.SQLAlchemy', 'SQLAlchemy', ([]...
import time def f(): [ # Must be split over multiple lines to see the error. # https://github.com/benfred/py-spy/pull/208 time.sleep(1) for _ in range(1000) ] f()
[ "time.sleep" ]
[((152, 165), 'time.sleep', 'time.sleep', (['(1)'], {}), '(1)\n', (162, 165), False, 'import time\n')]
# -*- coding: utf-8 -*- """ Microsoft-Windows-UAC-FileVirtualization GUID : c02afc2b-e24e-4449-ad76-bcc2c2575ead """ from construct import Int8sl, Int8ul, Int16ul, Int16sl, Int32sl, Int32ul, Int64sl, Int64ul, Bytes, Double, Float32l, Struct from etl.utils import WString, CString, SystemTime, Guid from etl.dtyp import S...
[ "construct.Bytes", "construct.Struct", "etl.parsers.etw.core.guid" ]
[((537, 562), 'construct.Struct', 'Struct', (["('Error' / Int32ul)"], {}), "('Error' / Int32ul)\n", (543, 562), False, 'from construct import Int8sl, Int8ul, Int16ul, Int16sl, Int32sl, Int32ul, Int64sl, Int64ul, Bytes, Double, Float32l, Struct\n'), ((739, 764), 'construct.Struct', 'Struct', (["('Error' / Int32ul)"], {}...
""" Processing data in win32 format. """ import glob import logging import math import os import subprocess import tempfile from fnmatch import fnmatch from multiprocessing import Pool, cpu_count from subprocess import DEVNULL, PIPE, Popen # Setup the logger FORMAT = "[%(asctime)s] %(levelname)s: %(message)s" logging....
[ "logging.basicConfig", "os.path.exists", "logging.getLogger", "os.makedirs", "math.pow", "subprocess.Popen", "os.rename", "os.path.join", "math.sqrt", "multiprocessing.cpu_count", "os.path.dirname", "subprocess.call", "tempfile.NamedTemporaryFile", "glob.glob" ]
[((312, 400), 'logging.basicConfig', 'logging.basicConfig', ([], {'level': 'logging.INFO', 'format': 'FORMAT', 'datefmt': '"""%Y-%m-%d %H:%M:%S"""'}), "(level=logging.INFO, format=FORMAT, datefmt=\n '%Y-%m-%d %H:%M:%S')\n", (331, 400), False, 'import logging\n'), ((405, 432), 'logging.getLogger', 'logging.getLogger'...
import asyncio import aiopg dsn = 'dbname=aiopg user=aiopg password=<PASSWORD> host=127.0.0.1' @asyncio.coroutine def test_select(): pool = yield from aiopg.create_pool(dsn) with (yield from pool.cursor()) as cur: yield from cur.execute("SELECT 1") ret = yield from cur.fetchone() asse...
[ "asyncio.get_event_loop", "aiopg.create_pool" ]
[((366, 390), 'asyncio.get_event_loop', 'asyncio.get_event_loop', ([], {}), '()\n', (388, 390), False, 'import asyncio\n'), ((158, 180), 'aiopg.create_pool', 'aiopg.create_pool', (['dsn'], {}), '(dsn)\n', (175, 180), False, 'import aiopg\n')]
from collections import defaultdict import io import hashlib from datetime import date, datetime from pyexcel_xls import get_data as xls_get import pandas import magic from contextlib import closing import csv from django.db import connection from io import StringIO import uuid from psycopg2.errors import UniqueViolati...
[ "care.facility.api.serializers.patient_external_test.PatientExternalTestSerializer", "care.facility.models.PatientExternalTest.CSV_MAKE_PRETTY.copy", "pandas.read_csv", "rest_framework.exceptions.ValidationError", "rest_framework.decorators.action", "pyexcel_xls.get_data", "core.fernet.FernetEncryption"...
[((1570, 1587), 'collections.defaultdict', 'defaultdict', (['list'], {}), '(list)\n', (1581, 1587), False, 'from collections import defaultdict\n'), ((1609, 1654), 'care.facility.models.PatientExternalTest.HEADER_CSV_MAPPING.keys', 'PatientExternalTest.HEADER_CSV_MAPPING.keys', ([], {}), '()\n', (1652, 1654), False, 'f...
"""Converts synonyms into SMILES for the data from Gerber's paper.""" # data/hsd11b1_validation/get_smiles_cactus.py from io import BytesIO import pandas as pd import pycurl def getsmiles_cactus(name): """Converts synonyms into SMILES strings. A function to use the public cactus (National Institutes of Canc...
[ "pycurl.Curl", "io.BytesIO", "pandas.read_csv" ]
[((615, 624), 'io.BytesIO', 'BytesIO', ([], {}), '()\n', (622, 624), False, 'from io import BytesIO\n'), ((633, 646), 'pycurl.Curl', 'pycurl.Curl', ([], {}), '()\n', (644, 646), False, 'import pycurl\n'), ((934, 961), 'pandas.read_csv', 'pd.read_csv', (['data'], {'sep': '"""\t"""'}), "(data, sep='\\t')\n", (945, 961), ...
#! /usr/bin/env python3 # -*- coding: utf-8 -*- # File : eval-referential.py # Author : <NAME>, <NAME> # Email : <EMAIL>, <EMAIL> # Date : 30.07.2019 # Last Modified Date: 16.10.2019 # Last Modified By : Chi Han, Jiayuan Mao # # This file is part of the VCML codebase # ...
[ "jacinle.io.load_json", "IPython.core.ultratb.FormattedTB", "jacinle.utils.meter.GroupMeters", "jacinle.random.randint", "jacinle.cli.argument.JacArgumentParser", "jacinle.random.choice_list", "jacinle.io.load", "numpy.fmin", "jacinle.utils.tqdm.tqdm_gofor" ]
[((890, 960), 'IPython.core.ultratb.FormattedTB', 'ultratb.FormattedTB', ([], {'mode': '"""Plain"""', 'color_scheme': '"""Linux"""', 'call_pdb': '(True)'}), "(mode='Plain', color_scheme='Linux', call_pdb=True)\n", (909, 960), False, 'from IPython.core import ultratb\n'), ((976, 995), 'jacinle.cli.argument.JacArgumentPa...
import pytest @pytest.fixture(scope="module") def client(looper, txnPoolNodeSet, client1, client1Connected): return client1Connected
[ "pytest.fixture" ]
[((17, 47), 'pytest.fixture', 'pytest.fixture', ([], {'scope': '"""module"""'}), "(scope='module')\n", (31, 47), False, 'import pytest\n')]
#!/usr/bin/env python # # Copyright (c) 2018 10X Genomics, Inc. All rights reserved. # # Utils for feature-barcoding technology import numpy as np import os import json import tenkit.safe_json as tk_safe_json def check_if_none_or_empty(matrix): if matrix is None or matrix.get_shape()[0] == 0 or matrix.get_shape(...
[ "os.path.isfile", "tenkit.safe_json.json_sanitize" ]
[((1193, 1214), 'os.path.isfile', 'os.path.isfile', (['fpath'], {}), '(fpath)\n', (1207, 1214), False, 'import os\n'), ((494, 532), 'tenkit.safe_json.json_sanitize', 'tk_safe_json.json_sanitize', (['input_dict'], {}), '(input_dict)\n', (520, 532), True, 'import tenkit.safe_json as tk_safe_json\n')]
import json import rimu from rimu import options def unexpectedError(_, message): raise Exception(f'unexpected callback: {message}') def test_render(): assert rimu.render('Hello World!') == '<p>Hello World!</p>' def test_jsonTests(): with open('./tests/rimu-tests.json') as f: data = json.load...
[ "json.load", "rimu.render", "rimu.RenderOptions" ]
[((172, 199), 'rimu.render', 'rimu.render', (['"""Hello World!"""'], {}), "('Hello World!')\n", (183, 199), False, 'import rimu\n'), ((311, 323), 'json.load', 'json.load', (['f'], {}), '(f)\n', (320, 323), False, 'import json\n'), ((599, 619), 'rimu.RenderOptions', 'rimu.RenderOptions', ([], {}), '()\n', (617, 619), Fa...
# -*- coding: utf-8 -*- """ app_test.py Tests the tkit.App class. Author: <NAME>; Oct 2017 License: MIT """ import tkit if __name__ == "__main__": # Create app test_app = tkit.App("Test App", 250, 100) # Create and customize menubar menubar = tkit.Menubar() menubar.add_menu("File") #test_men...
[ "tkit.Popup", "tkit.Menubar", "tkit.BrowseFile", "tkit.App" ]
[((183, 213), 'tkit.App', 'tkit.App', (['"""Test App"""', '(250)', '(100)'], {}), "('Test App', 250, 100)\n", (191, 213), False, 'import tkit\n'), ((263, 277), 'tkit.Menubar', 'tkit.Menubar', ([], {}), '()\n', (275, 277), False, 'import tkit\n'), ((652, 669), 'tkit.BrowseFile', 'tkit.BrowseFile', ([], {}), '()\n', (667...
""" Database models """ from typing import Tuple import attr import sqlalchemy as sa from .settings import DATCORE_STR, SIMCORE_S3_ID, SIMCORE_S3_STR #FIXME: W0611:Unused UUID imported from sqlalchemy.dialects.postgresql #from sqlalchemy.dialects.postgresql import UUID #FIXME: R0902: Too many instance attributes (...
[ "sqlalchemy.MetaData", "attr.s", "sqlalchemy.Column" ]
[((392, 405), 'sqlalchemy.MetaData', 'sa.MetaData', ([], {}), '()\n', (403, 405), True, 'import sqlalchemy as sa\n'), ((2050, 2075), 'attr.s', 'attr.s', ([], {'auto_attribs': '(True)'}), '(auto_attribs=True)\n', (2056, 2075), False, 'import attr\n'), ((487, 538), 'sqlalchemy.Column', 'sa.Column', (['"""file_uuid"""', '...
import pandas as pd from sklearn.pipeline import Pipeline from sklearn.ensemble import GradientBoostingClassifier from sklearn.metrics import confusion_matrix, roc_auc_score from category_encoders import MEstimateEncoder import numpy as np from collections import defaultdict import os from sklearn.metrics import roc_au...
[ "numpy.log", "sklearn.metrics.roc_auc_score", "numpy.arange", "numpy.histogram", "numpy.max", "numpy.empty", "numpy.min", "pandas.DataFrame", "sklearn.metrics.confusion_matrix", "numpy.abs", "sklearn.model_selection.train_test_split", "sklearn.pipeline.Pipeline", "sklearn.ensemble.GradientBo...
[((492, 539), 'sklearn.pipeline.Pipeline', 'Pipeline', (["[('encoder', enc), ('model', modelo)]"], {}), "([('encoder', enc), ('model', modelo)])\n", (500, 539), False, 'from sklearn.pipeline import Pipeline\n'), ((2336, 2365), 'sklearn.metrics.confusion_matrix', 'confusion_matrix', (['preds', 'true'], {}), '(preds, tru...
# Solution of; # Project Euler Problem 527: Randomized Binary Search # https://projecteuler.net/problem=527 # # A secret integer t is selected at random within the range 1 ≤ t ≤ n. The # goal is to guess the value of t by making repeated guesses, via integer g. # After a guess is made, there are three possible outco...
[ "timed.caller" ]
[((1894, 1928), 'timed.caller', 'timed.caller', (['dummy', 'n', 'i', 'prob_id'], {}), '(dummy, n, i, prob_id)\n', (1906, 1928), False, 'import timed\n')]
import asyncio from aiohttp import web from concurrent.futures import ThreadPoolExecutor, ProcessPoolExecutor from multiprocessing import Queue, Process import os from time import sleep async def handle(request): index = open("index.html", 'rb') content = index.read() return web.Response(body=content, co...
[ "aiohttp.web.run_app", "concurrent.futures.ThreadPoolExecutor", "multiprocessing.Process", "aiohttp.web.Response", "aiohttp.web.Application", "asyncio.wait", "time.sleep", "asyncio.Condition", "os.getpid", "multiprocessing.Queue", "asyncio.get_event_loop", "aiohttp.web.WebSocketResponse" ]
[((353, 372), 'asyncio.Condition', 'asyncio.Condition', ([], {}), '()\n', (370, 372), False, 'import asyncio\n'), ((2132, 2156), 'asyncio.get_event_loop', 'asyncio.get_event_loop', ([], {}), '()\n', (2154, 2156), False, 'import asyncio\n'), ((2168, 2201), 'concurrent.futures.ThreadPoolExecutor', 'ThreadPoolExecutor', (...
import pytest from fixtures import world from wecs.core import UID from wecs.core import NoSuchUID from wecs.core import Component @Component() class Reference: uid: UID def test_user_defined_names(world): entity = world.create_entity(name="foo") assert entity._uid.name == "foo" def test_automatic_n...
[ "wecs.core.Component", "fixtures.world.create_entity", "fixtures.world.flush_component_updates", "wecs.core.UID", "pytest.raises" ]
[((136, 147), 'wecs.core.Component', 'Component', ([], {}), '()\n', (145, 147), False, 'from wecs.core import Component\n'), ((229, 260), 'fixtures.world.create_entity', 'world.create_entity', ([], {'name': '"""foo"""'}), "(name='foo')\n", (248, 260), False, 'from fixtures import world\n'), ((346, 367), 'fixtures.world...
from selenium import webdriver from selenium.webdriver.chrome.options import Options from webdriver_manager.chrome import ChromeDriverManager from pages.home_page import HomePage from pages.profile_page import ProfilePage from pages.login_page import LoginPage from pages.registration_page import RegistrationPage from p...
[ "selenium.webdriver.chrome.options.Options", "pages.profile_page.ProfilePage", "pages.home_page.HomePage", "pages.login_page.LoginPage", "pages.navigation_bar.NavigationBar", "pages.article_page.ArticlePage", "pages.new_article_page.NewArticlePage", "webdriver_manager.chrome.ChromeDriverManager" ]
[((498, 507), 'selenium.webdriver.chrome.options.Options', 'Options', ([], {}), '()\n', (505, 507), False, 'from selenium.webdriver.chrome.options import Options\n'), ((908, 937), 'pages.home_page.HomePage', 'HomePage', ([], {'driver': 'self.browser'}), '(driver=self.browser)\n', (916, 937), False, 'from pages.home_pag...
import numpy as np import numpy.linalg as LA from .solve_R1 import problem_R1, Classo_R1, pathlasso_R1 from .solve_R2 import problem_R2, Classo_R2, pathlasso_R2 from .solve_R3 import problem_R3, Classo_R3, pathlasso_R3 from .solve_R4 import problem_R4, Classo_R4, pathlasso_R4 from .path_alg import solve_path, pathalgo...
[ "numpy.vdot", "numpy.array", "numpy.linspace", "numpy.mean" ]
[((9722, 9762), 'numpy.array', 'np.array', (['[(beta / ww) for beta in BETA]'], {}), '([(beta / ww) for beta in BETA])\n', (9730, 9762), True, 'import numpy as np\n'), ((9864, 9878), 'numpy.array', 'np.array', (['BETA'], {}), '(BETA)\n', (9872, 9878), True, 'import numpy as np\n'), ((5733, 5757), 'numpy.linspace', 'np....
from tensorpy import image_base classifications = image_base.classify_folder_images('./images') print("*** Displaying Image Classification Results as a list: ***") for classification in classifications: print(classification)
[ "tensorpy.image_base.classify_folder_images" ]
[((51, 96), 'tensorpy.image_base.classify_folder_images', 'image_base.classify_folder_images', (['"""./images"""'], {}), "('./images')\n", (84, 96), False, 'from tensorpy import image_base\n')]
import argparse import torch from torch.utils.data import DataLoader import sys, os sys.path.insert(0, os.path.join( os.path.dirname(os.path.realpath(__file__)), "../../")) from deep_audio_features.dataloading.dataloading import FeatureExtractorDataset from deep_audio_features.models.cnn import load_cnn from deep_a...
[ "deep_audio_features.models.cnn.load_cnn", "deep_audio_features.utils.model_editing.drop_layers", "argparse.ArgumentParser", "deep_audio_features.dataloading.dataloading.FeatureExtractorDataset", "deep_audio_features.lib.training.test", "os.path.realpath", "numpy.array", "torch.cuda.is_available", "...
[((1363, 1382), 'deep_audio_features.models.cnn.load_cnn', 'load_cnn', (['modelpath'], {}), '(modelpath)\n', (1371, 1382), False, 'from deep_audio_features.models.cnn import load_cnn\n'), ((1621, 1655), 'deep_audio_features.utils.model_editing.drop_layers', 'drop_layers', (['model', 'layers_dropped'], {}), '(model, lay...
import re import traceback from textwrap import dedent def camel_to_snake(value): s1 = re.sub('(.)([A-Z][a-z]+)', r'\1_\2', value) return re.sub('([a-z0-9])([A-Z])', r'\1_\2', s1).lower() def snake_to_camel(value): camel = '' words = value.split('_') for w in words: camel += w.title() ...
[ "re.sub", "re.escape", "traceback.extract_stack" ]
[((93, 137), 're.sub', 're.sub', (['"""(.)([A-Z][a-z]+)"""', '"""\\\\1_\\\\2"""', 'value'], {}), "('(.)([A-Z][a-z]+)', '\\\\1_\\\\2', value)\n", (99, 137), False, 'import re\n'), ((973, 995), 're.escape', 're.escape', (['replacement'], {}), '(replacement)\n', (982, 995), False, 'import re\n'), ((148, 190), 're.sub', 'r...
import sys import os import random klasorAdi = os.path.dirname(sys.argv[0]) dosyaIsmi = klasorAdi + "/test.txt" soruSayisi = 40 ogrenciSayisi = 60 d = {} dogruSayisi = {} yalisSayisi = {} bosSayisi = {} puan = {} def sinavHazirla(): for j in range(1, soruSayisi + 1): r1 = random.randint(1, 5) d[0, j] = chr(...
[ "os.path.isfile", "os.path.dirname", "random.randint" ]
[((47, 75), 'os.path.dirname', 'os.path.dirname', (['sys.argv[0]'], {}), '(sys.argv[0])\n', (62, 75), False, 'import os\n'), ((281, 301), 'random.randint', 'random.randint', (['(1)', '(5)'], {}), '(1, 5)\n', (295, 301), False, 'import random\n'), ((2184, 2209), 'os.path.isfile', 'os.path.isfile', (['dosyaIsmi'], {}), '...
from twisted.plugin import IPlugin from heufybot.moduleinterface import IBotModule from heufybot.modules.commandinterface import BotCommand from heufybot.utils.timeutils import now, timestamp from zope.interface import implements from datetime import datetime class TimeCommand(BotCommand): implements(IPlugin, IBo...
[ "heufybot.utils.timeutils.now", "zope.interface.implements" ]
[((297, 328), 'zope.interface.implements', 'implements', (['IPlugin', 'IBotModule'], {}), '(IPlugin, IBotModule)\n', (307, 328), False, 'from zope.interface import implements\n'), ((4317, 4322), 'heufybot.utils.timeutils.now', 'now', ([], {}), '()\n', (4320, 4322), False, 'from heufybot.utils.timeutils import now, time...
import pandas as pd import click import collections def kmer_suffix(kmer): return kmer[1:] def kmer_prefix(kmer): return kmer[:-1] def chunks(l, n): """Yield successive n-sized chunks from l.""" for i in range(0, len(l), n): yield l[i:i + n] def build_graph(kmers): graph = collection...
[ "click.option", "click.command", "collections.defaultdict", "pandas.read_csv" ]
[((1789, 1804), 'click.command', 'click.command', ([], {}), '()\n', (1802, 1804), False, 'import click\n'), ((1806, 1868), 'click.option', 'click.option', (['"""--fin"""'], {'type': 'str', 'default': '"""problem11_input.tsv"""'}), "('--fin', type=str, default='problem11_input.tsv')\n", (1818, 1868), False, 'import clic...
#!/usr/bin/env python3 import sys import re import time import datetime import os for module in sorted(sys.modules): print("%-20s : %s" % (module, sys.modules[module])) print('USER : ', os.environ['USER']) print('PWD : ', os.environ['PWD']) print('PYTHONPATH: ', os.environ.get('PYTHONPATH')) print(sy...
[ "os.environ.get" ]
[((281, 309), 'os.environ.get', 'os.environ.get', (['"""PYTHONPATH"""'], {}), "('PYTHONPATH')\n", (295, 309), False, 'import os\n')]
# SPDX-FileCopyrightText: 2020 <NAME> <<EMAIL>> # # SPDX-License-Identifier: Apache-2.0 from unittest import mock import pytest from codeprep.bpepkg.bpe_config import BpeConfig, BpeParam, BpeConfigNotSupported from codeprep.pipeline.bpelearner import run @mock.patch('codeprep.pipeline.bpelearner.Dataset', autospec...
[ "codeprep.pipeline.bpelearner.run", "codeprep.bpepkg.bpe_config.BpeConfig", "unittest.mock.patch", "pytest.raises" ]
[((261, 326), 'unittest.mock.patch', 'mock.patch', (['"""codeprep.pipeline.bpelearner.Dataset"""'], {'autospec': '(True)'}), "('codeprep.pipeline.bpelearner.Dataset', autospec=True)\n", (271, 326), False, 'from unittest import mock\n'), ((621, 686), 'unittest.mock.patch', 'mock.patch', (['"""codeprep.pipeline.bpelearne...
#!/usr/bin/env python3 import torrent_parser as tp import asyncio import contextlib import pathlib import argparse import pprint import hashlib import concurrent.futures import os.path import logging import tqdm class TorrentChecker(object): def __init__(self, datadir=pathlib.Path('.'), data_file_globs=["**"], ...
[ "logging.getLogger", "asyncio.CancelledError", "torrent_parser.parse_torrent_file", "argparse.ArgumentParser", "pathlib.Path", "pathlib.PurePath", "hashlib.sha1" ]
[((5286, 5351), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {'description': '"""Verify downloaded torrents"""'}), "(description='Verify downloaded torrents')\n", (5309, 5351), False, 'import argparse\n'), ((272, 289), 'pathlib.Path', 'pathlib.Path', (['"""."""'], {}), "('.')\n", (284, 289), False, 'impor...
import os _rootdir = os.getcwd() def find_rootdir(filenames = ('__main__.py', 'main.ipynb')): path = os.getcwd() while os.path.isdir(path): ls = os.listdir(path) if any([f in ls for f in filenames]): return os.path.abspath(path) else: path += '/..' # nothi...
[ "os.path.abspath", "os.listdir", "os.path.isdir", "os.getcwd" ]
[((22, 33), 'os.getcwd', 'os.getcwd', ([], {}), '()\n', (31, 33), False, 'import os\n'), ((107, 118), 'os.getcwd', 'os.getcwd', ([], {}), '()\n', (116, 118), False, 'import os\n'), ((130, 149), 'os.path.isdir', 'os.path.isdir', (['path'], {}), '(path)\n', (143, 149), False, 'import os\n'), ((371, 382), 'os.getcwd', 'os...
import math import random import numpy as np # 先生成一个随机的信源 def random_sources(): random_sources = random.randint(0, 16) print('这个随机数是', random_sources) return hanming(random_sources) # return bin(int(random_sources)...
[ "random.randint" ]
[((191, 212), 'random.randint', 'random.randint', (['(0)', '(16)'], {}), '(0, 16)\n', (205, 212), False, 'import random\n')]
from django.db import models from users.models import User class Assignment(models.Model): title = models.CharField(max_length=50) teacher = models.ForeignKey(User, on_delete=models.CASCADE) def __str__(self): return self.title class GradedAssignment(models.Model): student = models.ForeignK...
[ "django.db.models.FloatField", "django.db.models.ForeignKey", "django.db.models.ManyToManyField", "django.db.models.SmallIntegerField", "django.db.models.CharField" ]
[((105, 136), 'django.db.models.CharField', 'models.CharField', ([], {'max_length': '(50)'}), '(max_length=50)\n', (121, 136), False, 'from django.db import models\n'), ((151, 200), 'django.db.models.ForeignKey', 'models.ForeignKey', (['User'], {'on_delete': 'models.CASCADE'}), '(User, on_delete=models.CASCADE)\n', (16...
from fastapi import FastAPI from . import api app = FastAPI(debug=True) app.include_router(api.router)
[ "fastapi.FastAPI" ]
[((55, 74), 'fastapi.FastAPI', 'FastAPI', ([], {'debug': '(True)'}), '(debug=True)\n', (62, 74), False, 'from fastapi import FastAPI\n')]
#!/usr/bin/env python """ Perform compressed sensing analysis on a dax file using the homotopy approach. Return the results in hres image format and as a list of object locations. Hazen 09/12 """ import numpy import storm_analysis.sa_library.datareader as datareader import storm_analysis.sa_library.parameters as par...
[ "storm_analysis.sa_library.parameters.ParametersL1H", "storm_analysis.sa_library.datareader.inferReader", "argparse.ArgumentParser", "storm_analysis.L1H.setup_A_matrix.loadAMatrix", "storm_analysis.sa_library.writeinsight3.I3Writer" ]
[((660, 694), 'storm_analysis.sa_library.datareader.inferReader', 'datareader.inferReader', (['movie_name'], {}), '(movie_name)\n', (682, 694), True, 'import storm_analysis.sa_library.datareader as datareader\n'), ((823, 855), 'storm_analysis.sa_library.writeinsight3.I3Writer', 'writeinsight3.I3Writer', (['bin_name'], ...
""" Methods for handling DB creation and CRUD operations in Sqlite3. """ # Standard library imports import logging import sqlite3 # Local application imports from ism.exceptions.exceptions import UnrecognisedParameterisationCharacter from ism.interfaces.dao_interface import DAOInterface class Sqlite3DAO(DAOInterfac...
[ "logging.getLogger", "logging.error", "sqlite3.connect", "ism.exceptions.exceptions.UnrecognisedParameterisationCharacter" ]
[((617, 664), 'logging.getLogger', 'logging.getLogger', (['"""ism.sqlite3_dao.Sqlite3DAO"""'], {}), "('ism.sqlite3_dao.Sqlite3DAO')\n", (634, 664), False, 'import logging\n'), ((2365, 2394), 'sqlite3.connect', 'sqlite3.connect', (['self.db_path'], {}), '(self.db_path)\n', (2380, 2394), False, 'import sqlite3\n'), ((152...
# Author: btjanaka (<NAME>) # Problem: (UVa) 247 import sys from collections import defaultdict def kosaraju(g, g_rev): order = [] visited = set() def visit(u): visited.add(u) for v in g[u]: if v not in visited: visit(v) order.append(u) for u in g...
[ "collections.defaultdict" ]
[((865, 881), 'collections.defaultdict', 'defaultdict', (['set'], {}), '(set)\n', (876, 881), False, 'from collections import defaultdict\n'), ((883, 899), 'collections.defaultdict', 'defaultdict', (['set'], {}), '(set)\n', (894, 899), False, 'from collections import defaultdict\n')]
import pyautogui import time time.sleep(3) print(pyautogui.position())
[ "pyautogui.position", "time.sleep" ]
[((30, 43), 'time.sleep', 'time.sleep', (['(3)'], {}), '(3)\n', (40, 43), False, 'import time\n'), ((50, 70), 'pyautogui.position', 'pyautogui.position', ([], {}), '()\n', (68, 70), False, 'import pyautogui\n')]
from jmap.account.imap.imap_utf7 import imap_utf7_decode, imap_utf7_encode KNOWN_SPECIALS = set('\\HasChildren \\HasNoChildren \\NoSelect \\NoInferiors \\UnMarked \\Subscribed'.lower().split()) # special use or name magic ROLE_MAP = { 'inbox': 'inbox', 'drafts': 'drafts', 'draft': 'drafts', 'draft messages':...
[ "jmap.account.imap.imap_utf7.imap_utf7_encode" ]
[((2719, 2749), 'jmap.account.imap.imap_utf7.imap_utf7_encode', 'imap_utf7_encode', (["self['name']"], {}), "(self['name'])\n", (2735, 2749), False, 'from jmap.account.imap.imap_utf7 import imap_utf7_decode, imap_utf7_encode\n')]
from pyieee1905.ieee1905_tlv import IEEE1905_TLV from scapy.packet import Packet, bind_layers from scapy.fields import BitField, XByteField, XShortField, XShortEnumField from scapy.layers.l2 import Ether IEEE1905_MCAST = "01:80:c2:00:00:13" ieee1905_msg_type = { 0x0000:"TOPOLOGY_DISCOVERY_MESSAGE", 0x0001:...
[ "scapy.packet.bind_layers", "scapy.fields.BitField", "scapy.fields.XShortField", "scapy.fields.XByteField", "scapy.fields.XShortEnumField" ]
[((2890, 2937), 'scapy.packet.bind_layers', 'bind_layers', (['Ether', 'MultiAP_Message'], {'type': '(35130)'}), '(Ether, MultiAP_Message, type=35130)\n', (2901, 2937), False, 'from scapy.packet import Packet, bind_layers\n'), ((2939, 2981), 'scapy.packet.bind_layers', 'bind_layers', (['MultiAP_Message', 'IEEE1905_TLV']...
""" Command line management utilities for ITEEBot. This module's command line interface will act as the bot's entry point when installed. """ import click from . import configurator as conf from . import database as db from .bot import ITEEBot @click.group() def cli(): pass @click.command("init-config") @click.a...
[ "click.group", "click.option", "click.command", "click.Path" ]
[((247, 260), 'click.group', 'click.group', ([], {}), '()\n', (258, 260), False, 'import click\n'), ((283, 311), 'click.command', 'click.command', (['"""init-config"""'], {}), "('init-config')\n", (296, 311), False, 'import click\n'), ((770, 794), 'click.command', 'click.command', (['"""init-db"""'], {}), "('init-db')\...