code
stringlengths
22
1.05M
apis
listlengths
1
3.31k
extract_api
stringlengths
75
3.25M
# coding=utf-8 import tensorflow as tf v = tf.Variable(0, dtype=tf.float32, name='v3') # 在没有声明滑动平均模型时只有一个变量v,所以下面的语句只会输出v:0 for variables in tf.global_variables(): print(variables.name) ema = tf.train.ExponentialMovingAverage(0.99) # 加入命名空间中 maintain_averages_op = ema.apply(tf.global_variables()) # 在申明滑动平均模型之后,T...
[ "tensorflow.Variable", "tensorflow.Session", "tensorflow.train.Saver", "tensorflow.global_variables", "tensorflow.global_variables_initializer", "tensorflow.assign", "tensorflow.train.ExponentialMovingAverage" ]
[((44, 87), 'tensorflow.Variable', 'tf.Variable', (['(0)'], {'dtype': 'tf.float32', 'name': '"""v3"""'}), "(0, dtype=tf.float32, name='v3')\n", (55, 87), True, 'import tensorflow as tf\n'), ((143, 164), 'tensorflow.global_variables', 'tf.global_variables', ([], {}), '()\n', (162, 164), True, 'import tensorflow as tf\n'...
""" validataclass Copyright (c) 2021, binary butterfly GmbH and contributors Use of this source code is governed by an MIT-style license that can be found in the LICENSE file. """ from datetime import date from typing import Any from .string_validator import StringValidator from validataclass.exceptions import Invali...
[ "validataclass.exceptions.InvalidDateError", "datetime.date.fromisoformat" ]
[((1329, 1360), 'datetime.date.fromisoformat', 'date.fromisoformat', (['date_string'], {}), '(date_string)\n', (1347, 1360), False, 'from datetime import date\n'), ((1406, 1424), 'validataclass.exceptions.InvalidDateError', 'InvalidDateError', ([], {}), '()\n', (1422, 1424), False, 'from validataclass.exceptions import...
#!/usr/bin/env python # coding: utf-8 from __future__ import print_function from __future__ import absolute_import import os import re import shutil import stat import sys import tct from os.path import exists as ospe, join as ospj from tct import deepget params = tct.readjson(sys.argv[1]) binabspath = sys.argv[2] ...
[ "tct.make_snapshot_of_milestones", "os.path.exists", "tct.deepget", "shutil.copy2", "os.path.join", "os.chmod", "shutil.copytree", "tct.readjson", "os.path.split", "os.path.isdir", "tct.save_the_result", "sys.exit", "os.stat", "re.subn" ]
[((269, 294), 'tct.readjson', 'tct.readjson', (['sys.argv[1]'], {}), '(sys.argv[1])\n', (281, 294), False, 'import tct\n'), ((328, 361), 'tct.readjson', 'tct.readjson', (["params['factsfile']"], {}), "(params['factsfile'])\n", (340, 361), False, 'import tct\n'), ((375, 413), 'tct.readjson', 'tct.readjson', (["params['m...
import logging import random import time import src.support.outbound_routing as ob from src.support.creds import build_cred, build_proof_request, build_schema, build_credential_proposal, build_proof_proposal import src.support.settings as config # This file containst the functions that perform transaction-specific # ...
[ "src.support.creds.build_proof_request", "logging.debug", "src.support.outbound_routing.get_pres_ex_records", "src.support.creds.build_credential_proposal", "src.support.settings.agent_data.update_package_no", "src.support.outbound_routing.send_proof_proposal", "src.support.outbound_routing.get_schema",...
[((949, 1129), 'src.support.creds.build_credential_proposal', 'build_credential_proposal', (['config.agent_data.current_connection'], {'comment': '"""request for payment agreement credential"""', 'schema_name': '"""payment agreement"""', 'prop_schema': 'proposal'}), "(config.agent_data.current_connection, comment=\n ...
# # To run locally, execute: # # spark-submit --master local[2] wide_and_deep_example.py # S3_ROOT_DIR = 's3://{YOUR_S3_BUCKET}/{YOUR_S3_PATH}/' batch_size = 100 worker_count = 1 server_count = 1 import metaspore as ms spark = ms.spark.get_session(batch_size=batch_size, worker_count=wo...
[ "metaspore.spark.get_session", "pyspark.ml.evaluation.BinaryClassificationEvaluator", "metaspore.nn.WideAndDeepModule", "metaspore.PyTorchEstimator", "metaspore.input.read_s3_csv" ]
[((232, 333), 'metaspore.spark.get_session', 'ms.spark.get_session', ([], {'batch_size': 'batch_size', 'worker_count': 'worker_count', 'server_count': 'server_count'}), '(batch_size=batch_size, worker_count=worker_count,\n server_count=server_count)\n', (252, 333), True, 'import metaspore as ms\n'), ((468, 827), 'me...
# dockerpty. # # Copyright 2014 <NAME> <<EMAIL>> # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or ag...
[ "dockerpty.pty.RunOperation", "dockerpty.pty.ExecOperation", "dockerpty.pty.PseudoTerminal", "dockerpty.pty.exec_create" ]
[((950, 1064), 'dockerpty.pty.RunOperation', 'RunOperation', (['client', 'container'], {'interactive': 'interactive', 'stdout': 'stdout', 'stderr': 'stderr', 'stdin': 'stdin', 'logs': 'logs'}), '(client, container, interactive=interactive, stdout=stdout,\n stderr=stderr, stdin=stdin, logs=logs)\n', (962, 1064), Fals...
import numpy as np from astroquery.hitran import Hitran from astropy import units as un from astropy.constants import c, k_B, h, u def calc_solid_angle(radius,distance): ''' Convenience function to calculate solid angle from radius and distance, assuming a disk shape. Parameters ---------- radius ...
[ "numpy.sqrt" ]
[((945, 973), 'numpy.sqrt', 'np.sqrt', (['(solid_angle / np.pi)'], {}), '(solid_angle / np.pi)\n', (952, 973), True, 'import numpy as np\n')]
from random import shuffle#自带洗牌方法 from copy import deepcopy class Solution(object): def __init__(self, nums): """ :type nums: List[int] :type size: int """ self.nums=nums def reset(self): """ Resets the array to its original configurati...
[ "random.shuffle", "copy.deepcopy" ]
[((537, 556), 'copy.deepcopy', 'deepcopy', (['self.nums'], {}), '(self.nums)\n', (545, 556), False, 'from copy import deepcopy\n'), ((602, 615), 'random.shuffle', 'shuffle', (['nums'], {}), '(nums)\n', (609, 615), False, 'from random import shuffle\n')]
# Copyright (c) Facebook, Inc. and its affiliates. import numpy as np from termcolor import colored import logging import torch.nn as nn import torch.utils.data log = logging.getLogger(__name__) import torch import numpy as np import math class Dataset(torch.utils.data.Dataset): def __init__(self, x, y): ...
[ "logging.getLogger", "torch.nn.Tanh", "numpy.array", "torch.sum", "numpy.mean", "torch.mean", "torch.zeros_like", "numpy.random.permutation", "torch.Tensor", "torch.autograd.grad", "torch.cat", "torch.clamp", "torch.nn.Softplus", "math.ceil", "torch.log", "torch.stack", "torch.nn.Lin...
[((169, 196), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (186, 196), False, 'import logging\n'), ((4210, 4239), 'numpy.random.permutation', 'np.random.permutation', (['n_data'], {}), '(n_data)\n', (4231, 4239), True, 'import numpy as np\n'), ((1774, 1795), 'numpy.array', 'np.array', (...
import unittest import os import sys import argparse import numpy as np import audacity as aud print('Module file:') print(aud.__file__) SCRIPT_DIR = os.path.split(os.path.realpath(__file__))[0] PACKAGE_DIR = os.path.realpath(os.path.join(SCRIPT_DIR,'..')) DATA_DIR = os.path.join(PACKAGE_DIR, 'data') TEST_FILE_1 = ...
[ "audacity.Aup", "argparse.ArgumentParser", "os.path.join", "os.path.realpath", "unittest.main" ]
[((271, 304), 'os.path.join', 'os.path.join', (['PACKAGE_DIR', '"""data"""'], {}), "(PACKAGE_DIR, 'data')\n", (283, 304), False, 'import os\n'), ((320, 356), 'os.path.join', 'os.path.join', (['DATA_DIR', '"""test-1.aup"""'], {}), "(DATA_DIR, 'test-1.aup')\n", (332, 356), False, 'import os\n'), ((229, 259), 'os.path.joi...
# -*- coding: utf-8 -*- """ # @Time : 24/10/18 2:40 PM # @Author : <NAME> # @FileName: plot_result.py # @Software: PyCharm # @Github : https://github.com/hzm2016 """ import collections import matplotlib.pyplot as plt import numpy as np import pickle import copy as cp from baselines.deepq.assembly.src.value_funct...
[ "matplotlib.pyplot.subplots_adjust", "matplotlib.pyplot.savefig", "matplotlib.pyplot.xticks", "matplotlib.pyplot.ylabel", "matplotlib.pyplot.xlabel", "matplotlib.pyplot.plot", "pickle.load", "numpy.array", "matplotlib.pyplot.figure", "matplotlib.pyplot.yticks", "matplotlib.pyplot.tight_layout", ...
[((719, 773), 'numpy.array', 'np.array', (['[40, 40, 0, 5, 5, 5, 542, -36, 188, 5, 5, 5]'], {}), '([40, 40, 0, 5, 5, 5, 542, -36, 188, 5, 5, 5])\n', (727, 773), True, 'import numpy as np\n'), ((780, 844), 'numpy.array', 'np.array', (['[-40, -40, -40, -5, -5, -5, 538, -42, 192, -5, -5, -5]'], {}), '([-40, -40, -40, -5, ...
from repacolors import ColorScale from .colorbrewer import PALETTES as CBPALETTES PALETTES = { "ryb": ["#fe2713", "#fd5307", "#fb9900", "#fabc00", "#fefe34", "#d1e92c", "#66b032", "#0492ce", "#0347fe", "#3e01a4", "#8600af", "#a7194b"], "rybw3": ["#FE2712", "#FC600A", "#FB9902", "#FCCC1A", "#FEFE33", "#B2D732"...
[ "repacolors.ColorScale" ]
[((790, 808), 'repacolors.ColorScale', 'ColorScale', (['colors'], {}), '(colors)\n', (800, 808), False, 'from repacolors import ColorScale\n')]
# Copyright (c) 2019 <NAME> <<EMAIL>> # See the COPYRIGHT file for more information import subprocess import time def ping(guest_ip): out = subprocess.run(['ping', '-c 1', guest_ip], capture_output=True) return out.returncode == 0 def nmap_ssh(guest_ip): out = subprocess.run(['nmap', guest_ip, '-PN', ...
[ "subprocess.run", "time.time" ]
[((147, 210), 'subprocess.run', 'subprocess.run', (["['ping', '-c 1', guest_ip]"], {'capture_output': '(True)'}), "(['ping', '-c 1', guest_ip], capture_output=True)\n", (161, 210), False, 'import subprocess\n'), ((278, 350), 'subprocess.run', 'subprocess.run', (["['nmap', guest_ip, '-PN', '-p ssh']"], {'capture_output'...
import json from urllib.parse import urljoin from django.conf import settings import requests def get_ticket_endpoint(): return urljoin(settings.ZENDESK_BASE_URL, '/api/v2/tickets.json') def zendesk_auth(): return ( '{username}/token'.format(username=settings.ZENDESK_API_USERNAME), settings...
[ "json.dumps", "urllib.parse.urljoin" ]
[((135, 193), 'urllib.parse.urljoin', 'urljoin', (['settings.ZENDESK_BASE_URL', '"""/api/v2/tickets.json"""'], {}), "(settings.ZENDESK_BASE_URL, '/api/v2/tickets.json')\n", (142, 193), False, 'from urllib.parse import urljoin\n'), ((1054, 1073), 'json.dumps', 'json.dumps', (['payload'], {}), '(payload)\n', (1064, 1073)...
from django.conf import settings from rest_framework.routers import DefaultRouter, SimpleRouter from message_service.mailing.api.views import ( ClientsViewSet, MailingViewSet, MessageViewSet, ) from message_service.users.api.views import UserViewSet if settings.DEBUG: router = DefaultRouter() else: ...
[ "rest_framework.routers.SimpleRouter", "rest_framework.routers.DefaultRouter" ]
[((296, 311), 'rest_framework.routers.DefaultRouter', 'DefaultRouter', ([], {}), '()\n', (309, 311), False, 'from rest_framework.routers import DefaultRouter, SimpleRouter\n'), ((331, 345), 'rest_framework.routers.SimpleRouter', 'SimpleRouter', ([], {}), '()\n', (343, 345), False, 'from rest_framework.routers import De...
#! /usr/bin/env python from math import factorial import numpy as np # test passed def generate_poly(max_exponent,max_diff,symbol): f=np.zeros((max_diff+1, max_exponent+1), dtype=float) for k in range(max_diff+1): for i in range(max_exponent+1): if (i - k) >= 0: f[k,i] = factorial(i)*symbol**(i-k)/facto...
[ "math.factorial", "numpy.zeros" ]
[((137, 192), 'numpy.zeros', 'np.zeros', (['(max_diff + 1, max_exponent + 1)'], {'dtype': 'float'}), '((max_diff + 1, max_exponent + 1), dtype=float)\n', (145, 192), True, 'import numpy as np\n'), ((315, 331), 'math.factorial', 'factorial', (['(i - k)'], {}), '(i - k)\n', (324, 331), False, 'from math import factorial\...
# --- # jupyter: # jupytext: # formats: ipynb,py:light # text_representation: # extension: .py # format_name: light # format_version: '1.5' # jupytext_version: 1.11.0 # kernelspec: # display_name: Python 3 # language: python # name: python3 # --- import torch def train...
[ "torch.vstack", "torch.cuda.empty_cache", "torch.randn" ]
[((3316, 3336), 'torch.randn', 'torch.randn', (['(15, 1)'], {}), '((15, 1))\n', (3327, 3336), False, 'import torch\n'), ((3341, 3360), 'torch.randn', 'torch.randn', (['(9, 1)'], {}), '((9, 1))\n', (3352, 3360), False, 'import torch\n'), ((3362, 3382), 'torch.vstack', 'torch.vstack', (['(a, b)'], {}), '((a, b))\n', (337...
import sqlite3 import pandas as pd import re import random from bs4 import BeautifulSoup class Process: SEQ_LENGTH = 40 sql_transaction = [] dataset = [] cursor_train = None cursor_validation = None cursor_test = None # I know in advance that there are 199819620 rows NUM_ROWS = 19...
[ "random.shuffle", "sqlite3.connect", "pandas.read_csv", "bs4.BeautifulSoup", "re.sub" ]
[((610, 662), 'sqlite3.connect', 'sqlite3.connect', (["(database_dir + '/sequence_train.db')"], {}), "(database_dir + '/sequence_train.db')\n", (625, 662), False, 'import sqlite3\n'), ((829, 879), 'sqlite3.connect', 'sqlite3.connect', (["(database_dir + '/sequence_val.db')"], {}), "(database_dir + '/sequence_val.db')\n...
from django.conf.urls import url from . import views urlpatterns = [ #url(r'^view/(?P<pk>[0-9]+)', views.ArticleDetailView.as_view(), name = "detail"), #url(r'', views.ArticleIndexView.as_view(), name = "index"), url(r'login', views.Login.as_view()), url(r'logout', views.Logout.as_view()), ...
[ "django.conf.urls.url" ]
[((444, 464), 'django.conf.urls.url', 'url', (['""""""', 'views.Panel'], {}), "('', views.Panel)\n", (447, 464), False, 'from django.conf.urls import url\n')]
""" Copyright (c) Contributors to the Open 3D Engine Project. For complete copyright and license terms please see the LICENSE at the root of this distribution. SPDX-License-Identifier: Apache-2.0 OR MIT """ # setup path import azlmbr.legacy.general as general import azlmbr.bus as bus import azlmbr.editor as editor im...
[ "azlmbr.entity.EntityId", "azlmbr.editor.EditorEntityAPIBus", "azlmbr.editor.EditorComponentAPIBus", "azlmbr.legacy.general.find_editor_entity" ]
[((675, 713), 'azlmbr.legacy.general.find_editor_entity', 'general.find_editor_entity', (['"""WhiteBox"""'], {}), "('WhiteBox')\n", (701, 713), True, 'import azlmbr.legacy.general as general\n'), ((1471, 1585), 'azlmbr.editor.EditorComponentAPIBus', 'editor.EditorComponentAPIBus', (['bus.Broadcast', '"""AddComponentsOf...
from dask.distributed import Client import dask.dataframe as dd import pandas as pd import numpy as np import os import matplotlib.pyplot as plt import matplotlib.cm as cm from sklearn.manifold import TSNE from sklearn.decomposition import PCA from IPython.display import display, HTML from sklearn.cluster import KMeans...
[ "sklearn.metrics.f1_score", "numpy.unique", "pandas.read_csv", "os.path.join", "sklearn.metrics.recall_score", "matplotlib.pyplot.rcParams.update", "matplotlib.pyplot.figure", "matplotlib.pyplot.axes", "functools.partial", "matplotlib.pyplot.scatter", "matplotlib.cm.tab20", "pandas.DataFrame",...
[((2021, 2059), 'matplotlib.pyplot.rcParams.update', 'plt.rcParams.update', (["{'font.size': 14}"], {}), "({'font.size': 14})\n", (2040, 2059), True, 'import matplotlib.pyplot as plt\n'), ((2306, 2333), 'matplotlib.pyplot.figure', 'plt.figure', ([], {'figsize': '(10, 8)'}), '(figsize=(10, 8))\n', (2316, 2333), True, 'i...
#!/usr/bin/env python3 # encoding: utf-8 import torch.nn.functional as F from rls.algorithms.single.dqn import DQN from rls.common.decorator import iton from rls.utils.torch_utils import n_step_return class DDQN(DQN): """ Double DQN, https://arxiv.org/abs/1509.06461 Double DQN + LSTM, https...
[ "rls.utils.torch_utils.n_step_return" ]
[((1115, 1207), 'rls.utils.torch_utils.n_step_return', 'n_step_return', (['BATCH.reward', 'self.gamma', 'BATCH.done', 'q_target_next_max', 'BATCH.begin_mask'], {}), '(BATCH.reward, self.gamma, BATCH.done, q_target_next_max,\n BATCH.begin_mask)\n', (1128, 1207), False, 'from rls.utils.torch_utils import n_step_return...
#!/usr/bin/python3 import os os.system('wget \ https://opendata.arcgis.com/datasets/6ac5e325468c4cb9b905f1728d6fbf0f_0.csv \ -O hifld_hospital.csv')
[ "os.system" ]
[((29, 154), 'os.system', 'os.system', (['"""wget https://opendata.arcgis.com/datasets/6ac5e325468c4cb9b905f1728d6fbf0f_0.csv -O hifld_hospital.csv"""'], {}), "(\n 'wget https://opendata.arcgis.com/datasets/6ac5e325468c4cb9b905f1728d6fbf0f_0.csv -O hifld_hospital.csv'\n )\n", (38, 154), False, 'import os\n')]
import argparse import subprocess import sys import logging logger = logging.getLogger("helper") def azcli(command): process = subprocess.Popen(command, stdout=subprocess.PIPE, stderr=subprocess.PIPE) out,err = process.communicate() logger.debug(str(out,"utf-8")) exit_code = process.returncode if ...
[ "logging.getLogger", "subprocess.Popen", "sys.exit" ]
[((70, 97), 'logging.getLogger', 'logging.getLogger', (['"""helper"""'], {}), "('helper')\n", (87, 97), False, 'import logging\n'), ((133, 206), 'subprocess.Popen', 'subprocess.Popen', (['command'], {'stdout': 'subprocess.PIPE', 'stderr': 'subprocess.PIPE'}), '(command, stdout=subprocess.PIPE, stderr=subprocess.PIPE)\n...
from Vertex import Vertex import pygame from Colours import Colours class Grid: def createGrid(self, rows, width): grid = [] space = width // rows for x in range(rows): grid.append([]) for i in range(rows): vertex = Vertex(space, rows, x, i...
[ "pygame.display.update", "Vertex.Vertex", "pygame.draw.line" ]
[((964, 987), 'pygame.display.update', 'pygame.display.update', ([], {}), '()\n', (985, 987), False, 'import pygame\n'), ((510, 585), 'pygame.draw.line', 'pygame.draw.line', (['window', 'Colours.BLACK', '(0, x * space)', '(width, x * space)'], {}), '(window, Colours.BLACK, (0, x * space), (width, x * space))\n', (526, ...
# -*- coding: utf-8 -*- # ----------------------------------------------------------------------------- # (C) British Crown copyright. The Met Office. # All rights reserved. # # Redistribution and use in source and binary forms, with or without # modification, are permitted provided that the following conditions are me...
[ "iris.cube.CubeList", "numpy.ones", "improver.synthetic_data.set_up_test_cubes.set_up_variable_cube", "improver.precipitation_type.shower_condition_probability.ShowerConditionProbability", "numpy.array", "numpy.zeros", "pytest.raises", "pytest.fixture" ]
[((2584, 2617), 'pytest.fixture', 'pytest.fixture', ([], {'name': '"""test_cubes"""'}), "(name='test_cubes')\n", (2598, 2617), False, 'import pytest\n'), ((2745, 2755), 'iris.cube.CubeList', 'CubeList', ([], {}), '()\n', (2753, 2755), False, 'from iris.cube import CubeList\n'), ((9135, 9171), 'improver.precipitation_ty...
# pylint: disable=no-member, too-many-locals, no-self-use """Vessels File Upload """ import time from flask import request # from library.couch_database import CouchDatabase from library.postgresql_queries import PostgreSQL from library.couch_queries import Queries from library.common import Common from library...
[ "flask.request.args.get", "flask.request.files.getlist", "library.couch_queries.Queries", "library.aws_s3.AwsS3", "library.postgresql_queries.PostgreSQL", "time.time", "flask.request.headers.get" ]
[((523, 532), 'library.couch_queries.Queries', 'Queries', ([], {}), '()\n', (530, 532), False, 'from library.couch_queries import Queries\n'), ((558, 570), 'library.postgresql_queries.PostgreSQL', 'PostgreSQL', ([], {}), '()\n', (568, 570), False, 'from library.postgresql_queries import PostgreSQL\n'), ((592, 599), 'li...
import tensorflow as tf import numpy as np from tqdm.notebook import tqdm class System(): def __init__(self, num_part, dim, Ansatz=None, External=None, Internal=None, Sampler=None ): self...
[ "tensorflow.random.uniform", "tensorflow.reshape", "tensorflow.reduce_sum", "tensorflow.math.sqrt", "numpy.linalg.norm", "numpy.sum", "numpy.zeros", "numpy.random.uniform", "tensorflow.convert_to_tensor", "tqdm.notebook.tqdm", "tensorflow.norm" ]
[((842, 929), 'tensorflow.random.uniform', 'tf.random.uniform', (['(batch_size, dim)'], {'minval': '(-2)', 'maxval': '(2)', 'dtype': 'tf.dtypes.float64'}), '((batch_size, dim), minval=-2, maxval=2, dtype=tf.dtypes.\n float64)\n', (859, 929), True, 'import tensorflow as tf\n'), ((2594, 2617), 'numpy.zeros', 'np.zeros...
from setuptools import setup setup( name='torch-dimcheck', version='0.0.1', description='Dimensionality annotations for tensor parameters and return values', packages=['torch_dimcheck'], author='<NAME>', author_email='<EMAIL>', )
[ "setuptools.setup" ]
[((30, 236), 'setuptools.setup', 'setup', ([], {'name': '"""torch-dimcheck"""', 'version': '"""0.0.1"""', 'description': '"""Dimensionality annotations for tensor parameters and return values"""', 'packages': "['torch_dimcheck']", 'author': '"""<NAME>"""', 'author_email': '"""<EMAIL>"""'}), "(name='torch-dimcheck', ver...
#!/usr/bin/env python from os import ( path as os_path, mkdir as os_mkdir, getcwd ) from argparse import ArgumentParser from logging import ( Logger, getLogger ) from glob import glob from typing import ( Dict, ) from colored import fg, bg, attr from brs_utils import ( create_logger ) from...
[ "logging.getLogger", "os.path.exists", "colored.fg", "os.path.join", "os.mkdir", "colored.attr", "os.path.getmtime", "brs_utils.create_logger" ]
[((729, 748), 'logging.getLogger', 'getLogger', (['__name__'], {}), '(__name__)\n', (738, 748), False, 'from logging import Logger, getLogger\n'), ((2136, 2172), 'brs_utils.create_logger', 'create_logger', (['parser.prog', 'args.log'], {}), '(parser.prog, args.log)\n', (2149, 2172), False, 'from brs_utils import create...
import functools import heapq import logging from collections import deque from threading import Condition, RLock from typing import Any, Callable, List, NamedTuple, Optional from pytils.mixins import DaemonHandler from ._config.time import DEFAULT_TIME_SUPPLIER, TimeSupplier, TimeType, ZERO_DURATION __all__ = [ ...
[ "logging.getLogger", "collections.deque", "threading.RLock", "heapq.heappop", "heapq.heappush", "threading.Condition" ]
[((842, 875), 'logging.getLogger', 'logging.getLogger', (['"""pytils.clock"""'], {}), "('pytils.clock')\n", (859, 875), False, 'import logging\n'), ((1683, 1694), 'threading.Condition', 'Condition', ([], {}), '()\n', (1692, 1694), False, 'from threading import Condition, RLock\n'), ((1722, 1750), 'collections.deque', '...
"""Execute validated & constructed query on device. Accepts input from front end application, validates the input and returns errors if input is invalid. Passes validated parameters to construct.py, which is used to build & run the Netmiko connections or hyperglass-frr API calls, returns the output back to the front e...
[ "hyperglass.util.parse_exception", "hyperglass.exceptions.ResponseEmpty", "httpx.AsyncClient", "hyperglass.log.log.error", "hyperglass.exceptions.RestError", "hyperglass.log.log.debug" ]
[((947, 992), 'hyperglass.log.log.debug', 'log.debug', (['"""Query parameters: {}"""', 'self.query'], {}), "('Query parameters: {}', self.query)\n", (956, 992), False, 'from hyperglass.log import log\n'), ((2077, 2116), 'hyperglass.log.log.debug', 'log.debug', (['"""URL endpoint: {}"""', 'endpoint'], {}), "('URL endpoi...
import test_agent print('Logging in') Meerkat = test_agent.TestAgent(username='meerkat', password='<PASSWORD>', endpoint='/messages/') Pangolin = test_agent.TestAgent(username='pangolin', password='<PASSWORD>', endpoint='/messages/') Badger = test_agent.TestAgent(username='badger', password='<PASSWORD>', endpoint='/me...
[ "test_agent.TestAgent" ]
[((49, 140), 'test_agent.TestAgent', 'test_agent.TestAgent', ([], {'username': '"""meerkat"""', 'password': '"""<PASSWORD>"""', 'endpoint': '"""/messages/"""'}), "(username='meerkat', password='<PASSWORD>', endpoint=\n '/messages/')\n", (69, 140), False, 'import test_agent\n'), ((147, 239), 'test_agent.TestAgent', '...
""" Unit tests for the HR solver. """ import pytest from matching import Matching from matching import Player as Resident from matching.games import HospitalResident from matching.players import Hospital from .params import HOSPITAL_RESIDENT, make_game, make_prefs @HOSPITAL_RESIDENT def test_init(resident_names, h...
[ "matching.players.Hospital", "matching.games.HospitalResident", "matching.games.HospitalResident.create_from_dictionaries", "pytest.raises", "matching.Player" ]
[((1226, 1316), 'matching.games.HospitalResident.create_from_dictionaries', 'HospitalResident.create_from_dictionaries', (['resident_prefs', 'hospital_prefs', 'capacities_'], {}), '(resident_prefs, hospital_prefs,\n capacities_)\n', (1267, 1316), False, 'from matching.games import HospitalResident\n'), ((4191, 4211)...
import date import os def get_time_delta(kline_type = '1_day'): if kline_type.lower() == '1_day'.lower(): return 0 kline_array = kline_type.split("_") if len(kline_array) != 2: raise ValueError('KLine_type {0} not supported'.format(kline_type)) if kline_array[1].lower() == 'min'.lower()...
[ "date.create_kline_time_string" ]
[((1617, 1681), 'date.create_kline_time_string', 'date.create_kline_time_string', (['(9)', '(30)', '(time_delta * time_interval)'], {}), '(9, 30, time_delta * time_interval)\n', (1646, 1681), False, 'import date\n'), ((1769, 1833), 'date.create_kline_time_string', 'date.create_kline_time_string', (['(13)', '(0)', '(tim...
from collections import defaultdict def check_winner(cards): for card_index, card in cards.items(): for index in range(5): complete_line = all([x[1] for x in card[index]]) complete_column = all([card[x][index][1] for x in range(5)]) if complete_line or complete_column: ...
[ "collections.defaultdict" ]
[((508, 525), 'collections.defaultdict', 'defaultdict', (['list'], {}), '(list)\n', (519, 525), False, 'from collections import defaultdict\n')]
""" Demonstrates the hover functionality of mpldatacursor as well as point labels and a custom formatting function. Notice that overlapping points have both labels displayed. """ import string import matplotlib.pyplot as plt import numpy as np from mpldatacursor import datacursor np.random.seed(1977) x, y = np.random....
[ "numpy.random.random", "numpy.random.seed", "matplotlib.pyplot.subplots", "mpldatacursor.datacursor", "matplotlib.pyplot.show" ]
[((281, 301), 'numpy.random.seed', 'np.random.seed', (['(1977)'], {}), '(1977)\n', (295, 301), True, 'import numpy as np\n'), ((310, 335), 'numpy.random.random', 'np.random.random', (['(2, 26)'], {}), '((2, 26))\n', (326, 335), True, 'import numpy as np\n'), ((379, 393), 'matplotlib.pyplot.subplots', 'plt.subplots', ([...
import torch import torch.nn as nn class FilterResponseNorm(nn.Module): def __init__(self, num_features, eps=1e-6, use_TLU=True): super(FilterResponseNorm, self).__init__() self.num_features = num_features self.eps = eps self.use_TLU = use_TLU self.weight = nn.Parameter(t...
[ "torch.max", "torch.nn.init.uniform_", "torch.nn.init.zeros_", "torch.Tensor" ]
[((629, 658), 'torch.nn.init.uniform_', 'nn.init.uniform_', (['self.weight'], {}), '(self.weight)\n', (645, 658), True, 'import torch.nn as nn\n'), ((667, 692), 'torch.nn.init.zeros_', 'nn.init.zeros_', (['self.bias'], {}), '(self.bias)\n', (681, 692), True, 'import torch.nn as nn\n'), ((319, 345), 'torch.Tensor', 'tor...
from time import time def profile(funcao): def funcao_wrapper(*args, **kwargs): inicio = time() resultado = funcao(*args, **kwargs) fim = time() print(fim - inicio) return resultado return funcao_wrapper @profile def f(n): return 'Executei f {}'.format(n) print(f...
[ "time.time" ]
[((103, 109), 'time.time', 'time', ([], {}), '()\n', (107, 109), False, 'from time import time\n'), ((168, 174), 'time.time', 'time', ([], {}), '()\n', (172, 174), False, 'from time import time\n')]
import numpy as np # import matplotlib.pyplot as plt from scipy.cluster.vq import kmeans # def plothist(x): # vmin = x.min()-1 # vmax = x.max()+1 # bins = np.arange(vmin, vmax, (vmax - vmin)/50) # plt.hist(x, bins=bins) # plt.show() # def scatterpred(pred): # plt.scatter(pred[:,0], pred[:,1]) ...
[ "numpy.zeros", "scipy.cluster.vq.kmeans" ]
[((564, 578), 'numpy.zeros', 'np.zeros', (['nb_c'], {}), '(nb_c)\n', (572, 578), True, 'import numpy as np\n'), ((840, 854), 'numpy.zeros', 'np.zeros', (['nb_c'], {}), '(nb_c)\n', (848, 854), True, 'import numpy as np\n'), ((866, 883), 'numpy.zeros', 'np.zeros', (['c.shape'], {}), '(c.shape)\n', (874, 883), True, 'impo...
# -*- coding: utf-8 -*- """ Created on Tue Apr 20 13:32:20 2021 #--- ag csv results to single file based on dashboard_dbs #--- <NAME> (<EMAIL>) #--- Jul, 2021. #--- Dev-log in: https://github.com/Murilodsv/py-jules @author: muril """ # DEBUG import os; os.chdir('C:/Murilo/py-jules') #---------------...
[ "os.path.exists", "time.time", "util.df_csv", "os.getcwd" ]
[((1562, 1568), 'time.time', 'time', ([], {}), '()\n', (1566, 1568), False, 'from time import time\n'), ((1722, 1750), 'util.df_csv', 'u.df_csv', (["(wd + '/' + dash_nm)"], {}), "(wd + '/' + dash_nm)\n", (1730, 1750), True, 'import util as u\n'), ((1919, 1937), 'os.path.exists', 'os.path.exists', (['fn'], {}), '(fn)\n'...
import os from django.conf.urls.defaults import * # Uncomment the next two lines to enable the admin: # from django.contrib import admin # admin.autodiscover() urlpatterns = patterns('', # Example: # (r'^django_amf_example/', include('django_amf_example.foo.urls')), # Uncomment the admin/doc line below ...
[ "os.path.join" ]
[((832, 862), 'os.path.join', 'os.path.join', (['"""flex"""', '"""deploy"""'], {}), "('flex', 'deploy')\n", (844, 862), False, 'import os\n')]
# coding: utf-8 """ MolecularMatch MMPower MMPower API # noqa: E501 OpenAPI spec version: 1.0.0 Contact: <EMAIL> Generated by: https://github.com/swagger-api/swagger-codegen.git """ import pprint import re # noqa: F401 import six class PrivateTrial(object): """NOTE: This class is auto g...
[ "six.iteritems" ]
[((50583, 50616), 'six.iteritems', 'six.iteritems', (['self.swagger_types'], {}), '(self.swagger_types)\n', (50596, 50616), False, 'import six\n')]
import argparse import gym import gym_module_select from stable_baselines.common.vec_env import DummyVecEnv def init_parse_argument(): parser = argparse.ArgumentParser() parser.add_argument('-e', '--num-exp', help='num experiment episode', type=int, default=10) args = parser.parse_args() return args ...
[ "gym.make", "argparse.ArgumentParser", "stable_baselines.common.vec_env.DummyVecEnv" ]
[((357, 426), 'gym.make', 'gym.make', (['"""ModuleSelect-v1"""'], {'verbose': '(1)', 'save_log_flag': '(True)', 'log_num': '(7)'}), "('ModuleSelect-v1', verbose=1, save_log_flag=True, log_num=7)\n", (365, 426), False, 'import gym\n'), ((495, 522), 'stable_baselines.common.vec_env.DummyVecEnv', 'DummyVecEnv', (['[lambda...
import unittest from flask_script import Manager, Shell, Server from app import app, db from app.fake_populate import populate manager = Manager(app) def make_shell_context(): return dict(app=app) @manager.command def recreate_db(): """ Create the SQL database. """ db.drop_all() db.create_all...
[ "app.db.session.commit", "flask_script.Server", "flask_script.Manager", "flask_script.Shell", "app.fake_populate.populate", "app.read_licenses.read", "app.db.create_all", "app.db.drop_all", "unittest.TextTestRunner", "unittest.TestLoader" ]
[((138, 150), 'flask_script.Manager', 'Manager', (['app'], {}), '(app)\n', (145, 150), False, 'from flask_script import Manager, Shell, Server\n'), ((289, 302), 'app.db.drop_all', 'db.drop_all', ([], {}), '()\n', (300, 302), False, 'from app import app, db\n'), ((307, 322), 'app.db.create_all', 'db.create_all', ([], {}...
from makememe.generator.prompts.prompt import Prompt import datetime from PIL import Image from makememe.generator.design.image_manager import Image_Manager class Waiting(Prompt): name = "Waiting" description = "waiting" def __init__(self): self.instruction = """ ### Message:I've been waiting for...
[ "datetime.datetime.now", "makememe.generator.design.image_manager.Image_Manager.add_text", "PIL.Image.alpha_composite" ]
[((928, 1045), 'makememe.generator.design.image_manager.Image_Manager.add_text', 'Image_Manager.add_text', ([], {'base': 'base', 'text': "meme_text['subject']", 'position': '(600, 950)', 'font_size': '(40)', 'wrapped_width': '(20)'}), "(base=base, text=meme_text['subject'], position=(600,\n 950), font_size=40, wrapp...
from python_helper import log, Test, SettingHelper, RandomHelper, ObjectHelper, TestHelper, ReflectionHelper, Constant from python_framework import EncapsulateItWithGlobalException, GlobalException, ExceptionHandler, HttpStatus LOG_HELPER_SETTINGS = { log.LOG : False, log.INFO : True, log.SUCCESS : True, ...
[ "python_framework.GlobalException", "python_helper.ReflectionHelper.getClass", "python_helper.Test", "python_helper.TestHelper.getRaisedException", "python_framework.EncapsulateItWithGlobalException" ]
[((830, 864), 'python_framework.EncapsulateItWithGlobalException', 'EncapsulateItWithGlobalException', ([], {}), '()\n', (862, 864), False, 'from python_framework import EncapsulateItWithGlobalException, GlobalException, ExceptionHandler, HttpStatus\n'), ((931, 965), 'python_framework.EncapsulateItWithGlobalException',...
# -*- coding: utf-8 -*- # Generated by Django 1.11 on 2018-03-30 12:46 from __future__ import unicode_literals from django.conf import settings from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('app', '0005_auto_20180330_1813'), ] operations = [ ...
[ "django.db.models.ManyToManyField" ]
[((434, 536), 'django.db.models.ManyToManyField', 'models.ManyToManyField', ([], {'blank': '(True)', 'related_name': '"""incubator_follows"""', 'to': 'settings.AUTH_USER_MODEL'}), "(blank=True, related_name='incubator_follows', to=\n settings.AUTH_USER_MODEL)\n", (456, 536), False, 'from django.db import migrations,...
# encoding=utf-8 from airtest.core.win import Windows import unittest import numpy import time from testconf import try_remove SNAPSHOT = "win_snapshot.png" class TestWin(unittest.TestCase): @classmethod def setUpClass(cls): w = Windows() w.start_app("calc") time.sleep(1) cl...
[ "unittest.main", "airtest.core.win.Windows", "testconf.try_remove", "time.sleep" ]
[((825, 840), 'unittest.main', 'unittest.main', ([], {}), '()\n', (838, 840), False, 'import unittest\n'), ((250, 259), 'airtest.core.win.Windows', 'Windows', ([], {}), '()\n', (257, 259), False, 'from airtest.core.win import Windows\n'), ((296, 309), 'time.sleep', 'time.sleep', (['(1)'], {}), '(1)\n', (306, 309), Fals...
# Import required libraries import cv2 from os.path import os, dirname import tensorflow as tf import numpy as np from tqdm import tqdm import random # List of categories (directories names) CATEGORIES = ["bad_apple", "bad_grape", "bad_pear", "cherry", "good_apple", "good_avocado", "good_grape", "good_pear", "ripe_avo...
[ "random.shuffle", "os.path.os.listdir", "numpy.array", "tensorflow.keras.models.load_model", "os.path.os.path.join", "os.path.os.path.abspath" ]
[((618, 664), 'os.path.os.path.join', 'os.path.join', (['main_dir', '"""database"""', '"""training"""'], {}), "(main_dir, 'database', 'training')\n", (630, 664), False, 'from os.path import os, dirname\n'), ((679, 724), 'os.path.os.path.join', 'os.path.join', (['main_dir', '"""database"""', '"""testing"""'], {}), "(mai...
"""Covers import of data downloaded from the `Meadows online behavior platform <https://meadows-research.com/>`_. For information on available file types see the meadows `documentation on downloads <https://meadows-research.com/documentation\ /researcher/downloads/>`_. """ from os.path import basename import numpy fr...
[ "numpy.stack", "scipy.io.loadmat", "os.path.basename" ]
[((948, 962), 'scipy.io.loadmat', 'loadmat', (['fpath'], {}), '(fpath)\n', (955, 962), False, 'from scipy.io import loadmat\n'), ((1543, 1583), 'numpy.stack', 'numpy.stack', (['[data[v] for v in utv_vars]'], {}), '([data[v] for v in utv_vars])\n', (1554, 1583), False, 'import numpy\n'), ((3258, 3273), 'os.path.basename...
import unittest import torch from torchvision.models.resnet import BasicBlock, Bottleneck from nuscenes.prediction.models.backbone import ResNetBackbone, MobileNetBackbone class TestBackBones(unittest.TestCase): def count_layers(self, model): if isinstance(model[4][0], BasicBlock): n_convs ...
[ "nuscenes.prediction.models.backbone.MobileNetBackbone", "nuscenes.prediction.models.backbone.ResNetBackbone", "torch.ones" ]
[((597, 623), 'nuscenes.prediction.models.backbone.ResNetBackbone', 'ResNetBackbone', (['"""resnet18"""'], {}), "('resnet18')\n", (611, 623), False, 'from nuscenes.prediction.models.backbone import ResNetBackbone, MobileNetBackbone\n'), ((640, 666), 'nuscenes.prediction.models.backbone.ResNetBackbone', 'ResNetBackbone'...
#!/usr/bin/python3 # -*- coding: utf-8 -*- # @File : Qrbar_test.py import cv2 import numpy as np from pyzbar.pyzbar import decode img = cv2.imread('qrcode.png') for barcode in decode(img): print(barcode.data.decode('utf-8')) print(barcode.data) pts = np.array([barcode.polygon], np.int32) pts = pts....
[ "pyzbar.pyzbar.decode", "numpy.array", "cv2.imread" ]
[((141, 165), 'cv2.imread', 'cv2.imread', (['"""qrcode.png"""'], {}), "('qrcode.png')\n", (151, 165), False, 'import cv2\n'), ((181, 192), 'pyzbar.pyzbar.decode', 'decode', (['img'], {}), '(img)\n', (187, 192), False, 'from pyzbar.pyzbar import decode\n'), ((268, 305), 'numpy.array', 'np.array', (['[barcode.polygon]', ...
#!/usr/bin/env python """ """ from __future__ import print_function import argparse import sys from . import common from . import helper from . import vcs_tool PARAMS = {} PARAMS['this_script'] = common.get_script_name_from_filename(__file__) def setup_and_dispatch(): parser = argparse.ArgumentParser( ...
[ "sys.exit" ]
[((1295, 1306), 'sys.exit', 'sys.exit', (['(0)'], {}), '(0)\n', (1303, 1306), False, 'import sys\n')]
from django.db import models # from themall.models import Customer # Create your models here. class Seller(models.Model): email = models.OneToOneField('themall.Customer', on_delete=models.CASCADE, to_field='email') store_name = models.CharField(max_length=100) slug = models.SlugField(max_length=100) descrip...
[ "django.db.models.OneToOneField", "django.db.models.TextField", "django.db.models.SlugField", "django.db.models.CharField" ]
[((136, 225), 'django.db.models.OneToOneField', 'models.OneToOneField', (['"""themall.Customer"""'], {'on_delete': 'models.CASCADE', 'to_field': '"""email"""'}), "('themall.Customer', on_delete=models.CASCADE, to_field\n ='email')\n", (156, 225), False, 'from django.db import models\n'), ((236, 268), 'django.db.mode...
# Copyright (c) 2017-2021, <NAME>. All rights reserved. # For licensing, see https://github.com/mudita/MuditaOS/LICENSE.md import time import pytest from harness import log from harness.dom_parser_utils import * from harness.interface.defs import key_codes from bt_fixtures import * @pytest.mark.rt1051 @pytest.mark.us...
[ "pytest.mark.usefixtures", "time.sleep", "pytest.mark.skipif" ]
[((306, 347), 'pytest.mark.usefixtures', 'pytest.mark.usefixtures', (['"""bt_all_devices"""'], {}), "('bt_all_devices')\n", (329, 347), False, 'import pytest\n'), ((349, 384), 'pytest.mark.usefixtures', 'pytest.mark.usefixtures', (['"""bt_reset"""'], {}), "('bt_reset')\n", (372, 384), False, 'import pytest\n'), ((386, ...
import smtplib, ssl, os from email.mime.text import MIMEText from email.mime.multipart import MIMEMultipart from .html_template import emailHtml from .text_template import emailText port = 465 context = ssl.create_default_context() def sendEmail(emailData): adminUser = os.getenv("ADMIN_USERNAME") password =...
[ "os.getenv", "smtplib.SMTP_SSL", "ssl.create_default_context", "email.mime.multipart.MIMEMultipart", "email.mime.text.MIMEText" ]
[((206, 234), 'ssl.create_default_context', 'ssl.create_default_context', ([], {}), '()\n', (232, 234), False, 'import smtplib, ssl, os\n'), ((278, 305), 'os.getenv', 'os.getenv', (['"""ADMIN_USERNAME"""'], {}), "('ADMIN_USERNAME')\n", (287, 305), False, 'import smtplib, ssl, os\n'), ((321, 344), 'os.getenv', 'os.geten...
import requests import os import json import datetime ''' Pulls a dbml file from the API. User must manually add the file id, found in the 'response_ids.json' file generated from dbml_post_to_api.py ''' url='http://ec2-54-167-67-34.compute-1.amazonaws.com/api/dbmls' #url of the API id = '6192b1f31c2a512293f...
[ "requests.get" ]
[((422, 449), 'requests.get', 'requests.get', (['f"""{url}/{id}"""'], {}), "(f'{url}/{id}')\n", (434, 449), False, 'import requests\n')]
from openpyxl import load_workbook def getRowCount(file): wb = load_workbook(file) sheet = wb.active return sheet.max_row def getColumnCount(file): wb = load_workbook(file) sheet = wb.active return sheet.max_column def getCellData(file, cell): wb = load_workbook(file) sheet = wb.acti...
[ "openpyxl.load_workbook" ]
[((69, 88), 'openpyxl.load_workbook', 'load_workbook', (['file'], {}), '(file)\n', (82, 88), False, 'from openpyxl import load_workbook\n'), ((172, 191), 'openpyxl.load_workbook', 'load_workbook', (['file'], {}), '(file)\n', (185, 191), False, 'from openpyxl import load_workbook\n'), ((281, 300), 'openpyxl.load_workboo...
# pommerman/cli/run_battle.py # pommerman/agents/TensorFlowAgent/pit.py import atexit from datetime import datetime import os import random import sys import time import argparse import numpy as np from pommerman import helpers, make from TensorFlowAgent import TensorFlowAgent from pommerman import utility import ...
[ "tensorflow.reset_default_graph", "argparse.ArgumentParser", "tensorflow.Session", "tensorflow.train.Saver", "tensorflow.global_variables_initializer", "numpy.array", "TensorFlowAgent.TensorFlowAgent", "numpy.zeros" ]
[((1290, 1314), 'tensorflow.reset_default_graph', 'tf.reset_default_graph', ([], {}), '()\n', (1312, 1314), True, 'import tensorflow as tf\n'), ((1580, 1605), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {}), '()\n', (1603, 1605), False, 'import argparse\n'), ((701, 717), 'numpy.zeros', 'np.zeros', (['(1,...
import numpy as np import tensorflow as tf import unittest from xcenternet.model.evaluation.overlap import compute_overlap from xcenternet.model.evaluation.mean_average_precision import MAP class TestMeanAveragePrecision(unittest.TestCase): def setUp(self): self.map_bboxes = np.array( [ ...
[ "numpy.array", "tensorflow.constant", "unittest.main", "xcenternet.model.evaluation.overlap.compute_overlap", "xcenternet.model.evaluation.mean_average_precision.MAP" ]
[((4629, 4644), 'unittest.main', 'unittest.main', ([], {}), '()\n', (4642, 4644), False, 'import unittest\n'), ((291, 435), 'numpy.array', 'np.array', (['[[[20, 10, 80, 60], [10, 40, 40, 90], [0, 0, 100, 100]], [[0, 0, 10, 10], [\n 20, 20, 40, 90], [80, 20, 100, 50]]]'], {'dtype': 'np.float64'}), '([[[20, 10, 80, 60...
# Copyright 2017 The Sonnet Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable l...
[ "sonnet.examples.rmc_learn_to_execute.build_and_train", "tensorflow.ones", "sonnet.examples.learn_to_execute.LearnToExecute", "sonnet.nets.MLP", "tensorflow.test.main", "sonnet.RelationalMemory", "sonnet.examples.rmc_learn_to_execute.SequenceModel", "tensorflow.zeros" ]
[((3048, 3062), 'tensorflow.test.main', 'tf.test.main', ([], {}), '()\n', (3060, 3062), True, 'import tensorflow as tf\n'), ((1280, 1376), 'sonnet.RelationalMemory', 'snt.RelationalMemory', ([], {'mem_slots': '(2)', 'head_size': '(4)', 'num_heads': '(1)', 'num_blocks': '(1)', 'gate_style': '"""unit"""'}), "(mem_slots=2...
import os.path from typing import Sequence, Optional, Dict import numpy as np import pandas as pd from nk_sent2vec import Sent2Vec as _Sent2Vec from d3m import container, utils from d3m.primitive_interfaces.transformer import TransformerPrimitiveBase from d3m.primitive_interfaces.base import CallResult from d3m.contai...
[ "d3m.primitive_interfaces.base.CallResult", "numpy.array", "nk_sent2vec.Sent2Vec", "d3m.container.DataFrame" ]
[((5704, 5730), 'd3m.container.DataFrame', 'd3m_DataFrame', (['embedded_df'], {}), '(embedded_df)\n', (5717, 5730), True, 'from d3m.container import DataFrame as d3m_DataFrame\n'), ((5034, 5080), 'nk_sent2vec.Sent2Vec', '_Sent2Vec', ([], {'path': "self.volumes['sent2vec_model']"}), "(path=self.volumes['sent2vec_model']...
# ============================================================================ # # Copyright (c) 2007-2010 Integral Technology Solutions Pty Ltd, # All Rights Reserved. # # THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR # IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABIL...
[ "validation_helper.printHeader", "java.io.File" ]
[((1218, 1276), 'validation_helper.printHeader', 'helper.printHeader', (['"""[VALIDATING] admin server properties"""'], {}), "('[VALIDATING] admin server properties')\n", (1236, 1276), True, 'import validation_helper as helper\n'), ((6765, 6779), 'java.io.File', 'File', (['filename'], {}), '(filename)\n', (6769, 6779),...
import sys # import the GameState of the game from Game.GameStateConnect4 import GameState # import all agents from Agents.MCTS import MCTSTree from Agents.Random import RandomAgent from Agents.AlphaBeta import AlphaBetaAgent # creates the board string for connect4 (full of zeros) start_board_list = ["000000 ...
[ "Game.GameStateConnect4.GameState", "Agents.Random.RandomAgent", "Agents.MCTS.MCTSTree", "sys.exit", "Agents.AlphaBeta.AlphaBetaAgent" ]
[((663, 688), 'Game.GameStateConnect4.GameState', 'GameState', (['start_board', '(1)'], {}), '(start_board, 1)\n', (672, 688), False, 'from Game.GameStateConnect4 import GameState\n'), ((1115, 1130), 'Agents.MCTS.MCTSTree', 'MCTSTree', (['state'], {}), '(state)\n', (1123, 1130), False, 'from Agents.MCTS import MCTSTree...
import os import subprocess import tempfile try: from PyQt5.QtCore import QBuffer, QIODevice, Qt from PyQt5.QtGui import QImage except ImportError: from PySide2.QtCore import QBuffer, QIODevice, Qt from PySide2.QtGui import QImage from .texture_format import TextureFormat def imageToBytes(image): ...
[ "os.path.exists", "tempfile.gettempdir", "os.getpid", "PySide2.QtCore.QBuffer", "PySide2.QtGui.QImage", "os.remove" ]
[((331, 340), 'PySide2.QtCore.QBuffer', 'QBuffer', ([], {}), '()\n', (338, 340), False, 'from PySide2.QtCore import QBuffer, QIODevice, Qt\n'), ((944, 969), 'os.path.exists', 'os.path.exists', (['temp_path'], {}), '(temp_path)\n', (958, 969), False, 'import os\n'), ((618, 630), 'PySide2.QtGui.QImage', 'QImage', (['path...
########################################################################## # PyPipe - Copyright (C) AGrigis, 2017 # Distributed under the terms of the CeCILL-B license, as published by # the CEA-CNRS-INRIA. Refer to the LICENSE file or to # http://www.cecill.info/licences/Licence_CeCILL-B_V1-en.html # for details. ####...
[ "PySide2.QtGui.QFont", "PySide2.QtWidgets.QTreeWidgetItem" ]
[((490, 526), 'PySide2.QtGui.QFont', 'QtGui.QFont', (['""""""', '(9)', 'QtGui.QFont.Bold'], {}), "('', 9, QtGui.QFont.Bold)\n", (501, 526), False, 'from PySide2 import QtGui\n'), ((2512, 2589), 'PySide2.QtWidgets.QTreeWidgetItem', 'QtWidgets.QTreeWidgetItem', (['parent_item', "[module_name, 'None', 'None', 'None']"], {...
""" Module implements simple ORM for SQLite. Module excludes using many-to-many and one-to-many relationships. Trying to save the same object (update) with another aggregated object will rewrite old object! """ import os import sqlite3 from array import array from inspect import * import builtins import sys import lo...
[ "logging.basicConfig", "sqlite3.connect", "os.path.getsize", "logging.getLogger" ]
[((864, 937), 'logging.basicConfig', 'logging.basicConfig', ([], {'level': 'logging.DEBUG', 'filename': 'log_file', 'filemode': '"""a"""'}), "(level=logging.DEBUG, filename=log_file, filemode='a')\n", (883, 937), False, 'import logging\n'), ((983, 1015), 'logging.getLogger', 'logging.getLogger', (['"""main_logger"""'],...
from setuptools import setup, find_packages with open("README.md", "r") as fh: long_description = fh.read() setup( name="algorithms", version="0.1", description="Implements a few optimisation algorithms", long_description=long_description, long_description_content_type="text/markdown", ur...
[ "setuptools.find_packages" ]
[((379, 394), 'setuptools.find_packages', 'find_packages', ([], {}), '()\n', (392, 394), False, 'from setuptools import setup, find_packages\n')]
import pandas as pd import dimensionality_reduction_functions as dim_red from plotting_functions import colored_line_plot, colored_line_and_scatter_plot, colored_line_plot_projected_data # Number of PCA components ndim = 3 ####################################### EXAMPLE 4: CYCLOPROPYLIDENE BIFURCATION ###############...
[ "dimensionality_reduction_functions.pathreducer", "dimensionality_reduction_functions.transform_new_data", "plotting_functions.colored_line_and_scatter_plot", "pandas.DataFrame", "plotting_functions.colored_line_plot_projected_data" ]
[((882, 971), 'dimensionality_reduction_functions.pathreducer', 'dim_red.pathreducer', (['file', 'ndim'], {'stereo_atoms': 'stereo_atoms_B', 'input_type': '"""Distances"""'}), "(file, ndim, stereo_atoms=stereo_atoms_B, input_type=\n 'Distances')\n", (901, 971), True, 'import dimensionality_reduction_functions as dim...
''' This code compares the loc and iloc in pandas dataframe ''' __author__ = "<NAME>" __email__ = "<EMAIL>" import pandas as pd import timeit df_test = pd.DataFrame() tlist = [] tlist2 = [] ################ this code creates a dataframe df_test ################## ###############with two columns and 5000000 entrie...
[ "pandas.DataFrame" ]
[((156, 170), 'pandas.DataFrame', 'pd.DataFrame', ([], {}), '()\n', (168, 170), True, 'import pandas as pd\n')]
from functools import total_ordering from random import shuffle class Player: def __init__(self, name): self.name = name self.hand = [] def __str__(self): return self.name def play(self): return self.hand.pop() def receive(self, cards): for card in cards: ...
[ "random.shuffle" ]
[((1331, 1350), 'random.shuffle', 'shuffle', (['self.cards'], {}), '(self.cards)\n', (1338, 1350), False, 'from random import shuffle\n')]
import numpy as np import matplotlib.pyplot as plt from matplotlib.ticker import MaxNLocator cz2 = (0.7, 0.7, 0.7) cz = (0.3, 0.3, 0.3) cy = (0.7, 0.4, 0.12) ci = (0.1, 0.3, 0.5) ct = (0.7, 0.2, 0.1) ax = plt.figure(figsize=(5,4)).gca() ax.xaxis.set_major_locator(MaxNLocator(integer=True)) ax.yaxis.grid(True) ax.set_...
[ "matplotlib.pyplot.figure", "matplotlib.ticker.MaxNLocator", "matplotlib.pyplot.savefig", "matplotlib.pyplot.legend" ]
[((1091, 1139), 'matplotlib.pyplot.legend', 'plt.legend', ([], {'handles': '[bt, cat, cet, mat, mmt, ht]'}), '(handles=[bt, cat, cet, mat, mmt, ht])\n', (1101, 1139), True, 'import matplotlib.pyplot as plt\n'), ((1153, 1191), 'matplotlib.pyplot.savefig', 'plt.savefig', (['"""curvetest.png"""'], {'dpi': '(1500)'}), "('c...
#!/usr/bin/env python # -*- coding: utf-8 -*- """ DataWorkshop: application to handle data, e.g. generated from imageviewer Author: <NAME> Created: Sep. 23rd, 2015 """ from ...utils import datautils from ...utils import miscutils from ...utils import funutils from ...utils import resutils import wx import wx.lib.mix...
[ "wx.App" ]
[((1259, 1297), 'wx.App', 'wx.App', ([], {'redirect': 'logon', 'filename': '"""log"""'}), "(redirect=logon, filename='log')\n", (1265, 1297), False, 'import wx\n')]
"""Authorization token handling.""" import logging from functools import wraps from flask import g, request from requests import get from pydantic.error_wrappers import ValidationError from bayesian.utility.user_utils import get_user, UserException, UserNotFoundException from bayesian.utility.v2.sa_models import Header...
[ "logging.getLogger", "requests.get", "functools.wraps", "bayesian.utility.user_utils.get_user", "bayesian.exceptions.HTTPError", "flask.request.headers.get" ]
[((466, 493), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (483, 493), False, 'import logging\n'), ((757, 793), 'flask.request.headers.get', 'request.headers.get', (['"""Authorization"""'], {}), "('Authorization')\n", (776, 793), False, 'from flask import g, request\n'), ((1273, 1284), ...
import time from signal import pause import logging import RPi.GPIO as GPIO GPIO.setmode(GPIO.BCM) logger = logging.getLogger(__name__) map_edge_parse = {'falling':GPIO.FALLING, 'rising':GPIO.RISING, 'both':GPIO.BOTH} map_pull_parse = {'pull_up':GPIO.PUD_UP, 'pull_down':GPIO.PUD_DOWN, 'pull_off':GPIO.PUD_OFF} map_edg...
[ "logging.getLogger", "RPi.GPIO.add_event_detect", "RPi.GPIO.setup", "time.perf_counter", "time.sleep", "signal.pause", "RPi.GPIO.remove_event_detect", "RPi.GPIO.input", "RPi.GPIO.setmode" ]
[((76, 98), 'RPi.GPIO.setmode', 'GPIO.setmode', (['GPIO.BCM'], {}), '(GPIO.BCM)\n', (88, 98), True, 'import RPi.GPIO as GPIO\n'), ((109, 136), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (126, 136), False, 'import logging\n'), ((1899, 1918), 'time.perf_counter', 'time.perf_counter', ([...
from setuptools import setup setup(name='gtkpass', version='0.2.7', description='A GTK+ 3 program for the standard unix password manager', url='http://github.com/raghavsub/gtkpass', author='<NAME>', author_email='<EMAIL>', license='MIT', packages=['gtkpass'], entry_point...
[ "setuptools.setup" ]
[((30, 366), 'setuptools.setup', 'setup', ([], {'name': '"""gtkpass"""', 'version': '"""0.2.7"""', 'description': '"""A GTK+ 3 program for the standard unix password manager"""', 'url': '"""http://github.com/raghavsub/gtkpass"""', 'author': '"""<NAME>"""', 'author_email': '"""<EMAIL>"""', 'license': '"""MIT"""', 'packa...
import pytest import time from .utils import ( init_app, init_db, clean_db, add_flow, add_run, add_step, add_task, add_artifact, _test_list_resources, _test_single_resource, add_metadata, get_heartbeat_ts ) pytestmark = [pytest.mark.integration_tests] # Fixtures begin @pytest.fixture def cli(loop, aiohtt...
[ "time.time" ]
[((15496, 15507), 'time.time', 'time.time', ([], {}), '()\n', (15505, 15507), False, 'import time\n')]
import torch from torch.optim import lr_scheduler from tqdm import tqdm from torchsummary import summary from torch.utils.tensorboard import SummaryWriter from apex import amp from loss import dice from pathlib import Path from data import CaseDataset, load_case, save_pred, \ orient_crop_case, regions_crop_case, re...
[ "apex.amp.scale_loss", "torch.softmax", "numpy.array", "apex.amp.initialize", "transform.crop_pad", "torch.squeeze", "transform.pad", "numpy.arange", "torch.isnan", "torch.utils.tensorboard.SummaryWriter", "apex.amp.load_state_dict", "pathlib.Path", "data.resample_normalize_case", "numpy.t...
[((881, 903), 'transform.pad', 'pad', (['input', 'patch_size'], {}), '(input, patch_size)\n', (884, 903), False, 'from transform import pad, crop_pad, to_numpy, to_tensor, resize\n'), ((958, 998), 'numpy.array', 'np.array', (['[(i // 2) for i in patch_size]'], {}), '([(i // 2) for i in patch_size])\n', (966, 998), True...
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Thu Jan 09 22:25:07 2019 @author: arnaudhub """ #import pandas as pd from sqlalchemy import create_engine from sqlalchemy.sql import text import configparser,os from urllib import parse #import sql.connector config = configparser.ConfigParser() config.re...
[ "urllib.parse.quote_plus", "configparser.ConfigParser", "os.path.expanduser" ]
[((283, 310), 'configparser.ConfigParser', 'configparser.ConfigParser', ([], {}), '()\n', (308, 310), False, 'import configparser, os\n'), ((333, 375), 'os.path.expanduser', 'os.path.expanduser', (['"""~/Bureau/OBJDOMO.cnf"""'], {}), "('~/Bureau/OBJDOMO.cnf')\n", (351, 375), False, 'import configparser, os\n'), ((502, ...
from multiml import logger from multiml.task.pytorch import PytorchASNGNASTask from multiml.task.pytorch import PytorchASNGNASBlockTask from . import PytorchConnectionRandomSearchAgent from multiml.task.pytorch.datasets import StoreGateDataset, NumpyDataset import numpy as np class PytorchASNGNASAgent(PytorchConnect...
[ "multiml.task.pytorch.PytorchASNGNASBlockTask", "multiml.logger.info", "multiml.task.pytorch.PytorchASNGNASTask" ]
[((2035, 2607), 'multiml.task.pytorch.PytorchASNGNASTask', 'PytorchASNGNASTask', ([], {'asng_args': 'self.asng_args', 'subtasks': 'asng_block_list', 'variable_mapping': "self._connectiontask_args['variable_mapping']", 'saver': 'self._saver', 'device': "self._connectiontask_args['device']", 'gpu_ids': 'None', 'amp': '(F...
import os import pytest from petisco import FlaskApplication SWAGGER_DIR = os.path.dirname(os.path.abspath(__file__)) + "/application/" app = FlaskApplication(application_name="petisco", swagger_dir=SWAGGER_DIR).get_app() @pytest.fixture def client(): with app.app.test_client() as c: yield c @pytest...
[ "os.path.abspath", "petisco.FlaskApplication" ]
[((95, 120), 'os.path.abspath', 'os.path.abspath', (['__file__'], {}), '(__file__)\n', (110, 120), False, 'import os\n'), ((146, 215), 'petisco.FlaskApplication', 'FlaskApplication', ([], {'application_name': '"""petisco"""', 'swagger_dir': 'SWAGGER_DIR'}), "(application_name='petisco', swagger_dir=SWAGGER_DIR)\n", (16...
from sklearn.datasets import load_iris from sklearn.linear_model import LogisticRegression from flask import Blueprint, jsonify, request, flash, redirect, render_template from web_app.models import User from web_app.statsmodels import load_model from web_app.services.basilica_service import connection as basilica_con...
[ "sklearn.datasets.load_iris", "flask.render_template", "web_app.statsmodels.load_model", "sklearn.linear_model.LogisticRegression", "web_app.models.User.query.filter", "web_app.services.basilica_service.connection.embed_sentence", "flask.Blueprint" ]
[((344, 379), 'flask.Blueprint', 'Blueprint', (['"""stats_routes"""', '__name__'], {}), "('stats_routes', __name__)\n", (353, 379), False, 'from flask import Blueprint, jsonify, request, flash, redirect, render_template\n'), ((439, 465), 'sklearn.datasets.load_iris', 'load_iris', ([], {'return_X_y': '(True)'}), '(retur...
import argparse import os import pyprind import utils import treetk import treetk.rstdt def main(args): """ We use n-ary ctrees (ie., *.labeled.nary.ctree) to generate dtrees. Morey et al. (2018) demonstrate that scores evaluated on these dtrees are superficially lower than those on right-heavy binarize...
[ "os.listdir", "argparse.ArgumentParser", "treetk.ctree2dtree", "os.path.join", "treetk.sexp2tree", "treetk.rstdt.assign_heads", "pyprind.prog_bar" ]
[((402, 418), 'os.listdir', 'os.listdir', (['path'], {}), '(path)\n', (412, 418), False, 'import os\n'), ((844, 871), 'pyprind.prog_bar', 'pyprind.prog_bar', (['filenames'], {}), '(filenames)\n', (860, 871), False, 'import pyprind\n'), ((1723, 1748), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {}), '()\n...
from contextlib import closing import h5py import numpy as np def save_h5(outfile, dictionary): """ Saves passed dictionary to an h5 file Parameters ---------- outfile : string Name of output h5 file dictionary : dictionary Dictionary that will be saved """ def save_layer(...
[ "numpy.asarray", "h5py.File" ]
[((1659, 1687), 'h5py.File', 'h5py.File', (['feature_file', '"""r"""'], {}), "(feature_file, 'r')\n", (1668, 1687), False, 'import h5py\n'), ((622, 645), 'h5py.File', 'h5py.File', (['outfile', '"""w"""'], {}), "(outfile, 'w')\n", (631, 645), False, 'import h5py\n'), ((1830, 1848), 'numpy.asarray', 'np.asarray', (['f[ke...
# -*- coding: utf-8 -*- """ Helper functions for VariationalModel class """ from __future__ import print_function from __future__ import division import math import random import tensorflow as tf from tensorflow.contrib.legacy_seq2seq.python.ops import seq2seq as s2s def linearOutcomePrediction(zs, pa...
[ "tensorflow.contrib.legacy_seq2seq.python.ops.seq2seq.core_rnn_cell.EmbeddingWrapper", "tensorflow.sigmoid", "tensorflow.matmul", "tensorflow.contrib.legacy_seq2seq.python.ops.seq2seq.embedding_rnn_decoder", "tensorflow.contrib.legacy_seq2seq.python.ops.seq2seq.rnn.static_rnn", "tensorflow.contrib.legacy_...
[((648, 717), 'tensorflow.contrib.legacy_seq2seq.python.ops.seq2seq.variable_scope.variable_scope', 's2s.variable_scope.variable_scope', (["(scope or 'outcomepred')"], {'reuse': '(True)'}), "(scope or 'outcomepred', reuse=True)\n", (681, 717), True, 'from tensorflow.contrib.legacy_seq2seq.python.ops import seq2seq as s...
"""empty message Revision ID: b9ab1a9a2113 Revises: Create Date: 2021-11-28 22:41:01.160642 """ from alembic import op import sqlalchemy as sa # revision identifiers, used by Alembic. revision = 'b9ab1a9a2113' down_revision = None branch_labels = None depends_on = None def upgrade(): # ### commands auto gene...
[ "sqlalchemy.ForeignKeyConstraint", "sqlalchemy.Float", "alembic.op.drop_table", "sqlalchemy.PrimaryKeyConstraint", "sqlalchemy.Date", "sqlalchemy.Integer", "sqlalchemy.String" ]
[((1073, 1098), 'alembic.op.drop_table', 'op.drop_table', (['"""currates"""'], {}), "('currates')\n", (1086, 1098), False, 'from alembic import op\n'), ((1103, 1128), 'alembic.op.drop_table', 'op.drop_table', (['"""curpairs"""'], {}), "('curpairs')\n", (1116, 1128), False, 'from alembic import op\n'), ((575, 604), 'sql...
import os from pathlib import Path import pandas as pd from lime.lime_tabular import LimeTabularExplainer from ml_editor.data_processing import get_split_by_author FEATURE_DISPLAY_NAMES = { "num_questions": "물음표 빈도", "num_periods": "마침표 빈도", "num_commas": "쉼표 빈도", "num_exclam": "느낌표 빈도", "num_quot...
[ "pandas.read_csv", "pathlib.Path", "lime.lime_tabular.LimeTabularExplainer", "os.path.dirname", "ml_editor.data_processing.get_split_by_author" ]
[((1881, 1922), 'pathlib.Path', 'Path', (['"""../data/writers_with_features.csv"""'], {}), "('../data/writers_with_features.csv')\n", (1885, 1922), False, 'from pathlib import Path\n'), ((1932, 1966), 'pandas.read_csv', 'pd.read_csv', (['(curr_path / data_path)'], {}), '(curr_path / data_path)\n', (1943, 1966), True, '...
""" Tests for the test utils. """ import pytest from straitlets import Serializable, Integer from straitlets.test_utils import assert_serializables_equal def test_assert_serializables_equal(): class Foo(Serializable): x = Integer() y = Integer() class Bar(Serializable): x = Integer(...
[ "straitlets.Integer", "pytest.raises" ]
[((238, 247), 'straitlets.Integer', 'Integer', ([], {}), '()\n', (245, 247), False, 'from straitlets import Serializable, Integer\n'), ((260, 269), 'straitlets.Integer', 'Integer', ([], {}), '()\n', (267, 269), False, 'from straitlets import Serializable, Integer\n'), ((312, 321), 'straitlets.Integer', 'Integer', ([], ...
import numpy from fdm.geometry import create_close_point_finder def create_weights_distributor(close_point_finder): def distribute(point, value): close_points = close_point_finder(point) distance_sum = sum(close_points.values()) return dict( {p: (1. - distance/distance_sum)*va...
[ "fdm.geometry.create_close_point_finder", "numpy.copy", "numpy.zeros" ]
[((589, 607), 'numpy.copy', 'numpy.copy', (['matrix'], {}), '(matrix)\n', (599, 607), False, 'import numpy\n'), ((622, 640), 'numpy.copy', 'numpy.copy', (['vector'], {}), '(vector)\n', (632, 640), False, 'import numpy\n'), ((1565, 1585), 'numpy.copy', 'numpy.copy', (['matrix_a'], {}), '(matrix_a)\n', (1575, 1585), Fals...
import unittest from csound import output, orchestra from csound.orchestra import gen08 from data import constants as c from data import get class TestSounds(unittest.TestCase): def test_simple_soundwaves(self): # Get all data place = "Madrid" mad2t = get(c.T, location=place) ma...
[ "csound.orchestra.oscillator1", "data.get", "csound.orchestra.gen08", "csound.output.get_csd" ]
[((285, 309), 'data.get', 'get', (['c.T'], {'location': 'place'}), '(c.T, location=place)\n', (288, 309), False, 'from data import get\n'), ((325, 349), 'data.get', 'get', (['c.P'], {'location': 'place'}), '(c.P, location=place)\n', (328, 349), False, 'from data import get\n'), ((365, 389), 'data.get', 'get', (['c.W'],...
#!/usr/bin/env python """ Nicholas' Example API code for interacting with Alienvault API. This is just Example code written by NMA.IO. There isn't really much you can do with the API just yet, so this will be a work in progress. Grab your API key here: https://www.alienvault.com/documentation/usm-anywhere/api/ali...
[ "base64.b64encode" ]
[((828, 873), 'base64.b64encode', 'base64.b64encode', (["('%s:%s' % (apiuser, apikey))"], {}), "('%s:%s' % (apiuser, apikey))\n", (844, 873), False, 'import base64\n')]
from timebox.timebox import TimeBox from timebox.utils.exceptions import InvalidPandasIndexError import pandas as pd import numpy as np import unittest import os import logging class TestTimeBoxPandas(unittest.TestCase): def test_save_pandas(self): file_name = 'save_pandas.npb' df = pd.read_csv('t...
[ "os.path.exists", "pandas.read_csv", "timebox.timebox.TimeBox.save_pandas", "timebox.timebox.TimeBox", "numpy.array", "unittest.main", "pandas.to_datetime", "os.remove" ]
[((2124, 2139), 'unittest.main', 'unittest.main', ([], {}), '()\n', (2137, 2139), False, 'import unittest\n'), ((306, 377), 'pandas.read_csv', 'pd.read_csv', (['"""timebox/tests/data/ETH-USD_combined_utc.csv"""'], {'index_col': '(0)'}), "('timebox/tests/data/ETH-USD_combined_utc.csv', index_col=0)\n", (317, 377), True,...
from typing import Any from flaskapp.models import Activities, Logs, LogsToActivities, db from sqlalchemy.sql.expression import desc, func """ Este modulo contiene funciones para escribir y leer los logs desde el chatbot """ MAX_LOGS_PER_QUERY = 5 MAX_STR_SIZE_LOGS = 128 def write_log( chatid: int, intent: ...
[ "flaskapp.models.Activities.query.filter_by", "flaskapp.models.LogsToActivities", "flaskapp.models.db.session.commit", "sqlalchemy.sql.expression.desc", "flaskapp.models.Activities", "flaskapp.models.Logs", "sqlalchemy.sql.expression.func.max" ]
[((1232, 1251), 'flaskapp.models.db.session.commit', 'db.session.commit', ([], {}), '()\n', (1249, 1251), False, 'from flaskapp.models import Activities, Logs, LogsToActivities, db\n'), ((1843, 1862), 'flaskapp.models.db.session.commit', 'db.session.commit', ([], {}), '()\n', (1860, 1862), False, 'from flaskapp.models ...
""" Luhn Algorithm """ from typing import List def is_luhn(string: str) -> bool: """ Perform Luhn validation on input string Algorithm: * Double every other digit starting from 2nd last digit. * Subtract 9 if number is greater than 9. * Sum the numbers * >>> test_cases = [79927398710, ...
[ "doctest.testmod" ]
[((1181, 1198), 'doctest.testmod', 'doctest.testmod', ([], {}), '()\n', (1196, 1198), False, 'import doctest\n')]
""" ================== welly ================== """ from .project import Project from .well import Well from .header import Header from .curve import Curve from .synthetic import Synthetic from .location import Location from .crs import CRS from . import tools from . import quality def read_las(path, **kwargs): "...
[ "pkg_resources.get_distribution" ]
[((1709, 1735), 'pkg_resources.get_distribution', 'get_distribution', (['__name__'], {}), '(__name__)\n', (1725, 1735), False, 'from pkg_resources import get_distribution, DistributionNotFound\n')]
#!/usr/bin/env python3 """ Build the demos Usage: python setup.py build_ext -i """ import numpy as np from distutils.core import setup from Cython.Build import cythonize from setuptools.extension import Extension from os.path import join extending = Extension("extending", sources=['extending.py...
[ "Cython.Build.cythonize", "os.path.join", "numpy.get_include" ]
[((760, 781), 'Cython.Build.cythonize', 'cythonize', (['extensions'], {}), '(extensions)\n', (769, 781), False, 'from Cython.Build import cythonize\n'), ((361, 377), 'numpy.get_include', 'np.get_include', ([], {}), '()\n', (375, 377), True, 'import numpy as np\n'), ((534, 593), 'os.path.join', 'join', (['""".."""', '""...
#!/usr/bin/env python # encoding: utf-8 import signal import sys #import pandas as pd #import numpy as np def setGlobals(g): #print globals() globals().update(g) #print globals() def exit(): mtsExit(0) def quit(signum): sys.exit(signum) def quitNow(signum,frame): quit(signum) def initializ...
[ "signal.signal", "sys.exit" ]
[((244, 260), 'sys.exit', 'sys.exit', (['signum'], {}), '(signum)\n', (252, 260), False, 'import sys\n'), ((343, 380), 'signal.signal', 'signal.signal', (['signal.SIGINT', 'quitNow'], {}), '(signal.SIGINT, quitNow)\n', (356, 380), False, 'import signal\n'), ((385, 423), 'signal.signal', 'signal.signal', (['signal.SIGTE...
from enum import Enum import json from SCA11H.commands.base.PostCommand import PostCommand class Command(Enum): # Restore BCG factory settings.(BCG parameters, direction and running mode) Restore = ('restore', 'restore') # Restore default BCG parameters SetDefaultParameters = ('set_default_pars', 'r...
[ "json.dumps" ]
[((533, 570), 'json.dumps', 'json.dumps', (["{'cmd': command.value[0]}"], {}), "({'cmd': command.value[0]})\n", (543, 570), False, 'import json\n')]
#!/usr/bin/env python3 import pandas as pd ft_input='TEMP_DIR/tmp-predictions_reformatted_gexpnn20200320allCOHORTS.tsv' df = pd.read_csv(ft_input,sep='\t') # Get all tumors present in df (ACC, BRCA, ...) temp = df['Label'].unique() u_tumor = {} #k=tumor, v=1 for t in temp: t= t.split(":")[0] if t not in u_t...
[ "pandas.read_csv" ]
[((128, 159), 'pandas.read_csv', 'pd.read_csv', (['ft_input'], {'sep': '"""\t"""'}), "(ft_input, sep='\\t')\n", (139, 159), True, 'import pandas as pd\n')]