code stringlengths 22 1.05M | apis listlengths 1 3.31k | extract_api stringlengths 75 3.25M |
|---|---|---|
from unittest import TestCase
from CTCI.Ch2_Linked_Lists.common.SinglyLinkedList import Empty, Node
from CTCI.Ch2_Linked_Lists.exercises.CTCI_Ch2_Ex6 import PalindromeSinglyLinkedList, is_palindrome_brute_force
from CTCI.Ch2_Linked_Lists.exercises.CTCI_Ch2_Ex6 import is_palindrome_reverse
class TestPalindromeSinglyLi... | [
"CTCI.Ch2_Linked_Lists.common.SinglyLinkedList.Node",
"CTCI.Ch2_Linked_Lists.exercises.CTCI_Ch2_Ex6.is_palindrome_brute_force",
"CTCI.Ch2_Linked_Lists.exercises.CTCI_Ch2_Ex6.PalindromeSinglyLinkedList",
"CTCI.Ch2_Linked_Lists.exercises.CTCI_Ch2_Ex6.is_palindrome_reverse"
] | [((381, 409), 'CTCI.Ch2_Linked_Lists.exercises.CTCI_Ch2_Ex6.PalindromeSinglyLinkedList', 'PalindromeSinglyLinkedList', ([], {}), '()\n', (407, 409), False, 'from CTCI.Ch2_Linked_Lists.exercises.CTCI_Ch2_Ex6 import PalindromeSinglyLinkedList, is_palindrome_brute_force\n'), ((1752, 1759), 'CTCI.Ch2_Linked_Lists.common.Si... |
# utils for working with 3d-protein structures
import os
import numpy as np
import torch
from functools import wraps
from einops import rearrange, repeat
# import torch_sparse # only needed for sparse nth_deg adj calculation
# bio
from Bio import SeqIO
import itertools
import string
# sidechainnet
from sidechainnet... | [
"numpy.random.rand",
"torch.sin",
"torch.det",
"torch.searchsorted",
"torch.cdist",
"torch.min",
"numpy.array",
"torch.cos",
"numpy.linalg.norm",
"numpy.arange",
"torch.logical_or",
"numpy.mean",
"numpy.cross",
"torch.mean",
"einops.repeat",
"torch.logical_not",
"functools.wraps",
... | [((641, 660), 'sidechainnet.utils.sequence.ProteinVocabulary', 'ProteinVocabulary', ([], {}), '()\n', (658, 660), False, 'from sidechainnet.utils.sequence import ProteinVocabulary, ONE_TO_THREE_LETTER_MAP\n'), ((842, 898), 'torch.linspace', 'torch.linspace', (['(2)', '(20)'], {'steps': 'constants.DISTOGRAM_BUCKETS'}), ... |
"""
Interfaces with Verisure sensors.
For more details about this platform, please refer to the documentation at
documentation at https://home-assistant.io/components/verisure/
"""
import logging
from homeassistant.components.verisure import HUB as hub
from homeassistant.const import TEMP_CELSIUS
from homeassistant.h... | [
"logging.getLogger",
"homeassistant.components.verisure.HUB.climate_status.values",
"homeassistant.components.verisure.HUB.mouse_status.values",
"homeassistant.components.verisure.HUB.config.get",
"homeassistant.components.verisure.HUB.update_climate",
"homeassistant.components.verisure.HUB.update_mousede... | [((359, 386), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (376, 386), False, 'import logging\n'), ((525, 560), 'homeassistant.components.verisure.HUB.config.get', 'hub.config.get', (['"""thermometers"""', '"""1"""'], {}), "('thermometers', '1')\n", (539, 560), True, 'from homeassistant... |
import factory
from factory.django import DjangoModelFactory as Factory
from django.contrib.auth.models import Permission
from ..models import Blog
from articles.users.tests.factories import UserFactory
class Blogfactory(Factory):
user = user = factory.SubFactory(UserFactory)
title = factory.Faker('sentence... | [
"factory.SubFactory",
"factory.Faker",
"django.contrib.auth.models.Permission.objects.filter",
"articles.users.tests.factories.UserFactory"
] | [((253, 284), 'factory.SubFactory', 'factory.SubFactory', (['UserFactory'], {}), '(UserFactory)\n', (271, 284), False, 'import factory\n'), ((297, 334), 'factory.Faker', 'factory.Faker', (['"""sentence"""'], {'nb_words': '(3)'}), "('sentence', nb_words=3)\n", (310, 334), False, 'import factory\n'), ((353, 395), 'factor... |
import sqlite3
class Database:
# create book always if not exists
def __init__(self,db):
self.conn = sqlite3.connect(db)
self.cur = self.conn.execute("CREATE TABLE IF NOT EXISTS book (id INTEGER PRIMARY KEY, " +
"title TEXT, author TEXT, year INTEGER, isbn INTEGER)"... | [
"sqlite3.connect"
] | [((119, 138), 'sqlite3.connect', 'sqlite3.connect', (['db'], {}), '(db)\n', (134, 138), False, 'import sqlite3\n')] |
#!/usr/bin/env python
# ----------------------------------------------------------------------------
# Copyright 2018 ARM Ltd.
#
# SPDX-License-Identifier: Apache-2.0
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain ... | [
"os.path.isfile",
"pyclibrary.CParser",
"collections.namedtuple",
"cbor2.dump"
] | [((2216, 2259), 'collections.namedtuple', 'namedtuple', (['"""ConfigParam"""', "['Data', 'Name']"], {}), "('ConfigParam', ['Data', 'Name'])\n", (2226, 2259), False, 'from collections import namedtuple\n'), ((2274, 2327), 'collections.namedtuple', 'namedtuple', (['"""Certificate"""', "['Data', 'Format', 'Name']"], {}), ... |
from django.db import models
# Create your models here.
class Owner(models.Model):
Owner_id = models.AutoField
Owner_firstname = models.CharField(max_length=60)
Owner_lastname = models.CharField(max_length=60)
Owner_address = models.CharField(max_length=600)
Owner_email = models.CharField(max_lengt... | [
"django.db.models.DateField",
"django.db.models.IntegerField",
"django.db.models.BooleanField",
"django.db.models.ImageField",
"django.db.models.CharField"
] | [((138, 169), 'django.db.models.CharField', 'models.CharField', ([], {'max_length': '(60)'}), '(max_length=60)\n', (154, 169), False, 'from django.db import models\n'), ((191, 222), 'django.db.models.CharField', 'models.CharField', ([], {'max_length': '(60)'}), '(max_length=60)\n', (207, 222), False, 'from django.db im... |
import numpy as np
import tectosaur.util.gpu as gpu
from tectosaur.fmm.c2e import build_c2e
import logging
logger = logging.getLogger(__name__)
def make_tree(m, cfg, max_pts_per_cell):
tri_pts = m[0][m[1]]
centers = np.mean(tri_pts, axis = 1)
pt_dist = tri_pts - centers[:,np.newaxis,:]
Rs = np.max(np... | [
"logging.getLogger",
"numpy.mean",
"tectosaur.fmm.c2e.build_c2e",
"tectosaur.util.gpu.to_gpu",
"numpy.array",
"numpy.empty_like",
"numpy.linalg.norm"
] | [((118, 145), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (135, 145), False, 'import logging\n'), ((227, 251), 'numpy.mean', 'np.mean', (['tri_pts'], {'axis': '(1)'}), '(tri_pts, axis=1)\n', (234, 251), True, 'import numpy as np\n'), ((318, 349), 'numpy.linalg.norm', 'np.linalg.norm', ... |
#!/usr/bin/env python
# -*- coding: utf-8 -*-
#
# Project: Fable Input Output
# https://github.com/silx-kit/fabio
#
# Copyright (C) European Synchrotron Radiation Facility, Grenoble, France
#
# Principal author: <NAME> (<EMAIL>)
#
# This program is free software: you can redistribute it an... | [
"logging.getLogger",
"unittest.TestSuite",
"unittest.TextTestRunner",
"fabio.mpaimage.MpaImage"
] | [((1074, 1101), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (1091, 1101), False, 'import logging\n'), ((2354, 2374), 'unittest.TestSuite', 'unittest.TestSuite', ([], {}), '()\n', (2372, 2374), False, 'import unittest\n'), ((2480, 2505), 'unittest.TextTestRunner', 'unittest.TextTestRunn... |
# -*- coding: utf-8 -*-
import scrapy
import codecs
import sys
#리눅스상에서 utf-8 로 파일에 내용을 기록하려면 시스템 기본 인코딩으 ㄹutf-8 로 설정해야함
reload(sys)
sys.setdefaultencoding('utf8')
# scrapy 에서 spider 는 crawling/scrapping을 담당하는 핵심부분
#crawling/scrapping 절차에 대한 정의를 하는 부분
class CurrSpider(scrapy.Spider):
name = 'currSpider'
start_... | [
"codecs.open",
"sys.setdefaultencoding"
] | [((132, 162), 'sys.setdefaultencoding', 'sys.setdefaultencoding', (['"""utf8"""'], {}), "('utf8')\n", (154, 162), False, 'import sys\n'), ((565, 602), 'codecs.open', 'codecs.open', (['"""curr.csv"""', '"""w"""', '"""utf-8"""'], {}), "('curr.csv', 'w', 'utf-8')\n", (576, 602), False, 'import codecs\n')] |
import __init__
import os
#os.environ['LD_LIBRARY_PATH'] += ':/usr/local/cuda-11.1/bin64:/usr/local/cuda-11.2/bin64'
import numpy as np
import torch
import torch.multiprocessing as mp
import torch_geometric.datasets as GeoData
from torch_geometric.loader import DenseDataLoader
import torch_geometric.transforms as T... | [
"utils.metrics.AverageMeter",
"torch.nn.CrossEntropyLoss",
"torch.utils.data.distributed.DistributedSampler",
"logging.info",
"torch.utils.tensorboard.SummaryWriter",
"numpy.mean",
"config.OptInit",
"comm.is_main_process",
"comm.get_local_rank",
"numpy.isnan",
"parallel_wrapper.launch",
"utils... | [((799, 832), 'torch.utils.tensorboard.SummaryWriter', 'SummaryWriter', ([], {'log_dir': '"""log/mlp4"""'}), "(log_dir='log/mlp4')\n", (812, 832), False, 'from torch.utils.tensorboard import SummaryWriter\n'), ((2714, 2727), 'numpy.mean', 'np.mean', (['ious'], {}), '(ious)\n', (2721, 2727), True, 'import numpy as np\n'... |
import tkinter as tk
from tkinter import ttk
win = tk.Tk()
win.title("Python GUI")
win.resizable(False, False)
win.configure(background = "grey94")
a_label = ttk.Label(win, text = "Gib Deinen Namen ein:")
a_label.grid(column = 0, row = 0)
a_label.grid_configure(padx = 8, pady = 8)
def clickMe():
action.configure... | [
"tkinter.ttk.Button",
"tkinter.ttk.Entry",
"tkinter.ttk.Label",
"tkinter.StringVar",
"tkinter.Tk"
] | [((52, 59), 'tkinter.Tk', 'tk.Tk', ([], {}), '()\n', (57, 59), True, 'import tkinter as tk\n'), ((160, 204), 'tkinter.ttk.Label', 'ttk.Label', (['win'], {'text': '"""Gib Deinen Namen ein:"""'}), "(win, text='Gib Deinen Namen ein:')\n", (169, 204), False, 'from tkinter import ttk\n'), ((363, 377), 'tkinter.StringVar', '... |
# -*- coding: utf-8 -*-
# Form implementation generated from reading ui file 'src/gui/ui_paste_dialog.ui'
#
# Created by: PyQt5 UI code generator 5.11.2
#
# WARNING! All changes made in this file will be lost!
from PyQt5 import QtCore, QtGui, QtWidgets
class Ui_PasteDialog(object):
def setupUi(self, PasteDialog)... | [
"PyQt5.QtCore.QMetaObject.connectSlotsByName",
"PyQt5.QtWidgets.QRadioButton",
"PyQt5.QtCore.QRect",
"PyQt5.QtWidgets.QGridLayout",
"PyQt5.QtWidgets.QGroupBox",
"PyQt5.QtWidgets.QVBoxLayout",
"PyQt5.QtWidgets.QPushButton"
] | [((434, 468), 'PyQt5.QtWidgets.QGridLayout', 'QtWidgets.QGridLayout', (['PasteDialog'], {}), '(PasteDialog)\n', (455, 468), False, 'from PyQt5 import QtCore, QtGui, QtWidgets\n'), ((649, 681), 'PyQt5.QtWidgets.QGroupBox', 'QtWidgets.QGroupBox', (['PasteDialog'], {}), '(PasteDialog)\n', (668, 681), False, 'from PyQt5 im... |
from yalul.interpreters.environment import Environment
from yalul.interpreters.expressions.var_assignment_interpreter import VarAssignmentInterpreter
from yalul.interpreters.interpreter_errors import InterpreterErrors
class TestVarAssignmentInterpreter:
"""Test var assignment expression interpreter"""
def te... | [
"yalul.interpreters.expressions.var_assignment_interpreter.VarAssignmentInterpreter",
"yalul.interpreters.environment.Environment",
"yalul.interpreters.interpreter_errors.InterpreterErrors"
] | [((485, 504), 'yalul.interpreters.interpreter_errors.InterpreterErrors', 'InterpreterErrors', ([], {}), '()\n', (502, 504), False, 'from yalul.interpreters.interpreter_errors import InterpreterErrors\n'), ((519, 538), 'yalul.interpreters.environment.Environment', 'Environment', (['{}', '{}'], {}), '({}, {})\n', (530, 5... |
import boto3
from django.conf import settings
from backend.models import CloudWatchEvent
import json
class Events:
def __init__(self):
self.client = boto3.client('events', region_name=settings.NARUKO_REGION)
def list_rules(self):
response = []
for rules in self._list_rul... | [
"boto3.client",
"backend.models.CloudWatchEvent"
] | [((173, 231), 'boto3.client', 'boto3.client', (['"""events"""'], {'region_name': 'settings.NARUKO_REGION'}), "('events', region_name=settings.NARUKO_REGION)\n", (185, 231), False, 'import boto3\n'), ((2363, 2500), 'backend.models.CloudWatchEvent', 'CloudWatchEvent', ([], {'name': "response['Name']", 'schedule_expressio... |
#!/usr/bin/env python
# Manipulate sys.path to be able to import converscript from this local git
# repository.
import os
import sys
sys.path.append(os.path.join(os.path.dirname(__file__), "..", ".."))
from converscript import RiveScript
import json
bot = RiveScript()
bot.load_file("example.rive")
dep = bot.deparse... | [
"os.path.dirname",
"json.dumps",
"converscript.RiveScript"
] | [((259, 271), 'converscript.RiveScript', 'RiveScript', ([], {}), '()\n', (269, 271), False, 'from converscript import RiveScript\n'), ((329, 354), 'json.dumps', 'json.dumps', (['dep'], {'indent': '(2)'}), '(dep, indent=2)\n', (339, 354), False, 'import json\n'), ((163, 188), 'os.path.dirname', 'os.path.dirname', (['__f... |
# -*- coding: utf-8 -*-
from __future__ import unicode_literals
from django.db import models, migrations
class Migration(migrations.Migration):
dependencies = [
('modules', '0002_module_floor'),
]
operations = [
migrations.AddField(
model_name='module',
name='sho... | [
"django.db.models.BooleanField"
] | [((346, 380), 'django.db.models.BooleanField', 'models.BooleanField', ([], {'default': '(False)'}), '(default=False)\n', (365, 380), False, 'from django.db import models, migrations\n')] |
import abc
import argparse
import logging
import pathlib
from collections import namedtuple
from operator import itemgetter
import toml
class NotConfiguredError(Exception):
pass
class ParseError(Exception):
pass
class Model(abc.ABC):
"""Interface for model that can save/load parameters.
Each mod... | [
"logging.getLogger",
"collections.namedtuple",
"pathlib.Path",
"logging.Formatter",
"operator.itemgetter"
] | [((2561, 2579), 'pathlib.Path', 'pathlib.Path', (['path'], {}), '(path)\n', (2573, 2579), False, 'import pathlib\n'), ((5102, 5125), 'logging.getLogger', 'logging.getLogger', (['name'], {}), '(name)\n', (5119, 5125), False, 'import logging\n'), ((5258, 5360), 'logging.Formatter', 'logging.Formatter', (['"""%(levelname)... |
import os
import subprocess
import threading
mutex = threading.Lock()
def render_appleseed(target_file, base_color_tex, normal_tex, roughness_tex, metallic_tex, resolution, appleseed_path):
mutex.acquire()
try:
# Read the template file from disk.
with open("scene_template.appleseed", "r") as... | [
"threading.Lock",
"os.path.splitext",
"os.path.join",
"subprocess.check_call"
] | [((54, 70), 'threading.Lock', 'threading.Lock', ([], {}), '()\n', (68, 70), False, 'import threading\n'), ((1189, 1289), 'os.path.join', 'os.path.join', (['appleseed_path', '"""bin"""', "('appleseed.cli.exe' if os.name == 'nt' else 'appleseed.cli')"], {}), "(appleseed_path, 'bin', 'appleseed.cli.exe' if os.name == 'nt'... |
from ubuntui.utils import Padding
from ubuntui.widgets.hr import HR
from conjureup.app_config import app
from conjureup.ui.views.base import BaseView, SchemaFormView
from conjureup.ui.widgets.selectors import MenuSelectButtonList
class NewCredentialView(SchemaFormView):
title = "New Credential Creation"
def... | [
"ubuntui.widgets.hr.HR",
"conjureup.ui.widgets.selectors.MenuSelectButtonList",
"ubuntui.utils.Padding.line_break",
"conjureup.app_config.app.provider.cloud_type.upper"
] | [((375, 406), 'conjureup.app_config.app.provider.cloud_type.upper', 'app.provider.cloud_type.upper', ([], {}), '()\n', (404, 406), False, 'from conjureup.app_config import app\n'), ((1051, 1103), 'conjureup.ui.widgets.selectors.MenuSelectButtonList', 'MenuSelectButtonList', (['self.credentials', 'self.default'], {}), '... |
#!/usr/bin/env python3
import transactions
import taxmap
import db
import settings
import datetime
import argparse
import uuid
import pickle
import jsonpickle
import logging
import logging.handlers
import traceback
def main():
handler = logging.handlers.RotatingFileHandler('../main.log', maxBytes=33554432, backup... | [
"transactions.getTransactionCount",
"logging.info",
"db.findReport",
"argparse.ArgumentParser",
"transactions.getTransactionList",
"db.updateReportError",
"traceback.print_exc",
"jsonpickle.loads",
"logging.handlers.RotatingFileHandler",
"pickle.load",
"logging.warning",
"db.ReportOptions",
... | [((243, 333), 'logging.handlers.RotatingFileHandler', 'logging.handlers.RotatingFileHandler', (['"""../main.log"""'], {'maxBytes': '(33554432)', 'backupCount': '(10)'}), "('../main.log', maxBytes=33554432,\n backupCount=10)\n", (279, 333), False, 'import logging\n'), ((334, 494), 'logging.basicConfig', 'logging.basi... |
"""データセットをダウンロードするためのスクリプトです."""
# default packages
import logging
import pathlib
import traceback
import urllib.request as request
# third party
import pandas as pd
import tqdm as tqdm_std
# my packages
import src.data.directory as directory
# logger
logger = logging.getLogger(__name__)
class TqdmUpTo(tqdm_std.tq... | [
"logging.getLogger",
"logging.basicConfig",
"traceback.format_exc",
"pandas.read_csv",
"urllib.request.urlretrieve",
"src.data.directory.get_raw"
] | [((264, 291), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (281, 291), False, 'import logging\n'), ((1292, 1331), 'logging.basicConfig', 'logging.basicConfig', ([], {'level': 'logging.INFO'}), '(level=logging.INFO)\n', (1311, 1331), False, 'import logging\n'), ((1833, 1854), 'pandas.rea... |
# -*- coding: utf-8 -*-
# Form implementation generated from reading ui file 'F:\work\code\pyqt5\ui\main.ui'
#
# Created by: PyQt5 UI code generator 5.9
#
# WARNING! All changes made in this file will be lost!
from PyQt5 import QtCore, QtGui, QtWidgets
class Ui_MainWindow(object):
def setupUi(self, MainWindow):
... | [
"PyQt5.QtWidgets.QWidget",
"PyQt5.QtWidgets.QTextEdit",
"PyQt5.QtWidgets.QMainWindow",
"PyQt5.QtWidgets.QComboBox",
"PyQt5.QtCore.QMetaObject.connectSlotsByName",
"PyQt5.QtWidgets.QHBoxLayout",
"PyQt5.QtWidgets.QVBoxLayout",
"PyQt5.QtWidgets.QGridLayout",
"PyQt5.QtWidgets.QStatusBar",
"PyQt5.QtWid... | [((7848, 7880), 'PyQt5.QtWidgets.QApplication', 'QtWidgets.QApplication', (['sys.argv'], {}), '(sys.argv)\n', (7870, 7880), False, 'from PyQt5 import QtCore, QtGui, QtWidgets\n'), ((7898, 7921), 'PyQt5.QtWidgets.QMainWindow', 'QtWidgets.QMainWindow', ([], {}), '()\n', (7919, 7921), False, 'from PyQt5 import QtCore, QtG... |
# Copyright 2018 ZTE Corporation.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in ... | [
"logging.getLogger",
"lcm.ns.serializers.deprecated.ns_serializers._CreateNsReqSerializer",
"lcm.pub.utils.values.ignore_case_get",
"lcm.ns.biz.ns_create.CreateNSService",
"lcm.ns.serializers.deprecated.ns_serializers._CreateNsRespSerializer",
"lcm.ns.serializers.deprecated.ns_serializers._QueryNsRespSeri... | [((1305, 1332), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (1322, 1332), False, 'import logging\n'), ((1853, 1896), 'lcm.ns.serializers.deprecated.ns_serializers._QueryNsRespSerializer', '_QueryNsRespSerializer', ([], {'data': 'ret', 'many': '(True)'}), '(data=ret, many=True)\n', (187... |
#!/usr/bin/env python3
import javalang
def isPrimitive(obj):
return not hasattr(obj, '__dict__')
def extract_bad_function_from_text(src):
return extract_function_from_text(src, criterion='bad')
def extract_function_from_text(src, criterion='bad'):
def recursive_find_deepest_child_position(node_body, ... | [
"javalang.parse.parse"
] | [((2088, 2113), 'javalang.parse.parse', 'javalang.parse.parse', (['src'], {}), '(src)\n', (2108, 2113), False, 'import javalang\n')] |
import torch
import torch.nn as nn
from torchvision.datasets.vision import VisionDataset
from PIL import Image
import os, sys, math
import os.path
import torch
import json
import torch.utils.model_zoo as model_zoo
from Yolo_v2_pytorch.src.utils import *
from Yolo_v2_pytorch.src.yolo_net import Yolo
from Yolo_v2_pytorc... | [
"torch.nn.ReLU",
"torch.nn.Dropout",
"torch.sin",
"torch.exp",
"math.log",
"numpy.array",
"torch.cos",
"torch.sum",
"torch.nn.AvgPool2d",
"torch.nn.functional.softmax",
"torch.arange",
"torch.nn.BatchNorm2d",
"Yolo_v2_pytorch.src.yolo_tunning.YoloD",
"torch.nn.LayerNorm",
"torch.matmul",... | [((1750, 1839), 'torch.nn.Conv2d', 'nn.Conv2d', (['in_planes', 'out_planes'], {'kernel_size': '(3)', 'stride': 'stride', 'padding': '(1)', 'bias': '(False)'}), '(in_planes, out_planes, kernel_size=3, stride=stride, padding=1,\n bias=False)\n', (1759, 1839), True, 'import torch.nn as nn\n'), ((6545, 6631), 'torch.nn.... |
"""
Challenge Exercises for Chapter 1.
"""
import random
import timeit
from algs.table import DataTable, ExerciseNum, caption
from algs.counting import RecordedItem
def partition(A, lo, hi, idx):
"""
Partition using A[idx] as value. Note lo and hi are INCLUSIVE on both
ends and idx must be valid index. C... | [
"random.shuffle",
"algs.table.caption",
"ch01.largest_two.Match.advance",
"algs.counting.RecordedItem.report",
"algs.counting.RecordedItem.clear",
"algs.table.ExerciseNum",
"ch01.largest_two.Match",
"algs.counting.RecordedItem",
"algs.table.DataTable",
"random.randint"
] | [((2712, 2803), 'algs.table.DataTable', 'DataTable', (['[8, 15, 15]', "['N', 'counting_sort', 'counting_sort_improved']"], {'output': 'output'}), "([8, 15, 15], ['N', 'counting_sort', 'counting_sort_improved'],\n output=output)\n", (2721, 2803), False, 'from algs.table import DataTable, ExerciseNum, caption\n'), ((3... |
import torch
import pytest
# NOTE: also registers the KL divergence
from chmp.torch_utils import NormalModule, WeightsHS, fixed
def test_kl_divergence__gamma__log_normal():
p = torch.distributions.LogNormal(torch.zeros(2), torch.ones(2))
q = torch.distributions.Gamma(torch.ones(2), torch.ones(2))
torch.... | [
"pytest.approx",
"torch.mean",
"torch.distributions.kl_divergence",
"chmp.torch_utils.WeightsHS",
"torch.zeros",
"torch.ones"
] | [((314, 353), 'torch.distributions.kl_divergence', 'torch.distributions.kl_divergence', (['p', 'q'], {}), '(p, q)\n', (347, 353), False, 'import torch\n'), ((1303, 1339), 'chmp.torch_utils.WeightsHS', 'WeightsHS', (['[10, 20, 30]'], {'tau_0': '(1e-05)'}), '([10, 20, 30], tau_0=1e-05)\n', (1312, 1339), False, 'from chmp... |
import torch
from torch_geometric.data import Data
from graphnet.components.pool import group_pulses_to_dom, group_pulses_to_pmt, sum_pool_and_distribute
from graphnet.data.constants import FEATURES
from graphnet.models.detector.detector import Detector
class IceCube86(Detector):
"""`Detector` class for IceCube-... | [
"graphnet.components.pool.group_pulses_to_dom",
"torch.stack",
"torch.log10",
"graphnet.components.pool.group_pulses_to_pmt",
"graphnet.components.pool.sum_pool_and_distribute",
"torch.norm"
] | [((3121, 3146), 'graphnet.components.pool.group_pulses_to_dom', 'group_pulses_to_dom', (['data'], {}), '(data)\n', (3140, 3146), False, 'from graphnet.components.pool import group_pulses_to_dom, group_pulses_to_pmt, sum_pool_and_distribute\n'), ((3162, 3187), 'graphnet.components.pool.group_pulses_to_pmt', 'group_pulse... |
import dearpygui.dearpygui as dpg
import datetime as dt
import math
from registry import *
SUN_DATA.update_date()
# FUNCTIONS
def get_semi_circle_points( center, radius, angle_i, angle_f, segments = 360, closed = False ):
points_close = [[ center[0], center[1]-radius ] , center, [ c... | [
"dearpygui.dearpygui.set_axis_limits",
"dearpygui.dearpygui.set_axis_limits_auto",
"math.cos",
"dearpygui.dearpygui.window",
"dearpygui.dearpygui.add_drag_floatx",
"dearpygui.dearpygui.menu_bar",
"dearpygui.dearpygui.hide_item",
"dearpygui.dearpygui.add_drag_float",
"dearpygui.dearpygui.add_drawlist... | [((5563, 5661), 'dearpygui.dearpygui.configure_item', 'dpg.configure_item', (['(id_link + 1)'], {'p1': '[center[0] - r, center[1]]', 'p2': '[center[0] + r, center[1]]'}), '(id_link + 1, p1=[center[0] - r, center[1]], p2=[center[0\n ] + r, center[1]])\n', (5581, 5661), True, 'import dearpygui.dearpygui as dpg\n'), ((... |
import os, time, mimetypes, glob
from django.utils.translation import gettext_lazy as _
from django.urls import reverse
from task.const import *
from task.models import Task, detect_group
from rusel.base.config import Config
from rusel.base.forms import CreateGroupForm
from rusel.context import get_base_context
from ru... | [
"os.path.getsize",
"task.models.detect_group",
"rusel.base.forms.CreateGroupForm",
"django.utils.translation.gettext_lazy",
"task.models.Task.get_role_tasks",
"rusel.base.config.Config",
"rusel.utils.extract_get_params",
"os.scandir",
"django.urls.reverse",
"mimetypes.guess_type",
"os.path.getmt... | [((438, 462), 'rusel.base.config.Config', 'Config', (['config', 'cur_view'], {}), '(config, cur_view)\n', (444, 462), False, 'from rusel.base.config import Config\n'), ((1433, 1450), 'rusel.base.forms.CreateGroupForm', 'CreateGroupForm', ([], {}), '()\n', (1448, 1450), False, 'from rusel.base.forms import CreateGroupFo... |
# -*- coding: utf-8 -*-
# Copyright (c) 2017, Frappe Technologies Pvt. Ltd. and Contributors
# See license.txt
from __future__ import unicode_literals
import frappe
import unittest
class TestStockSettings(unittest.TestCase):
def setUp(self):
frappe.db.set_value("Stock Settings", None, "clean_description_html", 0)
... | [
"frappe.get_single",
"frappe.db.set_value"
] | [((247, 319), 'frappe.db.set_value', 'frappe.db.set_value', (['"""Stock Settings"""', 'None', '"""clean_description_html"""', '(0)'], {}), "('Stock Settings', None, 'clean_description_html', 0)\n", (266, 319), False, 'import frappe\n'), ((837, 872), 'frappe.get_single', 'frappe.get_single', (['"""Stock Settings"""'], {... |
#!/usr/bin/env python
import rospy
from std_msgs.msg import Float64
import random
possibleLayers = [140, 50, 80, 200, 100]
cur_position = 0.0
def position_callback(msg):
global cur_position
cur_position = msg.data
#Build the layers simulation, then publish material strengths. Lasts 100 seconds.
def runLayers... | [
"rospy.Subscriber",
"rospy.is_shutdown",
"rospy.init_node",
"rospy.Rate",
"rospy.Publisher",
"random.randint"
] | [((343, 365), 'random.randint', 'random.randint', (['(10)', '(20)'], {}), '(10, 20)\n', (357, 365), False, 'import random\n'), ((575, 635), 'rospy.Publisher', 'rospy.Publisher', (['"""material_strength"""', 'Float64'], {'queue_size': '(10)'}), "('material_strength', Float64, queue_size=10)\n", (590, 635), False, 'impor... |
import scrapy
import re
from scrapy.linkextractors import LinkExtractor
from scrapy.spiders import CrawlSpider, Rule
from ..items import ImagescraperItem
class ImageCrawlSpiderSpider(CrawlSpider):
name = "image_crawl_spider"
allowed_domains = ["books.toscrape.com"]
def start_requests(self):
url =... | [
"scrapy.Request",
"scrapy.linkextractors.LinkExtractor",
"re.match"
] | [((364, 387), 'scrapy.Request', 'scrapy.Request', ([], {'url': 'url'}), '(url=url)\n', (378, 387), False, 'import scrapy\n'), ((407, 440), 'scrapy.linkextractors.LinkExtractor', 'LinkExtractor', ([], {'allow': '"""catalogue/"""'}), "(allow='catalogue/')\n", (420, 440), False, 'from scrapy.linkextractors import LinkExtr... |
import os
import argparse
import subprocess
import random
import edlib
from typing import List
from collections import Counter
import stanza
class ExtractMetric(object):
"""used for precision recall"""
def __init__(self, nume=0, denom_p=0, denom_r=0, precision=0, recall=0, f1=0):
super(ExtractMetric, ... | [
"edlib.align",
"argparse.ArgumentParser",
"os.path.join",
"collections.Counter",
"stanza.Pipeline"
] | [((6056, 6120), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {'description': '"""Evaluate analysis metrics"""'}), "(description='Evaluate analysis metrics')\n", (6079, 6120), False, 'import argparse\n'), ((6576, 6619), 'os.path.join', 'os.path.join', (['args.exp_dir', '"""prototype.txt"""'], {}), "(args.e... |
""" The ARIMA model. """
import torch
import numpy as np
class ARIMA(torch.nn.Module):
"""ARIMA [summary]
"""
def __init__(self,
p: int = 0,
d: int = 0,
q: int = 0) -> None:
"""__init__ General ARIMA model constructor.
Args:
... | [
"numpy.random.normal",
"torch.pow",
"torch.diff",
"torch.no_grad",
"torch.zeros",
"torch.rand"
] | [((789, 802), 'torch.rand', 'torch.rand', (['p'], {}), '(p)\n', (799, 802), False, 'import torch\n'), ((889, 902), 'torch.rand', 'torch.rand', (['q'], {}), '(q)\n', (899, 902), False, 'import torch\n'), ((989, 1002), 'torch.rand', 'torch.rand', (['d'], {}), '(d)\n', (999, 1002), False, 'import torch\n'), ((1067, 1080),... |
#import
import os
#import torch
#import torch.nn as nn
import torch.utils.data as Data
#import torchvision
import matplotlib.pyplot as plt
import h5py
#from torch.autograd import Variable
import numpy as np
import torch
class rawdataDataset(Data.Dataset):
def __init__(self):
super(rawdataDataset, self)... | [
"os.path.join",
"h5py.File"
] | [((796, 833), 'os.path.join', 'os.path.join', (['opt.dataroot', 'opt.phase'], {}), '(opt.dataroot, opt.phase)\n', (808, 833), False, 'import os\n'), ((1025, 1053), 'h5py.File', 'h5py.File', (['self.A_paths', '"""r"""'], {}), "(self.A_paths, 'r')\n", (1034, 1053), False, 'import h5py\n'), ((1076, 1104), 'h5py.File', 'h5... |
"""
Copyright 2012-2019 <NAME>
Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to in writing, software
di... | [
"dateutil.tz.tzutc",
"unittest.mock.Mock",
"hqlib.metric_source.JunitTestReport",
"dateutil.tz.tzlocal"
] | [((921, 938), 'hqlib.metric_source.JunitTestReport', 'JunitTestReport', ([], {}), '()\n', (936, 938), False, 'from hqlib.metric_source import JunitTestReport\n'), ((1058, 1271), 'unittest.mock.Mock', 'Mock', ([], {'return_value': '"""<testsuites> <testsuite tests="12" failures="2" errors="0" skipped="1" disabled="0"> ... |
sort_functions = [
builtinsort, # see implementation above
insertion_sort, # see [[Insertion sort]]
insertion_sort_lowb, # ''insertion_sort'', where sequential search is replaced
# by lower_bound() function
qsort, # see [[Quicksort]]
qsortranlc... | [
"sys.setrecursionlimit"
] | [((719, 747), 'sys.setrecursionlimit', 'sys.setrecursionlimit', (['(10000)'], {}), '(10000)\n', (740, 747), False, 'import sys\n')] |
from flaskapp import app, db, mail
from flask import render_template, url_for
from flask import request, flash, redirect
# from flaskapp.model import User
from flaskapp.form import SurveyForm
from flask_mail import Message
@app.route('/', methods = ['POST', 'GET'])
def form():
form = SurveyForm()
if ... | [
"flask.render_template",
"flaskapp.app.route",
"flask.flash",
"flask.url_for",
"flaskapp.form.SurveyForm",
"flask_mail.Message",
"flaskapp.mail.send"
] | [((232, 271), 'flaskapp.app.route', 'app.route', (['"""/"""'], {'methods': "['POST', 'GET']"}), "('/', methods=['POST', 'GET'])\n", (241, 271), False, 'from flaskapp import app, db, mail\n'), ((884, 903), 'flaskapp.app.route', 'app.route', (['"""/thank"""'], {}), "('/thank')\n", (893, 903), False, 'from flaskapp import... |
import unittest
from kameramera import camera
class Camera(unittest.TestCase):
def setUp(self):
self.camera = camera.Camera(camera_id='canon_ae1')
def test_general_manufacturer(self):
self.assertEqual(self.camera.general.manufacturer, 'Canon')
def test_general_name(self):
se... | [
"kameramera.camera.Camera"
] | [((125, 161), 'kameramera.camera.Camera', 'camera.Camera', ([], {'camera_id': '"""canon_ae1"""'}), "(camera_id='canon_ae1')\n", (138, 161), False, 'from kameramera import camera\n')] |
import matplotlib.pyplot as plt
import numpy as np
import math
import cv2
kernel = np.ones((3, 3), np.int8)
# 去除雜訊
def eraseImage (image):
return cv2.erode(image, kernel, iterations = 1)
# 模糊圖片
def blurImage (image):
return cv2.GaussianBlur(image, (5, 5), 0)
# 銳利化圖片
# threshold1,2,較小的值為作為偵測邊界的最小值
def edgedImage... | [
"cv2.rectangle",
"numpy.ones",
"cv2.resize",
"cv2.erode",
"matplotlib.pyplot.plot",
"cv2.boundingRect",
"cv2.imshow",
"cv2.waitKey",
"cv2.destroyAllWindows",
"cv2.findContours",
"cv2.dilate",
"cv2.Canny",
"cv2.imread",
"cv2.GaussianBlur",
"matplotlib.pyplot.show"
] | [((84, 108), 'numpy.ones', 'np.ones', (['(3, 3)', 'np.int8'], {}), '((3, 3), np.int8)\n', (91, 108), True, 'import numpy as np\n'), ((150, 188), 'cv2.erode', 'cv2.erode', (['image', 'kernel'], {'iterations': '(1)'}), '(image, kernel, iterations=1)\n', (159, 188), False, 'import cv2\n'), ((231, 265), 'cv2.GaussianBlur',... |
import pprint
from nn_wtf.parameter_optimizers.neural_network_optimizer import NeuralNetworkOptimizer
__author__ = '<NAME> <<EMAIL>>'
class BruteForceOptimizer(NeuralNetworkOptimizer):
DEFAULT_LAYER_SIZES = (
(32, 48, 64), # (32, 48, 64, 80, 96, 128),
(32, 48, 64, 80, 96, 128),
(None, ... | [
"pprint.pprint",
"nn_wtf.parameter_optimizers.neural_network_optimizer.NeuralNetworkOptimizer.OptimizationParameters"
] | [((1758, 1791), 'pprint.pprint', 'pprint.pprint', (['results'], {'width': '(100)'}), '(results, width=100)\n', (1771, 1791), False, 'import pprint\n'), ((1509, 1584), 'nn_wtf.parameter_optimizers.neural_network_optimizer.NeuralNetworkOptimizer.OptimizationParameters', 'NeuralNetworkOptimizer.OptimizationParameters', ([... |
import numpy as np
import tensorflow as tf
import sys, os
sys.path.extend(['alg/', 'models/'])
from visualisation import plot_images
from encoder_no_shared import encoder, recon
from utils import init_variables, save_params, load_params, load_data
from eval_test_ll import construct_eval_func
dimZ = 50
dimH = 500
n_cha... | [
"eval_test_ll.construct_eval_func",
"generator.construct_gen",
"notmnist.load_notmnist",
"vae_laplace.init_fisher_accum",
"visualisation.plot_images",
"vae_si.update_si_reg",
"tensorflow.placeholder",
"tensorflow.Session",
"utils.load_params",
"os.path.isdir",
"sys.path.extend",
"os.mkdir",
... | [((58, 94), 'sys.path.extend', 'sys.path.extend', (["['alg/', 'models/']"], {}), "(['alg/', 'models/'])\n", (73, 94), False, 'import sys, os\n'), ((580, 608), 'config.config', 'config', (['data_name', 'n_channel'], {}), '(data_name, n_channel)\n', (586, 608), False, 'from config import config\n'), ((1757, 1810), 'tenso... |
import logging
import os
import cltl.combot.infra.config.local as local_config
logger = logging.getLogger(__name__)
K8_CONFIG_DIR = "/cltl_k8_config"
K8_CONFIG = "config/k8.config"
class K8LocalConfigurationContainer(local_config.LocalConfigurationContainer):
@staticmethod
def load_configuration(config_fi... | [
"logging.getLogger",
"os.listdir",
"os.path.join",
"cltl.combot.infra.config.local.LocalConfigurationContainer.load_configuration"
] | [((90, 117), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (107, 117), False, 'import logging\n'), ((773, 858), 'cltl.combot.infra.config.local.LocalConfigurationContainer.load_configuration', 'local_config.LocalConfigurationContainer.load_configuration', (['config_file', 'configs'], {})... |
import json
import apiai
import speech_recognition as sr
def speechRecognition():
recog = sr.Recognizer()
with sr.Microphone() as source:
print("It's your cue")
audio = recog.listen(source)
i = True
while i is True:
try:
text = recog.recognize_google(audio)
... | [
"speech_recognition.Recognizer",
"apiai.ApiAI",
"speech_recognition.Microphone"
] | [((96, 111), 'speech_recognition.Recognizer', 'sr.Recognizer', ([], {}), '()\n', (109, 111), True, 'import speech_recognition as sr\n'), ((650, 682), 'apiai.ApiAI', 'apiai.ApiAI', (['CLIENT_ACCESS_TOKEN'], {}), '(CLIENT_ACCESS_TOKEN)\n', (661, 682), False, 'import apiai\n'), ((121, 136), 'speech_recognition.Microphone'... |
# raise NotImplementedError("Did not check!")
"""MSCOCO Semantic Segmentation pretraining for VOC."""
import os
from tqdm import trange
from PIL import Image, ImageOps, ImageFilter
import numpy as np
import pickle
from gluoncv.data.segbase import SegmentationDataset
class COCOSegmentation(SegmentationDataset):
... | [
"os.path.exists",
"pickle.dump",
"os.path.join",
"pycocotools.coco.COCO",
"pickle.load",
"numpy.sum",
"numpy.zeros",
"os.path.expanduser"
] | [((471, 515), 'os.path.expanduser', 'os.path.expanduser', (['"""~/.mxnet/datasets/coco"""'], {}), "('~/.mxnet/datasets/coco')\n", (489, 515), False, 'import os\n'), ((1266, 1280), 'pycocotools.coco.COCO', 'COCO', (['ann_file'], {}), '(ann_file)\n', (1270, 1280), False, 'from pycocotools.coco import COCO\n'), ((1322, 13... |
# __author__ = 'sree'
import urllib2
from lxml import html
import requests
def get_page_tree(url=None):
page = requests.get(url=url, verify=False)
return html.fromstring(page.text)
def get_title(url=None):
tree = get_page_tree(url=url)
return tree.xpath('//title//text()')[0].strip().split(' -')[0]
d... | [
"lxml.html.fromstring",
"urllib2.quote",
"requests.get"
] | [((117, 152), 'requests.get', 'requests.get', ([], {'url': 'url', 'verify': '(False)'}), '(url=url, verify=False)\n', (129, 152), False, 'import requests\n'), ((164, 190), 'lxml.html.fromstring', 'html.fromstring', (['page.text'], {}), '(page.text)\n', (179, 190), False, 'from lxml import html\n'), ((668, 688), 'urllib... |
# encoding: utf-8
# Copyright 2009-2020 <NAME>.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# https://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or a... | [
"os.path.exists",
"xml.dom.minidom.parse",
"sqlite3.connect",
"os.path.join",
"os.path.isfile",
"os.path.dirname",
"xml.dom.minidom.parseString",
"os.path.isdir",
"os.path.basename",
"os.walk"
] | [((1103, 1119), 'os.walk', 'os.walk', (['dirpath'], {}), '(dirpath)\n', (1110, 1119), False, 'import os\n'), ((1515, 1531), 'os.walk', 'os.walk', (['dirpath'], {}), '(dirpath)\n', (1522, 1531), False, 'import os\n'), ((1953, 1969), 'os.walk', 'os.walk', (['dirpath'], {}), '(dirpath)\n', (1960, 1969), False, 'import os\... |
import datetime, hashlib, base64, traceback, os, re
import poshc2.server.database.DB as DB
from poshc2.Colours import Colours
from poshc2.server.Config import ModulesDirectory, DownloadsDirectory, ReportsDirectory
from poshc2.server.Implant import Implant
from poshc2.server.Core import decrypt, encrypt, default_respon... | [
"poshc2.server.Core.process_mimikatz",
"poshc2.server.database.DB.get_implanttype",
"base64.b64encode",
"poshc2.server.database.DB.get_implants_all",
"poshc2.server.database.DB.get_implantbyrandomuri",
"poshc2.server.database.DB.new_task",
"re.search",
"os.listdir",
"poshc2.Utils.randomuri",
"posh... | [((704, 727), 'datetime.datetime.now', 'datetime.datetime.now', ([], {}), '()\n', (725, 727), False, 'import datetime, hashlib, base64, traceback, os, re\n'), ((747, 768), 'poshc2.server.database.DB.get_implants_all', 'DB.get_implants_all', ([], {}), '()\n', (766, 768), True, 'import poshc2.server.database.DB as DB\n')... |
from django.urls import path, re_path
from . import views
urlpatterns = [
path('', views.index, name='index'),
path('countdown/', views.countdown, name='countdown'),
#re_path(r'.+', views.redir, name='redir'),
]
| [
"django.urls.path"
] | [((80, 115), 'django.urls.path', 'path', (['""""""', 'views.index'], {'name': '"""index"""'}), "('', views.index, name='index')\n", (84, 115), False, 'from django.urls import path, re_path\n'), ((121, 174), 'django.urls.path', 'path', (['"""countdown/"""', 'views.countdown'], {'name': '"""countdown"""'}), "('countdown/... |
# Copyright (C) 2013 Nippon Telegraph and Telephone Corporation.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by appli... | [
"flask.render_template",
"flask.request.host.split"
] | [((819, 842), 'flask.request.host.split', 'request.host.split', (['""":"""'], {}), "(':')\n", (837, 842), False, 'from flask import render_template, request\n'), ((858, 912), 'flask.render_template', 'render_template', (['"""topology.html"""'], {'host': 'host', 'port': 'port'}), "('topology.html', host=host, port=port)... |
import tensorflow as tf
import numpy as np
def clipped_error(x):
# Huber loss
try:
return tf.select(tf.abs(x) < 1.0, 0.5 * tf.square(x), tf.abs(x) - 0.5 )
except:
return tf.where(tf.abs(x) < 1.0, 0.5 * tf.square(x), tf.abs(x) - 0.5 )
def linear(input_, output_size, stddev=0.02, bias_star... | [
"tensorflow.variable_scope",
"tensorflow.random_normal_initializer",
"tensorflow.matmul",
"tensorflow.constant_initializer",
"tensorflow.square",
"tensorflow.abs"
] | [((425, 448), 'tensorflow.variable_scope', 'tf.variable_scope', (['name'], {}), '(name)\n', (442, 448), True, 'import tensorflow as tf\n'), ((553, 596), 'tensorflow.random_normal_initializer', 'tf.random_normal_initializer', ([], {'stddev': 'stddev'}), '(stddev=stddev)\n', (581, 596), True, 'import tensorflow as tf\n')... |
"""Cell parameter random initializations."""
from typing import Any, Dict
import numpy as np
from ..parameters import (
Height,
NewCellBendLowerLower,
NewCellBendLowerUpper,
NewCellBendOverallLower,
NewCellBendOverallUpper,
NewCellBendUpperLower,
NewCellBendUpperUpper,
NewCellLength1Me... | [
"numpy.sin",
"numpy.cos"
] | [((3328, 3341), 'numpy.cos', 'np.cos', (['angle'], {}), '(angle)\n', (3334, 3341), True, 'import numpy as np\n'), ((3397, 3410), 'numpy.sin', 'np.sin', (['angle'], {}), '(angle)\n', (3403, 3410), True, 'import numpy as np\n')] |
import qiskit
import numpy as np
import matplotlib.pyplot as plt
import json
from graph import *
# Random comment
P =1
def makeCircuit(inbits, outbits):
q = qiskit.QuantumRegister(inbits+outbits)
c = qiskit.ClassicalRegister(inbits+outbits)
qc = qiskit.QuantumCircuit(q, c)
q_input = [q[i] for i in ran... | [
"qiskit.ClassicalRegister",
"matplotlib.pyplot.title",
"qiskit.execute",
"matplotlib.pyplot.ylabel",
"matplotlib.pyplot.xlabel",
"json.dump",
"numpy.binary_repr",
"numpy.linspace",
"matplotlib.pyplot.errorbar",
"json.load",
"qiskit.QuantumCircuit",
"qiskit.QuantumRegister",
"matplotlib.pyplo... | [((162, 202), 'qiskit.QuantumRegister', 'qiskit.QuantumRegister', (['(inbits + outbits)'], {}), '(inbits + outbits)\n', (184, 202), False, 'import qiskit\n'), ((209, 251), 'qiskit.ClassicalRegister', 'qiskit.ClassicalRegister', (['(inbits + outbits)'], {}), '(inbits + outbits)\n', (233, 251), False, 'import qiskit\n'),... |
"""Simple code for training an RNN for motion prediction."""
import os
import sys
import time
import numpy as np
import torch
import torch.optim as optim
from torch.autograd import Variable
import mtfixb_model
import mtfixb_model2
import parseopts
def create_model(args, total_num_batches):
"""Create MT model a... | [
"numpy.sqrt",
"torch.from_numpy",
"parseopts.parse_args",
"torch.exp",
"parseopts.initial_arg_transform",
"mtfixb_model2.MTGRU_NoBias",
"mtfixb_model.MTGRU",
"mtfixb_model.DataIterator",
"mtfixb_model.OpenLoopGRU",
"sys.stdout.flush",
"torch.optim.SGD",
"mtfixb_model.DynamicsDict",
"numpy.in... | [((890, 1307), 'mtfixb_model.MTGRU', 'mtfixb_model.MTGRU', (['args.seq_length_out', 'args.decoder_size', 'args.decoder_size2', 'args.batch_size', 'total_num_batches', 'args.k', 'args.size_psi_hidden', 'args.size_psi_lowrank', 'args.bottleneck'], {'output_dim': 'args.human_size', 'input_dim': 'args.input_size', 'dropout... |
# Licensed to the Apache Software Foundation (ASF) under one
# or more contributor license agreements. See the NOTICE file
# distributed with this work for additional information
# regarding copyright ownership. The ASF licenses this file
# to you under the Apache License, Version 2.0 (the
# "License"); you may not u... | [
"tvm.relay.nn.conv2d_transpose",
"tvm.relay.transpose",
"tvm.te.compute",
"tvm.te.reduce_axis"
] | [((3445, 3589), 'tvm.te.compute', 'te.compute', (['(out_c, in_c, filter_h, filter_w)', '(lambda o, i, h, w: kernel[i][o][filter_h - 1 - h][filter_w - 1 - w])'], {'name': '"""kernel_transform"""'}), "((out_c, in_c, filter_h, filter_w), lambda o, i, h, w: kernel[i][\n o][filter_h - 1 - h][filter_w - 1 - w], name='kern... |
import argparse
import copy
import torch
from torchvision.datasets import MNIST, CIFAR10
import torchvision.transforms as TF
import torchelie as tch
import torchelie.loss.gan.hinge as gan_loss
from torchelie.recipes.gan import GANRecipe
import torchelie.callbacks as tcb
from torchelie.recipes import Recipe
parser =... | [
"copy.deepcopy",
"argparse.ArgumentParser",
"torchelie.transforms.AdaptPad",
"torchelie.callbacks.Log",
"torchelie.loss.gan.hinge.generated",
"torchvision.transforms.RandomHorizontalFlip",
"torchelie.callbacks.WindowedMetricAvg",
"torchelie.callbacks.Polyak",
"torchvision.datasets.CIFAR10",
"torch... | [((321, 346), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {}), '()\n', (344, 346), False, 'import argparse\n'), ((607, 671), 'torchvision.datasets.CIFAR10', 'CIFAR10', (['"""~/.cache/torch/cifar10"""'], {'download': '(True)', 'transform': 'tfms'}), "('~/.cache/torch/cifar10', download=True, transform=tfm... |
import xml.sax
import rdflib
from django.db import transaction
from hs_core.serialization import GenericResourceMeta
class RasterResourceMeta(GenericResourceMeta):
"""
Lightweight class for representing metadata of RasterResource instances.
"""
def __init__(self):
super(RasterResourceMeta, s... | [
"rdflib.namespace.Namespace",
"hs_core.serialization.GenericResourceMeta.ResourceMetaException"
] | [((876, 934), 'rdflib.namespace.Namespace', 'rdflib.namespace.Namespace', (['"""http://hydroshare.org/terms/"""'], {}), "('http://hydroshare.org/terms/')\n", (902, 934), False, 'import rdflib\n'), ((1387, 1433), 'hs_core.serialization.GenericResourceMeta.ResourceMetaException', 'GenericResourceMeta.ResourceMetaExceptio... |
import os
import os.path
import sqlite3
import logging
from typing import List
from gumtree_watchdog.types import Listing, Contract, ListingWithChatId
TConn = sqlite3.Connection
DB_PATH = os.environ.get('GUMTREE_DB')
def get_connection() -> TConn:
if not DB_PATH:
raise Exception("Please specify Sqlite3 db... | [
"sqlite3.connect",
"gumtree_watchdog.types.ListingWithChatId",
"logging.warning",
"os.environ.get",
"os.path.isfile",
"gumtree_watchdog.types.Contract",
"logging.info"
] | [((189, 217), 'os.environ.get', 'os.environ.get', (['"""GUMTREE_DB"""'], {}), "('GUMTREE_DB')\n", (203, 217), False, 'import os\n'), ((375, 399), 'sqlite3.connect', 'sqlite3.connect', (['DB_PATH'], {}), '(DB_PATH)\n', (390, 399), False, 'import sqlite3\n'), ((478, 501), 'os.path.isfile', 'os.path.isfile', (['DB_PATH'],... |
#section [initial]
def _parameters_accessors_checks():
from finess.params import Parameter, Accessor, Check, \
CheckGreaterEqual, CheckGreaterThan, \
CheckOneOf, EnumParameterType
parameters = []
checks = []
rhol = Parameter(variable_name = "rhol",
... | [
"finess.params.CheckGreaterThan",
"finess.params.Parameter"
] | [((280, 359), 'finess.params.Parameter', 'Parameter', ([], {'variable_name': '"""rhol"""', 'section': '"""initial"""', 'name': '"""rhol"""', 'type_': '"""double"""'}), "(variable_name='rhol', section='initial', name='rhol', type_='double')\n", (289, 359), False, 'from finess.params import Parameter, Accessor, Check, Ch... |
from setuptools import setup
import os
with open("README.md", "r", encoding="utf-8") as f:
long_description = f.read()
requirements = []
if os.path.isfile("./requirements.txt"):
with open("requirements.txt", "r") as f:
requirements = f.read()
requirements = [x for x in requirements.split("\n") if ... | [
"os.path.isfile",
"setuptools.setup"
] | [((146, 182), 'os.path.isfile', 'os.path.isfile', (['"""./requirements.txt"""'], {}), "('./requirements.txt')\n", (160, 182), False, 'import os\n'), ((330, 1000), 'setuptools.setup', 'setup', ([], {'name': '"""ma5_expert"""', 'version': '"""0.0.1"""', 'description': '"""MadAnalysis 5 interpreter for Expert mode"""', 'l... |
import time
import numpy as np
import vtk
from vtk.util import numpy_support
from svtk.lib.toolbox.integer import minmax
from svtk.lib.toolbox.idarray import IdArray
from svtk.lib.toolbox.numpy_helpers import normalize
import math as m
class VTKAnimationTimerCallback(object):
"""This class is called every few m... | [
"numpy.tile",
"numpy.allclose",
"math.ceil",
"numpy.cross",
"time.clock",
"math.floor",
"svtk.lib.toolbox.idarray.IdArray",
"vtk.util.numpy_support.numpy_to_vtkIdTypeArray",
"svtk.lib.toolbox.integer.minmax",
"numpy.append",
"numpy.array",
"numpy.matmul",
"vtk.util.numpy_support.numpy_to_vtk... | [((1209, 1221), 'time.clock', 'time.clock', ([], {}), '()\n', (1219, 1221), False, 'import time\n'), ((1264, 1276), 'time.clock', 'time.clock', ([], {}), '()\n', (1274, 1276), False, 'import time\n'), ((1322, 1334), 'time.clock', 'time.clock', ([], {}), '()\n', (1332, 1334), False, 'import time\n'), ((1361, 1373), 'tim... |
import time
import numpy as np
from tqdm import tqdm
from utils import RandomCNOT, RandomCNOTs
def SimulatedAnnealing(quantum_count, layer_count, solver, epochs=100, save_path=None, global_best_score=0):
#TODO:
best_score = 0
cnot = RandomCNOTs(quantum_count, layer_count)
sc, model = solver(cnot)
... | [
"utils.RandomCNOTs",
"numpy.random.randint",
"time.time",
"utils.RandomCNOT"
] | [((247, 286), 'utils.RandomCNOTs', 'RandomCNOTs', (['quantum_count', 'layer_count'], {}), '(quantum_count, layer_count)\n', (258, 286), False, 'from utils import RandomCNOT, RandomCNOTs\n'), ((591, 602), 'time.time', 'time.time', ([], {}), '()\n', (600, 602), False, 'import time\n'), ((2114, 2125), 'time.time', 'time.t... |
#!/usr/bin/env python
#
# Copyright 2013 <NAME>
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed ... | [
"testtools.ExpectedException",
"os.path.join",
"os.path.dirname",
"os.path.basename",
"jenkins_jobs.config.JJBConfig",
"jenkins_jobs.parser.YamlParser",
"tests.base.get_scenarios"
] | [((1172, 1258), 'tests.base.get_scenarios', 'base.get_scenarios', (['fixtures_path', '"""yaml"""', '"""json"""'], {'filter_func': '_exclude_scenarios'}), "(fixtures_path, 'yaml', 'json', filter_func=\n _exclude_scenarios)\n", (1190, 1258), False, 'from tests import base\n'), ((1879, 1930), 'tests.base.get_scenarios'... |
from bot_interface import *
import math
class SeijiBot(BotBase):
def __init__(self):
self.initialized = False
def initialize(self, gamestate):
gamestate.log("Initializing...")
#Getting UID
self.uid = gamestate.bot.uid
gamestate.log("This ship has uid " + str(self.ui... | [
"math.floor",
"math.sqrt",
"math.cos",
"math.atan2",
"math.sin"
] | [((1088, 1104), 'math.sqrt', 'math.sqrt', (['delta'], {}), '(delta)\n', (1097, 1104), False, 'import math\n'), ((1214, 1284), 'math.sqrt', 'math.sqrt', (['((obj1.posx - obj2.posx) ** 2 + (obj1.posy - obj2.posy) ** 2)'], {}), '((obj1.posx - obj2.posx) ** 2 + (obj1.posy - obj2.posy) ** 2)\n', (1223, 1284), False, 'import... |
from datetime import date, datetime, timedelta
import time
START_DATE = date(2021, 5, 25)
duration = timedelta(days=100)
def countdown():
event_delta = LAST_DAY_OF_SCHOOL - datetime.now()
print()
print("\tTime until school is out for summer 2021:", end="\n\n")
while event_delta.seconds > 0:
h... | [
"datetime.datetime.now",
"datetime.timedelta",
"datetime.date",
"time.sleep"
] | [((73, 90), 'datetime.date', 'date', (['(2021)', '(5)', '(25)'], {}), '(2021, 5, 25)\n', (77, 90), False, 'from datetime import date, datetime, timedelta\n'), ((102, 121), 'datetime.timedelta', 'timedelta', ([], {'days': '(100)'}), '(days=100)\n', (111, 121), False, 'from datetime import date, datetime, timedelta\n'), ... |
#!/usr/bin/env python
# -*- coding: utf-8 -*-
"""
__title__ = ''
__author__ = 'HaiFeng'
__mtime__ = '2016/8/16'
"""
import time
from py_at.EnumDefine import *
########################################################################
class OrderItem(object):
"""策略信号"""
#----------------------------------------------... | [
"time.time"
] | [((477, 488), 'time.time', 'time.time', ([], {}), '()\n', (486, 488), False, 'import time\n')] |
from __future__ import print_function, division
import numpy as np
from numpy import identity, dot, zeros, zeros_like
def rf_den_via_rf0(self, rf0, v):
""" Whole matrix of the interacting response via non-interacting response and interaction"""
rf = zeros_like(rf0)
I = identity(rf0.shape[1])
for ir,r in enume... | [
"numpy.identity",
"numpy.dot",
"numpy.zeros_like"
] | [((255, 270), 'numpy.zeros_like', 'zeros_like', (['rf0'], {}), '(rf0)\n', (265, 270), False, 'from numpy import identity, dot, zeros, zeros_like\n'), ((278, 300), 'numpy.identity', 'identity', (['rf0.shape[1]'], {}), '(rf0.shape[1])\n', (286, 300), False, 'from numpy import identity, dot, zeros, zeros_like\n'), ((364, ... |
import os
import logging
import numpy as np
from typing import Optional
import torch
from torch.utils.data import DataLoader
from ..eval import Metric
from .dataset import CHMMBaseDataset
from .dataset import collate_fn as default_collate_fn
logger = logging.getLogger(__name__)
OUT_RECALL = 0.9
OUT_PRECISION = 0.8
... | [
"logging.getLogger",
"torch.load",
"os.path.join",
"os.path.split",
"os.path.isfile",
"numpy.random.dirichlet",
"torch.save",
"torch.utils.data.DataLoader",
"torch.zeros"
] | [((254, 281), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (271, 281), False, 'import logging\n'), ((14610, 14650), 'numpy.random.dirichlet', 'np.random.dirichlet', (['(init_counts + 1e-10)'], {}), '(init_counts + 1e-10)\n', (14629, 14650), True, 'import numpy as np\n'), ((2409, 2454), ... |
from random import Random
from collections_extended import setlist
# The version of seeding to use for random
SEED_VERSION = 2
# Common alphabets to use
ALPHANUM = '0123456789abcdefghijklmnopqrstuvwxyzABCDEFGHIJKLMNOPQRSTUVWXYZ'
BASE58 = '123456789ABCDEFGHJKLMNPQRSTUVWXYZabcdefghijkmnopqrstuvwxyz'
def shuffle(key... | [
"random.Random",
"collections_extended.setlist"
] | [((336, 347), 'random.Random', 'Random', (['key'], {}), '(key)\n', (342, 347), False, 'from random import Random\n'), ((711, 722), 'random.Random', 'Random', (['key'], {}), '(key)\n', (717, 722), False, 'from random import Random\n'), ((6443, 6460), 'collections_extended.setlist', 'setlist', (['alphabet'], {}), '(alpha... |
# -*- coding: utf-8 -*-
"""
pbkdf2 hashing handler module.
"""
import hashlib
import re
import pyrin.configuration.services as config_services
import pyrin.security.utils.services as security_utils_services
from pyrin.security.hashing.decorators import hashing
from pyrin.security.hashing.handlers.base import Hashing... | [
"pyrin.configuration.services.get",
"pyrin.security.hashing.decorators.hashing",
"pyrin.security.utils.services.get_bytes",
"re.compile"
] | [((488, 497), 'pyrin.security.hashing.decorators.hashing', 'hashing', ([], {}), '()\n', (495, 497), False, 'from pyrin.security.hashing.decorators import hashing\n'), ((790, 848), 're.compile', 're.compile', (['"""^\\\\$PBKDF2\\\\$[^$]+\\\\$[\\\\d]+\\\\$[\\\\d]+\\\\$(.+)$"""'], {}), "('^\\\\$PBKDF2\\\\$[^$]+\\\\$[\\\\d... |
# -*- coding: utf-8 -*-
"""
Created on Wed Jun 28 13:03:05 2017
@author: <NAME>
"""
import cntk as C
import _cntk_py
import cntk.layers
import cntk.initializer
import cntk.losses
import cntk.metrics
import cntk.logging
import cntk.io.transforms as xforms
import cntk.io
import cntk.train
import os
import numpy as np
i... | [
"os.path.exists",
"yolo2.Yolo2Metric",
"_cntk_py.set_computation_network_trace_level",
"cntk.Trainer",
"os.path.join",
"cntk.learners.momentum_sgd",
"yolo2.Yolo2Error",
"cntk.io.transforms.color",
"cntk.learners.momentum_schedule",
"cntk.input_variable",
"cntk.io.transforms.scale",
"numpy.arra... | [((469, 501), 'os.path.join', 'os.path.join', (['abs_path', '"""Models"""'], {}), "(abs_path, 'Models')\n", (481, 501), False, 'import os\n'), ((429, 454), 'os.path.abspath', 'os.path.abspath', (['__file__'], {}), '(__file__)\n', (444, 454), False, 'import os\n'), ((2267, 2326), 'cntk.input_variable', 'C.input_variable... |
from django.core.mail import EmailMessage
from django.conf import settings
def send_email(name, date, email):
txt = """
<html>
<body>
<table cellpadding='0' cellspacing='0' width='100%' border='0'>
<tbody>
<tr>
<td style='word-wrap:break-word;font-size:0px;padding:0px;padding-bottom:10px' alig... | [
"django.core.mail.EmailMessage"
] | [((2019, 2071), 'django.core.mail.EmailMessage', 'EmailMessage', (['settings.EMAIL_TITLE', 'txt'], {'to': '(email,)'}), '(settings.EMAIL_TITLE, txt, to=(email,))\n', (2031, 2071), False, 'from django.core.mail import EmailMessage\n')] |
import KNN as K
K.clearScreen()
dataTraining= K.loadData("dataTraining.txt")
X=dataTraining[:,0:3]
initial_centroids=K.listToArray([[3, 3,3],[6, 2,4],[8,5,7]])
idx=K.KMean_Run(X,initial_centroids,5)
K.SaveData(K.concatenateVectors(X,idx))
K.plotKNN2(X,idx)
| [
"KNN.concatenateVectors",
"KNN.plotKNN2",
"KNN.listToArray",
"KNN.loadData",
"KNN.KMean_Run",
"KNN.clearScreen"
] | [((20, 35), 'KNN.clearScreen', 'K.clearScreen', ([], {}), '()\n', (33, 35), True, 'import KNN as K\n'), ((50, 80), 'KNN.loadData', 'K.loadData', (['"""dataTraining.txt"""'], {}), "('dataTraining.txt')\n", (60, 80), True, 'import KNN as K\n'), ((124, 172), 'KNN.listToArray', 'K.listToArray', (['[[3, 3, 3], [6, 2, 4], [8... |
#!/usr/bin/env python
import pyqtgraph as pg
from pyqtgraph import ViewBox
from hummingbird.graphics.plotter_args import PlotBoxArgs
from hummingbird.graphics.state_plot import StatePlot
class StatePlotBox:
def __init__(self, window, args):
""" Create a new plotbox wrapper object
Arguments:
... | [
"hummingbird.graphics.state_plot.StatePlot",
"pyqtgraph.intColor",
"pyqtgraph.mkPen"
] | [((2115, 2149), 'pyqtgraph.mkPen', 'pg.mkPen', ([], {'color': 'color', 'width': 'width'}), '(color=color, width=width)\n', (2123, 2149), True, 'import pyqtgraph as pg\n'), ((2531, 2565), 'hummingbird.graphics.state_plot.StatePlot', 'StatePlot', (['self.plotbox', 'plot_args'], {}), '(self.plotbox, plot_args)\n', (2540, ... |
import json
#start
print('start')
with open('quizoutput.txt') as f:
lines = f.readlines()
print('loaded quiz data')
print('changing to json')
json_output = json.loads(lines[0])
print(json_output)
with open('quizoutput.txt', 'w') as f:
f.write(json_output)
# for item in json_output:
# print(item['ques... | [
"json.loads"
] | [((166, 186), 'json.loads', 'json.loads', (['lines[0]'], {}), '(lines[0])\n', (176, 186), False, 'import json\n')] |
#!/usr/bin/python
import requests
import boto3
import time
import geopy.distance
import xml.etree.ElementTree as ET
import itertools
import sys
import pickle
S3_BUCKET = "panku-gdzie-jestes-latest-storage"
class LatestPositionStorage(object):
def __init__(self, service):
self.objectName = "%s.latest" % service... | [
"boto3.client",
"boto3.session.Session",
"requests.Session",
"pickle.dumps",
"sys.exc_info",
"boto3.resource",
"xml.etree.ElementTree.fromstring",
"time.time"
] | [((373, 393), 'boto3.resource', 'boto3.resource', (['"""s3"""'], {}), "('s3')\n", (387, 393), False, 'import boto3\n'), ((716, 736), 'boto3.resource', 'boto3.resource', (['"""s3"""'], {}), "('s3')\n", (730, 736), False, 'import boto3\n'), ((817, 840), 'pickle.dumps', 'pickle.dumps', (['positions'], {}), '(positions)\n'... |
# -*- coding: utf-8 -*-
# Generated by Django 1.10.1 on 2017-12-08 18:42
from __future__ import unicode_literals
from django.db import migrations, models
import django.db.models.deletion
def update_questionnaires(apps, schema_editor):
"""
Forward migration function to normalize settings into VMSettings and C... | [
"django.db.migrations.RunPython"
] | [((1361, 1404), 'django.db.migrations.RunPython', 'migrations.RunPython', (['update_questionnaires'], {}), '(update_questionnaires)\n', (1381, 1404), False, 'from django.db import migrations, models\n')] |
import os
import unittest
import json
import filecmp
from genofunk.sequence_utils import *
this_dir = os.path.dirname(os.path.dirname(os.path.abspath(__file__)))
class TestSequenceUtils(unittest.TestCase):
def test_get_coordinates_from_json_simple_pairs(self):
json_value = {
"start": 30,
... | [
"os.path.abspath"
] | [((136, 161), 'os.path.abspath', 'os.path.abspath', (['__file__'], {}), '(__file__)\n', (151, 161), False, 'import os\n')] |
# -*- coding: utf-8 -*-
"""
Created on Thu Aug 13 09:52:47 2015
@author: wirkert
"""
import unittest
import os
import numpy as np
import msi.msimanipulations as msimani
from msi.io.nrrdreader import NrrdReader
from msi.io.nrrdwriter import NrrdWriter
from msi.test import helpers
class TestNrrdWriter(unittest.TestC... | [
"msi.io.nrrdreader.NrrdReader",
"os.remove",
"os.path.isfile",
"numpy.array",
"msi.msimanipulations.calculate_mean_spectrum",
"msi.test.helpers.getFakeMsi",
"msi.io.nrrdwriter.NrrdWriter"
] | [((430, 450), 'msi.test.helpers.getFakeMsi', 'helpers.getFakeMsi', ([], {}), '()\n', (448, 450), False, 'from msi.test import helpers\n'), ((574, 604), 'os.remove', 'os.remove', (['self.fileUriToWrite'], {}), '(self.fileUriToWrite)\n', (583, 604), False, 'import os\n'), ((666, 686), 'msi.io.nrrdwriter.NrrdWriter', 'Nrr... |
import h2o
from h2o.base import Keyed
from h2o.exceptions import H2OValueError
from h2o.job import H2OJob
from h2o.model import ModelBase
from h2o.utils.typechecks import assert_is_type, is_type
class H2OAutoMLBaseMixin:
def predict(self, test_data):
"""
Predict on a dataset.
:param ... | [
"h2o.get_model",
"h2o.model.ModelBase.download_mojo",
"h2o.utils.typechecks.is_type",
"h2o.download_pojo",
"h2o.utils.typechecks.assert_is_type",
"h2o.H2OFrame",
"h2o.assign",
"h2o.automl.get_leaderboard"
] | [((9792, 9836), 'h2o.utils.typechecks.assert_is_type', 'assert_is_type', (['extensions', 'None', 'str', '[str]'], {}), '(extensions, None, str, [str])\n', (9806, 9836), False, 'from h2o.utils.typechecks import assert_is_type, is_type\n'), ((1841, 1932), 'h2o.download_pojo', 'h2o.download_pojo', (['self.leader', 'path']... |
# coding: utf-8
from pymongo import MongoClient
import conf
class MongoQuery(object):
def __init__(self):
self._conn = MongoClient(conf.mongodb_conn_str)
self._db = self._conn.geokb
def query(self, grounded, limit=15, sort_keys=None):
col = self._db[grounded['from']]
docs = co... | [
"pymongo.MongoClient"
] | [((133, 167), 'pymongo.MongoClient', 'MongoClient', (['conf.mongodb_conn_str'], {}), '(conf.mongodb_conn_str)\n', (144, 167), False, 'from pymongo import MongoClient\n')] |
#
# Copyright (C) 2014 Dell, Inc.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in wri... | [
"logging.getLogger",
"os.path.exists",
"dcm.agent.logger.log_to_dcm_console_configuration_management_error",
"dcm.agent.plugins.api.exceptions.AgentPluginParameterBadValueException",
"dcm.agent.logger.log_to_dcm_console_configuration_management_output",
"dcm.agent.plugins.api.base.PluginReply",
"json.du... | [((948, 975), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (965, 975), False, 'import logging\n'), ((3059, 3096), 'os.path.join', 'os.path.join', (['chef_dir', '"""runList.cfg"""'], {}), "(chef_dir, 'runList.cfg')\n", (3071, 3096), False, 'import os\n'), ((4736, 4767), 'configparser.Saf... |
from toolz import get
PLOT_VALIDATORS = [
(
{"line", "scatter", "bar"},
lambda x: ("x" not in x) or ("y" not in x),
"XY plot does not have X and Y.",
),
(
{"histogram"},
lambda x: ("step" in x) and ("bins" in x),
"Histogram cannot have STEP and BINS.",
),... | [
"toolz.get"
] | [((1666, 1683), 'toolz.get', 'get', (['"""step"""', 'x', '(1)'], {}), "('step', x, 1)\n", (1669, 1683), False, 'from toolz import get\n'), ((1787, 1804), 'toolz.get', 'get', (['"""bins"""', 'x', '(1)'], {}), "('bins', x, 1)\n", (1790, 1804), False, 'from toolz import get\n'), ((551, 566), 'toolz.get', 'get', (['"""x"""... |
from hoyolab import main
from os import environ
from os.path import exists
import atoma
def init_environ(d):
environ['HOYOLAB_JSON_PATH'] = '{}/hoyolab.json'.format(d)
environ['HOYOLAB_ATOM_PATH'] = '{}/hoyolab.xml'.format(d)
environ['HOYOLAB_JSON_URL'] = 'hoyolab.json'
environ['HOYOLAB_ATOM_URL'] = '... | [
"os.path.exists",
"atoma.parse_json_feed_file",
"hoyolab.main",
"atoma.parse_atom_file"
] | [((571, 577), 'hoyolab.main', 'main', ([], {}), '()\n', (575, 577), False, 'from hoyolab import main\n'), ((590, 607), 'os.path.exists', 'exists', (['json_path'], {}), '(json_path)\n', (596, 607), False, 'from os.path import exists\n'), ((619, 636), 'os.path.exists', 'exists', (['atom_path'], {}), '(atom_path)\n', (625... |
from django.urls import path
from . import views as sg
urlpatterns = [
path('artist', sg.artist),
path('genre', sg.genre),
path('release', sg.release),
path('track', sg.track),
path('', sg.sausage_grinder_index),
]
| [
"django.urls.path"
] | [((77, 102), 'django.urls.path', 'path', (['"""artist"""', 'sg.artist'], {}), "('artist', sg.artist)\n", (81, 102), False, 'from django.urls import path\n'), ((108, 131), 'django.urls.path', 'path', (['"""genre"""', 'sg.genre'], {}), "('genre', sg.genre)\n", (112, 131), False, 'from django.urls import path\n'), ((137, ... |
from FitnessPlot import FitnessPlot
'''
for n in range(1,6):
plot = FitnessPlot(folder_prefix='data_top{}'.format(n))
plot.plot_all_workers()
plot.plot_workers_as_average()
'''
plot = FitnessPlot(folder_prefix='data_top1', num_workers=16)
worker_dict = plot.create_worker_dict()
#plot.plot_all_workers()
#... | [
"FitnessPlot.FitnessPlot"
] | [((199, 253), 'FitnessPlot.FitnessPlot', 'FitnessPlot', ([], {'folder_prefix': '"""data_top1"""', 'num_workers': '(16)'}), "(folder_prefix='data_top1', num_workers=16)\n", (210, 253), False, 'from FitnessPlot import FitnessPlot\n')] |
import logging
def setup_logger():
formatter = logging.Formatter('%(asctime)s.%(msecs)03d %(levelname)s %(message)s',
'%Y-%m-%d %H:%M:%S')
logging.basicConfig(level=logging.INFO,
format='%(asctime)s.%(msecs)03d %(levelname)-8s %(message)s',
... | [
"logging.basicConfig",
"logging.Formatter",
"logging.getLogger"
] | [((54, 149), 'logging.Formatter', 'logging.Formatter', (['"""%(asctime)s.%(msecs)03d %(levelname)s %(message)s"""', '"""%Y-%m-%d %H:%M:%S"""'], {}), "('%(asctime)s.%(msecs)03d %(levelname)s %(message)s',\n '%Y-%m-%d %H:%M:%S')\n", (71, 149), False, 'import logging\n'), ((184, 324), 'logging.basicConfig', 'logging.ba... |
#!/usr/bin/env python3
# -*- coding: utf-8 -*-
"""ThreeScale Proxies Rule interface for APIs."""
from .base import ThreeScale
import logging
import requests
import xmltodict
import json
logger = logging.getLogger(__name__)
class Proxies(ThreeScale):
"""ThreeScale Proxies create, update."""
response = None... | [
"logging.getLogger",
"requests.post",
"xmltodict.parse",
"requests.patch",
"json.dumps",
"requests.get",
"requests.put"
] | [((198, 225), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (215, 225), False, 'import logging\n'), ((2457, 2496), 'requests.patch', 'requests.patch', (['_url'], {'data': 'request_body'}), '(_url, data=request_body)\n', (2471, 2496), False, 'import requests\n'), ((3491, 3524), 'requests.... |
# Created with tutorials:
# https://www.digitalocean.com/community/tutorials/how-to-structure-large-flask-applications
# http://flask.pocoo.org/docs/0.12/tutorial
from flask import Flask, g, render_template
from flask_sqlalchemy import SQLAlchemy
import sqlite3
# Define WSGI application object.
app = Flask(__name__... | [
"flask.render_template",
"flask_sqlalchemy.SQLAlchemy",
"flask.Flask"
] | [((306, 321), 'flask.Flask', 'Flask', (['__name__'], {}), '(__name__)\n', (311, 321), False, 'from flask import Flask, g, render_template\n'), ((451, 466), 'flask_sqlalchemy.SQLAlchemy', 'SQLAlchemy', (['app'], {}), '(app)\n', (461, 466), False, 'from flask_sqlalchemy import SQLAlchemy\n'), ((524, 551), 'flask.render_t... |
# coding=utf-8
# *** WARNING: this file was generated by the Pulumi Terraform Bridge (tfgen) Tool. ***
# *** Do not edit by hand unless you're certain you know what you are doing! ***
import warnings
import pulumi
import pulumi.runtime
from typing import Any, Mapping, Optional, Sequence, Union, overload
from . import ... | [
"pulumi.get",
"pulumi.getter",
"pulumi.set",
"pulumi.InvokeOptions",
"pulumi.runtime.invoke"
] | [((1788, 1817), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""isLocal"""'}), "(name='isLocal')\n", (1801, 1817), False, 'import pulumi\n'), ((1914, 1945), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""loginName"""'}), "(name='loginName')\n", (1927, 1945), False, 'import pulumi\n'), ((766, 802), 'pulumi.set'... |
import logging
import os
import tensorflow as tf
from punc_recover.models.punc_transformer import PuncTransformer
from punc_recover.tester.base_tester import BaseTester
from utils.text_featurizers import TextFeaturizer
class PuncTester(BaseTester):
""" Trainer for CTC Models """
def __init__(self,
... | [
"utils.text_featurizers.TextFeaturizer",
"os.listdir",
"tensorflow.equal",
"tensorflow.reduce_sum",
"tensorflow.keras.metrics.Mean",
"os.path.join",
"tensorflow.keras.metrics.sparse_categorical_accuracy",
"tensorflow.math.equal",
"tensorflow.reduce_mean",
"punc_recover.models.punc_transformer.Punc... | [((507, 543), 'utils.text_featurizers.TextFeaturizer', 'TextFeaturizer', (["config['punc_vocab']"], {}), "(config['punc_vocab'])\n", (521, 543), False, 'from utils.text_featurizers import TextFeaturizer\n'), ((573, 612), 'utils.text_featurizers.TextFeaturizer', 'TextFeaturizer', (["config['punc_biaodian']"], {}), "(con... |
# encoding: utf-8
import datetime
from south.db import db
from south.v2 import SchemaMigration
from django.db import models
class Migration(SchemaMigration):
def forwards(self, orm):
# Adding field 'ProductVersion.cc_version'
db.add_column('core_productversion', 'cc_version', self.gf('django.db.... | [
"south.db.db.delete_column"
] | [((652, 705), 'south.db.db.delete_column', 'db.delete_column', (['"""core_productversion"""', '"""cc_version"""'], {}), "('core_productversion', 'cc_version')\n", (668, 705), False, 'from south.db import db\n'), ((761, 807), 'south.db.db.delete_column', 'db.delete_column', (['"""core_product"""', '"""cc_version"""'], {... |
import matplotlib.pyplot as pl
import os
import numpy as np
from ticle.data.dataHandler import normalizeData,load_file
from ticle.analysis.analysis import get_phases,normalize_phase
pl.rc('xtick', labelsize='x-small')
pl.rc('ytick', labelsize='x-small')
pl.rc('font', family='serif')
pl.rcParams.update({'font.size': 2... | [
"os.makedirs",
"matplotlib.pyplot.ylabel",
"matplotlib.pyplot.xlabel",
"matplotlib.pyplot.plot",
"ticle.data.dataHandler.load_file",
"os.getcwd",
"ticle.data.dataHandler.normalizeData",
"matplotlib.pyplot.rcParams.update",
"matplotlib.pyplot.figure",
"ticle.analysis.analysis.get_phases",
"numpy.... | [((184, 219), 'matplotlib.pyplot.rc', 'pl.rc', (['"""xtick"""'], {'labelsize': '"""x-small"""'}), "('xtick', labelsize='x-small')\n", (189, 219), True, 'import matplotlib.pyplot as pl\n'), ((220, 255), 'matplotlib.pyplot.rc', 'pl.rc', (['"""ytick"""'], {'labelsize': '"""x-small"""'}), "('ytick', labelsize='x-small')\n"... |
#!/usr/bin/python
import shlex
import simplejson
from putil.rabbitmq.rabbitmqadmin import Management, make_parser, LISTABLE, DELETABLE
class RabbitManagementUtil(object):
def __init__(self, config, options=None, sysname=None):
"""
Given a config object (system CFG or rabbit mgmt config), extra... | [
"shlex.split",
"putil.rabbitmq.rabbitmqadmin.make_parser",
"simplejson.loads",
"putil.rabbitmq.rabbitmqadmin.Management"
] | [((773, 786), 'putil.rabbitmq.rabbitmqadmin.make_parser', 'make_parser', ([], {}), '()\n', (784, 786), False, 'from putil.rabbitmq.rabbitmqadmin import Management, make_parser, LISTABLE, DELETABLE\n'), ((4489, 4518), 'putil.rabbitmq.rabbitmqadmin.Management', 'Management', (['options', 'args[1:]'], {}), '(options, args... |
#!/usr/bin/env python3
# -*- coding: utf-8 -*-
"""
ExchangeConnector fixEngine
Copyright (c) 2020 <NAME>
Permission is hereby granted, free of charge, to any person obtaining a copy
of this software and associated documentation files (the "Software"), to deal
in the Software without restriction, including without lim... | [
"configparser.SafeConfigParser",
"asyncio.open_connection",
"connectionHandler.FIXConnectionHandler.__init__"
] | [((1479, 1555), 'connectionHandler.FIXConnectionHandler.__init__', 'FIXConnectionHandler.__init__', (['self', 'config', 'reader', 'writer', 'messageListener'], {}), '(self, config, reader, writer, messageListener)\n', (1508, 1555), False, 'from connectionHandler import FIXConnectionHandler, SocketConnectionState\n'), (... |
from jadi import component
from aj.api.http import url, HttpPlugin
from aj.auth import authorize
from aj.api.endpoint import endpoint, EndpointError
import aj
import gevent
@component(HttpPlugin)
class Handler(HttpPlugin):
def __init__(self, context):
self.context = context
@url(r'/api/session_list/l... | [
"gevent.sleep",
"aj.api.http.url",
"jadi.component",
"aj.api.endpoint.endpoint"
] | [((176, 197), 'jadi.component', 'component', (['HttpPlugin'], {}), '(HttpPlugin)\n', (185, 197), False, 'from jadi import component\n'), ((295, 324), 'aj.api.http.url', 'url', (['"""/api/session_list/list"""'], {}), "('/api/session_list/list')\n", (298, 324), False, 'from aj.api.http import url, HttpPlugin\n'), ((331, ... |
import pandas as pd
import os
#opt = itertools.islice(ls, len(ls))
#st = map(lambda x : )
def parsecode(txt):
df = pd.read_csv(os.getcwd() + '\\OMDB.csv')
ls = df['Code'].to_list()
code = []
q = 0
for i in range(len(ls)):
text = txt
if ls[i] in text:
n = text.find(ls[... | [
"pandas.read",
"os.getcwd"
] | [((135, 146), 'os.getcwd', 'os.getcwd', ([], {}), '()\n', (144, 146), False, 'import os\n'), ((1379, 1401), 'pandas.read', 'pd.read', (['qry'], {'con': 'conn'}), '(qry, con=conn)\n', (1386, 1401), True, 'import pandas as pd\n')] |
import socket
import unittest
from eats.webdriver import PytractorWebDriver
from eats.tests.common import SimpleWebServerProcess as SimpleServer
def _get_local_ip_addr():
s = socket.socket(socket.AF_INET, socket.SOCK_DGRAM)
s.connect(("gmail.com",80))
local_ip_addr = s.getsockname()[0]
s.close()
re... | [
"eats.tests.common.SimpleWebServerProcess",
"socket.socket"
] | [((180, 228), 'socket.socket', 'socket.socket', (['socket.AF_INET', 'socket.SOCK_DGRAM'], {}), '(socket.AF_INET, socket.SOCK_DGRAM)\n', (193, 228), False, 'import socket\n'), ((454, 468), 'eats.tests.common.SimpleWebServerProcess', 'SimpleServer', ([], {}), '()\n', (466, 468), True, 'from eats.tests.common import Simpl... |