code stringlengths 22 1.05M | apis listlengths 1 3.31k | extract_api stringlengths 75 3.25M |
|---|---|---|
import tqdm
import torch
from lav.lav_privileged import LAV
from lav.utils.datasets import get_data_loader
from lav.utils.logger import Logger
def main(args):
dmd = LAV(args)
data_loader = get_data_loader('bev', args)
logger = Logger('lav_bev', args)
save_dir = logger.save_dir
torch.manual_seed(a... | [
"torch.manual_seed",
"argparse.ArgumentParser",
"tqdm.tqdm",
"lav.utils.logger.Logger",
"lav.lav_privileged.LAV",
"lav.utils.datasets.get_data_loader"
] | [((171, 180), 'lav.lav_privileged.LAV', 'LAV', (['args'], {}), '(args)\n', (174, 180), False, 'from lav.lav_privileged import LAV\n'), ((199, 227), 'lav.utils.datasets.get_data_loader', 'get_data_loader', (['"""bev"""', 'args'], {}), "('bev', args)\n", (214, 227), False, 'from lav.utils.datasets import get_data_loader\... |
#!/usr/bin/env python3
# async_requests.py
"""Asynchronously get links embedded in multiple pages' HTML."""
import asyncio
import logging
import re
import sys
# from typing import IO # Use pathlib instead
import urllib.error
import urllib.parse
import aiofiles
import aiohttp
from aiohttp import ClientSession
impor... | [
"logging.basicConfig",
"aiohttp.ClientSession",
"logging.getLogger",
"asyncio.WindowsSelectorEventLoopPolicy",
"re.compile",
"pathlib.Path",
"aiofiles.open",
"asyncio.gather"
] | [((393, 539), 'logging.basicConfig', 'logging.basicConfig', ([], {'format': '"""%(asctime)s %(levelname)s:%(name)s: %(message)s"""', 'level': 'logging.DEBUG', 'datefmt': '"""%H:%M:%S"""', 'stream': 'sys.stderr'}), "(format=\n '%(asctime)s %(levelname)s:%(name)s: %(message)s', level=logging.DEBUG,\n datefmt='%H:%M... |
# -*- coding: utf-8 -*-
from django.shortcuts import render
from qiniuyun.backend import QiniuPush
from qiniuyun.models import ImageAtQiniu
from .indexView import ImgList
from DJangoHotel.models import RoomInfo
def roomInfo(request):
rooms=RoomInfo.objects.all()
imgObjs=ImageAtQiniu.objects.all()
imgUrls=... | [
"django.shortcuts.render",
"qiniuyun.models.ImageAtQiniu.objects.all",
"qiniuyun.backend.QiniuPush.private_download_url",
"DJangoHotel.models.RoomInfo.objects.all"
] | [((246, 268), 'DJangoHotel.models.RoomInfo.objects.all', 'RoomInfo.objects.all', ([], {}), '()\n', (266, 268), False, 'from DJangoHotel.models import RoomInfo\n'), ((281, 307), 'qiniuyun.models.ImageAtQiniu.objects.all', 'ImageAtQiniu.objects.all', ([], {}), '()\n', (305, 307), False, 'from qiniuyun.models import Image... |
import pyb
import sensor
import image
import time
# Status LED
led = pyb.LED(3)
# Configure camera
sensor.reset()
sensor.set_contrast(3)
sensor.set_gainceiling(16)
sensor.set_framesize(sensor.QVGA)
sensor.set_pixformat(sensor.GRAYSCALE)
# Get center x, y of camera image
WIDTH = sensor.width()
HEIGHT = sensor.height(... | [
"sensor.set_contrast",
"sensor.set_gainceiling",
"sensor.set_pixformat",
"time.clock",
"sensor.set_framesize",
"image.HaarCascade",
"sensor.width",
"sensor.reset",
"sensor.height",
"sensor.snapshot",
"pyb.LED"
] | [((70, 80), 'pyb.LED', 'pyb.LED', (['(3)'], {}), '(3)\n', (77, 80), False, 'import pyb\n'), ((101, 115), 'sensor.reset', 'sensor.reset', ([], {}), '()\n', (113, 115), False, 'import sensor\n'), ((116, 138), 'sensor.set_contrast', 'sensor.set_contrast', (['(3)'], {}), '(3)\n', (135, 138), False, 'import sensor\n'), ((13... |
from tortoise import fields, models
class User(models.Model):
""" Model user """
username = fields.CharField(max_length=100, unique=True)
password = fields.CharField(max_length=100)
email = fields.CharField(max_length=100, unique=True)
first_name = fields.CharField(max_length=100)
last_name = ... | [
"tortoise.fields.CharField",
"tortoise.fields.DatetimeField",
"tortoise.fields.BooleanField",
"tortoise.fields.TextField"
] | [((102, 147), 'tortoise.fields.CharField', 'fields.CharField', ([], {'max_length': '(100)', 'unique': '(True)'}), '(max_length=100, unique=True)\n', (118, 147), False, 'from tortoise import fields, models\n'), ((163, 195), 'tortoise.fields.CharField', 'fields.CharField', ([], {'max_length': '(100)'}), '(max_length=100)... |
import webbrowser
from ontology.exception.exception import SDKException
from click import (
argument,
pass_context
)
from .main import main
from punica.box.repo_box import Box
from punica.utils.output import echo_cli_exception
from punica.exception.punica_exception import PunicaException
@main.command('unb... | [
"click.argument",
"punica.utils.output.echo_cli_exception",
"punica.box.repo_box.Box",
"webbrowser.open"
] | [((326, 355), 'click.argument', 'argument', (['"""box_name"""'], {'nargs': '(1)'}), "('box_name', nargs=1)\n", (334, 355), False, 'from click import argument, pass_context\n'), ((481, 508), 'punica.box.repo_box.Box', 'Box', (["ctx.obj['PROJECT_DIR']"], {}), "(ctx.obj['PROJECT_DIR'])\n", (484, 508), False, 'from punica.... |
import sys
from datetime import date, timedelta
import requests
def date_gen(d1, d2):
# d1 = date(2020, 5, 1)
# d2 = date(2020, 5, 31)
delta = d2 - d1
return [(d1 + timedelta(days=i)).strftime('%Y%m%d') for i in range(delta.days + 1)]
def download_by_dates(date_list):
for date_to_download in da... | [
"requests.get",
"datetime.timedelta",
"datetime.date",
"sys.stdout.flush",
"sys.stdout.write"
] | [((625, 655), 'requests.get', 'requests.get', (['url'], {'stream': '(True)'}), '(url, stream=True)\n', (637, 655), False, 'import requests\n'), ((1321, 1337), 'datetime.date', 'date', (['(2020)', '(3)', '(1)'], {}), '(2020, 3, 1)\n', (1325, 1337), False, 'from datetime import date, timedelta\n'), ((1342, 1358), 'dateti... |
from heapq import heapify, heappush, heappop
from collections import defaultdict
import math
def shortest_path(M, start, goal):
frontier = {start}
explored = set()
came_from = dict()
f_costs = get_initial_f_costs(M, start, goal) # heapq type
g_costs = get_initial_g_costs(start) # defaultdict typ... | [
"math.pow",
"heapq.heappop",
"collections.defaultdict",
"heapq.heappush",
"heapq.heapify"
] | [((1937, 1953), 'heapq.heapify', 'heapify', (['f_costs'], {}), '(f_costs)\n', (1944, 1953), False, 'from heapq import heapify, heappush, heappop\n'), ((2012, 2057), 'heapq.heappush', 'heappush', (['f_costs', '(f_cost_start_node, start)'], {}), '(f_costs, (f_cost_start_node, start))\n', (2020, 2057), False, 'from heapq ... |
# -*- coding: utf-8 -*-
import cherrypy
from jinja2 import Template
import mock
from tests.utils import BaseToolsTest
from lib.tool.allowed_methods import AllowedMethodsTool
from lib.tool.cpemail import EmailTool
from lib.tool.template import Jinja2Tool
class TestAllowedMethods(BaseToolsTest):
_cp_config = {
... | [
"lib.tool.allowed_methods.AllowedMethodsTool",
"lib.tool.template.Jinja2Tool"
] | [((509, 529), 'lib.tool.allowed_methods.AllowedMethodsTool', 'AllowedMethodsTool', ([], {}), '()\n', (527, 529), False, 'from lib.tool.allowed_methods import AllowedMethodsTool\n'), ((1072, 1084), 'lib.tool.template.Jinja2Tool', 'Jinja2Tool', ([], {}), '()\n', (1082, 1084), False, 'from lib.tool.template import Jinja2T... |
import os
import logging
import stackprinter
from celery import Celery, Task
from celery.schedules import crontab
from django.apps import apps, AppConfig
from django.conf import settings
if not settings.configured:
# set the default Django settings module for the 'celery' program.
os.environ.setdefault(
... | [
"os.environ.setdefault",
"celery.Celery",
"django.apps.apps.get_app_configs",
"stackprinter.format",
"celery.schedules.crontab",
"logging.error"
] | [((407, 430), 'celery.Celery', 'Celery', (['"""instanotifier"""'], {}), "('instanotifier')\n", (413, 430), False, 'from celery import Celery, Task\n'), ((293, 365), 'os.environ.setdefault', 'os.environ.setdefault', (['"""DJANGO_SETTINGS_MODULE"""', '"""config.settings.local"""'], {}), "('DJANGO_SETTINGS_MODULE', 'confi... |
from snovault import (
AuditFailure,
audit_checker,
)
@audit_checker('ReferenceEpigenome', frame=['related_datasets',
'related_datasets.replicates',
'related_datasets.replicates.library',
... | [
"snovault.AuditFailure",
"snovault.audit_checker"
] | [((65, 378), 'snovault.audit_checker', 'audit_checker', (['"""ReferenceEpigenome"""'], {'frame': "['related_datasets', 'related_datasets.replicates',\n 'related_datasets.replicates.library',\n 'related_datasets.replicates.library.biosample',\n 'related_datasets.replicates.library.biosample.donor',\n 'relate... |
import openmc
from scipy import interpolate
import matplotlib.pyplot as plt
from matplotlib.colors import LogNorm
from matplotlib import ticker
import matplotx
import numpy as np
import scipy.ndimage as ndimage
def reshape_values_to_mesh_shape(tally, values):
mesh_filter = tally.find_filter(filter_type=openmc.Mes... | [
"matplotlib.pyplot.savefig",
"matplotlib.pyplot.gca",
"matplotlib.pyplot.xlabel",
"matplotlib.pyplot.sca",
"matplotx.ylabel_top",
"numpy.array",
"openmc.StatePoint",
"matplotlib.pyplot.style.context",
"numpy.linspace",
"matplotlib.pyplot.subplots",
"matplotlib.pyplot.scatter",
"scipy.ndimage.g... | [((2780, 2823), 'openmc.StatePoint', 'openmc.StatePoint', ([], {'filepath': 'statepoint_file'}), '(filepath=statepoint_file)\n', (2797, 2823), False, 'import openmc\n'), ((1908, 1930), 'numpy.array', 'np.array', (["data['mean']"], {}), "(data['mean'])\n", (1916, 1930), True, 'import numpy as np\n'), ((2302, 2353), 'sci... |
# -*- coding: utf-8 -*-
"""
Provides base logging functions
:copyright: © 2018 by <NAME>
:license: MIT, see LICENSE for more details.
"""
from logging import getLevelName, INFO, WARN, ERROR, DEBUG
from multiprocessing import current_process
from time import strftime
from core.common import load_config
from... | [
"logging.getLevelName",
"time.strftime",
"multiprocessing.current_process",
"core.common.load_config"
] | [((623, 636), 'core.common.load_config', 'load_config', ([], {}), '()\n', (634, 636), False, 'from core.common import load_config\n'), ((551, 568), 'multiprocessing.current_process', 'current_process', ([], {}), '()\n', (566, 568), False, 'from multiprocessing import current_process\n'), ((1828, 1849), 'time.strftime',... |
from interfaces.interface import Publisher, stock_list
from utils.notifier import NotificationUtils
from service.loader import LoadStock
from utils import cache_util, common_constants
class WatchTower(Publisher):
state = 0
observser_list = list()
def __init__(self, stock_val):
self.stock_val = st... | [
"utils.cache_util.create_cache_client",
"service.loader.LoadStock",
"utils.notifier.NotificationUtils"
] | [((635, 700), 'utils.notifier.NotificationUtils', 'NotificationUtils', ([], {'user_email': 'user_email', 'custom_message': 'messages'}), '(user_email=user_email, custom_message=messages)\n', (652, 700), False, 'from utils.notifier import NotificationUtils\n'), ((2588, 2620), 'utils.cache_util.create_cache_client', 'cac... |
"""This module tests Exceptions functionality in stereomideval module"""
import pytest
import numpy as np
from stereomideval.dataset import Dataset
from stereomideval.exceptions import ImageSizeNotEqual, PathNotFound, InvalidSceneName
def test_catch_invalid_image_sizes():
"""Test catching invalid image sizes"""
... | [
"stereomideval.dataset.Dataset.get_training_scene_list",
"stereomideval.dataset.Dataset.get_scene_list",
"stereomideval.exceptions.ImageSizeNotEqual.validate",
"numpy.zeros",
"pytest.raises",
"stereomideval.exceptions.PathNotFound.validate"
] | [((333, 349), 'numpy.zeros', 'np.zeros', (['(5, 5)'], {}), '((5, 5))\n', (341, 349), True, 'import numpy as np\n'), ((364, 380), 'numpy.zeros', 'np.zeros', (['(5, 6)'], {}), '((5, 6))\n', (372, 380), True, 'import numpy as np\n'), ((390, 422), 'pytest.raises', 'pytest.raises', (['ImageSizeNotEqual'], {}), '(ImageSizeNo... |
# -*- coding: utf8 -*-
# ============LICENSE_START=======================================================
# org.onap.vvp/validation-scripts
# ===================================================================
# Copyright © 2017 AT&T Intellectual Property. All rights reserved.
# ========================================... | [
"pytest.skip",
"tests.cached_yaml.load",
"re.compile"
] | [((1890, 1937), 're.compile', 're.compile', (['"""int_(?P<network_role>.+)_network$"""'], {}), "('int_(?P<network_role>.+)_network$')\n", (1900, 1937), False, 'import re\n'), ((2248, 2295), 're.compile', 're.compile', (['"""int_(?P<network_role>.+)_network$"""'], {}), "('int_(?P<network_role>.+)_network$')\n", (2258, 2... |
import json
from bson import ObjectId
from pymongo import ReturnDocument
from .exceptions import DBException
class DBActionsMixin:
def __init__(self, model, db):
self._model_cls = model
self._db = db
def add(self, item):
db_obj = self._collection.insert_one(item.prepare_for_db())
... | [
"bson.ObjectId",
"json.dumps"
] | [((1785, 1801), 'json.dumps', 'json.dumps', (['self'], {}), '(self)\n', (1795, 1801), False, 'import json\n'), ((1699, 1716), 'bson.ObjectId', 'ObjectId', (['self.id'], {}), '(self.id)\n', (1707, 1716), False, 'from bson import ObjectId\n')] |
# -*- coding: utf-8 -*-
from random import Random
#from core.dataloader import DataLoader
from torch.utils.data import DataLoader
import numpy as np
from math import *
import logging
from scipy import stats
import numpy as np
from pyemd import emd
from collections import OrderedDict
import time
import pickle, random
fr... | [
"numpy.identity",
"pyemd.emd",
"collections.OrderedDict",
"scipy.stats.entropy",
"random.Random",
"pickle.load",
"logging.info",
"numpy.sum",
"numpy.zeros",
"numpy.random.seed",
"numpy.concatenate",
"torch.utils.data.DataLoader",
"time.time"
] | [((20762, 20773), 'time.time', 'time.time', ([], {}), '()\n', (20771, 20773), False, 'import time\n'), ((941, 949), 'random.Random', 'Random', ([], {}), '()\n', (947, 949), False, 'from random import Random\n'), ((1168, 1188), 'numpy.random.seed', 'np.random.seed', (['seed'], {}), '(seed)\n', (1182, 1188), True, 'impor... |
import os
import shutil
import Ni__eam__born_exp_rose as configuration
from collections import OrderedDict
def cleanup_simulation_directories():
sim_directories = [
'Ni_fcc_vac.lmps_min_pos',
'Ni_fcc.lmps_elastic',
'Ni_fcc_100_unit.lmps_min_all',
'Ni_fcc_111_s.lmps_min_pos',
... | [
"collections.OrderedDict",
"pypospack.pyposmat.engines.PyposmatEngine",
"Ni__eam__born_exp_rose.write_configuration_file",
"os.path.isdir",
"shutil.rmtree"
] | [((889, 902), 'collections.OrderedDict', 'OrderedDict', ([], {}), '()\n', (900, 902), False, 'from collections import OrderedDict\n'), ((1537, 1550), 'collections.OrderedDict', 'OrderedDict', ([], {}), '()\n', (1548, 1550), False, 'from collections import OrderedDict\n'), ((1574, 1587), 'collections.OrderedDict', 'Orde... |
from django.contrib import admin
from stade.core.models import Challenge
from .task import TaskInline
@admin.register(Challenge)
class ChallengeAdmin(admin.ModelAdmin):
inlines = [TaskInline]
| [
"django.contrib.admin.register"
] | [((107, 132), 'django.contrib.admin.register', 'admin.register', (['Challenge'], {}), '(Challenge)\n', (121, 132), False, 'from django.contrib import admin\n')] |
#!/usr/bin/env python3
"""problem_059.py
Problem 59: XOR decryption
Each character on a computer is assigned a unique code and the preferred
standard is ASCII (American Standard Code for Information Interchange). For
example, uppercase A = 65, asterisk (*) = 42, and lowercase k = 107.
A modern encryption method is ... | [
"common.fileio.ints_from_file",
"common.alphabet.letter_char_lower"
] | [((2445, 2483), 'common.fileio.ints_from_file', 'fio.ints_from_file', (['FILE_NAME'], {'sep': '""","""'}), "(FILE_NAME, sep=',')\n", (2463, 2483), True, 'import common.fileio as fio\n'), ((2744, 2770), 'common.alphabet.letter_char_lower', 'alpha.letter_char_lower', (['i'], {}), '(i)\n', (2767, 2770), True, 'import comm... |
import numpy as np
import glob
import geo
import time
import pdb
start_time = time.time()
dataDir='./data/'
# get CrIS files
cris_sdr_files = sorted(glob.glob(dataDir+'SCRIS*'))
cris_geo_files = sorted(glob.glob(dataDir+'GCRSO*'))
# get VIIRS files
viirs_sdr_files = sorted(glob.glob(dataDir+'SVM15*'))
viirs_geo_f... | [
"geo.read_viirs_sdr",
"matplotlib.pyplot.savefig",
"geo.read_cris_sdr",
"geo.match_cris_viirs",
"numpy.zeros_like",
"numpy.ndindex",
"numpy.append",
"matplotlib.cm.ScalarMappable",
"glob.glob",
"matplotlib.colors.Normalize",
"geo.RAE2ENU",
"matplotlib.pyplot.get_cmap",
"geo.ENU2ECEF",
"tim... | [((79, 90), 'time.time', 'time.time', ([], {}), '()\n', (88, 90), False, 'import time\n'), ((457, 492), 'geo.read_viirs_geo', 'geo.read_viirs_geo', (['viirs_geo_files'], {}), '(viirs_geo_files)\n', (475, 492), False, 'import geo\n'), ((528, 563), 'geo.read_viirs_sdr', 'geo.read_viirs_sdr', (['viirs_sdr_files'], {}), '(... |
"""Aggregate hosts data set."""
from censys.search import SearchClient
c = SearchClient()
# The aggregate method constructs a report using a query, an aggregation field, and the
# number of buckets to bin.
report = c.v2.hosts.aggregate(
"service.service_name: HTTP",
"services.port",
num_buckets=5,
)
print... | [
"censys.search.SearchClient"
] | [((76, 90), 'censys.search.SearchClient', 'SearchClient', ([], {}), '()\n', (88, 90), False, 'from censys.search import SearchClient\n')] |
import aiohttp_jinja2
from aiohttp import web
from riego.db import get_db
from riego.web.security import raise_permission
router = web.RouteTableDef()
def setup_routes_events(app):
app.add_routes(router)
@router.get("/events", name='events')
@aiohttp_jinja2.template('events/index.html')
async... | [
"aiohttp_jinja2.template",
"riego.db.get_db",
"riego.web.security.raise_permission",
"aiohttp.web.RouteTableDef"
] | [((141, 160), 'aiohttp.web.RouteTableDef', 'web.RouteTableDef', ([], {}), '()\n', (158, 160), False, 'from aiohttp import web\n'), ((269, 313), 'aiohttp_jinja2.template', 'aiohttp_jinja2.template', (['"""events/index.html"""'], {}), "('events/index.html')\n", (292, 313), False, 'import aiohttp_jinja2\n'), ((863, 907), ... |
# coding: utf-8
from __future__ import division, print_function, absolute_import
import pkg_resources
__version__ = pkg_resources.get_distribution('pypcl').version
from pypcl.common import *
| [
"pkg_resources.get_distribution"
] | [((119, 158), 'pkg_resources.get_distribution', 'pkg_resources.get_distribution', (['"""pypcl"""'], {}), "('pypcl')\n", (149, 158), False, 'import pkg_resources\n')] |
import tkinter as tk
import tkinter.font as tkFont
class History(tk.Frame):
def __init__(self, master):
self.master=master
tk.Frame.__init__(self, master=master)
self.listOperations =[]
self.history = []
self.LEFT = 0
self.OP = 1
self.RIGHT = 2
self.E... | [
"tkinter.font.Font",
"tkinter.Frame.__init__"
] | [((144, 182), 'tkinter.Frame.__init__', 'tk.Frame.__init__', (['self'], {'master': 'master'}), '(self, master=master)\n', (161, 182), True, 'import tkinter as tk\n'), ((831, 851), 'tkinter.font.Font', 'tkFont.Font', ([], {'size': '(10)'}), '(size=10)\n', (842, 851), True, 'import tkinter.font as tkFont\n'), ((637, 657)... |
"""
Contains the code necessary to extract a list of optimal compression values from a csv file containing
columns corresponding to {compression_type}_{level}, {variable}, {time}, and {DSSIM}
It would be best to open the csv file once, and get a list of all variables, levels, and timesteps
so I don't read the csv file... | [
"csv.DictWriter",
"numpy.unique",
"argparse.ArgumentParser",
"re.compile",
"os.path.isfile",
"numpy.argsort",
"lcr_global_vars.varlist",
"csv.reader",
"re.search"
] | [((8948, 8964), 'numpy.unique', 'np.unique', (['times'], {}), '(times)\n', (8957, 8964), True, 'import numpy as np\n'), ((10290, 10306), 'numpy.unique', 'np.unique', (['times'], {}), '(times)\n', (10299, 10306), True, 'import numpy as np\n'), ((13106, 13131), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {... |
# -*- coding: utf-8 -*-
"""
===============================================================================
Horsager et al. (2009): Predicting temporal sensitivity
===============================================================================
This example shows how to use the
:py:class:`~pulse2percept.models.Horsager... | [
"pulse2percept.stimuli.BiphasicPulseTrain",
"pulse2percept.stimuli.MonophasicPulse",
"numpy.ceil",
"matplotlib.pyplot.xticks",
"matplotlib.pyplot.semilogx",
"matplotlib.pyplot.ylabel",
"numpy.arange",
"matplotlib.pyplot.xlabel",
"pulse2percept.models.Horsager2009Temporal",
"pulse2percept.stimuli.B... | [((862, 881), 'pulse2percept.datasets.load_horsager2009', 'load_horsager2009', ([], {}), '()\n', (879, 881), False, 'from pulse2percept.datasets import load_horsager2009\n'), ((1248, 1325), 'pulse2percept.datasets.load_horsager2009', 'load_horsager2009', ([], {'subjects': '"""S05"""', 'electrodes': '"""C3"""', 'stim_ty... |
#!/usr/bin/python3
import mysql.connector
from mysql.connector import Error
from mysql.connector import errorcode
from os import environ
import json
DB_HOST = environ.get('DB_HOST')
DB_NAME = environ.get('DB_NAME')
DB_USER = environ.get('DB_USER')
DB_PASSWORD = environ.get('DB_PASSWORD')
if DB_PASSWORD is not None:
... | [
"json.dumps",
"os.environ.get"
] | [((160, 182), 'os.environ.get', 'environ.get', (['"""DB_HOST"""'], {}), "('DB_HOST')\n", (171, 182), False, 'from os import environ\n'), ((193, 215), 'os.environ.get', 'environ.get', (['"""DB_NAME"""'], {}), "('DB_NAME')\n", (204, 215), False, 'from os import environ\n'), ((226, 248), 'os.environ.get', 'environ.get', (... |
from __future__ import print_function
import argparse
import glob
import io
import os
import subprocess as sp
import sys
from collections import defaultdict
from itertools import chain
import networkx as nx
import requests
import yaml
from conda_build import api
# ----------------------------------------------------... | [
"itertools.chain",
"conda_build.api.Config",
"io.open",
"networkx.dfs_successors",
"sys.exit",
"os.walk",
"os.path.exists",
"argparse.ArgumentParser",
"conda_build.api.get_output_file_paths",
"networkx.DiGraph",
"networkx.simple_cycles",
"subprocess.check_call",
"conda_build.api.render",
"... | [((1125, 1256), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {'description': '"""Identify and build all ggd recipes that are not currently in any ggd conda channel"""'}), "(description=\n 'Identify and build all ggd recipes that are not currently in any ggd conda channel'\n )\n", (1148, 1256), False... |
from django.test import TestCase
from corehq import toggles
from corehq.motech.dhis2.tasks import send_datasets_for_all_domains
class TestSendDatasetsForAllDomains(TestCase):
domain_name = 'does-not-exist'
def setUp(self):
toggles.DHIS2_INTEGRATION.set(
self.domain_name,
ena... | [
"corehq.toggles.DHIS2_INTEGRATION.set",
"corehq.motech.dhis2.tasks.send_datasets_for_all_domains"
] | [((244, 346), 'corehq.toggles.DHIS2_INTEGRATION.set', 'toggles.DHIS2_INTEGRATION.set', (['self.domain_name'], {'enabled': '(True)', 'namespace': 'toggles.NAMESPACE_DOMAIN'}), '(self.domain_name, enabled=True, namespace=\n toggles.NAMESPACE_DOMAIN)\n', (273, 346), False, 'from corehq import toggles\n'), ((421, 524), ... |
'''
@Author: fxm
@Date: Dec 27, 2020.
@Title: UI class.
'''
import sys
sys.path.append('..')
import pygame
import logging
from pygame.locals import *
import pygame.gfxdraw
from collections import namedtuple
from Framework.Net import dotdict
from Othello.Board import Board
log = logging.getLogger(__name__)
... | [
"logging.getLogger",
"pygame.mouse.get_pressed",
"pygame.draw.line",
"pygame.gfxdraw.aacircle",
"pygame.event.get",
"pygame.display.flip",
"Othello.Board.Board",
"pygame.mouse.get_pos",
"pygame.draw.rect",
"pygame.gfxdraw.filled_circle",
"sys.path.append",
"pygame.font.SysFont"
] | [((83, 104), 'sys.path.append', 'sys.path.append', (['""".."""'], {}), "('..')\n", (98, 104), False, 'import sys\n'), ((292, 319), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (309, 319), False, 'import logging\n'), ((2467, 2580), 'pygame.draw.rect', 'pygame.draw.rect', (['screen', 'BLA... |
# -*- coding: UTF-8 -*-
# main.py
# Root file for EmotionDetection program.
# Prints out command line menu and handles user choices
from __future__ import print_function
from EmotionDetection import WordMap
from EmotionDetection import EvaluateText
from EmotionDetection import GUI
try:
input = raw_input
except N... | [
"EmotionDetection.EvaluateText.evaluate",
"EmotionDetection.GUI.Evaluator",
"sys.setdefaultencoding",
"EmotionDetection.WordMap.buildWordMap"
] | [((363, 393), 'sys.setdefaultencoding', 'sys.setdefaultencoding', (['"""utf8"""'], {}), "('utf8')\n", (385, 393), False, 'import sys\n'), ((2684, 2699), 'EmotionDetection.GUI.Evaluator', 'GUI.Evaluator', ([], {}), '()\n', (2697, 2699), False, 'from EmotionDetection import GUI\n'), ((2544, 2575), 'EmotionDetection.Evalu... |
import os
def print_tables(tables, width=os.get_terminal_size().columns, spaces=3, index_width=3):
space = ' ' * spaces
lines = []
if index_width <= 0:
index_width = -1
tables.sort(key=lambda t: max(len(t), max([len(r) for r in t])), reverse=True)
for table in tables:
max_length = max([len(x) f... | [
"os.get_terminal_size"
] | [((43, 65), 'os.get_terminal_size', 'os.get_terminal_size', ([], {}), '()\n', (63, 65), False, 'import os\n')] |
import datetime as dt
from uploader.utils import NULL
from typing import List
__ESCAPE_SYMBOLS_MAPPING = {"'": r"''"}
def __value_empty(value) -> bool:
return value == NULL or value is None or not value or (isinstance(value, str) and value.isspace())
def __escaped_symbols() -> dict:
if not hasattr(__escape... | [
"datetime.datetime",
"datetime.datetime.strptime"
] | [((536, 570), 'datetime.datetime', 'dt.datetime', (['(1899)', '(12)', '(31)', '(0)', '(0)', '(0)'], {}), '(1899, 12, 31, 0, 0, 0)\n', (547, 570), True, 'import datetime as dt\n'), ((1196, 1231), 'datetime.datetime.strptime', 'dt.datetime.strptime', (['value', 'format'], {}), '(value, format)\n', (1216, 1231), True, 'im... |
import destination_search
'''
test용 값
gpsX = 127.0816985 # 동경 - 경도
gpsY = 37.5642135 # 북위 - 위도
time = 20
1. 좌표, 시간 입력 -> 주변 버스 정류소 ID 반환
2. 정류소 ID 입력 -> 버스 노선 획득
3. 획득한 버스 노선에서 출발지, 목적지 시간 계산하여 도착 버스 정류장 반환
'''
print("gps_x : ", end='')
gps_x = input()
print("gps_y : ", end='')
gps_y = input()
print("t... | [
"destination_search.DestinationStation"
] | [((429, 486), 'destination_search.DestinationStation', 'destination_search.DestinationStation', (['gps_x', 'gps_y', 'time'], {}), '(gps_x, gps_y, time)\n', (466, 486), False, 'import destination_search\n')] |
from torch.utils.data import Dataset
import cv2
class SegmentationDataSet(Dataset):
def __init__(self, image_paths, mask_paths, transform):
self.image_paths = image_paths
self.mask_paths = mask_paths
self.transforms = transform
def __len__(self):
return len(self.image_paths)
... | [
"cv2.imread",
"cv2.cvtColor"
] | [((600, 622), 'cv2.imread', 'cv2.imread', (['image_path'], {}), '(image_path)\n', (610, 622), False, 'import cv2\n'), ((639, 677), 'cv2.cvtColor', 'cv2.cvtColor', (['image', 'cv2.COLOR_BGR2RGB'], {}), '(image, cv2.COLOR_BGR2RGB)\n', (651, 677), False, 'import cv2\n'), ((700, 735), 'cv2.imread', 'cv2.imread', (['self.ma... |
# -*- coding: utf-8 -*-
from odoo import api, fields, models, _
from odoo.exceptions import UserError
from odoo.tools.misc import format_date
class AccrualAccountingWizard(models.TransientModel):
_name = 'account.accrual.accounting.wizard'
_description = 'Create accrual entry.'
date = fields.Date(require... | [
"odoo._",
"odoo.fields.Monetary",
"odoo.api.constrains",
"odoo.fields.Date",
"odoo.fields.Float",
"odoo.fields.Many2one",
"odoo.api.depends",
"odoo.tools.misc.format_date",
"odoo.fields.Selection",
"odoo.fields.Many2many"
] | [((301, 327), 'odoo.fields.Date', 'fields.Date', ([], {'required': '(True)'}), '(required=True)\n', (312, 327), False, 'from odoo import api, fields, models, _\n'), ((345, 390), 'odoo.fields.Many2one', 'fields.Many2one', (['"""res.company"""'], {'required': '(True)'}), "('res.company', required=True)\n", (360, 390), Fa... |
#!/usr/bin/env python3
# -*- coding: utf-8 -*-
"""
Created on Sun Mar 11 21:22:59 2018
@author: pami4
"""
#CUDA_VISIBLE_DEVICES=0 python
from pycocotools.coco import COCO
import coco
import numpy as np
from matplotlib import pyplot as plt
import visualize
import custom_utils
config = coco.CocoConfig()
config.GPU_CO... | [
"CustomDataGenerator.CustomDatasetIterator_MaskRCNN",
"CustomDataset.CocoDataset",
"numpy.sum",
"numpy.array",
"model.MaskRCNN",
"coco.CocoConfig",
"matplotlib.pyplot.show"
] | [((289, 306), 'coco.CocoConfig', 'coco.CocoConfig', ([], {}), '()\n', (304, 306), False, 'import coco\n'), ((361, 388), 'CustomDataset.CocoDataset', 'CustomDataset.CocoDataset', ([], {}), '()\n', (386, 388), False, 'import CustomDataset\n'), ((668, 778), 'CustomDataGenerator.CustomDatasetIterator_MaskRCNN', 'CustomData... |
from importlib.metadata import PackageNotFoundError, version
try:
__version__ = version("x_mlps")
except PackageNotFoundError:
# package is not installed
__version__ = None
| [
"importlib.metadata.version"
] | [((85, 102), 'importlib.metadata.version', 'version', (['"""x_mlps"""'], {}), "('x_mlps')\n", (92, 102), False, 'from importlib.metadata import PackageNotFoundError, version\n')] |
from django.db.models.signals import post_save
from django.dispatch import receiver
from company.models import Company
from company.tasks import deploy_new_company
@receiver(post_save, sender=Company)
def company_created(sender, instance, created, **kwargs):
if created:
deploy_new_company.delay(instance.... | [
"django.dispatch.receiver",
"company.tasks.deploy_new_company.delay"
] | [((168, 203), 'django.dispatch.receiver', 'receiver', (['post_save'], {'sender': 'Company'}), '(post_save, sender=Company)\n', (176, 203), False, 'from django.dispatch import receiver\n'), ((286, 323), 'company.tasks.deploy_new_company.delay', 'deploy_new_company.delay', (['instance.id'], {}), '(instance.id)\n', (310, ... |
# -*- coding:utf-8 -*-
from sc import SupercellSWF
import os
def findAllFile(basePath, endsWith):
for root, ds, fs in os.walk(basePath):
for f in fs:
if not f.startswith('.') and f.endswith(endsWith):
fullname = os.path.join(root, f)
yield fullname
| [
"os.path.join",
"os.walk"
] | [((123, 140), 'os.walk', 'os.walk', (['basePath'], {}), '(basePath)\n', (130, 140), False, 'import os\n'), ((253, 274), 'os.path.join', 'os.path.join', (['root', 'f'], {}), '(root, f)\n', (265, 274), False, 'import os\n')] |
#####################################################
# #
# Source file of the MadFlow plugin #
# Use only with consent of its authors. #
# #
# authors: S.Carrazza, J.Cruz-Martinez, ... | [
"os.path.realpath",
"sys.path.insert"
] | [((638, 667), 'sys.path.insert', 'sys.path.insert', (['(0)', 'root_path'], {}), '(0, root_path)\n', (653, 667), False, 'import sys\n'), ((604, 630), 'os.path.realpath', 'os.path.realpath', (['__file__'], {}), '(__file__)\n', (620, 630), False, 'import os\n')] |
import abc
import time
from datetime import datetime
from pathlib import Path
from typing import Any, Dict, List, Optional, cast
from loguru import logger
from rich import print
from dubdub import (
Binary,
Grouping,
Literal,
Node,
Token,
TokenType,
Unary,
Visitor,
dataclass,
)
fro... | [
"loguru.logger.debug",
"dubdub.parser.Parser",
"pathlib.Path.cwd",
"rich.print",
"dubdub.scanner.Scanner",
"dubdub.env.Environment",
"typing.cast"
] | [((533, 543), 'pathlib.Path.cwd', 'Path.cwd', ([], {}), '()\n', (541, 543), False, 'from pathlib import Path\n'), ((660, 673), 'dubdub.env.Environment', 'Environment', ([], {}), '()\n', (671, 673), False, 'from dubdub.env import Environment\n'), ((7550, 7579), 'dubdub.scanner.Scanner', 'Scanner', ([], {'source': 'neste... |
from pdpyras import APISession
from requests.sessions import session
from kafka import KafkaConsumer
import json
import sys
print("Starting the alerts listener")
# Kafka configurations
kafka_broker = 'kafka:9092'
alerts_topic = "high-temp-alerts-u5-1631586083-13100517045407420899"
# PageDuty configurations
api_token... | [
"sys.exc_info",
"json.dumps"
] | [((1833, 1852), 'json.dumps', 'json.dumps', (['payload'], {}), '(payload)\n', (1843, 1852), False, 'import json\n'), ((1925, 1939), 'sys.exc_info', 'sys.exc_info', ([], {}), '()\n', (1937, 1939), False, 'import sys\n')] |
from decimal import Decimal
from src import profitable_music_genre
EXPECTED_RESULT = [
(Decimal('826.65'), 'Rock'),
(Decimal('382.14'), 'Latin'),
(Decimal('261.36'), 'Metal'),
(Decimal('241.56'), 'Alternative & Punk'),
(Decimal('93.53'), 'TV Shows'),
(Decimal('79.20'), 'Jazz'),
(Decimal('6... | [
"src.profitable_music_genre.most_profitable_music_genre",
"decimal.Decimal"
] | [((94, 111), 'decimal.Decimal', 'Decimal', (['"""826.65"""'], {}), "('826.65')\n", (101, 111), False, 'from decimal import Decimal\n'), ((127, 144), 'decimal.Decimal', 'Decimal', (['"""382.14"""'], {}), "('382.14')\n", (134, 144), False, 'from decimal import Decimal\n'), ((161, 178), 'decimal.Decimal', 'Decimal', (['""... |
import multiprocessing as mp
import os
import queue
import signal
import torch
class BaseCallback:
def __init__(self, verbose: int = 0):
self.model = None
self.verbose = verbose
def init_callback(self, model):
self.model = model
def on_training_start(self):
pass
def ... | [
"multiprocessing.Queue",
"os.kill",
"multiprocessing.Event"
] | [((1120, 1130), 'multiprocessing.Queue', 'mp.Queue', ([], {}), '()\n', (1128, 1130), True, 'import multiprocessing as mp\n'), ((1165, 1175), 'multiprocessing.Queue', 'mp.Queue', ([], {}), '()\n', (1173, 1175), True, 'import multiprocessing as mp\n'), ((1209, 1219), 'multiprocessing.Queue', 'mp.Queue', ([], {}), '()\n',... |
import os
import subprocess
import tempfile
import time
import wave
import simpleaudio
def extract(file):
ptmv_tempdir = os.path.join(tempfile.gettempdir(), "ptmv")
if not os.path.exists(ptmv_tempdir): os.makedirs(ptmv_tempdir)
snd_file = ptmv_tempdir + str(int(time.time())) + ".wav"
command = "ffmpeg -i " + file ... | [
"os.path.exists",
"wave.open",
"simpleaudio.WaveObject.from_wave_read",
"os.makedirs",
"tempfile.gettempdir",
"time.time"
] | [((538, 553), 'wave.open', 'wave.open', (['file'], {}), '(file)\n', (547, 553), False, 'import wave\n'), ((136, 157), 'tempfile.gettempdir', 'tempfile.gettempdir', ([], {}), '()\n', (155, 157), False, 'import tempfile\n'), ((175, 203), 'os.path.exists', 'os.path.exists', (['ptmv_tempdir'], {}), '(ptmv_tempdir)\n', (189... |
# -*- coding: utf-8 -*-
"""
Created on Thu Dec 30 11:53:32 2021
@author: HP
"""
PATH = "Forest.png"
from PIL import Image
sprites = [Image.new("RGBA", (32, 32), (0,0,0,0)) for x in range(8*6)]
current = 0
with Image.open(PATH) as im:
curI = 0
curJ = 0
for curI in range(0,256,32):
for curJ in range(... | [
"PIL.Image.new",
"PIL.Image.open"
] | [((134, 175), 'PIL.Image.new', 'Image.new', (['"""RGBA"""', '(32, 32)', '(0, 0, 0, 0)'], {}), "('RGBA', (32, 32), (0, 0, 0, 0))\n", (143, 175), False, 'from PIL import Image\n'), ((211, 227), 'PIL.Image.open', 'Image.open', (['PATH'], {}), '(PATH)\n', (221, 227), False, 'from PIL import Image\n')] |
from minibench import Benchmark
import time
class PauseBenchmark(Benchmark):
times = 10
def bench_one_hundredth(self):
time.sleep(.01)
def bench_one_tenth(self):
time.sleep(.1)
| [
"time.sleep"
] | [((139, 155), 'time.sleep', 'time.sleep', (['(0.01)'], {}), '(0.01)\n', (149, 155), False, 'import time\n'), ((195, 210), 'time.sleep', 'time.sleep', (['(0.1)'], {}), '(0.1)\n', (205, 210), False, 'import time\n')] |
# Check http://piwigo.com/
from urllib.parse import urlparse
from os import path
from plugins.cloudimport.cloud_platform import File, Folder
from plugins.cloudimport.extensions.cloud_library import CloudLibrary
class Platform(CloudLibrary):
def __init__(self):
super().__init__('Piwigo', 'http://{server_url... | [
"urllib.parse.urlparse",
"plugins.cloudimport.cloud_platform.Folder",
"os.path.splitext",
"plugins.cloudimport.cloud_platform.File",
"os.path.index"
] | [((994, 1007), 'urllib.parse.urlparse', 'urlparse', (['url'], {}), '(url)\n', (1002, 1007), False, 'from urllib.parse import urlparse\n'), ((1926, 1984), 'plugins.cloudimport.cloud_platform.Folder', 'Folder', (["result['name']", "result['url']", "result['nb_images']"], {}), "(result['name'], result['url'], result['nb_i... |
"""
<Reinforcement Learning and Control>(Year 2020)
by <NAME>
@ Intelligent Driving Lab, Tsinghua University
OCP example for lane keeping problem in a circle road
[Method]
Model predictive control
"""
from casadi import *
from config import DynamicsConfig
import math
from dynamics import... | [
"dynamics.VehicleDynamics"
] | [((1087, 1104), 'dynamics.VehicleDynamics', 'VehicleDynamics', ([], {}), '()\n', (1102, 1104), False, 'from dynamics import VehicleDynamics\n')] |
#!/usr/bin/env python
"""Create benchmark for k nearest neighbor on unit sphere in R^k."""
# Scroll down to line 90 to "Adjust this" to add your experiment
import random
import numpy as np
import os.path
import logging
import sys
import Queue as queue
import h5py
import time
logging.basicConfig(format='%(asctime)s %... | [
"logging.basicConfig",
"random.uniform",
"Queue.PriorityQueue",
"time.time",
"h5py.File",
"numpy.array",
"numpy.zeros",
"numpy.linalg.norm",
"logging.info"
] | [((279, 391), 'logging.basicConfig', 'logging.basicConfig', ([], {'format': '"""%(asctime)s %(levelname)s %(message)s"""', 'level': 'logging.DEBUG', 'stream': 'sys.stdout'}), "(format='%(asctime)s %(levelname)s %(message)s', level=\n logging.DEBUG, stream=sys.stdout)\n", (298, 391), False, 'import logging\n'), ((407... |
from setuptools import setup
setup(name='pyrosettacolabsetup',
version='0.5',
description='Mounts Google Drive for PyRosetta use in Google Colaboratory',
url='https://github.com/kathyle9/pyrosettacolabsetup',
author='kathyle9',
author_email='<EMAIL>',
license='MIT',
packages=[... | [
"setuptools.setup"
] | [((30, 335), 'setuptools.setup', 'setup', ([], {'name': '"""pyrosettacolabsetup"""', 'version': '"""0.5"""', 'description': '"""Mounts Google Drive for PyRosetta use in Google Colaboratory"""', 'url': '"""https://github.com/kathyle9/pyrosettacolabsetup"""', 'author': '"""kathyle9"""', 'author_email': '"""<EMAIL>"""', '... |
"""The pyccl package contains all of the submodules that are implemented in
individual files in CCL.
"""
# flake8: noqa
from pkg_resources import get_distribution, DistributionNotFound
try:
__version__ = get_distribution(__name__).version
except DistributionNotFound:
# package is not installed
pass
# Sets ... | [
"os.path.abspath",
"os.environ.get",
"pkg_resources.get_distribution"
] | [((428, 457), 'os.environ.get', 'environ.get', (['"""CCL_PARAM_FILE"""'], {}), "('CCL_PARAM_FILE')\n", (439, 457), False, 'from os import environ, path\n'), ((570, 600), 'os.environ.get', 'environ.get', (['"""CLASS_PARAM_DIR"""'], {}), "('CLASS_PARAM_DIR')\n", (581, 600), False, 'from os import environ, path\n'), ((208... |
# coding: utf-8
# Copyright (c) 2016, 2022, Oracle and/or its affiliates. All rights reserved.
# This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may c... | [
"oci.util.formatted_flat_dict",
"oci.util.value_allowed_none_or_none_sentinel"
] | [((35018, 35043), 'oci.util.formatted_flat_dict', 'formatted_flat_dict', (['self'], {}), '(self)\n', (35037, 35043), False, 'from oci.util import formatted_flat_dict, NONE_SENTINEL, value_allowed_none_or_none_sentinel\n'), ((19160, 19219), 'oci.util.value_allowed_none_or_none_sentinel', 'value_allowed_none_or_none_sent... |
#!/usr/bin/env python
# pylint: disable=disallowed-name
# pylint: disable=missing-class-docstring,
# pylint: disable=missing-function-docstring,
# pylint: disable=no-self-use
import unittest
from paramobject import ParametrizedObject, parameter, Parameter
class TestParametrizedObject(unittest.TestCase):
def te... | [
"unittest.main",
"paramobject.Parameter",
"paramobject.parameter"
] | [((4314, 4329), 'unittest.main', 'unittest.main', ([], {}), '()\n', (4327, 4329), False, 'import unittest\n'), ((413, 434), 'paramobject.Parameter', 'Parameter', ([], {'default': '(42)'}), '(default=42)\n', (422, 434), False, 'from paramobject import ParametrizedObject, parameter, Parameter\n'), ((453, 464), 'paramobje... |
"""
Copyright 2017-2020 Government of Canada - Public Services and Procurement Canada - buyandsell.gc.ca
Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
... | [
"logging.basicConfig",
"logging.getLogger",
"os.makedirs",
"os.path.join",
"von_anchor.frill.inis2dict",
"os.path.realpath",
"app.cache.MEM_CACHE.delete",
"app.cache.MEM_CACHE.get",
"app.cache.MEM_CACHE.set"
] | [((1022, 1054), 'os.makedirs', 'makedirs', (['dir_log'], {'exist_ok': '(True)'}), '(dir_log, exist_ok=True)\n', (1030, 1054), False, 'from os import makedirs\n'), ((1070, 1100), 'os.path.join', 'join', (['dir_log', '"""von_tails.log"""'], {}), "(dir_log, 'von_tails.log')\n", (1074, 1100), False, 'from os.path import di... |
from django.contrib import admin
from .models import Comment
class CommentAdmin(admin.ModelAdmin):
list_display = ("article", "to_comment", "status", "owner", "content")
admin.site.register(Comment, CommentAdmin)
| [
"django.contrib.admin.site.register"
] | [((178, 220), 'django.contrib.admin.site.register', 'admin.site.register', (['Comment', 'CommentAdmin'], {}), '(Comment, CommentAdmin)\n', (197, 220), False, 'from django.contrib import admin\n')] |
from flask import Flask
import os
from application.model.entity.aula import Aula
from application.model.entity.disciplina import Disciplina
app = Flask(__name__, static_folder=os.path.abspath("application/view/static"), template_folder=os.path.abspath("application/view/templates"))
aula1 = Aula(1, "Aula 1"... | [
"application.model.entity.aula.Aula",
"application.model.entity.disciplina.Disciplina",
"os.path.abspath"
] | [((304, 369), 'application.model.entity.aula.Aula', 'Aula', (['(1)', '"""Aula 1"""', '"""Introdução ao Linux"""', '"""Sistemas Operacionais"""'], {}), "(1, 'Aula 1', 'Introdução ao Linux', 'Sistemas Operacionais')\n", (308, 369), False, 'from application.model.entity.aula import Aula\n'), ((379, 464), 'application.mode... |
from typing import Callable
from putput.presets import displaCy
from putput.presets import iob2
from putput.presets import luis
from putput.presets import stochastic
def get_preset(preset: str) -> Callable:
"""A factory that gets a 'preset' Callable.
Args:
preset: the preset's name.
Returns:
... | [
"putput.presets.displaCy.preset",
"putput.presets.luis.preset",
"putput.presets.stochastic.preset",
"putput.presets.iob2.preset"
] | [((1513, 1526), 'putput.presets.iob2.preset', 'iob2.preset', ([], {}), '()\n', (1524, 1526), False, 'from putput.presets import iob2\n'), ((1571, 1588), 'putput.presets.displaCy.preset', 'displaCy.preset', ([], {}), '()\n', (1586, 1588), False, 'from putput.presets import displaCy\n'), ((1629, 1642), 'putput.presets.lu... |
from hqca.core import *
import numpy as np
from hqca.tools import *
class SingleQubitHamiltonian(Hamiltonian):
def __init__(self,sq=True,
**kw
):
self._order = 1
self._model = 'sq'
self._qubOp = ''
self.No_tot = 1
self.Ne_tot = 1
self.real = T... | [
"numpy.array",
"numpy.zeros",
"numpy.linalg.eigvalsh"
] | [((953, 988), 'numpy.zeros', 'np.zeros', (['(2, 2)'], {'dtype': 'np.complex_'}), '((2, 2), dtype=np.complex_)\n', (961, 988), True, 'import numpy as np\n'), ((1300, 1315), 'numpy.array', 'np.array', (['[mat]'], {}), '([mat])\n', (1308, 1315), True, 'import numpy as np\n'), ((1252, 1275), 'numpy.linalg.eigvalsh', 'np.li... |
# Author: <NAME>, <EMAIL>
# Dec 02, 2020
# Copyright 2020 <NAME>
import numpy as np
import pandas as pd
from matplotlib import pyplot as plt
from scipy.spatial import distance as dist
import scipy.io
import pickle
## TAPTC Dataset
group_list = [1,2]
instance_list = [0, 1, 2]
ratio_deadline_list = [1, 2, 3, 4]
rob... | [
"pandas.read_csv"
] | [((705, 761), 'pandas.read_csv', 'pd.read_csv', (['file_name'], {'sep': '""" """', 'header': 'None', 'skiprows': '(1)'}), "(file_name, sep=' ', header=None, skiprows=1)\n", (716, 761), True, 'import pandas as pd\n'), ((901, 957), 'pandas.read_csv', 'pd.read_csv', (['file_name'], {'sep': '""" """', 'header': 'None', 'sk... |
#!/usr/bin/python2
import os
from plumbum import local, FG
from plumbum.cmd import git
# the commits already tested
HISTORY = '/home/cmr/benches/data'
BUILDDIR = '/mnt/rustb'
BENCH_OVERRIDE = '/home/cmr/benches/bench-override.txt'
def run(hash):
local['benchit.py'][hash] & FG
for hash in open(BENCH_OVERRIDE).... | [
"os.listdir",
"os.path.join"
] | [((413, 433), 'os.listdir', 'os.listdir', (['BUILDDIR'], {}), '(BUILDDIR)\n', (423, 433), False, 'import os\n'), ((461, 485), 'os.path.join', 'os.path.join', (['HISTORY', 'd'], {}), '(HISTORY, d)\n', (473, 485), False, 'import os\n')] |
import RPi.GPIO as GPIO
import time
import datetime
from ReadWriteConfig import *
import Adafruit_ADS1x15
now = datetime.datetime.now()
print("Starting CheckWater.py", str(now))
adc = Adafruit_ADS1x15.ADS1015() # Pick Sensors
GAIN = 0 #import gain for adc reading
SS_COUNT = 0
with open("PlantMgr.xml", "r") as f:
... | [
"RPi.GPIO.cleanup",
"RPi.GPIO.setup",
"RPi.GPIO.output",
"time.sleep",
"RPi.GPIO.PWM",
"datetime.datetime.now",
"Adafruit_ADS1x15.ADS1015",
"RPi.GPIO.setmode"
] | [((113, 136), 'datetime.datetime.now', 'datetime.datetime.now', ([], {}), '()\n', (134, 136), False, 'import datetime\n'), ((187, 213), 'Adafruit_ADS1x15.ADS1015', 'Adafruit_ADS1x15.ADS1015', ([], {}), '()\n', (211, 213), False, 'import Adafruit_ADS1x15\n'), ((671, 693), 'RPi.GPIO.setmode', 'GPIO.setmode', (['GPIO.BCM'... |
import discord
import ctftime
import os
import random
from discord.ext import commands, tasks
from datetime import datetime
# Token generated from https://discord.com/developers/applications
# Keep this private, if exposed generate new one
TOKEN = ''
# Bot channel ID was grabbed from Settings > Appearance > Developer... | [
"random.choice",
"discord.ext.commands.Bot",
"datetime.datetime.now",
"ctftime.get_events",
"discord.ext.tasks.loop",
"ctftime.days_to_secs"
] | [((408, 440), 'discord.ext.commands.Bot', 'commands.Bot', ([], {'command_prefix': '"""!"""'}), "(command_prefix='!')\n", (420, 440), False, 'from discord.ext import commands, tasks\n'), ((5482, 5504), 'discord.ext.tasks.loop', 'tasks.loop', ([], {'minutes': '(30)'}), '(minutes=30)\n', (5492, 5504), False, 'from discord... |
import os
import sys
NAME = 'multipla'
PACKAGE = __import__(NAME)
AUTHOR, EMAIL = PACKAGE.__author__.rsplit(' ', 1)
with open('docs/index.rst', 'r') as INDEX:
DESCRIPTION = INDEX.readline()
with open('README.rst', 'r') as README:
LONG_DESCRIPTION = README.read()
URL = 'https://github.com/monkeython/%s' % NA... | [
"setuptools.setup"
] | [((796, 819), 'setuptools.setup', 'setuptools.setup', ([], {}), '(**EGG)\n', (812, 819), False, 'import setuptools\n')] |
# Generic-DiscordBot
# author: github/adibarra
# imports
import os
import time
import uuid
import enum
import glob
import traceback
from zipfile import ZipFile
from DB_prefsloader import PreferenceLoader
class Importance(enum.IntEnum):
""" Enum to keep track of logger message importance """
CRIT = 0
WARN... | [
"traceback.format_exc",
"time.strftime",
"os.path.isfile",
"os.path.realpath",
"os.umask",
"os.stat"
] | [((1689, 1713), 'os.path.isfile', 'os.path.isfile', (['filePath'], {}), '(filePath)\n', (1703, 1713), False, 'import os\n'), ((1196, 1207), 'os.umask', 'os.umask', (['(0)'], {}), '(0)\n', (1204, 1207), False, 'import os\n'), ((1316, 1340), 'os.umask', 'os.umask', (['original_umask'], {}), '(original_umask)\n', (1324, 1... |
from django.urls import path
from django.conf.urls import url
from store import views
from .views import getStore, addStore, updateStore, deleteStore
urlpatterns = [
url(r'getStore/', views.getStore),
url(r'addStore/', views.addStore),
url(r'updateStore/', views.updateStore),
url(r'deleteStore/', views... | [
"django.conf.urls.url"
] | [((171, 203), 'django.conf.urls.url', 'url', (['"""getStore/"""', 'views.getStore'], {}), "('getStore/', views.getStore)\n", (174, 203), False, 'from django.conf.urls import url\n'), ((210, 242), 'django.conf.urls.url', 'url', (['"""addStore/"""', 'views.addStore'], {}), "('addStore/', views.addStore)\n", (213, 242), F... |
# -*- coding: utf-8 -*-
"""
MIT License
Copyright (c) 2020 <NAME>, SE; tamalone1
"""
import unittest
from PyNite import FEModel3D
import sys
from io import StringIO
class Test_Spring_Elements(unittest.TestCase):
''' Tests of spring members.'''
def setUp(self):
# Suppress printed output temporarily
... | [
"io.StringIO",
"PyNite.FEModel3D"
] | [((340, 350), 'io.StringIO', 'StringIO', ([], {}), '()\n', (348, 350), False, 'from io import StringIO\n'), ((679, 690), 'PyNite.FEModel3D', 'FEModel3D', ([], {}), '()\n', (688, 690), False, 'from PyNite import FEModel3D\n')] |
"""
Likelihood maximization script. This program is designed to be entirely separable from ATESA in that it can be called
manually to perform likelihood maximization to user specifications and with arbitrary input files; however, it is
required by ATESA's aimless shooting information error convergence criterion.
"""
i... | [
"scipy.stats.linregress",
"numpy.sqrt",
"matplotlib.pyplot.ylabel",
"math.floor",
"numpy.array",
"argparse.Namespace",
"os.path.exists",
"numpy.histogram",
"numpy.mean",
"argparse.ArgumentParser",
"matplotlib.pyplot.xlabel",
"numpy.asarray",
"numpy.max",
"numpy.min",
"sys.stdout.flush",
... | [((2677, 2699), 'sys.stdout.write', 'sys.stdout.write', (['text'], {}), '(text)\n', (2693, 2699), False, 'import sys\n'), ((2704, 2722), 'sys.stdout.flush', 'sys.stdout.flush', ([], {}), '()\n', (2720, 2722), False, 'import sys\n'), ((26829, 26904), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {'descripti... |
#!/usr/bin/env python
# -*- coding: utf-8 -*-
# Splunk specific dependencies
import sys, os
from splunklib.searchcommands import dispatch, GeneratingCommand, Configuration, Option, validators, splunklib_logger as logger
# Command specific dependencies
import requests
from requests.auth import HTTPDigestAuth
import js... | [
"os.environ.keys",
"requests.post",
"splunklib.searchcommands.validators.Boolean",
"requests.get",
"splunklib.searchcommands.dispatch",
"splunklib.searchcommands.validators.Integer",
"splunklib.searchcommands.Option",
"splunklib.searchcommands.Configuration"
] | [((565, 596), 'splunklib.searchcommands.Configuration', 'Configuration', ([], {'type': '"""reporting"""'}), "(type='reporting')\n", (578, 596), False, 'from splunklib.searchcommands import dispatch, GeneratingCommand, Configuration, Option, validators, splunklib_logger as logger\n'), ((4507, 4571), 'splunklib.searchcom... |
"""Tests for RandoPony admin views and functionality.
"""
from datetime import datetime
import unittest
from unittest.mock import patch
from pyramid import testing
from pyramid_mailer import get_mailer
from sqlalchemy import create_engine
from randopony.models.meta import (
Base,
DBSession,
)
class TestCore... | [
"pyramid.testing.setUp",
"pyramid_mailer.get_mailer",
"randopony.views.admin.core.email_to_organizer",
"randopony.models.meta.Base.metadata.create_all",
"datetime.datetime",
"randopony.models.Brevet.get_current",
"pyramid.testing.tearDown",
"sqlalchemy.create_engine",
"randopony.views.admin.core.ema... | [((670, 685), 'pyramid.testing.setUp', 'testing.setUp', ([], {}), '()\n', (683, 685), False, 'from pyramid import testing\n'), ((703, 729), 'sqlalchemy.create_engine', 'create_engine', (['"""sqlite://"""'], {}), "('sqlite://')\n", (716, 729), False, 'from sqlalchemy import create_engine\n'), ((738, 770), 'randopony.mod... |
"""
WSGI entrypoint.
"""
from nunaserver.server import app
if __name__ == "__main__":
app.run()
| [
"nunaserver.server.app.run"
] | [((91, 100), 'nunaserver.server.app.run', 'app.run', ([], {}), '()\n', (98, 100), False, 'from nunaserver.server import app\n')] |
import os
import pytest
import csv_diff
import logging
import torch
from unit_tests.t_utils import remove_tmp_dir, create_tmp_dir, __data_testing_dir__, __tmp_dir__
from ivadomed.loader import utils as imed_loader_utils
from ivadomed.loader import loader as imed_loader
logger = logging.getLogger(__name__)
def setup_f... | [
"logging.getLogger",
"torch.unique",
"ivadomed.loader.loader.dropout_input",
"os.path.join",
"pytest.mark.parametrize",
"unit_tests.t_utils.remove_tmp_dir",
"torch.tensor",
"ivadomed.loader.utils.BidsDataframe",
"torch.rand",
"unit_tests.t_utils.create_tmp_dir"
] | [((279, 306), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (296, 306), False, 'import logging\n'), ((1820, 2081), 'pytest.mark.parametrize', 'pytest.mark.parametrize', (['"""loader_parameters"""', "[{'path_data': [__data_testing_dir__], 'target_suffix': ['_seg-manual'],\n 'extensions... |
from pymongo import MongoClient
from bson.objectid import ObjectId
from datetime import datetime as dt
import pprint
client = MongoClient()
db = client['mongo_db_lab']
defs = db['definitions']
def random_word_requester():
'''
This function should return a random word and its definition and also
log in the... | [
"pymongo.MongoClient",
"pprint.pprint",
"datetime.datetime.utcnow"
] | [((126, 139), 'pymongo.MongoClient', 'MongoClient', ([], {}), '()\n', (137, 139), False, 'from pymongo import MongoClient\n'), ((1150, 1170), 'pprint.pprint', 'pprint.pprint', (['duped'], {}), '(duped)\n', (1163, 1170), False, 'import pprint\n'), ((628, 639), 'datetime.datetime.utcnow', 'dt.utcnow', ([], {}), '()\n', (... |
# Generated by Django 2.1.7 on 2019-05-20 19:58
from django.db import migrations, models
class Migration(migrations.Migration):
dependencies = [
('leads', '0011_auto_20190520_1217'),
]
operations = [
migrations.AddField(
model_name='lead',
name='first_name',
... | [
"django.db.models.CharField"
] | [((334, 404), 'django.db.models.CharField', 'models.CharField', ([], {'max_length': '(255)', 'null': '(True)', 'verbose_name': '"""First name"""'}), "(max_length=255, null=True, verbose_name='First name')\n", (350, 404), False, 'from django.db import migrations, models\n'), ((525, 594), 'django.db.models.CharField', 'm... |
import sys
import torch
from args import get_argparser, parse_args, get_aligner, get_bbox
from os.path import join
if __name__ == '__main__':
parser = get_argparser()
parser.add_argument('--align_start',
help='align without vector voting the 2nd & 3rd sections, otherwise copy them', action='store_true')
args... | [
"args.get_argparser",
"args.get_bbox",
"os.path.join",
"args.get_aligner",
"args.parse_args"
] | [((154, 169), 'args.get_argparser', 'get_argparser', ([], {}), '()\n', (167, 169), False, 'from args import get_argparser, parse_args, get_aligner, get_bbox\n'), ((323, 341), 'args.parse_args', 'parse_args', (['parser'], {}), '(parser)\n', (333, 341), False, 'from args import get_argparser, parse_args, get_aligner, get... |
from cms.extensions.toolbar import ExtensionToolbar
from cms.utils import get_language_list
from django.utils.encoding import force_text
from django.utils.translation import get_language_info
class TitleExtensionToolbar(ExtensionToolbar):
model = None
insert_after = None
def get_item_position(self, menu)... | [
"django.utils.translation.get_language_info",
"django.utils.encoding.force_text",
"cms.utils.get_language_list"
] | [((896, 935), 'cms.utils.get_language_list', 'get_language_list', (['self.current_site.pk'], {}), '(self.current_site.pk)\n', (913, 935), False, 'from cms.utils import get_language_list\n'), ((1940, 1973), 'django.utils.translation.get_language_info', 'get_language_info', (['title.language'], {}), '(title.language)\n',... |
import datetime, os, pkg_resources, re, setuptools_scm
from .. import __name__ as package_name
try:
if int(os.environ.get("_ASTROPATH_VERSION_NO_GIT", 0)):
env_var_no_git = True
raise LookupError
env_var_no_git = False
astropathversion = "v"+setuptools_scm.get_version(root="../..", relative_to=__file__)
... | [
"re.match",
"os.environ.get",
"setuptools_scm.get_version",
"datetime.date.today",
"pkg_resources.get_distribution"
] | [((586, 716), 're.match', 're.match', (['"""v(?P<version>[0-9]+(?:\\\\.[0-9]+)*)(?P<dev>\\\\.dev[0-9]+\\\\+g[0-9a-f]+)?(?P<date>\\\\.d[0-9]+)?"""', 'astropathversion'], {}), "(\n 'v(?P<version>[0-9]+(?:\\\\.[0-9]+)*)(?P<dev>\\\\.dev[0-9]+\\\\+g[0-9a-f]+)?(?P<date>\\\\.d[0-9]+)?'\n , astropathversion)\n", (594, 71... |
###
### Precios de casas en boston
###
from sklearn.datasets import load_boston
from sklearn.model_selection import train_test_split
from sklearn.linear_model import LinearRegression, Ridge
from sklearn.neighbors import KNeighborsRegressor
boston = load_boston()
# Visualiza el nombre de los valores de los datos.
pri... | [
"sklearn.model_selection.train_test_split",
"sklearn.neighbors.KNeighborsRegressor",
"sklearn.datasets.load_boston",
"sklearn.linear_model.Ridge",
"sklearn.linear_model.LinearRegression"
] | [((251, 264), 'sklearn.datasets.load_boston', 'load_boston', ([], {}), '()\n', (262, 264), False, 'from sklearn.datasets import load_boston\n'), ((440, 484), 'sklearn.model_selection.train_test_split', 'train_test_split', (['boston.data', 'boston.target'], {}), '(boston.data, boston.target)\n', (456, 484), False, 'from... |
"""Ui."""
import logging
import logging.config
import os.path
from datetime import datetime
from PyQt5 import QtCore, QtGui
from .lib import (
EquityChart,
OptimizatimizedResultsTable,
OptimizationTable,
Portfolio,
QuotesChart,
ResultsTable,
Settings,
Symbol,
TradesTable,
get_... | [
"logging.getLogger",
"PyQt5.QtCore.QDir.currentPath",
"PyQt5.QtCore.pyqtSignal",
"PyQt5.QtGui.QPushButton",
"PyQt5.QtGui.QStandardItemModel",
"PyQt5.QtGui.QStandardItem",
"PyQt5.QtGui.QHBoxLayout",
"PyQt5.QtGui.QTableView",
"PyQt5.QtGui.QLabel",
"PyQt5.QtGui.QFormLayout",
"PyQt5.QtGui.QVBoxLayou... | [((411, 438), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (428, 438), False, 'import logging\n'), ((576, 601), 'PyQt5.QtCore.pyqtSignal', 'QtCore.pyqtSignal', (['object'], {}), '(object)\n', (593, 601), False, 'from PyQt5 import QtCore, QtGui\n'), ((777, 802), 'PyQt5.QtCore.pyqtSignal'... |
import os
import shutil
import re
from collections import OrderedDict
import subprocess
import numpy as np
import atexit
class Result:
checkpoint = None
log = None
tarball = None
board = None
if __name__ == '__main__':
results = OrderedDict()
def load_files():
files = os.listdir()
... | [
"collections.OrderedDict",
"os.listdir",
"os.path.getsize",
"numpy.unique",
"subprocess.Popen",
"re.match",
"os.path.join",
"os.path.isdir",
"re.finditer",
"shutil.rmtree",
"os.path.islink",
"os.walk",
"os.remove"
] | [((251, 264), 'collections.OrderedDict', 'OrderedDict', ([], {}), '()\n', (262, 264), False, 'from collections import OrderedDict\n'), ((304, 316), 'os.listdir', 'os.listdir', ([], {}), '()\n', (314, 316), False, 'import os\n'), ((1110, 1131), 'os.path.islink', 'os.path.islink', (['start'], {}), '(start)\n', (1124, 113... |
import os
import time
import logging
from typing import List, Dict
from collections import deque
# from search import app
from elasticsearch import Elasticsearch
from elasticsearch.helpers import parallel_bulk
from datetime import datetime
# from config import ELASTICSEARCH_URL
import threading
from .utils import Uti... | [
"logging.debug",
"elasticsearch.Elasticsearch",
"elasticsearch.helpers.parallel_bulk",
"os.environ.get",
"cueSearch.services.GlobalDimensionServices.nonGlobalDimensionForIndexing",
"time.time",
"threading.Thread",
"logging.info",
"logging.error"
] | [((362, 423), 'os.environ.get', 'os.environ.get', (['"""ELASTICSEARCH_URL"""', '"""http://localhost:9200/"""'], {}), "('ELASTICSEARCH_URL', 'http://localhost:9200/')\n", (376, 423), False, 'import os\n'), ((1400, 1441), 'elasticsearch.Elasticsearch', 'Elasticsearch', ([], {'hosts': '[esHost]', 'timeout': '(30)'}), '(ho... |
"""
A simple example of NATDiscovery between ByteBlower ports.
To discover the public IP address we will send a single packet
upstream, capture this packet at the WAN side and finally
pick it apart.
This example demonstrates:
* How to transmit a single custom packet.
* How to cap... | [
"byteblowerll.byteblower.ByteBlower.InstanceGet",
"time.sleep"
] | [((979, 1014), 'byteblowerll.byteblower.ByteBlower.InstanceGet', 'byteblower.ByteBlower.InstanceGet', ([], {}), '()\n', (1012, 1014), True, 'import byteblowerll.byteblower as byteblower\n'), ((2387, 2403), 'time.sleep', 'time.sleep', (['(0.01)'], {}), '(0.01)\n', (2397, 2403), False, 'import time\n')] |
from rest_framework import serializers
from game.serializers.question_serializer import QuestionSerializer
class QuestionWithAnswerSerializer(QuestionSerializer):
correct_answer = serializers.CharField()
| [
"rest_framework.serializers.CharField"
] | [((187, 210), 'rest_framework.serializers.CharField', 'serializers.CharField', ([], {}), '()\n', (208, 210), False, 'from rest_framework import serializers\n')] |
from sklearn import svm
from ..data_wrappers import reject
import numpy as np
from scipy.stats import multivariate_normal
from sklearn.mixture import GMM
from sklearn.neighbors import KernelDensity
class DensityEstimators(object):
def __init__(self):
self.models = {}
self.unknown = {}
self... | [
"numpy.sqrt",
"numpy.linalg.pinv",
"scipy.stats.multivariate_normal",
"numpy.array",
"numpy.cov",
"numpy.arange",
"numpy.rank",
"sklearn.neighbors.KernelDensity",
"numpy.subtract",
"numpy.dot",
"numpy.vstack",
"numpy.eye",
"numpy.random.multivariate_normal",
"numpy.alen",
"sklearn.svm.SV... | [((604, 629), 'sklearn.svm.SVC', 'svm.SVC', ([], {'probability': '(True)'}), '(probability=True)\n', (611, 629), False, 'from sklearn import svm\n'), ((709, 734), 'numpy.vstack', 'np.vstack', (['(X_kno, X_unk)'], {}), '((X_kno, X_unk))\n', (718, 734), True, 'import numpy as np\n'), ((1689, 1701), 'numpy.unique', 'np.un... |
from pyparsing import LineEnd, LineStart, SkipTo, Regex
from regparser.grammar import atomic, unified
section = (
atomic.section_marker.copy().leaveWhitespace()
+ unified.part_section
+ SkipTo(LineEnd())
)
par = (
atomic.section.copy().leaveWhitespace()
+ unified.depth1_p
+ SkipTo(LineEnd()... | [
"regparser.grammar.atomic.section.copy",
"pyparsing.Regex",
"pyparsing.LineEnd",
"regparser.grammar.atomic.section_marker.copy",
"regparser.grammar.atomic.appendix_marker.copy",
"pyparsing.LineStart",
"regparser.grammar.atomic.paragraph_marker.copy"
] | [((1039, 1050), 'pyparsing.LineStart', 'LineStart', ([], {}), '()\n', (1048, 1050), False, 'from pyparsing import LineEnd, LineStart, SkipTo, Regex\n'), ((208, 217), 'pyparsing.LineEnd', 'LineEnd', ([], {}), '()\n', (215, 217), False, 'from pyparsing import LineEnd, LineStart, SkipTo, Regex\n'), ((311, 320), 'pyparsing... |
import argparse
def commandLineArgs():
parser = argparse.ArgumentParser(description='Process some integers.')
parser.add_argument("--restrict-types",
dest="restrict_types",
default=False,
action="store_true")
parser.add_argument("--tes... | [
"argparse.ArgumentParser"
] | [((53, 114), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {'description': '"""Process some integers."""'}), "(description='Process some integers.')\n", (76, 114), False, 'import argparse\n')] |
"""
fakedata.py
====================================
Generate artificial pupil-data.
"""
import numpy as np
import scipy.stats as stats
from .baseline import *
from .pupil import *
def generate_pupil_data(event_onsets, fs=1000, pad=5000, baseline_lowpass=0.2,
evoked_response_perc=0.02, respon... | [
"numpy.mean",
"numpy.ceil",
"numpy.ones",
"numpy.random.rand",
"numpy.any",
"numpy.array",
"numpy.zeros",
"numpy.linspace",
"numpy.random.randn",
"numpy.concatenate",
"numpy.zeros_like",
"numpy.arange",
"scipy.stats.truncnorm.rvs"
] | [((3093, 3111), 'numpy.any', 'np.any', (['(npars <= 0)'], {}), '(npars <= 0)\n', (3099, 3111), True, 'import numpy as np\n'), ((3162, 3180), 'numpy.any', 'np.any', (['(tmaxs <= 0)'], {}), '(tmaxs <= 0)\n', (3168, 3180), True, 'import numpy as np\n'), ((3481, 3492), 'numpy.zeros', 'np.zeros', (['n'], {}), '(n)\n', (3489... |
"""
<NAME>
Advent of Code Day 5
Challenge 1
"""
import sys
import hashlib
def md5_func(string):
md5result = hashlib.md5()
md5result.update(string.encode('utf-8'))
return md5result.hexdigest()
INTEGER_ID = 0
PASSWORD = ""
if len(sys.argv) < 2:
print("Please pass the puzzle input as a command line arg... | [
"sys.stdout.flush",
"hashlib.md5"
] | [((590, 608), 'sys.stdout.flush', 'sys.stdout.flush', ([], {}), '()\n', (606, 608), False, 'import sys\n'), ((114, 127), 'hashlib.md5', 'hashlib.md5', ([], {}), '()\n', (125, 127), False, 'import hashlib\n')] |
# Generated by Django 3.0.5 on 2020-12-07 16:56
import django.db.models.deletion
from django.conf import settings
from django.db import migrations, models
class Migration(migrations.Migration):
initial = True
dependencies = [
migrations.swappable_dependency(settings.AUTH_USER_MODEL),
]
ope... | [
"django.db.models.TextField",
"django.db.models.IntegerField",
"django.db.models.ForeignKey",
"django.db.models.ManyToManyField",
"django.db.models.SlugField",
"django.db.models.AutoField",
"django.db.models.DateTimeField",
"django.db.migrations.swappable_dependency",
"django.db.models.CharField"
] | [((247, 304), 'django.db.migrations.swappable_dependency', 'migrations.swappable_dependency', (['settings.AUTH_USER_MODEL'], {}), '(settings.AUTH_USER_MODEL)\n', (278, 304), False, 'from django.db import migrations, models\n'), ((437, 530), 'django.db.models.AutoField', 'models.AutoField', ([], {'auto_created': '(True)... |
"""proj URL Configuration
The `urlpatterns` list routes URLs to views. For more information please see:
https://docs.djangoproject.com/en/3.1/topics/http/urls/
Examples:
Function views
1. Add an import: from my_app import views
2. Add a URL to urlpatterns: path('', views.home, name='home')
Class-based vi... | [
"django.urls.include",
"books.views.BookDetailView.as_view",
"books.views.DeleteBookView.as_view",
"books.views.UpdateBookView.as_view",
"books.views.CreateBookView.as_view",
"django.conf.urls.static.static",
"proj.auth_views.MyLoginView.as_view",
"django.views.generic.RedirectView.as_view",
"books.... | [((3301, 3362), 'django.conf.urls.static.static', 'static', (['settings.MEDIA_URL'], {'document_root': 'settings.MEDIA_ROOT'}), '(settings.MEDIA_URL, document_root=settings.MEDIA_ROOT)\n', (3307, 3362), False, 'from django.conf.urls.static import static\n'), ((1224, 1255), 'django.urls.path', 'path', (['"""admin/"""', ... |
"""
Default audio settings.
"""
import numpy as np
from modules.socket.settings import PACKAGE_SIZE
# Number of sound channels.
CHANNELS = 2
# The size of the streaming buffer, that needs to fit into the socket buffer.
CHUNK_SIZE = PACKAGE_SIZE // CHANNELS // np.dtype(np.int16).itemsize
# Sound device frame rate. ... | [
"numpy.dtype"
] | [((264, 282), 'numpy.dtype', 'np.dtype', (['np.int16'], {}), '(np.int16)\n', (272, 282), True, 'import numpy as np\n')] |
import ast
import base64
import jinja2
import logging
import random
import datetime
from functools import reduce
import werkzeug
from odoo import SUPERUSER_ID
from odoo import api, http
from odoo.exceptions import UserError
from odoo.http import request
from odoo.modules import get_module_resource
from odoo.addons.web... | [
"logging.getLogger",
"odoo.http.request.render",
"random.randrange",
"odoo.http.request.env",
"odoo.http.request.make_response",
"odoo.modules.get_module_resource",
"base64.b64encode",
"base64.b64decode",
"odoo.http.route",
"werkzeug.wrappers.Response",
"datetime.datetime.now",
"ast.literal_ev... | [((377, 404), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (394, 404), False, 'import logging\n'), ((448, 471), 'datetime.datetime.now', 'datetime.datetime.now', ([], {}), '()\n', (469, 471), False, 'import datetime\n'), ((543, 637), 'odoo.http.route', 'http.route', (['[\'/free\', \'/fr... |
"""
Unit and regression test for the neuralxc package.
"""
import copy
import os
import sys
from abc import ABC, abstractmethod
import dill as pickle
import matplotlib.pyplot as plt
import numpy as np
import pytest
# Import package, test suite, and other packages as needed
import neuralxc as xc
from neuralxc.constan... | [
"numpy.allclose",
"neuralxc.utils.SiestaDensityGetter",
"neuralxc.formatter.SpeciesGrouper",
"os.path.join",
"neuralxc.formatter.Formatter",
"numpy.sum",
"numpy.array",
"neuralxc.ml.transformer.GroupedVarianceThreshold",
"pytest.mark.skipif",
"os.path.abspath",
"neuralxc.ml.transformer.GroupedSt... | [((3614, 3670), 'pytest.mark.skipif', 'pytest.mark.skipif', (['(not ase_found)'], {'reason': '"""requires ase"""'}), "(not ase_found, reason='requires ase')\n", (3632, 3670), False, 'import pytest\n'), ((658, 683), 'os.path.abspath', 'os.path.abspath', (['__file__'], {}), '(__file__)\n', (673, 683), False, 'import os\n... |
# -*- coding: utf-8 -*-
# Generated by Django 1.11.11 on 2019-01-25 09:24
from __future__ import unicode_literals
from django.db import migrations
class Migration(migrations.Migration):
dependencies = [
('focus', '0003_auto_20190125_1721'),
]
operations = [
migrations.AlterModelOptions(... | [
"django.db.migrations.AlterModelOptions"
] | [((291, 400), 'django.db.migrations.AlterModelOptions', 'migrations.AlterModelOptions', ([], {'name': '"""article"""', 'options': "{'verbose_name': '文章', 'verbose_name_plural': '文章'}"}), "(name='article', options={'verbose_name': '文章',\n 'verbose_name_plural': '文章'})\n", (319, 400), False, 'from django.db import mig... |
import numpy as np
from PIL import Image
nets = ["caffenet", "googlenet", "vggf", "vgg16", "vgg19"]
def load(nets):
res = []
for net in nets:
data_path = "perturbations/perturbation_%s.npy" % net
imgs = np.load(data_path, allow_pickle=True, encoding="latin1")
# print(imgs.shape)
... | [
"numpy.uint8",
"PIL.Image.new",
"numpy.transpose",
"numpy.load"
] | [((717, 784), 'PIL.Image.new', 'Image.new', (['"""RGB"""', '(n * width + interval * (n - 1), height)', '"""white"""'], {}), "('RGB', (n * width + interval * (n - 1), height), 'white')\n", (726, 784), False, 'from PIL import Image\n'), ((230, 286), 'numpy.load', 'np.load', (['data_path'], {'allow_pickle': '(True)', 'enc... |
# coding=utf-8
# Copyright 2021 The Google Research Authors.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicab... | [
"numpy.tile",
"jax.numpy.cos",
"jax.numpy.sqrt",
"jax.numpy.asarray",
"numpy.array",
"jax.numpy.maximum",
"jax.numpy.minimum"
] | [((3342, 3377), 'jax.numpy.asarray', 'jnp.asarray', (['ret'], {'dtype': 'jnp.float32'}), '(ret, dtype=jnp.float32)\n', (3353, 3377), True, 'import jax.numpy as jnp\n'), ((3575, 3605), 'numpy.tile', 'np.tile', (['x[-1]', '(batch_pad, 1)'], {}), '(x[-1], (batch_pad, 1))\n', (3582, 3605), True, 'import numpy as np\n'), ((... |
# --------------------------------------------------------
# Swin Transformer
# Copyright (c) 2021 Microsoft
# Licensed under The MIT License [see LICENSE for details]
# Written by <NAME>
# --------------------------------------------------------
import numpy as np
import torch
import torch.distributed as dist
from dat... | [
"torch.utils.data.DistributedSampler",
"torchvision.transforms.CenterCrop",
"torch.distributed.get_rank",
"datasets.CustomDataset",
"torchvision.transforms.RandomCrop",
"torchvision.transforms.Normalize",
"torch.utils.data.DataLoader",
"timm.data.transforms._pil_interp",
"torchvision.transforms.ToTe... | [((1275, 1296), 'torch.distributed.get_world_size', 'dist.get_world_size', ([], {}), '()\n', (1294, 1296), True, 'import torch.distributed as dist\n'), ((1315, 1330), 'torch.distributed.get_rank', 'dist.get_rank', ([], {}), '()\n', (1328, 1330), True, 'import torch.distributed as dist\n'), ((1971, 2167), 'torch.utils.d... |