code
stringlengths
22
1.05M
apis
listlengths
1
3.31k
extract_api
stringlengths
75
3.25M
import tqdm import torch from lav.lav_privileged import LAV from lav.utils.datasets import get_data_loader from lav.utils.logger import Logger def main(args): dmd = LAV(args) data_loader = get_data_loader('bev', args) logger = Logger('lav_bev', args) save_dir = logger.save_dir torch.manual_seed(a...
[ "torch.manual_seed", "argparse.ArgumentParser", "tqdm.tqdm", "lav.utils.logger.Logger", "lav.lav_privileged.LAV", "lav.utils.datasets.get_data_loader" ]
[((171, 180), 'lav.lav_privileged.LAV', 'LAV', (['args'], {}), '(args)\n', (174, 180), False, 'from lav.lav_privileged import LAV\n'), ((199, 227), 'lav.utils.datasets.get_data_loader', 'get_data_loader', (['"""bev"""', 'args'], {}), "('bev', args)\n", (214, 227), False, 'from lav.utils.datasets import get_data_loader\...
#!/usr/bin/env python3 # async_requests.py """Asynchronously get links embedded in multiple pages' HTML.""" import asyncio import logging import re import sys # from typing import IO # Use pathlib instead import urllib.error import urllib.parse import aiofiles import aiohttp from aiohttp import ClientSession impor...
[ "logging.basicConfig", "aiohttp.ClientSession", "logging.getLogger", "asyncio.WindowsSelectorEventLoopPolicy", "re.compile", "pathlib.Path", "aiofiles.open", "asyncio.gather" ]
[((393, 539), 'logging.basicConfig', 'logging.basicConfig', ([], {'format': '"""%(asctime)s %(levelname)s:%(name)s: %(message)s"""', 'level': 'logging.DEBUG', 'datefmt': '"""%H:%M:%S"""', 'stream': 'sys.stderr'}), "(format=\n '%(asctime)s %(levelname)s:%(name)s: %(message)s', level=logging.DEBUG,\n datefmt='%H:%M...
# -*- coding: utf-8 -*- from django.shortcuts import render from qiniuyun.backend import QiniuPush from qiniuyun.models import ImageAtQiniu from .indexView import ImgList from DJangoHotel.models import RoomInfo def roomInfo(request): rooms=RoomInfo.objects.all() imgObjs=ImageAtQiniu.objects.all() imgUrls=...
[ "django.shortcuts.render", "qiniuyun.models.ImageAtQiniu.objects.all", "qiniuyun.backend.QiniuPush.private_download_url", "DJangoHotel.models.RoomInfo.objects.all" ]
[((246, 268), 'DJangoHotel.models.RoomInfo.objects.all', 'RoomInfo.objects.all', ([], {}), '()\n', (266, 268), False, 'from DJangoHotel.models import RoomInfo\n'), ((281, 307), 'qiniuyun.models.ImageAtQiniu.objects.all', 'ImageAtQiniu.objects.all', ([], {}), '()\n', (305, 307), False, 'from qiniuyun.models import Image...
import pyb import sensor import image import time # Status LED led = pyb.LED(3) # Configure camera sensor.reset() sensor.set_contrast(3) sensor.set_gainceiling(16) sensor.set_framesize(sensor.QVGA) sensor.set_pixformat(sensor.GRAYSCALE) # Get center x, y of camera image WIDTH = sensor.width() HEIGHT = sensor.height(...
[ "sensor.set_contrast", "sensor.set_gainceiling", "sensor.set_pixformat", "time.clock", "sensor.set_framesize", "image.HaarCascade", "sensor.width", "sensor.reset", "sensor.height", "sensor.snapshot", "pyb.LED" ]
[((70, 80), 'pyb.LED', 'pyb.LED', (['(3)'], {}), '(3)\n', (77, 80), False, 'import pyb\n'), ((101, 115), 'sensor.reset', 'sensor.reset', ([], {}), '()\n', (113, 115), False, 'import sensor\n'), ((116, 138), 'sensor.set_contrast', 'sensor.set_contrast', (['(3)'], {}), '(3)\n', (135, 138), False, 'import sensor\n'), ((13...
from tortoise import fields, models class User(models.Model): """ Model user """ username = fields.CharField(max_length=100, unique=True) password = fields.CharField(max_length=100) email = fields.CharField(max_length=100, unique=True) first_name = fields.CharField(max_length=100) last_name = ...
[ "tortoise.fields.CharField", "tortoise.fields.DatetimeField", "tortoise.fields.BooleanField", "tortoise.fields.TextField" ]
[((102, 147), 'tortoise.fields.CharField', 'fields.CharField', ([], {'max_length': '(100)', 'unique': '(True)'}), '(max_length=100, unique=True)\n', (118, 147), False, 'from tortoise import fields, models\n'), ((163, 195), 'tortoise.fields.CharField', 'fields.CharField', ([], {'max_length': '(100)'}), '(max_length=100)...
import webbrowser from ontology.exception.exception import SDKException from click import ( argument, pass_context ) from .main import main from punica.box.repo_box import Box from punica.utils.output import echo_cli_exception from punica.exception.punica_exception import PunicaException @main.command('unb...
[ "click.argument", "punica.utils.output.echo_cli_exception", "punica.box.repo_box.Box", "webbrowser.open" ]
[((326, 355), 'click.argument', 'argument', (['"""box_name"""'], {'nargs': '(1)'}), "('box_name', nargs=1)\n", (334, 355), False, 'from click import argument, pass_context\n'), ((481, 508), 'punica.box.repo_box.Box', 'Box', (["ctx.obj['PROJECT_DIR']"], {}), "(ctx.obj['PROJECT_DIR'])\n", (484, 508), False, 'from punica....
import sys from datetime import date, timedelta import requests def date_gen(d1, d2): # d1 = date(2020, 5, 1) # d2 = date(2020, 5, 31) delta = d2 - d1 return [(d1 + timedelta(days=i)).strftime('%Y%m%d') for i in range(delta.days + 1)] def download_by_dates(date_list): for date_to_download in da...
[ "requests.get", "datetime.timedelta", "datetime.date", "sys.stdout.flush", "sys.stdout.write" ]
[((625, 655), 'requests.get', 'requests.get', (['url'], {'stream': '(True)'}), '(url, stream=True)\n', (637, 655), False, 'import requests\n'), ((1321, 1337), 'datetime.date', 'date', (['(2020)', '(3)', '(1)'], {}), '(2020, 3, 1)\n', (1325, 1337), False, 'from datetime import date, timedelta\n'), ((1342, 1358), 'dateti...
from heapq import heapify, heappush, heappop from collections import defaultdict import math def shortest_path(M, start, goal): frontier = {start} explored = set() came_from = dict() f_costs = get_initial_f_costs(M, start, goal) # heapq type g_costs = get_initial_g_costs(start) # defaultdict typ...
[ "math.pow", "heapq.heappop", "collections.defaultdict", "heapq.heappush", "heapq.heapify" ]
[((1937, 1953), 'heapq.heapify', 'heapify', (['f_costs'], {}), '(f_costs)\n', (1944, 1953), False, 'from heapq import heapify, heappush, heappop\n'), ((2012, 2057), 'heapq.heappush', 'heappush', (['f_costs', '(f_cost_start_node, start)'], {}), '(f_costs, (f_cost_start_node, start))\n', (2020, 2057), False, 'from heapq ...
# -*- coding: utf-8 -*- import cherrypy from jinja2 import Template import mock from tests.utils import BaseToolsTest from lib.tool.allowed_methods import AllowedMethodsTool from lib.tool.cpemail import EmailTool from lib.tool.template import Jinja2Tool class TestAllowedMethods(BaseToolsTest): _cp_config = { ...
[ "lib.tool.allowed_methods.AllowedMethodsTool", "lib.tool.template.Jinja2Tool" ]
[((509, 529), 'lib.tool.allowed_methods.AllowedMethodsTool', 'AllowedMethodsTool', ([], {}), '()\n', (527, 529), False, 'from lib.tool.allowed_methods import AllowedMethodsTool\n'), ((1072, 1084), 'lib.tool.template.Jinja2Tool', 'Jinja2Tool', ([], {}), '()\n', (1082, 1084), False, 'from lib.tool.template import Jinja2T...
import os import logging import stackprinter from celery import Celery, Task from celery.schedules import crontab from django.apps import apps, AppConfig from django.conf import settings if not settings.configured: # set the default Django settings module for the 'celery' program. os.environ.setdefault( ...
[ "os.environ.setdefault", "celery.Celery", "django.apps.apps.get_app_configs", "stackprinter.format", "celery.schedules.crontab", "logging.error" ]
[((407, 430), 'celery.Celery', 'Celery', (['"""instanotifier"""'], {}), "('instanotifier')\n", (413, 430), False, 'from celery import Celery, Task\n'), ((293, 365), 'os.environ.setdefault', 'os.environ.setdefault', (['"""DJANGO_SETTINGS_MODULE"""', '"""config.settings.local"""'], {}), "('DJANGO_SETTINGS_MODULE', 'confi...
from snovault import ( AuditFailure, audit_checker, ) @audit_checker('ReferenceEpigenome', frame=['related_datasets', 'related_datasets.replicates', 'related_datasets.replicates.library', ...
[ "snovault.AuditFailure", "snovault.audit_checker" ]
[((65, 378), 'snovault.audit_checker', 'audit_checker', (['"""ReferenceEpigenome"""'], {'frame': "['related_datasets', 'related_datasets.replicates',\n 'related_datasets.replicates.library',\n 'related_datasets.replicates.library.biosample',\n 'related_datasets.replicates.library.biosample.donor',\n 'relate...
import openmc from scipy import interpolate import matplotlib.pyplot as plt from matplotlib.colors import LogNorm from matplotlib import ticker import matplotx import numpy as np import scipy.ndimage as ndimage def reshape_values_to_mesh_shape(tally, values): mesh_filter = tally.find_filter(filter_type=openmc.Mes...
[ "matplotlib.pyplot.savefig", "matplotlib.pyplot.gca", "matplotlib.pyplot.xlabel", "matplotlib.pyplot.sca", "matplotx.ylabel_top", "numpy.array", "openmc.StatePoint", "matplotlib.pyplot.style.context", "numpy.linspace", "matplotlib.pyplot.subplots", "matplotlib.pyplot.scatter", "scipy.ndimage.g...
[((2780, 2823), 'openmc.StatePoint', 'openmc.StatePoint', ([], {'filepath': 'statepoint_file'}), '(filepath=statepoint_file)\n', (2797, 2823), False, 'import openmc\n'), ((1908, 1930), 'numpy.array', 'np.array', (["data['mean']"], {}), "(data['mean'])\n", (1916, 1930), True, 'import numpy as np\n'), ((2302, 2353), 'sci...
# -*- coding: utf-8 -*- """ Provides base logging functions :copyright: © 2018 by <NAME> :license: MIT, see LICENSE for more details. """ from logging import getLevelName, INFO, WARN, ERROR, DEBUG from multiprocessing import current_process from time import strftime from core.common import load_config from...
[ "logging.getLevelName", "time.strftime", "multiprocessing.current_process", "core.common.load_config" ]
[((623, 636), 'core.common.load_config', 'load_config', ([], {}), '()\n', (634, 636), False, 'from core.common import load_config\n'), ((551, 568), 'multiprocessing.current_process', 'current_process', ([], {}), '()\n', (566, 568), False, 'from multiprocessing import current_process\n'), ((1828, 1849), 'time.strftime',...
from interfaces.interface import Publisher, stock_list from utils.notifier import NotificationUtils from service.loader import LoadStock from utils import cache_util, common_constants class WatchTower(Publisher): state = 0 observser_list = list() def __init__(self, stock_val): self.stock_val = st...
[ "utils.cache_util.create_cache_client", "service.loader.LoadStock", "utils.notifier.NotificationUtils" ]
[((635, 700), 'utils.notifier.NotificationUtils', 'NotificationUtils', ([], {'user_email': 'user_email', 'custom_message': 'messages'}), '(user_email=user_email, custom_message=messages)\n', (652, 700), False, 'from utils.notifier import NotificationUtils\n'), ((2588, 2620), 'utils.cache_util.create_cache_client', 'cac...
"""This module tests Exceptions functionality in stereomideval module""" import pytest import numpy as np from stereomideval.dataset import Dataset from stereomideval.exceptions import ImageSizeNotEqual, PathNotFound, InvalidSceneName def test_catch_invalid_image_sizes(): """Test catching invalid image sizes""" ...
[ "stereomideval.dataset.Dataset.get_training_scene_list", "stereomideval.dataset.Dataset.get_scene_list", "stereomideval.exceptions.ImageSizeNotEqual.validate", "numpy.zeros", "pytest.raises", "stereomideval.exceptions.PathNotFound.validate" ]
[((333, 349), 'numpy.zeros', 'np.zeros', (['(5, 5)'], {}), '((5, 5))\n', (341, 349), True, 'import numpy as np\n'), ((364, 380), 'numpy.zeros', 'np.zeros', (['(5, 6)'], {}), '((5, 6))\n', (372, 380), True, 'import numpy as np\n'), ((390, 422), 'pytest.raises', 'pytest.raises', (['ImageSizeNotEqual'], {}), '(ImageSizeNo...
# -*- coding: utf8 -*- # ============LICENSE_START======================================================= # org.onap.vvp/validation-scripts # =================================================================== # Copyright © 2017 AT&T Intellectual Property. All rights reserved. # ========================================...
[ "pytest.skip", "tests.cached_yaml.load", "re.compile" ]
[((1890, 1937), 're.compile', 're.compile', (['"""int_(?P<network_role>.+)_network$"""'], {}), "('int_(?P<network_role>.+)_network$')\n", (1900, 1937), False, 'import re\n'), ((2248, 2295), 're.compile', 're.compile', (['"""int_(?P<network_role>.+)_network$"""'], {}), "('int_(?P<network_role>.+)_network$')\n", (2258, 2...
import json from bson import ObjectId from pymongo import ReturnDocument from .exceptions import DBException class DBActionsMixin: def __init__(self, model, db): self._model_cls = model self._db = db def add(self, item): db_obj = self._collection.insert_one(item.prepare_for_db()) ...
[ "bson.ObjectId", "json.dumps" ]
[((1785, 1801), 'json.dumps', 'json.dumps', (['self'], {}), '(self)\n', (1795, 1801), False, 'import json\n'), ((1699, 1716), 'bson.ObjectId', 'ObjectId', (['self.id'], {}), '(self.id)\n', (1707, 1716), False, 'from bson import ObjectId\n')]
# -*- coding: utf-8 -*- from random import Random #from core.dataloader import DataLoader from torch.utils.data import DataLoader import numpy as np from math import * import logging from scipy import stats import numpy as np from pyemd import emd from collections import OrderedDict import time import pickle, random fr...
[ "numpy.identity", "pyemd.emd", "collections.OrderedDict", "scipy.stats.entropy", "random.Random", "pickle.load", "logging.info", "numpy.sum", "numpy.zeros", "numpy.random.seed", "numpy.concatenate", "torch.utils.data.DataLoader", "time.time" ]
[((20762, 20773), 'time.time', 'time.time', ([], {}), '()\n', (20771, 20773), False, 'import time\n'), ((941, 949), 'random.Random', 'Random', ([], {}), '()\n', (947, 949), False, 'from random import Random\n'), ((1168, 1188), 'numpy.random.seed', 'np.random.seed', (['seed'], {}), '(seed)\n', (1182, 1188), True, 'impor...
import os import shutil import Ni__eam__born_exp_rose as configuration from collections import OrderedDict def cleanup_simulation_directories(): sim_directories = [ 'Ni_fcc_vac.lmps_min_pos', 'Ni_fcc.lmps_elastic', 'Ni_fcc_100_unit.lmps_min_all', 'Ni_fcc_111_s.lmps_min_pos', ...
[ "collections.OrderedDict", "pypospack.pyposmat.engines.PyposmatEngine", "Ni__eam__born_exp_rose.write_configuration_file", "os.path.isdir", "shutil.rmtree" ]
[((889, 902), 'collections.OrderedDict', 'OrderedDict', ([], {}), '()\n', (900, 902), False, 'from collections import OrderedDict\n'), ((1537, 1550), 'collections.OrderedDict', 'OrderedDict', ([], {}), '()\n', (1548, 1550), False, 'from collections import OrderedDict\n'), ((1574, 1587), 'collections.OrderedDict', 'Orde...
from django.contrib import admin from stade.core.models import Challenge from .task import TaskInline @admin.register(Challenge) class ChallengeAdmin(admin.ModelAdmin): inlines = [TaskInline]
[ "django.contrib.admin.register" ]
[((107, 132), 'django.contrib.admin.register', 'admin.register', (['Challenge'], {}), '(Challenge)\n', (121, 132), False, 'from django.contrib import admin\n')]
#!/usr/bin/env python3 """problem_059.py Problem 59: XOR decryption Each character on a computer is assigned a unique code and the preferred standard is ASCII (American Standard Code for Information Interchange). For example, uppercase A = 65, asterisk (*) = 42, and lowercase k = 107. A modern encryption method is ...
[ "common.fileio.ints_from_file", "common.alphabet.letter_char_lower" ]
[((2445, 2483), 'common.fileio.ints_from_file', 'fio.ints_from_file', (['FILE_NAME'], {'sep': '""","""'}), "(FILE_NAME, sep=',')\n", (2463, 2483), True, 'import common.fileio as fio\n'), ((2744, 2770), 'common.alphabet.letter_char_lower', 'alpha.letter_char_lower', (['i'], {}), '(i)\n', (2767, 2770), True, 'import comm...
import numpy as np import glob import geo import time import pdb start_time = time.time() dataDir='./data/' # get CrIS files cris_sdr_files = sorted(glob.glob(dataDir+'SCRIS*')) cris_geo_files = sorted(glob.glob(dataDir+'GCRSO*')) # get VIIRS files viirs_sdr_files = sorted(glob.glob(dataDir+'SVM15*')) viirs_geo_f...
[ "geo.read_viirs_sdr", "matplotlib.pyplot.savefig", "geo.read_cris_sdr", "geo.match_cris_viirs", "numpy.zeros_like", "numpy.ndindex", "numpy.append", "matplotlib.cm.ScalarMappable", "glob.glob", "matplotlib.colors.Normalize", "geo.RAE2ENU", "matplotlib.pyplot.get_cmap", "geo.ENU2ECEF", "tim...
[((79, 90), 'time.time', 'time.time', ([], {}), '()\n', (88, 90), False, 'import time\n'), ((457, 492), 'geo.read_viirs_geo', 'geo.read_viirs_geo', (['viirs_geo_files'], {}), '(viirs_geo_files)\n', (475, 492), False, 'import geo\n'), ((528, 563), 'geo.read_viirs_sdr', 'geo.read_viirs_sdr', (['viirs_sdr_files'], {}), '(...
"""Aggregate hosts data set.""" from censys.search import SearchClient c = SearchClient() # The aggregate method constructs a report using a query, an aggregation field, and the # number of buckets to bin. report = c.v2.hosts.aggregate( "service.service_name: HTTP", "services.port", num_buckets=5, ) print...
[ "censys.search.SearchClient" ]
[((76, 90), 'censys.search.SearchClient', 'SearchClient', ([], {}), '()\n', (88, 90), False, 'from censys.search import SearchClient\n')]
import aiohttp_jinja2 from aiohttp import web from riego.db import get_db from riego.web.security import raise_permission router = web.RouteTableDef() def setup_routes_events(app): app.add_routes(router) @router.get("/events", name='events') @aiohttp_jinja2.template('events/index.html') async...
[ "aiohttp_jinja2.template", "riego.db.get_db", "riego.web.security.raise_permission", "aiohttp.web.RouteTableDef" ]
[((141, 160), 'aiohttp.web.RouteTableDef', 'web.RouteTableDef', ([], {}), '()\n', (158, 160), False, 'from aiohttp import web\n'), ((269, 313), 'aiohttp_jinja2.template', 'aiohttp_jinja2.template', (['"""events/index.html"""'], {}), "('events/index.html')\n", (292, 313), False, 'import aiohttp_jinja2\n'), ((863, 907), ...
# coding: utf-8 from __future__ import division, print_function, absolute_import import pkg_resources __version__ = pkg_resources.get_distribution('pypcl').version from pypcl.common import *
[ "pkg_resources.get_distribution" ]
[((119, 158), 'pkg_resources.get_distribution', 'pkg_resources.get_distribution', (['"""pypcl"""'], {}), "('pypcl')\n", (149, 158), False, 'import pkg_resources\n')]
import tkinter as tk import tkinter.font as tkFont class History(tk.Frame): def __init__(self, master): self.master=master tk.Frame.__init__(self, master=master) self.listOperations =[] self.history = [] self.LEFT = 0 self.OP = 1 self.RIGHT = 2 self.E...
[ "tkinter.font.Font", "tkinter.Frame.__init__" ]
[((144, 182), 'tkinter.Frame.__init__', 'tk.Frame.__init__', (['self'], {'master': 'master'}), '(self, master=master)\n', (161, 182), True, 'import tkinter as tk\n'), ((831, 851), 'tkinter.font.Font', 'tkFont.Font', ([], {'size': '(10)'}), '(size=10)\n', (842, 851), True, 'import tkinter.font as tkFont\n'), ((637, 657)...
""" Contains the code necessary to extract a list of optimal compression values from a csv file containing columns corresponding to {compression_type}_{level}, {variable}, {time}, and {DSSIM} It would be best to open the csv file once, and get a list of all variables, levels, and timesteps so I don't read the csv file...
[ "csv.DictWriter", "numpy.unique", "argparse.ArgumentParser", "re.compile", "os.path.isfile", "numpy.argsort", "lcr_global_vars.varlist", "csv.reader", "re.search" ]
[((8948, 8964), 'numpy.unique', 'np.unique', (['times'], {}), '(times)\n', (8957, 8964), True, 'import numpy as np\n'), ((10290, 10306), 'numpy.unique', 'np.unique', (['times'], {}), '(times)\n', (10299, 10306), True, 'import numpy as np\n'), ((13106, 13131), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {...
# -*- coding: utf-8 -*- """ =============================================================================== Horsager et al. (2009): Predicting temporal sensitivity =============================================================================== This example shows how to use the :py:class:`~pulse2percept.models.Horsager...
[ "pulse2percept.stimuli.BiphasicPulseTrain", "pulse2percept.stimuli.MonophasicPulse", "numpy.ceil", "matplotlib.pyplot.xticks", "matplotlib.pyplot.semilogx", "matplotlib.pyplot.ylabel", "numpy.arange", "matplotlib.pyplot.xlabel", "pulse2percept.models.Horsager2009Temporal", "pulse2percept.stimuli.B...
[((862, 881), 'pulse2percept.datasets.load_horsager2009', 'load_horsager2009', ([], {}), '()\n', (879, 881), False, 'from pulse2percept.datasets import load_horsager2009\n'), ((1248, 1325), 'pulse2percept.datasets.load_horsager2009', 'load_horsager2009', ([], {'subjects': '"""S05"""', 'electrodes': '"""C3"""', 'stim_ty...
#!/usr/bin/python3 import mysql.connector from mysql.connector import Error from mysql.connector import errorcode from os import environ import json DB_HOST = environ.get('DB_HOST') DB_NAME = environ.get('DB_NAME') DB_USER = environ.get('DB_USER') DB_PASSWORD = environ.get('DB_PASSWORD') if DB_PASSWORD is not None: ...
[ "json.dumps", "os.environ.get" ]
[((160, 182), 'os.environ.get', 'environ.get', (['"""DB_HOST"""'], {}), "('DB_HOST')\n", (171, 182), False, 'from os import environ\n'), ((193, 215), 'os.environ.get', 'environ.get', (['"""DB_NAME"""'], {}), "('DB_NAME')\n", (204, 215), False, 'from os import environ\n'), ((226, 248), 'os.environ.get', 'environ.get', (...
from __future__ import print_function import argparse import glob import io import os import subprocess as sp import sys from collections import defaultdict from itertools import chain import networkx as nx import requests import yaml from conda_build import api # ----------------------------------------------------...
[ "itertools.chain", "conda_build.api.Config", "io.open", "networkx.dfs_successors", "sys.exit", "os.walk", "os.path.exists", "argparse.ArgumentParser", "conda_build.api.get_output_file_paths", "networkx.DiGraph", "networkx.simple_cycles", "subprocess.check_call", "conda_build.api.render", "...
[((1125, 1256), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {'description': '"""Identify and build all ggd recipes that are not currently in any ggd conda channel"""'}), "(description=\n 'Identify and build all ggd recipes that are not currently in any ggd conda channel'\n )\n", (1148, 1256), False...
from django.test import TestCase from corehq import toggles from corehq.motech.dhis2.tasks import send_datasets_for_all_domains class TestSendDatasetsForAllDomains(TestCase): domain_name = 'does-not-exist' def setUp(self): toggles.DHIS2_INTEGRATION.set( self.domain_name, ena...
[ "corehq.toggles.DHIS2_INTEGRATION.set", "corehq.motech.dhis2.tasks.send_datasets_for_all_domains" ]
[((244, 346), 'corehq.toggles.DHIS2_INTEGRATION.set', 'toggles.DHIS2_INTEGRATION.set', (['self.domain_name'], {'enabled': '(True)', 'namespace': 'toggles.NAMESPACE_DOMAIN'}), '(self.domain_name, enabled=True, namespace=\n toggles.NAMESPACE_DOMAIN)\n', (273, 346), False, 'from corehq import toggles\n'), ((421, 524), ...
''' @Author: fxm @Date: Dec 27, 2020. @Title: UI class. ''' import sys sys.path.append('..') import pygame import logging from pygame.locals import * import pygame.gfxdraw from collections import namedtuple from Framework.Net import dotdict from Othello.Board import Board log = logging.getLogger(__name__) ...
[ "logging.getLogger", "pygame.mouse.get_pressed", "pygame.draw.line", "pygame.gfxdraw.aacircle", "pygame.event.get", "pygame.display.flip", "Othello.Board.Board", "pygame.mouse.get_pos", "pygame.draw.rect", "pygame.gfxdraw.filled_circle", "sys.path.append", "pygame.font.SysFont" ]
[((83, 104), 'sys.path.append', 'sys.path.append', (['""".."""'], {}), "('..')\n", (98, 104), False, 'import sys\n'), ((292, 319), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (309, 319), False, 'import logging\n'), ((2467, 2580), 'pygame.draw.rect', 'pygame.draw.rect', (['screen', 'BLA...
# -*- coding: UTF-8 -*- # main.py # Root file for EmotionDetection program. # Prints out command line menu and handles user choices from __future__ import print_function from EmotionDetection import WordMap from EmotionDetection import EvaluateText from EmotionDetection import GUI try: input = raw_input except N...
[ "EmotionDetection.EvaluateText.evaluate", "EmotionDetection.GUI.Evaluator", "sys.setdefaultencoding", "EmotionDetection.WordMap.buildWordMap" ]
[((363, 393), 'sys.setdefaultencoding', 'sys.setdefaultencoding', (['"""utf8"""'], {}), "('utf8')\n", (385, 393), False, 'import sys\n'), ((2684, 2699), 'EmotionDetection.GUI.Evaluator', 'GUI.Evaluator', ([], {}), '()\n', (2697, 2699), False, 'from EmotionDetection import GUI\n'), ((2544, 2575), 'EmotionDetection.Evalu...
import os def print_tables(tables, width=os.get_terminal_size().columns, spaces=3, index_width=3): space = ' ' * spaces lines = [] if index_width <= 0: index_width = -1 tables.sort(key=lambda t: max(len(t), max([len(r) for r in t])), reverse=True) for table in tables: max_length = max([len(x) f...
[ "os.get_terminal_size" ]
[((43, 65), 'os.get_terminal_size', 'os.get_terminal_size', ([], {}), '()\n', (63, 65), False, 'import os\n')]
import datetime as dt from uploader.utils import NULL from typing import List __ESCAPE_SYMBOLS_MAPPING = {"'": r"''"} def __value_empty(value) -> bool: return value == NULL or value is None or not value or (isinstance(value, str) and value.isspace()) def __escaped_symbols() -> dict: if not hasattr(__escape...
[ "datetime.datetime", "datetime.datetime.strptime" ]
[((536, 570), 'datetime.datetime', 'dt.datetime', (['(1899)', '(12)', '(31)', '(0)', '(0)', '(0)'], {}), '(1899, 12, 31, 0, 0, 0)\n', (547, 570), True, 'import datetime as dt\n'), ((1196, 1231), 'datetime.datetime.strptime', 'dt.datetime.strptime', (['value', 'format'], {}), '(value, format)\n', (1216, 1231), True, 'im...
import destination_search ''' test용 값 gpsX = 127.0816985 # 동경 - 경도 gpsY = 37.5642135 # 북위 - 위도 time = 20 1. 좌표, 시간 입력 -> 주변 버스 정류소 ID 반환 2. 정류소 ID 입력 -> 버스 노선 획득 3. 획득한 버스 노선에서 출발지, 목적지 시간 계산하여 도착 버스 정류장 반환 ''' print("gps_x : ", end='') gps_x = input() print("gps_y : ", end='') gps_y = input() print("t...
[ "destination_search.DestinationStation" ]
[((429, 486), 'destination_search.DestinationStation', 'destination_search.DestinationStation', (['gps_x', 'gps_y', 'time'], {}), '(gps_x, gps_y, time)\n', (466, 486), False, 'import destination_search\n')]
from torch.utils.data import Dataset import cv2 class SegmentationDataSet(Dataset): def __init__(self, image_paths, mask_paths, transform): self.image_paths = image_paths self.mask_paths = mask_paths self.transforms = transform def __len__(self): return len(self.image_paths) ...
[ "cv2.imread", "cv2.cvtColor" ]
[((600, 622), 'cv2.imread', 'cv2.imread', (['image_path'], {}), '(image_path)\n', (610, 622), False, 'import cv2\n'), ((639, 677), 'cv2.cvtColor', 'cv2.cvtColor', (['image', 'cv2.COLOR_BGR2RGB'], {}), '(image, cv2.COLOR_BGR2RGB)\n', (651, 677), False, 'import cv2\n'), ((700, 735), 'cv2.imread', 'cv2.imread', (['self.ma...
# -*- coding: utf-8 -*- from odoo import api, fields, models, _ from odoo.exceptions import UserError from odoo.tools.misc import format_date class AccrualAccountingWizard(models.TransientModel): _name = 'account.accrual.accounting.wizard' _description = 'Create accrual entry.' date = fields.Date(require...
[ "odoo._", "odoo.fields.Monetary", "odoo.api.constrains", "odoo.fields.Date", "odoo.fields.Float", "odoo.fields.Many2one", "odoo.api.depends", "odoo.tools.misc.format_date", "odoo.fields.Selection", "odoo.fields.Many2many" ]
[((301, 327), 'odoo.fields.Date', 'fields.Date', ([], {'required': '(True)'}), '(required=True)\n', (312, 327), False, 'from odoo import api, fields, models, _\n'), ((345, 390), 'odoo.fields.Many2one', 'fields.Many2one', (['"""res.company"""'], {'required': '(True)'}), "('res.company', required=True)\n", (360, 390), Fa...
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Sun Mar 11 21:22:59 2018 @author: pami4 """ #CUDA_VISIBLE_DEVICES=0 python from pycocotools.coco import COCO import coco import numpy as np from matplotlib import pyplot as plt import visualize import custom_utils config = coco.CocoConfig() config.GPU_CO...
[ "CustomDataGenerator.CustomDatasetIterator_MaskRCNN", "CustomDataset.CocoDataset", "numpy.sum", "numpy.array", "model.MaskRCNN", "coco.CocoConfig", "matplotlib.pyplot.show" ]
[((289, 306), 'coco.CocoConfig', 'coco.CocoConfig', ([], {}), '()\n', (304, 306), False, 'import coco\n'), ((361, 388), 'CustomDataset.CocoDataset', 'CustomDataset.CocoDataset', ([], {}), '()\n', (386, 388), False, 'import CustomDataset\n'), ((668, 778), 'CustomDataGenerator.CustomDatasetIterator_MaskRCNN', 'CustomData...
from importlib.metadata import PackageNotFoundError, version try: __version__ = version("x_mlps") except PackageNotFoundError: # package is not installed __version__ = None
[ "importlib.metadata.version" ]
[((85, 102), 'importlib.metadata.version', 'version', (['"""x_mlps"""'], {}), "('x_mlps')\n", (92, 102), False, 'from importlib.metadata import PackageNotFoundError, version\n')]
from django.db.models.signals import post_save from django.dispatch import receiver from company.models import Company from company.tasks import deploy_new_company @receiver(post_save, sender=Company) def company_created(sender, instance, created, **kwargs): if created: deploy_new_company.delay(instance....
[ "django.dispatch.receiver", "company.tasks.deploy_new_company.delay" ]
[((168, 203), 'django.dispatch.receiver', 'receiver', (['post_save'], {'sender': 'Company'}), '(post_save, sender=Company)\n', (176, 203), False, 'from django.dispatch import receiver\n'), ((286, 323), 'company.tasks.deploy_new_company.delay', 'deploy_new_company.delay', (['instance.id'], {}), '(instance.id)\n', (310, ...
# -*- coding:utf-8 -*- from sc import SupercellSWF import os def findAllFile(basePath, endsWith): for root, ds, fs in os.walk(basePath): for f in fs: if not f.startswith('.') and f.endswith(endsWith): fullname = os.path.join(root, f) yield fullname
[ "os.path.join", "os.walk" ]
[((123, 140), 'os.walk', 'os.walk', (['basePath'], {}), '(basePath)\n', (130, 140), False, 'import os\n'), ((253, 274), 'os.path.join', 'os.path.join', (['root', 'f'], {}), '(root, f)\n', (265, 274), False, 'import os\n')]
##################################################### # # # Source file of the MadFlow plugin # # Use only with consent of its authors. # # # # authors: S.Carrazza, J.Cruz-Martinez, ...
[ "os.path.realpath", "sys.path.insert" ]
[((638, 667), 'sys.path.insert', 'sys.path.insert', (['(0)', 'root_path'], {}), '(0, root_path)\n', (653, 667), False, 'import sys\n'), ((604, 630), 'os.path.realpath', 'os.path.realpath', (['__file__'], {}), '(__file__)\n', (620, 630), False, 'import os\n')]
import abc import time from datetime import datetime from pathlib import Path from typing import Any, Dict, List, Optional, cast from loguru import logger from rich import print from dubdub import ( Binary, Grouping, Literal, Node, Token, TokenType, Unary, Visitor, dataclass, ) fro...
[ "loguru.logger.debug", "dubdub.parser.Parser", "pathlib.Path.cwd", "rich.print", "dubdub.scanner.Scanner", "dubdub.env.Environment", "typing.cast" ]
[((533, 543), 'pathlib.Path.cwd', 'Path.cwd', ([], {}), '()\n', (541, 543), False, 'from pathlib import Path\n'), ((660, 673), 'dubdub.env.Environment', 'Environment', ([], {}), '()\n', (671, 673), False, 'from dubdub.env import Environment\n'), ((7550, 7579), 'dubdub.scanner.Scanner', 'Scanner', ([], {'source': 'neste...
from pdpyras import APISession from requests.sessions import session from kafka import KafkaConsumer import json import sys print("Starting the alerts listener") # Kafka configurations kafka_broker = 'kafka:9092' alerts_topic = "high-temp-alerts-u5-1631586083-13100517045407420899" # PageDuty configurations api_token...
[ "sys.exc_info", "json.dumps" ]
[((1833, 1852), 'json.dumps', 'json.dumps', (['payload'], {}), '(payload)\n', (1843, 1852), False, 'import json\n'), ((1925, 1939), 'sys.exc_info', 'sys.exc_info', ([], {}), '()\n', (1937, 1939), False, 'import sys\n')]
from decimal import Decimal from src import profitable_music_genre EXPECTED_RESULT = [ (Decimal('826.65'), 'Rock'), (Decimal('382.14'), 'Latin'), (Decimal('261.36'), 'Metal'), (Decimal('241.56'), 'Alternative & Punk'), (Decimal('93.53'), 'TV Shows'), (Decimal('79.20'), 'Jazz'), (Decimal('6...
[ "src.profitable_music_genre.most_profitable_music_genre", "decimal.Decimal" ]
[((94, 111), 'decimal.Decimal', 'Decimal', (['"""826.65"""'], {}), "('826.65')\n", (101, 111), False, 'from decimal import Decimal\n'), ((127, 144), 'decimal.Decimal', 'Decimal', (['"""382.14"""'], {}), "('382.14')\n", (134, 144), False, 'from decimal import Decimal\n'), ((161, 178), 'decimal.Decimal', 'Decimal', (['""...
import multiprocessing as mp import os import queue import signal import torch class BaseCallback: def __init__(self, verbose: int = 0): self.model = None self.verbose = verbose def init_callback(self, model): self.model = model def on_training_start(self): pass def ...
[ "multiprocessing.Queue", "os.kill", "multiprocessing.Event" ]
[((1120, 1130), 'multiprocessing.Queue', 'mp.Queue', ([], {}), '()\n', (1128, 1130), True, 'import multiprocessing as mp\n'), ((1165, 1175), 'multiprocessing.Queue', 'mp.Queue', ([], {}), '()\n', (1173, 1175), True, 'import multiprocessing as mp\n'), ((1209, 1219), 'multiprocessing.Queue', 'mp.Queue', ([], {}), '()\n',...
import os import subprocess import tempfile import time import wave import simpleaudio def extract(file): ptmv_tempdir = os.path.join(tempfile.gettempdir(), "ptmv") if not os.path.exists(ptmv_tempdir): os.makedirs(ptmv_tempdir) snd_file = ptmv_tempdir + str(int(time.time())) + ".wav" command = "ffmpeg -i " + file ...
[ "os.path.exists", "wave.open", "simpleaudio.WaveObject.from_wave_read", "os.makedirs", "tempfile.gettempdir", "time.time" ]
[((538, 553), 'wave.open', 'wave.open', (['file'], {}), '(file)\n', (547, 553), False, 'import wave\n'), ((136, 157), 'tempfile.gettempdir', 'tempfile.gettempdir', ([], {}), '()\n', (155, 157), False, 'import tempfile\n'), ((175, 203), 'os.path.exists', 'os.path.exists', (['ptmv_tempdir'], {}), '(ptmv_tempdir)\n', (189...
# -*- coding: utf-8 -*- """ Created on Thu Dec 30 11:53:32 2021 @author: HP """ PATH = "Forest.png" from PIL import Image sprites = [Image.new("RGBA", (32, 32), (0,0,0,0)) for x in range(8*6)] current = 0 with Image.open(PATH) as im: curI = 0 curJ = 0 for curI in range(0,256,32): for curJ in range(...
[ "PIL.Image.new", "PIL.Image.open" ]
[((134, 175), 'PIL.Image.new', 'Image.new', (['"""RGBA"""', '(32, 32)', '(0, 0, 0, 0)'], {}), "('RGBA', (32, 32), (0, 0, 0, 0))\n", (143, 175), False, 'from PIL import Image\n'), ((211, 227), 'PIL.Image.open', 'Image.open', (['PATH'], {}), '(PATH)\n', (221, 227), False, 'from PIL import Image\n')]
from minibench import Benchmark import time class PauseBenchmark(Benchmark): times = 10 def bench_one_hundredth(self): time.sleep(.01) def bench_one_tenth(self): time.sleep(.1)
[ "time.sleep" ]
[((139, 155), 'time.sleep', 'time.sleep', (['(0.01)'], {}), '(0.01)\n', (149, 155), False, 'import time\n'), ((195, 210), 'time.sleep', 'time.sleep', (['(0.1)'], {}), '(0.1)\n', (205, 210), False, 'import time\n')]
# Check http://piwigo.com/ from urllib.parse import urlparse from os import path from plugins.cloudimport.cloud_platform import File, Folder from plugins.cloudimport.extensions.cloud_library import CloudLibrary class Platform(CloudLibrary): def __init__(self): super().__init__('Piwigo', 'http://{server_url...
[ "urllib.parse.urlparse", "plugins.cloudimport.cloud_platform.Folder", "os.path.splitext", "plugins.cloudimport.cloud_platform.File", "os.path.index" ]
[((994, 1007), 'urllib.parse.urlparse', 'urlparse', (['url'], {}), '(url)\n', (1002, 1007), False, 'from urllib.parse import urlparse\n'), ((1926, 1984), 'plugins.cloudimport.cloud_platform.Folder', 'Folder', (["result['name']", "result['url']", "result['nb_images']"], {}), "(result['name'], result['url'], result['nb_i...
""" <Reinforcement Learning and Control>(Year 2020) by <NAME> @ Intelligent Driving Lab, Tsinghua University OCP example for lane keeping problem in a circle road [Method] Model predictive control """ from casadi import * from config import DynamicsConfig import math from dynamics import...
[ "dynamics.VehicleDynamics" ]
[((1087, 1104), 'dynamics.VehicleDynamics', 'VehicleDynamics', ([], {}), '()\n', (1102, 1104), False, 'from dynamics import VehicleDynamics\n')]
#!/usr/bin/env python """Create benchmark for k nearest neighbor on unit sphere in R^k.""" # Scroll down to line 90 to "Adjust this" to add your experiment import random import numpy as np import os.path import logging import sys import Queue as queue import h5py import time logging.basicConfig(format='%(asctime)s %...
[ "logging.basicConfig", "random.uniform", "Queue.PriorityQueue", "time.time", "h5py.File", "numpy.array", "numpy.zeros", "numpy.linalg.norm", "logging.info" ]
[((279, 391), 'logging.basicConfig', 'logging.basicConfig', ([], {'format': '"""%(asctime)s %(levelname)s %(message)s"""', 'level': 'logging.DEBUG', 'stream': 'sys.stdout'}), "(format='%(asctime)s %(levelname)s %(message)s', level=\n logging.DEBUG, stream=sys.stdout)\n", (298, 391), False, 'import logging\n'), ((407...
from setuptools import setup setup(name='pyrosettacolabsetup', version='0.5', description='Mounts Google Drive for PyRosetta use in Google Colaboratory', url='https://github.com/kathyle9/pyrosettacolabsetup', author='kathyle9', author_email='<EMAIL>', license='MIT', packages=[...
[ "setuptools.setup" ]
[((30, 335), 'setuptools.setup', 'setup', ([], {'name': '"""pyrosettacolabsetup"""', 'version': '"""0.5"""', 'description': '"""Mounts Google Drive for PyRosetta use in Google Colaboratory"""', 'url': '"""https://github.com/kathyle9/pyrosettacolabsetup"""', 'author': '"""kathyle9"""', 'author_email': '"""<EMAIL>"""', '...
"""The pyccl package contains all of the submodules that are implemented in individual files in CCL. """ # flake8: noqa from pkg_resources import get_distribution, DistributionNotFound try: __version__ = get_distribution(__name__).version except DistributionNotFound: # package is not installed pass # Sets ...
[ "os.path.abspath", "os.environ.get", "pkg_resources.get_distribution" ]
[((428, 457), 'os.environ.get', 'environ.get', (['"""CCL_PARAM_FILE"""'], {}), "('CCL_PARAM_FILE')\n", (439, 457), False, 'from os import environ, path\n'), ((570, 600), 'os.environ.get', 'environ.get', (['"""CLASS_PARAM_DIR"""'], {}), "('CLASS_PARAM_DIR')\n", (581, 600), False, 'from os import environ, path\n'), ((208...
# coding: utf-8 # Copyright (c) 2016, 2022, Oracle and/or its affiliates. All rights reserved. # This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may c...
[ "oci.util.formatted_flat_dict", "oci.util.value_allowed_none_or_none_sentinel" ]
[((35018, 35043), 'oci.util.formatted_flat_dict', 'formatted_flat_dict', (['self'], {}), '(self)\n', (35037, 35043), False, 'from oci.util import formatted_flat_dict, NONE_SENTINEL, value_allowed_none_or_none_sentinel\n'), ((19160, 19219), 'oci.util.value_allowed_none_or_none_sentinel', 'value_allowed_none_or_none_sent...
#!/usr/bin/env python # pylint: disable=disallowed-name # pylint: disable=missing-class-docstring, # pylint: disable=missing-function-docstring, # pylint: disable=no-self-use import unittest from paramobject import ParametrizedObject, parameter, Parameter class TestParametrizedObject(unittest.TestCase): def te...
[ "unittest.main", "paramobject.Parameter", "paramobject.parameter" ]
[((4314, 4329), 'unittest.main', 'unittest.main', ([], {}), '()\n', (4327, 4329), False, 'import unittest\n'), ((413, 434), 'paramobject.Parameter', 'Parameter', ([], {'default': '(42)'}), '(default=42)\n', (422, 434), False, 'from paramobject import ParametrizedObject, parameter, Parameter\n'), ((453, 464), 'paramobje...
""" Copyright 2017-2020 Government of Canada - Public Services and Procurement Canada - buyandsell.gc.ca Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 ...
[ "logging.basicConfig", "logging.getLogger", "os.makedirs", "os.path.join", "von_anchor.frill.inis2dict", "os.path.realpath", "app.cache.MEM_CACHE.delete", "app.cache.MEM_CACHE.get", "app.cache.MEM_CACHE.set" ]
[((1022, 1054), 'os.makedirs', 'makedirs', (['dir_log'], {'exist_ok': '(True)'}), '(dir_log, exist_ok=True)\n', (1030, 1054), False, 'from os import makedirs\n'), ((1070, 1100), 'os.path.join', 'join', (['dir_log', '"""von_tails.log"""'], {}), "(dir_log, 'von_tails.log')\n", (1074, 1100), False, 'from os.path import di...
from django.contrib import admin from .models import Comment class CommentAdmin(admin.ModelAdmin): list_display = ("article", "to_comment", "status", "owner", "content") admin.site.register(Comment, CommentAdmin)
[ "django.contrib.admin.site.register" ]
[((178, 220), 'django.contrib.admin.site.register', 'admin.site.register', (['Comment', 'CommentAdmin'], {}), '(Comment, CommentAdmin)\n', (197, 220), False, 'from django.contrib import admin\n')]
from flask import Flask import os from application.model.entity.aula import Aula from application.model.entity.disciplina import Disciplina app = Flask(__name__, static_folder=os.path.abspath("application/view/static"), template_folder=os.path.abspath("application/view/templates")) aula1 = Aula(1, "Aula 1"...
[ "application.model.entity.aula.Aula", "application.model.entity.disciplina.Disciplina", "os.path.abspath" ]
[((304, 369), 'application.model.entity.aula.Aula', 'Aula', (['(1)', '"""Aula 1"""', '"""Introdução ao Linux"""', '"""Sistemas Operacionais"""'], {}), "(1, 'Aula 1', 'Introdução ao Linux', 'Sistemas Operacionais')\n", (308, 369), False, 'from application.model.entity.aula import Aula\n'), ((379, 464), 'application.mode...
from typing import Callable from putput.presets import displaCy from putput.presets import iob2 from putput.presets import luis from putput.presets import stochastic def get_preset(preset: str) -> Callable: """A factory that gets a 'preset' Callable. Args: preset: the preset's name. Returns: ...
[ "putput.presets.displaCy.preset", "putput.presets.luis.preset", "putput.presets.stochastic.preset", "putput.presets.iob2.preset" ]
[((1513, 1526), 'putput.presets.iob2.preset', 'iob2.preset', ([], {}), '()\n', (1524, 1526), False, 'from putput.presets import iob2\n'), ((1571, 1588), 'putput.presets.displaCy.preset', 'displaCy.preset', ([], {}), '()\n', (1586, 1588), False, 'from putput.presets import displaCy\n'), ((1629, 1642), 'putput.presets.lu...
from hqca.core import * import numpy as np from hqca.tools import * class SingleQubitHamiltonian(Hamiltonian): def __init__(self,sq=True, **kw ): self._order = 1 self._model = 'sq' self._qubOp = '' self.No_tot = 1 self.Ne_tot = 1 self.real = T...
[ "numpy.array", "numpy.zeros", "numpy.linalg.eigvalsh" ]
[((953, 988), 'numpy.zeros', 'np.zeros', (['(2, 2)'], {'dtype': 'np.complex_'}), '((2, 2), dtype=np.complex_)\n', (961, 988), True, 'import numpy as np\n'), ((1300, 1315), 'numpy.array', 'np.array', (['[mat]'], {}), '([mat])\n', (1308, 1315), True, 'import numpy as np\n'), ((1252, 1275), 'numpy.linalg.eigvalsh', 'np.li...
# Author: <NAME>, <EMAIL> # Dec 02, 2020 # Copyright 2020 <NAME> import numpy as np import pandas as pd from matplotlib import pyplot as plt from scipy.spatial import distance as dist import scipy.io import pickle ## TAPTC Dataset group_list = [1,2] instance_list = [0, 1, 2] ratio_deadline_list = [1, 2, 3, 4] rob...
[ "pandas.read_csv" ]
[((705, 761), 'pandas.read_csv', 'pd.read_csv', (['file_name'], {'sep': '""" """', 'header': 'None', 'skiprows': '(1)'}), "(file_name, sep=' ', header=None, skiprows=1)\n", (716, 761), True, 'import pandas as pd\n'), ((901, 957), 'pandas.read_csv', 'pd.read_csv', (['file_name'], {'sep': '""" """', 'header': 'None', 'sk...
#!/usr/bin/python2 import os from plumbum import local, FG from plumbum.cmd import git # the commits already tested HISTORY = '/home/cmr/benches/data' BUILDDIR = '/mnt/rustb' BENCH_OVERRIDE = '/home/cmr/benches/bench-override.txt' def run(hash): local['benchit.py'][hash] & FG for hash in open(BENCH_OVERRIDE)....
[ "os.listdir", "os.path.join" ]
[((413, 433), 'os.listdir', 'os.listdir', (['BUILDDIR'], {}), '(BUILDDIR)\n', (423, 433), False, 'import os\n'), ((461, 485), 'os.path.join', 'os.path.join', (['HISTORY', 'd'], {}), '(HISTORY, d)\n', (473, 485), False, 'import os\n')]
import RPi.GPIO as GPIO import time import datetime from ReadWriteConfig import * import Adafruit_ADS1x15 now = datetime.datetime.now() print("Starting CheckWater.py", str(now)) adc = Adafruit_ADS1x15.ADS1015() # Pick Sensors GAIN = 0 #import gain for adc reading SS_COUNT = 0 with open("PlantMgr.xml", "r") as f: ...
[ "RPi.GPIO.cleanup", "RPi.GPIO.setup", "RPi.GPIO.output", "time.sleep", "RPi.GPIO.PWM", "datetime.datetime.now", "Adafruit_ADS1x15.ADS1015", "RPi.GPIO.setmode" ]
[((113, 136), 'datetime.datetime.now', 'datetime.datetime.now', ([], {}), '()\n', (134, 136), False, 'import datetime\n'), ((187, 213), 'Adafruit_ADS1x15.ADS1015', 'Adafruit_ADS1x15.ADS1015', ([], {}), '()\n', (211, 213), False, 'import Adafruit_ADS1x15\n'), ((671, 693), 'RPi.GPIO.setmode', 'GPIO.setmode', (['GPIO.BCM'...
import discord import ctftime import os import random from discord.ext import commands, tasks from datetime import datetime # Token generated from https://discord.com/developers/applications # Keep this private, if exposed generate new one TOKEN = '' # Bot channel ID was grabbed from Settings > Appearance > Developer...
[ "random.choice", "discord.ext.commands.Bot", "datetime.datetime.now", "ctftime.get_events", "discord.ext.tasks.loop", "ctftime.days_to_secs" ]
[((408, 440), 'discord.ext.commands.Bot', 'commands.Bot', ([], {'command_prefix': '"""!"""'}), "(command_prefix='!')\n", (420, 440), False, 'from discord.ext import commands, tasks\n'), ((5482, 5504), 'discord.ext.tasks.loop', 'tasks.loop', ([], {'minutes': '(30)'}), '(minutes=30)\n', (5492, 5504), False, 'from discord...
import os import sys NAME = 'multipla' PACKAGE = __import__(NAME) AUTHOR, EMAIL = PACKAGE.__author__.rsplit(' ', 1) with open('docs/index.rst', 'r') as INDEX: DESCRIPTION = INDEX.readline() with open('README.rst', 'r') as README: LONG_DESCRIPTION = README.read() URL = 'https://github.com/monkeython/%s' % NA...
[ "setuptools.setup" ]
[((796, 819), 'setuptools.setup', 'setuptools.setup', ([], {}), '(**EGG)\n', (812, 819), False, 'import setuptools\n')]
# Generic-DiscordBot # author: github/adibarra # imports import os import time import uuid import enum import glob import traceback from zipfile import ZipFile from DB_prefsloader import PreferenceLoader class Importance(enum.IntEnum): """ Enum to keep track of logger message importance """ CRIT = 0 WARN...
[ "traceback.format_exc", "time.strftime", "os.path.isfile", "os.path.realpath", "os.umask", "os.stat" ]
[((1689, 1713), 'os.path.isfile', 'os.path.isfile', (['filePath'], {}), '(filePath)\n', (1703, 1713), False, 'import os\n'), ((1196, 1207), 'os.umask', 'os.umask', (['(0)'], {}), '(0)\n', (1204, 1207), False, 'import os\n'), ((1316, 1340), 'os.umask', 'os.umask', (['original_umask'], {}), '(original_umask)\n', (1324, 1...
from django.urls import path from django.conf.urls import url from store import views from .views import getStore, addStore, updateStore, deleteStore urlpatterns = [ url(r'getStore/', views.getStore), url(r'addStore/', views.addStore), url(r'updateStore/', views.updateStore), url(r'deleteStore/', views...
[ "django.conf.urls.url" ]
[((171, 203), 'django.conf.urls.url', 'url', (['"""getStore/"""', 'views.getStore'], {}), "('getStore/', views.getStore)\n", (174, 203), False, 'from django.conf.urls import url\n'), ((210, 242), 'django.conf.urls.url', 'url', (['"""addStore/"""', 'views.addStore'], {}), "('addStore/', views.addStore)\n", (213, 242), F...
# -*- coding: utf-8 -*- """ MIT License Copyright (c) 2020 <NAME>, SE; tamalone1 """ import unittest from PyNite import FEModel3D import sys from io import StringIO class Test_Spring_Elements(unittest.TestCase): ''' Tests of spring members.''' def setUp(self): # Suppress printed output temporarily ...
[ "io.StringIO", "PyNite.FEModel3D" ]
[((340, 350), 'io.StringIO', 'StringIO', ([], {}), '()\n', (348, 350), False, 'from io import StringIO\n'), ((679, 690), 'PyNite.FEModel3D', 'FEModel3D', ([], {}), '()\n', (688, 690), False, 'from PyNite import FEModel3D\n')]
""" Likelihood maximization script. This program is designed to be entirely separable from ATESA in that it can be called manually to perform likelihood maximization to user specifications and with arbitrary input files; however, it is required by ATESA's aimless shooting information error convergence criterion. """ i...
[ "scipy.stats.linregress", "numpy.sqrt", "matplotlib.pyplot.ylabel", "math.floor", "numpy.array", "argparse.Namespace", "os.path.exists", "numpy.histogram", "numpy.mean", "argparse.ArgumentParser", "matplotlib.pyplot.xlabel", "numpy.asarray", "numpy.max", "numpy.min", "sys.stdout.flush", ...
[((2677, 2699), 'sys.stdout.write', 'sys.stdout.write', (['text'], {}), '(text)\n', (2693, 2699), False, 'import sys\n'), ((2704, 2722), 'sys.stdout.flush', 'sys.stdout.flush', ([], {}), '()\n', (2720, 2722), False, 'import sys\n'), ((26829, 26904), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {'descripti...
#!/usr/bin/env python # -*- coding: utf-8 -*- # Splunk specific dependencies import sys, os from splunklib.searchcommands import dispatch, GeneratingCommand, Configuration, Option, validators, splunklib_logger as logger # Command specific dependencies import requests from requests.auth import HTTPDigestAuth import js...
[ "os.environ.keys", "requests.post", "splunklib.searchcommands.validators.Boolean", "requests.get", "splunklib.searchcommands.dispatch", "splunklib.searchcommands.validators.Integer", "splunklib.searchcommands.Option", "splunklib.searchcommands.Configuration" ]
[((565, 596), 'splunklib.searchcommands.Configuration', 'Configuration', ([], {'type': '"""reporting"""'}), "(type='reporting')\n", (578, 596), False, 'from splunklib.searchcommands import dispatch, GeneratingCommand, Configuration, Option, validators, splunklib_logger as logger\n'), ((4507, 4571), 'splunklib.searchcom...
"""Tests for RandoPony admin views and functionality. """ from datetime import datetime import unittest from unittest.mock import patch from pyramid import testing from pyramid_mailer import get_mailer from sqlalchemy import create_engine from randopony.models.meta import ( Base, DBSession, ) class TestCore...
[ "pyramid.testing.setUp", "pyramid_mailer.get_mailer", "randopony.views.admin.core.email_to_organizer", "randopony.models.meta.Base.metadata.create_all", "datetime.datetime", "randopony.models.Brevet.get_current", "pyramid.testing.tearDown", "sqlalchemy.create_engine", "randopony.views.admin.core.ema...
[((670, 685), 'pyramid.testing.setUp', 'testing.setUp', ([], {}), '()\n', (683, 685), False, 'from pyramid import testing\n'), ((703, 729), 'sqlalchemy.create_engine', 'create_engine', (['"""sqlite://"""'], {}), "('sqlite://')\n", (716, 729), False, 'from sqlalchemy import create_engine\n'), ((738, 770), 'randopony.mod...
""" WSGI entrypoint. """ from nunaserver.server import app if __name__ == "__main__": app.run()
[ "nunaserver.server.app.run" ]
[((91, 100), 'nunaserver.server.app.run', 'app.run', ([], {}), '()\n', (98, 100), False, 'from nunaserver.server import app\n')]
import os import pytest import csv_diff import logging import torch from unit_tests.t_utils import remove_tmp_dir, create_tmp_dir, __data_testing_dir__, __tmp_dir__ from ivadomed.loader import utils as imed_loader_utils from ivadomed.loader import loader as imed_loader logger = logging.getLogger(__name__) def setup_f...
[ "logging.getLogger", "torch.unique", "ivadomed.loader.loader.dropout_input", "os.path.join", "pytest.mark.parametrize", "unit_tests.t_utils.remove_tmp_dir", "torch.tensor", "ivadomed.loader.utils.BidsDataframe", "torch.rand", "unit_tests.t_utils.create_tmp_dir" ]
[((279, 306), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (296, 306), False, 'import logging\n'), ((1820, 2081), 'pytest.mark.parametrize', 'pytest.mark.parametrize', (['"""loader_parameters"""', "[{'path_data': [__data_testing_dir__], 'target_suffix': ['_seg-manual'],\n 'extensions...
from pymongo import MongoClient from bson.objectid import ObjectId from datetime import datetime as dt import pprint client = MongoClient() db = client['mongo_db_lab'] defs = db['definitions'] def random_word_requester(): ''' This function should return a random word and its definition and also log in the...
[ "pymongo.MongoClient", "pprint.pprint", "datetime.datetime.utcnow" ]
[((126, 139), 'pymongo.MongoClient', 'MongoClient', ([], {}), '()\n', (137, 139), False, 'from pymongo import MongoClient\n'), ((1150, 1170), 'pprint.pprint', 'pprint.pprint', (['duped'], {}), '(duped)\n', (1163, 1170), False, 'import pprint\n'), ((628, 639), 'datetime.datetime.utcnow', 'dt.utcnow', ([], {}), '()\n', (...
# Generated by Django 2.1.7 on 2019-05-20 19:58 from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('leads', '0011_auto_20190520_1217'), ] operations = [ migrations.AddField( model_name='lead', name='first_name', ...
[ "django.db.models.CharField" ]
[((334, 404), 'django.db.models.CharField', 'models.CharField', ([], {'max_length': '(255)', 'null': '(True)', 'verbose_name': '"""First name"""'}), "(max_length=255, null=True, verbose_name='First name')\n", (350, 404), False, 'from django.db import migrations, models\n'), ((525, 594), 'django.db.models.CharField', 'm...
import sys import torch from args import get_argparser, parse_args, get_aligner, get_bbox from os.path import join if __name__ == '__main__': parser = get_argparser() parser.add_argument('--align_start', help='align without vector voting the 2nd & 3rd sections, otherwise copy them', action='store_true') args...
[ "args.get_argparser", "args.get_bbox", "os.path.join", "args.get_aligner", "args.parse_args" ]
[((154, 169), 'args.get_argparser', 'get_argparser', ([], {}), '()\n', (167, 169), False, 'from args import get_argparser, parse_args, get_aligner, get_bbox\n'), ((323, 341), 'args.parse_args', 'parse_args', (['parser'], {}), '(parser)\n', (333, 341), False, 'from args import get_argparser, parse_args, get_aligner, get...
from cms.extensions.toolbar import ExtensionToolbar from cms.utils import get_language_list from django.utils.encoding import force_text from django.utils.translation import get_language_info class TitleExtensionToolbar(ExtensionToolbar): model = None insert_after = None def get_item_position(self, menu)...
[ "django.utils.translation.get_language_info", "django.utils.encoding.force_text", "cms.utils.get_language_list" ]
[((896, 935), 'cms.utils.get_language_list', 'get_language_list', (['self.current_site.pk'], {}), '(self.current_site.pk)\n', (913, 935), False, 'from cms.utils import get_language_list\n'), ((1940, 1973), 'django.utils.translation.get_language_info', 'get_language_info', (['title.language'], {}), '(title.language)\n',...
import datetime, os, pkg_resources, re, setuptools_scm from .. import __name__ as package_name try: if int(os.environ.get("_ASTROPATH_VERSION_NO_GIT", 0)): env_var_no_git = True raise LookupError env_var_no_git = False astropathversion = "v"+setuptools_scm.get_version(root="../..", relative_to=__file__) ...
[ "re.match", "os.environ.get", "setuptools_scm.get_version", "datetime.date.today", "pkg_resources.get_distribution" ]
[((586, 716), 're.match', 're.match', (['"""v(?P<version>[0-9]+(?:\\\\.[0-9]+)*)(?P<dev>\\\\.dev[0-9]+\\\\+g[0-9a-f]+)?(?P<date>\\\\.d[0-9]+)?"""', 'astropathversion'], {}), "(\n 'v(?P<version>[0-9]+(?:\\\\.[0-9]+)*)(?P<dev>\\\\.dev[0-9]+\\\\+g[0-9a-f]+)?(?P<date>\\\\.d[0-9]+)?'\n , astropathversion)\n", (594, 71...
### ### Precios de casas en boston ### from sklearn.datasets import load_boston from sklearn.model_selection import train_test_split from sklearn.linear_model import LinearRegression, Ridge from sklearn.neighbors import KNeighborsRegressor boston = load_boston() # Visualiza el nombre de los valores de los datos. pri...
[ "sklearn.model_selection.train_test_split", "sklearn.neighbors.KNeighborsRegressor", "sklearn.datasets.load_boston", "sklearn.linear_model.Ridge", "sklearn.linear_model.LinearRegression" ]
[((251, 264), 'sklearn.datasets.load_boston', 'load_boston', ([], {}), '()\n', (262, 264), False, 'from sklearn.datasets import load_boston\n'), ((440, 484), 'sklearn.model_selection.train_test_split', 'train_test_split', (['boston.data', 'boston.target'], {}), '(boston.data, boston.target)\n', (456, 484), False, 'from...
"""Ui.""" import logging import logging.config import os.path from datetime import datetime from PyQt5 import QtCore, QtGui from .lib import ( EquityChart, OptimizatimizedResultsTable, OptimizationTable, Portfolio, QuotesChart, ResultsTable, Settings, Symbol, TradesTable, get_...
[ "logging.getLogger", "PyQt5.QtCore.QDir.currentPath", "PyQt5.QtCore.pyqtSignal", "PyQt5.QtGui.QPushButton", "PyQt5.QtGui.QStandardItemModel", "PyQt5.QtGui.QStandardItem", "PyQt5.QtGui.QHBoxLayout", "PyQt5.QtGui.QTableView", "PyQt5.QtGui.QLabel", "PyQt5.QtGui.QFormLayout", "PyQt5.QtGui.QVBoxLayou...
[((411, 438), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (428, 438), False, 'import logging\n'), ((576, 601), 'PyQt5.QtCore.pyqtSignal', 'QtCore.pyqtSignal', (['object'], {}), '(object)\n', (593, 601), False, 'from PyQt5 import QtCore, QtGui\n'), ((777, 802), 'PyQt5.QtCore.pyqtSignal'...
import os import shutil import re from collections import OrderedDict import subprocess import numpy as np import atexit class Result: checkpoint = None log = None tarball = None board = None if __name__ == '__main__': results = OrderedDict() def load_files(): files = os.listdir() ...
[ "collections.OrderedDict", "os.listdir", "os.path.getsize", "numpy.unique", "subprocess.Popen", "re.match", "os.path.join", "os.path.isdir", "re.finditer", "shutil.rmtree", "os.path.islink", "os.walk", "os.remove" ]
[((251, 264), 'collections.OrderedDict', 'OrderedDict', ([], {}), '()\n', (262, 264), False, 'from collections import OrderedDict\n'), ((304, 316), 'os.listdir', 'os.listdir', ([], {}), '()\n', (314, 316), False, 'import os\n'), ((1110, 1131), 'os.path.islink', 'os.path.islink', (['start'], {}), '(start)\n', (1124, 113...
import os import time import logging from typing import List, Dict from collections import deque # from search import app from elasticsearch import Elasticsearch from elasticsearch.helpers import parallel_bulk from datetime import datetime # from config import ELASTICSEARCH_URL import threading from .utils import Uti...
[ "logging.debug", "elasticsearch.Elasticsearch", "elasticsearch.helpers.parallel_bulk", "os.environ.get", "cueSearch.services.GlobalDimensionServices.nonGlobalDimensionForIndexing", "time.time", "threading.Thread", "logging.info", "logging.error" ]
[((362, 423), 'os.environ.get', 'os.environ.get', (['"""ELASTICSEARCH_URL"""', '"""http://localhost:9200/"""'], {}), "('ELASTICSEARCH_URL', 'http://localhost:9200/')\n", (376, 423), False, 'import os\n'), ((1400, 1441), 'elasticsearch.Elasticsearch', 'Elasticsearch', ([], {'hosts': '[esHost]', 'timeout': '(30)'}), '(ho...
""" A simple example of NATDiscovery between ByteBlower ports. To discover the public IP address we will send a single packet upstream, capture this packet at the WAN side and finally pick it apart. This example demonstrates: * How to transmit a single custom packet. * How to cap...
[ "byteblowerll.byteblower.ByteBlower.InstanceGet", "time.sleep" ]
[((979, 1014), 'byteblowerll.byteblower.ByteBlower.InstanceGet', 'byteblower.ByteBlower.InstanceGet', ([], {}), '()\n', (1012, 1014), True, 'import byteblowerll.byteblower as byteblower\n'), ((2387, 2403), 'time.sleep', 'time.sleep', (['(0.01)'], {}), '(0.01)\n', (2397, 2403), False, 'import time\n')]
from rest_framework import serializers from game.serializers.question_serializer import QuestionSerializer class QuestionWithAnswerSerializer(QuestionSerializer): correct_answer = serializers.CharField()
[ "rest_framework.serializers.CharField" ]
[((187, 210), 'rest_framework.serializers.CharField', 'serializers.CharField', ([], {}), '()\n', (208, 210), False, 'from rest_framework import serializers\n')]
from sklearn import svm from ..data_wrappers import reject import numpy as np from scipy.stats import multivariate_normal from sklearn.mixture import GMM from sklearn.neighbors import KernelDensity class DensityEstimators(object): def __init__(self): self.models = {} self.unknown = {} self...
[ "numpy.sqrt", "numpy.linalg.pinv", "scipy.stats.multivariate_normal", "numpy.array", "numpy.cov", "numpy.arange", "numpy.rank", "sklearn.neighbors.KernelDensity", "numpy.subtract", "numpy.dot", "numpy.vstack", "numpy.eye", "numpy.random.multivariate_normal", "numpy.alen", "sklearn.svm.SV...
[((604, 629), 'sklearn.svm.SVC', 'svm.SVC', ([], {'probability': '(True)'}), '(probability=True)\n', (611, 629), False, 'from sklearn import svm\n'), ((709, 734), 'numpy.vstack', 'np.vstack', (['(X_kno, X_unk)'], {}), '((X_kno, X_unk))\n', (718, 734), True, 'import numpy as np\n'), ((1689, 1701), 'numpy.unique', 'np.un...
from pyparsing import LineEnd, LineStart, SkipTo, Regex from regparser.grammar import atomic, unified section = ( atomic.section_marker.copy().leaveWhitespace() + unified.part_section + SkipTo(LineEnd()) ) par = ( atomic.section.copy().leaveWhitespace() + unified.depth1_p + SkipTo(LineEnd()...
[ "regparser.grammar.atomic.section.copy", "pyparsing.Regex", "pyparsing.LineEnd", "regparser.grammar.atomic.section_marker.copy", "regparser.grammar.atomic.appendix_marker.copy", "pyparsing.LineStart", "regparser.grammar.atomic.paragraph_marker.copy" ]
[((1039, 1050), 'pyparsing.LineStart', 'LineStart', ([], {}), '()\n', (1048, 1050), False, 'from pyparsing import LineEnd, LineStart, SkipTo, Regex\n'), ((208, 217), 'pyparsing.LineEnd', 'LineEnd', ([], {}), '()\n', (215, 217), False, 'from pyparsing import LineEnd, LineStart, SkipTo, Regex\n'), ((311, 320), 'pyparsing...
import argparse def commandLineArgs(): parser = argparse.ArgumentParser(description='Process some integers.') parser.add_argument("--restrict-types", dest="restrict_types", default=False, action="store_true") parser.add_argument("--tes...
[ "argparse.ArgumentParser" ]
[((53, 114), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {'description': '"""Process some integers."""'}), "(description='Process some integers.')\n", (76, 114), False, 'import argparse\n')]
""" fakedata.py ==================================== Generate artificial pupil-data. """ import numpy as np import scipy.stats as stats from .baseline import * from .pupil import * def generate_pupil_data(event_onsets, fs=1000, pad=5000, baseline_lowpass=0.2, evoked_response_perc=0.02, respon...
[ "numpy.mean", "numpy.ceil", "numpy.ones", "numpy.random.rand", "numpy.any", "numpy.array", "numpy.zeros", "numpy.linspace", "numpy.random.randn", "numpy.concatenate", "numpy.zeros_like", "numpy.arange", "scipy.stats.truncnorm.rvs" ]
[((3093, 3111), 'numpy.any', 'np.any', (['(npars <= 0)'], {}), '(npars <= 0)\n', (3099, 3111), True, 'import numpy as np\n'), ((3162, 3180), 'numpy.any', 'np.any', (['(tmaxs <= 0)'], {}), '(tmaxs <= 0)\n', (3168, 3180), True, 'import numpy as np\n'), ((3481, 3492), 'numpy.zeros', 'np.zeros', (['n'], {}), '(n)\n', (3489...
""" <NAME> Advent of Code Day 5 Challenge 1 """ import sys import hashlib def md5_func(string): md5result = hashlib.md5() md5result.update(string.encode('utf-8')) return md5result.hexdigest() INTEGER_ID = 0 PASSWORD = "" if len(sys.argv) < 2: print("Please pass the puzzle input as a command line arg...
[ "sys.stdout.flush", "hashlib.md5" ]
[((590, 608), 'sys.stdout.flush', 'sys.stdout.flush', ([], {}), '()\n', (606, 608), False, 'import sys\n'), ((114, 127), 'hashlib.md5', 'hashlib.md5', ([], {}), '()\n', (125, 127), False, 'import hashlib\n')]
# Generated by Django 3.0.5 on 2020-12-07 16:56 import django.db.models.deletion from django.conf import settings from django.db import migrations, models class Migration(migrations.Migration): initial = True dependencies = [ migrations.swappable_dependency(settings.AUTH_USER_MODEL), ] ope...
[ "django.db.models.TextField", "django.db.models.IntegerField", "django.db.models.ForeignKey", "django.db.models.ManyToManyField", "django.db.models.SlugField", "django.db.models.AutoField", "django.db.models.DateTimeField", "django.db.migrations.swappable_dependency", "django.db.models.CharField" ]
[((247, 304), 'django.db.migrations.swappable_dependency', 'migrations.swappable_dependency', (['settings.AUTH_USER_MODEL'], {}), '(settings.AUTH_USER_MODEL)\n', (278, 304), False, 'from django.db import migrations, models\n'), ((437, 530), 'django.db.models.AutoField', 'models.AutoField', ([], {'auto_created': '(True)...
"""proj URL Configuration The `urlpatterns` list routes URLs to views. For more information please see: https://docs.djangoproject.com/en/3.1/topics/http/urls/ Examples: Function views 1. Add an import: from my_app import views 2. Add a URL to urlpatterns: path('', views.home, name='home') Class-based vi...
[ "django.urls.include", "books.views.BookDetailView.as_view", "books.views.DeleteBookView.as_view", "books.views.UpdateBookView.as_view", "books.views.CreateBookView.as_view", "django.conf.urls.static.static", "proj.auth_views.MyLoginView.as_view", "django.views.generic.RedirectView.as_view", "books....
[((3301, 3362), 'django.conf.urls.static.static', 'static', (['settings.MEDIA_URL'], {'document_root': 'settings.MEDIA_ROOT'}), '(settings.MEDIA_URL, document_root=settings.MEDIA_ROOT)\n', (3307, 3362), False, 'from django.conf.urls.static import static\n'), ((1224, 1255), 'django.urls.path', 'path', (['"""admin/"""', ...
""" Default audio settings. """ import numpy as np from modules.socket.settings import PACKAGE_SIZE # Number of sound channels. CHANNELS = 2 # The size of the streaming buffer, that needs to fit into the socket buffer. CHUNK_SIZE = PACKAGE_SIZE // CHANNELS // np.dtype(np.int16).itemsize # Sound device frame rate. ...
[ "numpy.dtype" ]
[((264, 282), 'numpy.dtype', 'np.dtype', (['np.int16'], {}), '(np.int16)\n', (272, 282), True, 'import numpy as np\n')]
import ast import base64 import jinja2 import logging import random import datetime from functools import reduce import werkzeug from odoo import SUPERUSER_ID from odoo import api, http from odoo.exceptions import UserError from odoo.http import request from odoo.modules import get_module_resource from odoo.addons.web...
[ "logging.getLogger", "odoo.http.request.render", "random.randrange", "odoo.http.request.env", "odoo.http.request.make_response", "odoo.modules.get_module_resource", "base64.b64encode", "base64.b64decode", "odoo.http.route", "werkzeug.wrappers.Response", "datetime.datetime.now", "ast.literal_ev...
[((377, 404), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (394, 404), False, 'import logging\n'), ((448, 471), 'datetime.datetime.now', 'datetime.datetime.now', ([], {}), '()\n', (469, 471), False, 'import datetime\n'), ((543, 637), 'odoo.http.route', 'http.route', (['[\'/free\', \'/fr...
""" Unit and regression test for the neuralxc package. """ import copy import os import sys from abc import ABC, abstractmethod import dill as pickle import matplotlib.pyplot as plt import numpy as np import pytest # Import package, test suite, and other packages as needed import neuralxc as xc from neuralxc.constan...
[ "numpy.allclose", "neuralxc.utils.SiestaDensityGetter", "neuralxc.formatter.SpeciesGrouper", "os.path.join", "neuralxc.formatter.Formatter", "numpy.sum", "numpy.array", "neuralxc.ml.transformer.GroupedVarianceThreshold", "pytest.mark.skipif", "os.path.abspath", "neuralxc.ml.transformer.GroupedSt...
[((3614, 3670), 'pytest.mark.skipif', 'pytest.mark.skipif', (['(not ase_found)'], {'reason': '"""requires ase"""'}), "(not ase_found, reason='requires ase')\n", (3632, 3670), False, 'import pytest\n'), ((658, 683), 'os.path.abspath', 'os.path.abspath', (['__file__'], {}), '(__file__)\n', (673, 683), False, 'import os\n...
# -*- coding: utf-8 -*- # Generated by Django 1.11.11 on 2019-01-25 09:24 from __future__ import unicode_literals from django.db import migrations class Migration(migrations.Migration): dependencies = [ ('focus', '0003_auto_20190125_1721'), ] operations = [ migrations.AlterModelOptions(...
[ "django.db.migrations.AlterModelOptions" ]
[((291, 400), 'django.db.migrations.AlterModelOptions', 'migrations.AlterModelOptions', ([], {'name': '"""article"""', 'options': "{'verbose_name': '文章', 'verbose_name_plural': '文章'}"}), "(name='article', options={'verbose_name': '文章',\n 'verbose_name_plural': '文章'})\n", (319, 400), False, 'from django.db import mig...
import numpy as np from PIL import Image nets = ["caffenet", "googlenet", "vggf", "vgg16", "vgg19"] def load(nets): res = [] for net in nets: data_path = "perturbations/perturbation_%s.npy" % net imgs = np.load(data_path, allow_pickle=True, encoding="latin1") # print(imgs.shape) ...
[ "numpy.uint8", "PIL.Image.new", "numpy.transpose", "numpy.load" ]
[((717, 784), 'PIL.Image.new', 'Image.new', (['"""RGB"""', '(n * width + interval * (n - 1), height)', '"""white"""'], {}), "('RGB', (n * width + interval * (n - 1), height), 'white')\n", (726, 784), False, 'from PIL import Image\n'), ((230, 286), 'numpy.load', 'np.load', (['data_path'], {'allow_pickle': '(True)', 'enc...
# coding=utf-8 # Copyright 2021 The Google Research Authors. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicab...
[ "numpy.tile", "jax.numpy.cos", "jax.numpy.sqrt", "jax.numpy.asarray", "numpy.array", "jax.numpy.maximum", "jax.numpy.minimum" ]
[((3342, 3377), 'jax.numpy.asarray', 'jnp.asarray', (['ret'], {'dtype': 'jnp.float32'}), '(ret, dtype=jnp.float32)\n', (3353, 3377), True, 'import jax.numpy as jnp\n'), ((3575, 3605), 'numpy.tile', 'np.tile', (['x[-1]', '(batch_pad, 1)'], {}), '(x[-1], (batch_pad, 1))\n', (3582, 3605), True, 'import numpy as np\n'), ((...
# -------------------------------------------------------- # Swin Transformer # Copyright (c) 2021 Microsoft # Licensed under The MIT License [see LICENSE for details] # Written by <NAME> # -------------------------------------------------------- import numpy as np import torch import torch.distributed as dist from dat...
[ "torch.utils.data.DistributedSampler", "torchvision.transforms.CenterCrop", "torch.distributed.get_rank", "datasets.CustomDataset", "torchvision.transforms.RandomCrop", "torchvision.transforms.Normalize", "torch.utils.data.DataLoader", "timm.data.transforms._pil_interp", "torchvision.transforms.ToTe...
[((1275, 1296), 'torch.distributed.get_world_size', 'dist.get_world_size', ([], {}), '()\n', (1294, 1296), True, 'import torch.distributed as dist\n'), ((1315, 1330), 'torch.distributed.get_rank', 'dist.get_rank', ([], {}), '()\n', (1328, 1330), True, 'import torch.distributed as dist\n'), ((1971, 2167), 'torch.utils.d...