code
stringlengths
22
1.05M
apis
listlengths
1
3.31k
extract_api
stringlengths
75
3.25M
# Generated by Django 3.1.12 on 2021-08-18 21:24 from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('reo', '0115_auto_20210810_1550'), ] operations = [ migrations.AddField( model_name='chpmodel', name='supplementary_...
[ "django.db.models.FloatField" ]
[((358, 398), 'django.db.models.FloatField', 'models.FloatField', ([], {'blank': '(True)', 'null': '(True)'}), '(blank=True, null=True)\n', (375, 398), False, 'from django.db import migrations, models\n'), ((550, 590), 'django.db.models.FloatField', 'models.FloatField', ([], {'blank': '(True)', 'null': '(True)'}), '(bl...
import factory class ServiceCategoryFactory(factory.django.DjangoModelFactory): class Meta: model = "core.ServiceCategory" name = factory.Sequence(lambda n: f"Service Category {n}") slug = factory.Sequence(lambda n: f"service-category-{n}") description = factory.Faker("sentence") icon = "...
[ "factory.Faker", "factory.Sequence", "factory.SubFactory" ]
[((149, 200), 'factory.Sequence', 'factory.Sequence', (["(lambda n: f'Service Category {n}')"], {}), "(lambda n: f'Service Category {n}')\n", (165, 200), False, 'import factory\n'), ((212, 263), 'factory.Sequence', 'factory.Sequence', (["(lambda n: f'service-category-{n}')"], {}), "(lambda n: f'service-category-{n}')\n...
# Generated by Django 2.1.7 on 2019-03-22 01:36 from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('nxos_config_import', '0009_auto_20190322_1408'), ] operations = [ migrations.AddField( model_name='objectconfigurationstatus', ...
[ "django.db.models.URLField" ]
[((366, 397), 'django.db.models.URLField', 'models.URLField', ([], {'default': '"""none"""'}), "(default='none')\n", (381, 397), False, 'from django.db import migrations, models\n')]
import logging import psycopg2 import psycopg2.extras import socket import sys import time from cluster_under_test import * class DbRetriable: """ Wrapper around psycopg2, which offers convenient retry functionality. If connection to postgres is lost during query execution or between queries, retry wi...
[ "psycopg2.connect", "sys.stdout.flush", "time.sleep", "sys.stdout.write" ]
[((1299, 1360), 'psycopg2.connect', 'psycopg2.connect', ([], {'host': 'resolved'}), '(host=resolved, **self.other_connection_args)\n', (1315, 1360), False, 'import psycopg2\n'), ((2585, 2602), 'time.sleep', 'time.sleep', (['delay'], {}), '(delay)\n', (2595, 2602), False, 'import time\n'), ((2413, 2434), 'sys.stdout.wri...
#!/usr/bin/python3 import json from argparse import ArgumentParser def get_args(): p = ArgumentParser(description='Merge CLOSURE xdconf.ini files') p.add_argument('-f', '--files', required=True, type=str, help='Input files') p.add_argument('-o', '--outfile', required=False, type=str, default='xdconf.ini',...
[ "json.load", "json.dump", "argparse.ArgumentParser" ]
[((97, 157), 'argparse.ArgumentParser', 'ArgumentParser', ([], {'description': '"""Merge CLOSURE xdconf.ini files"""'}), "(description='Merge CLOSURE xdconf.ini files')\n", (111, 157), False, 'from argparse import ArgumentParser\n'), ((1406, 1437), 'json.dump', 'json.dump', (['data', 'outf'], {'indent': '(2)'}), '(data...
''' Copyright (c) The Dojo Foundation 2011. All Rights Reserved. Copyright (c) IBM Corporation 2008, 2011. All Rights Reserved. ''' # std lib import ConfigParser # tornado import tornado.web from base import AuthBase class IniAuth(AuthBase): cookieName = 'coweb.auth.ini.username' def __init__(self, container, ...
[ "ConfigParser.ConfigParser" ]
[((1083, 1110), 'ConfigParser.ConfigParser', 'ConfigParser.ConfigParser', ([], {}), '()\n', (1108, 1110), False, 'import ConfigParser\n')]
from django.conf import settings from django_filters.rest_framework import DjangoFilterBackend from django_filters.rest_framework import UUIDFilter from rest_framework.permissions import IsAuthenticated from contentcuration.celery import app from contentcuration.models import Channel from contentcuration.models import...
[ "contentcuration.celery.app.control.revoke", "django_filters.rest_framework.UUIDFilter", "contentcuration.celery.app.AsyncResult", "contentcuration.models.Task.objects.all", "contentcuration.models.Task.objects.filter", "contentcuration.models.Channel.objects.all" ]
[((563, 598), 'django_filters.rest_framework.UUIDFilter', 'UUIDFilter', ([], {'method': '"""filter_channel"""'}), "(method='filter_channel')\n", (573, 598), False, 'from django_filters.rest_framework import UUIDFilter\n'), ((1036, 1054), 'contentcuration.models.Task.objects.all', 'Task.objects.all', ([], {}), '()\n', (...
from django.shortcuts import render from django.http import HttpResponse, JsonResponse from django.core import serializers from .models import Event from .serializers import EventSerializer from django.db.models import Q import datetime def get_events(request): startDate = request.GET.get('start') endDate = r...
[ "django.http.JsonResponse", "datetime.datetime.strptime", "datetime.date.today", "django.db.models.Q", "datetime.timedelta" ]
[((352, 373), 'datetime.date.today', 'datetime.date.today', ([], {}), '()\n', (371, 373), False, 'import datetime\n'), ((382, 403), 'datetime.date.today', 'datetime.date.today', ([], {}), '()\n', (401, 403), False, 'import datetime\n'), ((796, 799), 'django.db.models.Q', 'Q', ([], {}), '()\n', (797, 799), False, 'from ...
# ------------------------------------ # Copyright (c) Microsoft Corporation. # Licensed under the MIT License. # ------------------------------------ from azure.keyvault.certificates._shared import HttpChallengeCache from azure.keyvault.certificates._shared.client_base import DEFAULT_VERSION from devtools_testutils im...
[ "azure.keyvault.certificates._shared.HttpChallengeCache.clear", "pytest.skip" ]
[((640, 666), 'azure.keyvault.certificates._shared.HttpChallengeCache.clear', 'HttpChallengeCache.clear', ([], {}), '()\n', (664, 666), False, 'from azure.keyvault.certificates._shared import HttpChallengeCache\n'), ((1445, 1518), 'pytest.skip', 'pytest.skip', (['"""This test only uses the default API version for live ...
import logging from pyramid.httpexceptions import HTTPNotImplemented from pyramid.renderers import render, render_to_response log = logging.getLogger(__name__) class RestView(object): renderers = {} def __init__(self, request): self.request = request self.params = request.params se...
[ "logging.getLogger", "pyramid.renderers.render", "pyramid.httpexceptions.HTTPNotImplemented", "pyramid.renderers.render_to_response" ]
[((134, 161), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (151, 161), False, 'import logging\n'), ((527, 550), 'pyramid.renderers.render', 'render', (['*args'], {}), '(*args, **kwargs)\n', (533, 550), False, 'from pyramid.renderers import render, render_to_response\n'), ((647, 682), 'p...
# Let's make an database # # this is like the worst code ever # # just to make an test DB for burn the subs import time from datetime import datetime from bts.dataBaseClass import Sub def main(): fileName = open("subscriberListTest.txt") print("making:") for entry in fileName: en...
[ "bts.dataBaseClass.Sub.create", "time.sleep", "datetime.datetime.utcnow" ]
[((360, 377), 'datetime.datetime.utcnow', 'datetime.utcnow', ([], {}), '()\n', (375, 377), False, 'from datetime import datetime\n'), ((407, 512), 'bts.dataBaseClass.Sub.create', 'Sub.create', ([], {'userName': 'entry', 'entryTime': 'dateTime', 'status': '(2)', 'fontSize': '(72)', 'positionX': '(1000)', 'positionY': '(...
from getpass import getpass from nxapi_plumbing import Device from lxml import etree from pprint import pprint as pp # Disable Self-signed Certificate Warnings import requests from requests.packages.urllib3.exceptions import InsecureRequestWarning requests.packages.urllib3.disable_warnings(InsecureRequestWarning) '''...
[ "requests.packages.urllib3.disable_warnings", "getpass.getpass", "lxml.etree.tostring" ]
[((249, 315), 'requests.packages.urllib3.disable_warnings', 'requests.packages.urllib3.disable_warnings', (['InsecureRequestWarning'], {}), '(InsecureRequestWarning)\n', (291, 315), False, 'import requests\n'), ((1227, 1236), 'getpass.getpass', 'getpass', ([], {}), '()\n', (1234, 1236), False, 'from getpass import getp...
import os import os.path import webapp2 import logging from webapp2 import WSGIApplication, Route from google.appengine.api import users # hack until we can make this public cache = dict() class Content(webapp2.RequestHandler): def get(self, *args, **kwargs): urlPath = args[0] root = os.path.split(__file__...
[ "webapp2.Route", "os.path.join", "logging.exception", "os.path.isfile", "os.path.split" ]
[((341, 380), 'os.path.join', 'os.path.join', (['root', '"""404"""', '"""index.html"""'], {}), "(root, '404', 'index.html')\n", (353, 380), False, 'import os\n'), ((1781, 1813), 'webapp2.Route', 'Route', (['"""/<:.*>"""'], {'handler': 'Content'}), "('/<:.*>', handler=Content)\n", (1786, 1813), False, 'from webapp2 impo...
# This file is used to define an AML pipeline for training the teacher on new labeled data import json import shutil import os from azureml.core import Workspace, Run, Experiment, Datastore from azureml.data.data_reference import DataReference from azureml.pipeline.core.schedule import ScheduleRecurrence, Schedule fr...
[ "azureml.core.runconfig.CondaDependencies.create", "azureml.train.hyperdrive.loguniform", "azureml.core.compute.AmlCompute.provisioning_configuration", "azureml.core.compute.AmlCompute", "azureml.pipeline.core.PipelineData", "utils.azure.disable_pipeline", "azureml.pipeline.steps.HyperDriveStep", "azu...
[((1277, 1329), 'azureml.core.Workspace.from_config', 'Workspace.from_config', ([], {'path': 'config_json', 'auth': 'svc_pr'}), '(path=config_json, auth=svc_pr)\n', (1298, 1329), False, 'from azureml.core import Workspace, Run, Experiment, Datastore\n'), ((1677, 1718), 'os.makedirs', 'os.makedirs', (['script_folder'], ...
""" ------------------------------------------------------------------------- Library of RTL queues ------------------------------------------------------------------------- Author : <NAME> Date : Mar 23, 2019 """ from pymtl3 import * from pymtl3.stdlib.ifcs import DeqIfcRTL, EnqIfcRTL from pymtl3.stdlib.rtl impor...
[ "pymtl3.stdlib.ifcs.DeqIfcRTL", "pymtl3.stdlib.rtl.Mux", "pymtl3.stdlib.ifcs.EnqIfcRTL", "pymtl3.stdlib.rtl.RegisterFile" ]
[((3348, 3368), 'pymtl3.stdlib.ifcs.EnqIfcRTL', 'EnqIfcRTL', (['EntryType'], {}), '(EntryType)\n', (3357, 3368), False, 'from pymtl3.stdlib.ifcs import DeqIfcRTL, EnqIfcRTL\n'), ((3385, 3405), 'pymtl3.stdlib.ifcs.DeqIfcRTL', 'DeqIfcRTL', (['EntryType'], {}), '(EntryType)\n', (3394, 3405), False, 'from pymtl3.stdlib.ifc...
# coding=utf-8 # # Copyright Zucker # # Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in w...
[ "zipfile.ZipFile", "re.compile", "copy.deepcopy", "os.path.islink", "os.walk", "os.path.exists", "os.listdir", "os.readlink", "pathlib.Path", "sys.argv.append", "os.path.isdir", "subprocess.call", "os.mkdir", "shutil.Error", "shutil.copystat", "os.path.getsize", "subprocess.check_cal...
[((30736, 30765), 'sys.argv.append', 'sys.argv.append', (['"""ZuckerDemo"""'], {}), "('ZuckerDemo')\n", (30751, 30765), False, 'import sys\n'), ((30839, 30850), 'os.getcwd', 'os.getcwd', ([], {}), '()\n', (30848, 30850), False, 'import os\n'), ((30879, 30914), 'os.path.join', 'os.path.join', (['currentPath', '"""output...
# Licensed to the Apache Software Foundation (ASF) under one # or more contributor license agreements. See the NOTICE file # distributed with this work for additional information # regarding copyright ownership. The ASF licenses this file # to you under the Apache License, Version 2.0 (the # "License"); you may not u...
[ "pytest.mark.parametrize", "unittest.mock.MagicMock", "tests.test_utils.mock_operators.MockOperator" ]
[((1135, 1650), 'pytest.mark.parametrize', 'pytest.mark.parametrize', (['"""region_name, aws_partition,keywords,expected_value"""', "[('eu-central-1', 'aws', {}, {'region_name': 'eu-central-1', 'aws_domain':\n 'aws.amazon.com'}), ('cn-north-1', 'aws-cn', {}, {'region_name':\n 'cn-north-1', 'aws_domain': 'amazonaw...
from django.db import models from articles.models import Article from users.models import User class Comment(models.Model): id = models.IntegerField(primary_key=True, editable=False, auto_created=True) author = models.ForeignKey(User, ...
[ "django.db.models.ForeignKey", "django.db.models.DateField", "django.db.models.TextField", "django.db.models.IntegerField" ]
[((136, 208), 'django.db.models.IntegerField', 'models.IntegerField', ([], {'primary_key': '(True)', 'editable': '(False)', 'auto_created': '(True)'}), '(primary_key=True, editable=False, auto_created=True)\n', (155, 208), False, 'from django.db import models\n'), ((280, 354), 'django.db.models.ForeignKey', 'models.For...
# -*- coding: utf-8 -*- import requests_mock from chaoslib.run import EventHandlerRegistry from logzero import logger from chaosgrafana.controls.loki import configure_control def test_sending_to_loki(): with requests_mock.Mocker() as m: m.post("http://localhost.test:3100/loki/api/v1/push", status_code=20...
[ "chaoslib.run.EventHandlerRegistry", "requests_mock.Mocker", "logzero.logger.error", "chaosgrafana.controls.loki.configure_control" ]
[((215, 237), 'requests_mock.Mocker', 'requests_mock.Mocker', ([], {}), '()\n', (235, 237), False, 'import requests_mock\n'), ((343, 365), 'chaoslib.run.EventHandlerRegistry', 'EventHandlerRegistry', ([], {}), '()\n', (363, 365), False, 'from chaoslib.run import EventHandlerRegistry\n'), ((374, 586), 'chaosgrafana.cont...
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """数据模型""" __author__ = 'stone' import re import LianJiaWangConfig class LianJiaWangModel(object): def __init__(self): self.title = '' self.des = '' self.time = '' """命名 bottom 是为了跟 HTML 代码中的名字保持一致并方便批量赋值;但其实的意义是 features""" ...
[ "re.findall", "re.compile" ]
[((968, 994), 're.compile', 're.compile', (['"""href="(.+?)\\""""'], {}), '(\'href="(.+?)"\')\n', (978, 994), False, 'import re\n'), ((1016, 1044), 're.findall', 're.findall', (['pattern', 'striped'], {}), '(pattern, striped)\n', (1026, 1044), False, 'import re\n')]
#!/usr/bin/env python3 from xmlrpc.client import Boolean from sqlalchemy import false, true import yaml import os import rospy from ros_pepper_pkg.srv import * from std_msgs.msg import String from std_msgs.msg import Bool def callback(message, tts_service, publisher): """ Topic callback. Parameters --...
[ "yaml.full_load", "rospy.Publisher", "rospy.init_node", "rospy.get_param", "rospy.ServiceProxy", "os.path.join", "rospy.spin", "os.path.abspath", "rospy.logdebug", "rospy.wait_for_service" ]
[((749, 787), 'rospy.logdebug', 'rospy.logdebug', (['"""Tts interface READY."""'], {}), "('Tts interface READY.')\n", (763, 787), False, 'import rospy\n'), ((792, 804), 'rospy.spin', 'rospy.spin', ([], {}), '()\n', (802, 804), False, 'import rospy\n'), ((1114, 1140), 'rospy.init_node', 'rospy.init_node', (['node_name']...
import sys import random from test_base import * class TestBlockLD(TestBase): def generate(self): self.clear_tag() for n in range(50000): store_not_load = random.randint(0,1) tag = random.randint(0, 15) index = random.randint(0,self.sets_p-1) taddr = self.get_addr(tag,index) ...
[ "random.randint" ]
[((176, 196), 'random.randint', 'random.randint', (['(0)', '(1)'], {}), '(0, 1)\n', (190, 196), False, 'import random\n'), ((208, 229), 'random.randint', 'random.randint', (['(0)', '(15)'], {}), '(0, 15)\n', (222, 229), False, 'import random\n'), ((244, 278), 'random.randint', 'random.randint', (['(0)', '(self.sets_p -...
import sys import json import nltk import nltk from nltk.corpus import stopwords from nltk.tokenize import word_tokenize, sent_tokenize class Partsofspeech(): def pos(txt): tokens = nltk.word_tokenize(txt) return (nltk.pos_tag(tokens)) # stop_words = set(stopwords.words('english'))
[ "nltk.pos_tag", "nltk.word_tokenize" ]
[((202, 225), 'nltk.word_tokenize', 'nltk.word_tokenize', (['txt'], {}), '(txt)\n', (220, 225), False, 'import nltk\n'), ((242, 262), 'nltk.pos_tag', 'nltk.pos_tag', (['tokens'], {}), '(tokens)\n', (254, 262), False, 'import nltk\n')]
#!/bin/python3 # author: <NAME> import sys import time basic_format = '{desc} {n:02d}/{total:02d} [{elapsed}/{remaining}] [{item_duration}/item]' reverse_format = '{n:02d}/{total:02d} {desc} [{elapsed}/{remaining}] [{item_duration}/item]' class Line(object): def __init__(self, iterable=None, total=None, format=...
[ "time.time" ]
[((902, 913), 'time.time', 'time.time', ([], {}), '()\n', (911, 913), False, 'import time\n'), ((995, 1006), 'time.time', 'time.time', ([], {}), '()\n', (1004, 1006), False, 'import time\n'), ((1263, 1274), 'time.time', 'time.time', ([], {}), '()\n', (1272, 1274), False, 'import time\n')]
import game_framework import pico2d import start_state pico2d.open_canvas() game_framework.run(start_state) pico2d.close_canvas()
[ "game_framework.run", "pico2d.close_canvas", "pico2d.open_canvas" ]
[((57, 77), 'pico2d.open_canvas', 'pico2d.open_canvas', ([], {}), '()\n', (75, 77), False, 'import pico2d\n'), ((78, 109), 'game_framework.run', 'game_framework.run', (['start_state'], {}), '(start_state)\n', (96, 109), False, 'import game_framework\n'), ((110, 131), 'pico2d.close_canvas', 'pico2d.close_canvas', ([], {...
import os import subprocess from pathlib import Path from ..views.viewhelper import delay_refresh_detail from ..helper.config import config def edit(filepath: Path, loop): if isinstance(filepath, str): filepath = Path(filepath) editor = os.environ.get('EDITOR', 'vi').lower() # vim if editor ==...
[ "subprocess.check_output", "pathlib.Path", "pathlib.Path.cwd", "subprocess.check_call", "os.environ.get", "os.chdir", "subprocess.call", "os.system" ]
[((1274, 1288), 'os.system', 'os.system', (['cmd'], {}), '(cmd)\n', (1283, 1288), False, 'import os\n'), ((1349, 1363), 'os.system', 'os.system', (['cmd'], {}), '(cmd)\n', (1358, 1363), False, 'import os\n'), ((227, 241), 'pathlib.Path', 'Path', (['filepath'], {}), '(filepath)\n', (231, 241), False, 'from pathlib impor...
import json import sys import os from algoliasearch.search_client import SearchClient # This script updates the product list in products/data/product_list.json # This is done by downloading the data from algolia index and replacing the product_list.json with the new content. def main(): product_list_filepa...
[ "algoliasearch.search_client.SearchClient.create", "json.dump", "os.getcwd" ]
[((458, 528), 'algoliasearch.search_client.SearchClient.create', 'SearchClient.create', (['"""9SXIDIVU1E"""', "os.environ['ALGOLIA_ADMIN_API_KEY']"], {}), "('9SXIDIVU1E', os.environ['ALGOLIA_ADMIN_API_KEY'])\n", (477, 528), False, 'from algoliasearch.search_client import SearchClient\n'), ((325, 336), 'os.getcwd', 'os....
#!/usr/bin/env python import asyncio import io import http.server import json import numpy import os import queue import re import sys import threading import websockets from obspy.io.quakeml.core import Pickler from obspy.core.stream import Stream from obspy.core.utcdatetime import UTCDateTime from obspy.core.event....
[ "logging.getLogger", "threading.Thread.__init__", "obspy.core.stream.Stream", "logging.StreamHandler", "obspy.core.event.catalog.Catalog", "json.JSONEncoder.default", "re.compile", "asyncio.get_event_loop", "asyncio.new_event_loop", "json.dumps", "asyncio.Queue", "re.match", "io.BytesIO", ...
[((471, 501), 'logging.getLogger', 'logging.getLogger', (['"""viewobspy"""'], {}), "('viewobspy')\n", (488, 501), False, 'import logging\n'), ((550, 573), 'logging.StreamHandler', 'logging.StreamHandler', ([], {}), '()\n', (571, 573), False, 'import logging\n'), ((13750, 13783), 're.compile', 're.compile', (['"""eventi...
#!/usr/bin/env python # Created by "Thieu" at 00:16, 15/03/2022 ----------% # Email: <EMAIL> % # Github: https://github.com/thieu1995 % ...
[ "mealpy.bio_based.BBO.OriginalBBO", "pytest.mark.parametrize", "numpy.sum", "pytest.raises", "pytest.fixture", "mealpy.bio_based.BBO.BaseBBO" ]
[((485, 515), 'pytest.fixture', 'pytest.fixture', ([], {'scope': '"""module"""'}), "(scope='module')\n", (499, 515), False, 'import pytest\n'), ((3242, 3489), 'pytest.mark.parametrize', 'pytest.mark.parametrize', (['"""problem, p_m, system_code"""', "[(problem, None, 0), (problem, 'hello', 0), (problem, -1.0, 0), (prob...
from io import StringIO from hypothesis import given, strategies from intentionally_blank import api def test_format_empty_with_empty_string(): with StringIO() as in_file, StringIO() as out_file: api.format_from_file_to_file(in_file, out_file, format_names=["identity"]) assert len(out_file.getva...
[ "hypothesis.strategies.text", "io.StringIO", "intentionally_blank.api.format_from_file_to_file" ]
[((157, 167), 'io.StringIO', 'StringIO', ([], {}), '()\n', (165, 167), False, 'from io import StringIO\n'), ((180, 190), 'io.StringIO', 'StringIO', ([], {}), '()\n', (188, 190), False, 'from io import StringIO\n'), ((212, 286), 'intentionally_blank.api.format_from_file_to_file', 'api.format_from_file_to_file', (['in_fi...
""" Mask R-CNN Train on the toy Balloon dataset and implement color splash effect. Copyright (c) 2018 Matterport, Inc. Licensed under the MIT License (see LICENSE for details) Written by <NAME> ------------------------------------------------------------""" import os import sys import json import numpy as...
[ "mrcnn.model.MaskRCNN", "os.path.join", "numpy.array", "mrcnn.visualize.display_instances", "os.path.abspath", "sys.path.append", "matplotlib.pyplot.subplots" ]
[((541, 566), 'os.path.abspath', 'os.path.abspath', (['"""../../"""'], {}), "('../../')\n", (556, 566), False, 'import os\n'), ((590, 615), 'sys.path.append', 'sys.path.append', (['ROOT_DIR'], {}), '(ROOT_DIR)\n', (605, 615), False, 'import sys\n'), ((1185, 1215), 'os.path.join', 'os.path.join', (['ROOT_DIR', '"""logs"...
import os, csv path = 'F:\Movies-TV' with open('C:\wsl\local-movies\db\movies.csv', 'w', newline='') as csvfile: writer = csv.writer(csvfile) for root,dirs, files in os.walk(path): for folders in dirs: if folders == "Subs" or folders == "Subtitles" or folders == "Other" or folders == "subtitles": ...
[ "csv.writer", "os.walk" ]
[((126, 145), 'csv.writer', 'csv.writer', (['csvfile'], {}), '(csvfile)\n', (136, 145), False, 'import os, csv\n'), ((172, 185), 'os.walk', 'os.walk', (['path'], {}), '(path)\n', (179, 185), False, 'import os, csv\n')]
import io import logging from functools import cached_property from .base_experiment import BaseExperiment from .utils import insert_suffix class TqdmToLogger(io.StringIO): buffer = '' def __init__(self, logger, level=logging.DEBUG): super().__init__() self.logger = logger self.level = ...
[ "logging.Formatter", "logging.FileHandler", "logging.info" ]
[((1041, 1076), 'logging.FileHandler', 'logging.FileHandler', (['name'], {'mode': '"""w"""'}), "(name, mode='w')\n", (1060, 1076), False, 'import logging\n'), ((1278, 1313), 'logging.info', 'logging.info', (['"""Logging in %s"""', 'name'], {}), "('Logging in %s', name)\n", (1290, 1313), False, 'import logging\n'), ((11...
#!/usr/bin/python3 # Entry-point for running from the CLI when not installed via Pip, Pip will handle the console_scripts entry_points's from setup.py # It's recommended to use `pip3 install changedetection.io` and start with `changedetection.py` instead, it will be linkd to your global path. # or Docker. # Read more ...
[ "mb_changedetectionio.mb_changedetection.main" ]
[((457, 482), 'mb_changedetectionio.mb_changedetection.main', 'mb_changedetection.main', ([], {}), '()\n', (480, 482), False, 'from mb_changedetectionio import mb_changedetection\n')]
'''Autogenerated by get_gl_extensions script, do not edit!''' from OpenGL import platform as _p from OpenGL.GL import glget EXTENSION_NAME = 'GL_INGR_color_clamp' _p.unpack_constants( """GL_RED_MIN_CLAMP_INGR 0x8560 GL_GREEN_MIN_CLAMP_INGR 0x8561 GL_BLUE_MIN_CLAMP_INGR 0x8562 GL_ALPHA_MIN_CLAMP_INGR 0x8563 GL_RED_MAX_C...
[ "OpenGL.extensions.hasGLExtension", "OpenGL.GL.glget.addGLGetConstant" ]
[((444, 495), 'OpenGL.GL.glget.addGLGetConstant', 'glget.addGLGetConstant', (['GL_RED_MIN_CLAMP_INGR', '(1,)'], {}), '(GL_RED_MIN_CLAMP_INGR, (1,))\n', (466, 495), False, 'from OpenGL.GL import glget\n'), ((498, 551), 'OpenGL.GL.glget.addGLGetConstant', 'glget.addGLGetConstant', (['GL_GREEN_MIN_CLAMP_INGR', '(1,)'], {}...
# -*- coding: utf-8 -*- """ Created on Tue Nov 10 22:12:12 2020 @author: vxr131730 """ import glob import os import sys import random import time import numpy as np import cv2 from test import * from casadi import * from numpy import random as npr from casadi.tools import * try: sys.path.appe...
[ "time.sleep", "carla.Client", "carla.Color", "numpy.transpose", "glob.glob" ]
[((633, 664), 'carla.Client', 'carla.Client', (['"""localhost"""', '(2000)'], {}), "('localhost', 2000)\n", (645, 664), False, 'import carla\n'), ((4756, 4773), 'numpy.transpose', 'np.transpose', (['lbx'], {}), '(lbx)\n', (4768, 4773), True, 'import numpy as np\n'), ((4785, 4802), 'numpy.transpose', 'np.transpose', (['...
""" Collection of functions to calculate lag correlations and significance following Ebisuzaki 97 JCLIM """ def phaseran(recblk, nsurr,ax): """ Phaseran by <NAME>: http://www.mathworks.nl/matlabcentral/fileexchange/32621-phase-randomization/content/phaseran.m Args: recblk (2D array): Row: time sample....
[ "numpy.tile", "numpy.mean", "numpy.flip", "numpy.random.rand", "numpy.flipud", "numpy.fft.fft", "numpy.ndim", "numpy.exp", "numpy.zeros", "ds21grl.misc.AxRoll", "numpy.concatenate", "numpy.std", "numpy.moveaxis", "numpy.percentile", "numpy.fft.ifft", "numpy.arange" ]
[((1166, 1184), 'ds21grl.misc.AxRoll', 'AxRoll', (['recblk', 'ax'], {}), '(recblk, ax)\n', (1172, 1184), False, 'from ds21grl.misc import AxRoll\n'), ((1453, 1478), 'numpy.arange', 'np.arange', (['(1)', '(len_ser + 1)'], {}), '(1, len_ser + 1)\n', (1462, 1478), True, 'import numpy as np\n'), ((1491, 1520), 'numpy.arang...
from django.conf.urls import url from . import views urlpatterns = [ url(r'^$', views.index, name="home"), url(r'success/(?P<id>\d+)', views.success, name="success"), url(r'register$', views.register, name="register"), url(r'login$', views.login, name="login"), url(r'users$', views.users, name="use...
[ "django.conf.urls.url" ]
[((74, 109), 'django.conf.urls.url', 'url', (['"""^$"""', 'views.index'], {'name': '"""home"""'}), "('^$', views.index, name='home')\n", (77, 109), False, 'from django.conf.urls import url\n'), ((116, 174), 'django.conf.urls.url', 'url', (['"""success/(?P<id>\\\\d+)"""', 'views.success'], {'name': '"""success"""'}), "(...
import argparse import os import mlflow import numpy as np import pandas as pd import torch import torch.optim as optim from matplotlib import pyplot as plt from matplotlib.backends.backend_agg import FigureCanvasAgg from mlflow import log_metric, log_param, get_artifact_uri from skimage.io import imsave from sklearn.m...
[ "mlflow.create_experiment", "matplotlib.pyplot.grid", "numpy.sqrt", "matplotlib.pyplot.ylabel", "mlflow.log_param", "torch.cuda.is_available", "argparse.Namespace", "numpy.arange", "sklearn.model_selection.ParameterGrid", "transform.transforms", "torch.set_grad_enabled", "argparse.ArgumentPars...
[((6499, 6585), 'sampler.TomoBatchSampler', 'TomoBatchSampler', ([], {'batch_size': 'args.batch_size', 'data_frame': 'dataset_train.data_frame'}), '(batch_size=args.batch_size, data_frame=dataset_train.\n data_frame)\n', (6515, 6585), False, 'from sampler import TomoBatchSampler\n'), ((6687, 6800), 'torch.utils.data...
from server_commands.argument_helpers import TunableInstanceParam, get_tunable_instance import services import sims4.commands ZONE_MODIFIER_CAP = 3 @sims4.commands.Command('zone_modifier.add_zone_modifier', command_type=sims4.commands.CommandType.DebugOnly) def add_zone_modifier(zone_modifier:TunableInstanceParam(sims...
[ "services.current_zone_id", "server_commands.argument_helpers.get_tunable_instance", "services.get_zone_modifier_service", "services.get_persistence_service", "server_commands.argument_helpers.TunableInstanceParam" ]
[((507, 541), 'services.get_persistence_service', 'services.get_persistence_service', ([], {}), '()\n', (539, 541), False, 'import services\n'), ((1593, 1627), 'services.get_persistence_service', 'services.get_persistence_service', ([], {}), '()\n', (1625, 1627), False, 'import services\n'), ((2411, 2445), 'services.ge...
# This file is part of Ansible # # Ansible is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation, either version 3 of the License, or # (at your option) any later version. # # Ansible is distributed in the hope that ...
[ "os.path.dirname" ]
[((13231, 13256), 'os.path.dirname', 'os.path.dirname', (['__file__'], {}), '(__file__)\n', (13246, 13256), False, 'import os\n'), ((13760, 13785), 'os.path.dirname', 'os.path.dirname', (['__file__'], {}), '(__file__)\n', (13775, 13785), False, 'import os\n'), ((14466, 14491), 'os.path.dirname', 'os.path.dirname', (['_...
import numpy as np def point_to_seg(x1, x2) -> np.ndarray: ''' Method: ------- Transform 2 points into a parametrized segment. Implicitely phi is in [-pi/2; pi/2], it is the oriented angle the segment makes with the horizontal line passing through its middle c. ''' c = (x1[:2] + x2[:2]...
[ "numpy.abs", "numpy.hstack", "numpy.sum", "numpy.cos", "numpy.sign", "numpy.sin" ]
[((793, 815), 'numpy.hstack', 'np.hstack', (['[c, r, phi]'], {}), '([c, r, phi])\n', (802, 815), True, 'import numpy as np\n'), ((395, 417), 'numpy.sum', 'np.sum', (['((x2 - x1) ** 2)'], {}), '((x2 - x1) ** 2)\n', (401, 417), True, 'import numpy as np\n'), ((472, 494), 'numpy.sum', 'np.sum', (['((x2 - x1) ** 2)'], {}),...
#!/usr/bin/python from __future__ import absolute_import from __future__ import print_function import sys, os import syslog try: import time import daemon import pwd from . import pid djangopath = os.path.join(os.path.dirname(sys.argv[0]), "../../") sys.path.append(djangopath) os.environ...
[ "main.models.ProjectUpload.objects.filter", "pwd.getpwnam", "os.getuid", "time.sleep", "syslog.syslog", "os.path.dirname", "sys.exc_info", "sys.path.append" ]
[((278, 305), 'sys.path.append', 'sys.path.append', (['djangopath'], {}), '(djangopath)\n', (293, 305), False, 'import sys, os\n'), ((878, 922), 'main.models.ProjectUpload.objects.filter', 'ProjectUpload.objects.filter', ([], {'complete': '(False)'}), '(complete=False)\n', (906, 922), False, 'from main.models import Pr...
# -*- coding: utf-8 -*- # Copyright (C) 2010-2016 <NAME> All rights reserved # Langstrasse 4, A--2244 Spannberg, Austria. <EMAIL> # **************************************************************************** # This module is part of the package GTW.__test__. # # This module is licensed under the terms of the BSD 3-Cla...
[ "_TFL.Regexp.Re_Replacer" ]
[((1502, 1576), '_TFL.Regexp.Re_Replacer', 'Re_Replacer', (['"""\'start\', \'\\\\d{4}-\\\\d{2}-\\\\d{2}\'"""', '"""\'start\', <date instance>"""'], {}), '("\'start\', \'\\\\d{4}-\\\\d{2}-\\\\d{2}\'", "\'start\', <date instance>")\n', (1513, 1576), False, 'from _TFL.Regexp import Re_Replacer, re\n')]
#!/usr/bin/env python # -*- coding: utf-8 -*- """ sfftk.unittests.test_readers This testing module should have no side-effects because it only reads. """ from __future__ import division, print_function import glob import os import struct import sys import unittest import numpy import random_words import __init__ a...
[ "random_words.RandomWords", "os.path.join", "os.path.basename", "unittest.main", "glob.glob", "os.remove" ]
[((544, 570), 'random_words.RandomWords', 'random_words.RandomWords', ([], {}), '()\n', (568, 570), False, 'import random_words\n'), ((17550, 17565), 'unittest.main', 'unittest.main', ([], {}), '()\n', (17563, 17565), False, 'import unittest\n'), ((693, 760), 'os.path.join', 'os.path.join', (['tests.TEST_DATA_PATH', '"...
#!/usr/bin/env python3 # _*_ coding: utf-8 _*_ ### # Project : SubLime # FileName : util.py # ----------------------------------------------------------------------------- # Author : sham # E-Mail : <EMAIL> # -------------------------------------------------------------------------...
[ "logging.getLogger", "os.path.splitext", "sublime.util.get_exe_dir", "csv.reader" ]
[((439, 466), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (456, 466), False, 'import logging\n'), ((3886, 3912), 'os.path.splitext', 'os.path.splitext', (['filepath'], {}), '(filepath)\n', (3902, 3912), False, 'import os\n'), ((2500, 2518), 'sublime.util.get_exe_dir', 'util.get_exe_dir...
# -*- coding: utf-8 -*- """ @author:XuMing(<EMAIL>), <NAME>(<EMAIL>) @description: Graph classify """ import numpy from sklearn.metrics import f1_score, accuracy_score from sklearn.multiclass import OneVsRestClassifier from sklearn.preprocessing import MultiLabelBinarizer class TopKRanker(OneVsRestClassifier): d...
[ "sklearn.metrics.f1_score", "numpy.asarray", "numpy.random.seed", "sklearn.preprocessing.MultiLabelBinarizer", "sklearn.metrics.accuracy_score" ]
[((702, 727), 'numpy.asarray', 'numpy.asarray', (['all_labels'], {}), '(all_labels)\n', (715, 727), False, 'import numpy\n'), ((894, 933), 'sklearn.preprocessing.MultiLabelBinarizer', 'MultiLabelBinarizer', ([], {'sparse_output': '(True)'}), '(sparse_output=True)\n', (913, 933), False, 'from sklearn.preprocessing impor...
############################################################################### # Module: ddl_statement # Purpose: Parent class for DDL (Data Definition Language) statements # # Notes: # ############################################################################### import data_pipeline.constants.const as const ...
[ "data_pipeline.constants.const.COMMASPACE.join" ]
[((889, 918), 'data_pipeline.constants.const.COMMASPACE.join', 'const.COMMASPACE.join', (['params'], {}), '(params)\n', (910, 918), True, 'import data_pipeline.constants.const as const\n')]
# Copyright 2016 Google Inc. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, ...
[ "google.cloud.logging.entries.ProtobufEntry.from_api_repr", "google.cloud.logging.entries.StructEntry.from_api_repr", "google.cloud.logging.entries.TextEntry.from_api_repr" ]
[((1481, 1531), 'google.cloud.logging.entries.TextEntry.from_api_repr', 'TextEntry.from_api_repr', (['resource', 'client', 'loggers'], {}), '(resource, client, loggers)\n', (1504, 1531), False, 'from google.cloud.logging.entries import TextEntry\n'), ((1583, 1635), 'google.cloud.logging.entries.StructEntry.from_api_rep...
import os import pandas as pd import numpy as np def inp_mortality_tot(): df=pd.read_csv('total_country.csv') mortal_rate=[] cured_rate=[] i=0 while(i<len(df)): res=df.iloc[i]['Deaths']/df.iloc[i]['Confirmed'] res_2=df.iloc[i]['Cured']/df.iloc[i]['Confirmed'] mortal_rate.append(res) cured_rate.append...
[ "os.listdir", "pandas.read_csv" ]
[((80, 112), 'pandas.read_csv', 'pd.read_csv', (['"""total_country.csv"""'], {}), "('total_country.csv')\n", (91, 112), True, 'import pandas as pd\n'), ((482, 509), 'os.listdir', 'os.listdir', (['"""details_state"""'], {}), "('details_state')\n", (492, 509), False, 'import os\n'), ((966, 1002), 'pandas.read_csv', 'pd.r...
# encoding: utf-8 from collections import OrderedDict import string from pydatacube.pydatacube import _DataCube import px_reader # A bit scandinavian specific default_translate = dict(zip( u"äöä -", u"aoa__" )) class Sluger(object): def __init__(self, translate=default_translate): self.given_out = {} self.tra...
[ "pydatacube.pydatacube._DataCube", "px_reader.Px", "collections.OrderedDict" ]
[((834, 859), 'px_reader.Px', 'px_reader.Px', (['pcaxis_data'], {}), '(pcaxis_data)\n', (846, 859), False, 'import px_reader\n'), ((868, 881), 'collections.OrderedDict', 'OrderedDict', ([], {}), '()\n', (879, 881), False, 'from collections import OrderedDict\n'), ((894, 907), 'collections.OrderedDict', 'OrderedDict', (...
# Licensed to the Apache Software Foundation (ASF) under one # or more contributor license agreements. See the NOTICE file # distributed with this work for additional information # regarding copyright ownership. The ASF licenses this file # to you under the Apache License, Version 2.0 (the # "License"); you may not u...
[ "numpy.ones", "tvm.relay.backend.contrib.ethosu.codegen._create_cascader", "tvm.contrib.ethosu.cascader.EthosuDeviceConfig", "pytest.importorskip", "tvm.contrib.ethosu.cascader.MemoryRegion", "tvm.IRModule", "tvm.relay.transform.InferType", "tvm.relay.analysis.free_vars", "tvm.relay.var" ]
[((906, 940), 'pytest.importorskip', 'pytest.importorskip', (['"""ethosu.vela"""'], {}), "('ethosu.vela')\n", (925, 940), False, 'import pytest\n'), ((1329, 1461), 'tvm.contrib.ethosu.cascader.MemoryRegion', 'MemoryRegion', ([], {'name': '"""SRAM"""', 'size': '(10 ** 6)', 'read_bandwidth': '(16)', 'write_bandwidth': '(...
# encoding: utf-8 ''' Created on Dec 18, 2018 @author: <NAME> ''' import time from array import * from ctypes import * from sys import exit from multiprocessing import Process from multiprocessing import Queue import numpy as np class EmotivDeviceReader(object): ''' classdocs This class is used to read ...
[ "multiprocessing.Process", "time.sleep", "numpy.array", "sys.exit", "multiprocessing.Queue" ]
[((9695, 9708), 'time.sleep', 'time.sleep', (['(5)'], {}), '(5)\n', (9705, 9708), False, 'import time\n'), ((494, 511), 'multiprocessing.Queue', 'Queue', ([], {'maxsize': '(-1)'}), '(maxsize=-1)\n', (499, 511), False, 'from multiprocessing import Queue\n'), ((8523, 8548), 'multiprocessing.Process', 'Process', ([], {'ta...
import scipy import matplotlib.pyplot as plt import numpy as np x = [ 0.001, 0.019, 0.039, 0.058, 0.080, 0.098, 0.119, 0.139, 0.159, 0.180, 0.198, 0.249, 0.298, 0.349, 0.398, 0.419, 0.439, 0.460, 0.479, 0.499, 0.519, 0.540, 0.558, 0.578, 0.598, 0.649, 0.698, 0.749, 0.798, 0.819, 0.839, 0.859, 0.879...
[ "numpy.polyfit", "numpy.poly1d", "matplotlib.pyplot.subplots", "matplotlib.pyplot.show" ]
[((724, 738), 'matplotlib.pyplot.subplots', 'plt.subplots', ([], {}), '()\n', (736, 738), True, 'import matplotlib.pyplot as plt\n'), ((1037, 1047), 'matplotlib.pyplot.show', 'plt.show', ([], {}), '()\n', (1045, 1047), True, 'import matplotlib.pyplot as plt\n'), ((827, 862), 'numpy.polyfit', 'np.polyfit', (['x', 'y', '...
#!/usr/bin/env python from setuptools import setup import backtracepython setup( name='backtracepython', version=backtracepython.version_string, description='Backtrace error reporting tool for Python', author='<NAME>', author_email='<EMAIL>', packages=['backtracepython'], test_suite="test...
[ "setuptools.setup" ]
[((77, 370), 'setuptools.setup', 'setup', ([], {'name': '"""backtracepython"""', 'version': 'backtracepython.version_string', 'description': '"""Backtrace error reporting tool for Python"""', 'author': '"""<NAME>"""', 'author_email': '"""<EMAIL>"""', 'packages': "['backtracepython']", 'test_suite': '"""tests"""', 'url'...
# # MIT License # # Copyright (c) 2020 Airbyte # # Permission is hereby granted, free of charge, to any person obtaining a copy # of this software and associated documentation files (the "Software"), to deal # in the Software without restriction, including without limitation the rights # to use, copy, modify, merge, pu...
[ "boto3.client", "airbyte_cdk.logger.AirbyteLogger", "time.sleep", "boto3.resource", "json.load" ]
[((1435, 1450), 'airbyte_cdk.logger.AirbyteLogger', 'AirbyteLogger', ([], {}), '()\n', (1448, 1450), False, 'from airbyte_cdk.logger import AirbyteLogger\n'), ((2201, 2360), 'boto3.client', 'boto3.client', (['"""s3"""'], {'aws_access_key_id': "credentials['aws_access_key_id']", 'aws_secret_access_key': "credentials['aw...
from serverless_crud.model import BaseModel try: from troposphere import Sub except ImportError: class Sub: def __init__(self, name): self.name = name def __str__(self): return self.name class PolicyBuilder: def __init__(self, statements=None): statements...
[ "troposphere.Sub" ]
[((1630, 1727), 'troposphere.Sub', 'Sub', (['f"""arn:aws:dynamodb:${{AWS::Region}}:${{AWS::AccountId}}:table/{model._meta.table_name}"""'], {}), "(f'arn:aws:dynamodb:${{AWS::Region}}:${{AWS::AccountId}}:table/{model._meta.table_name}'\n )\n", (1633, 1727), False, 'from troposphere import Sub\n'), ((1814, 1919), 'tro...
# Installation script from distutils.core import setup, Extension setup( name='neat-python', version='0.1', description='A NEAT (NeuroEvolution of Augmenting Topologies) implementation', packages=['neat', 'neat/iznn', 'neat/nn', 'neat/ctrnn', 'neat/ifnn'], #ext_modules=[ # Ex...
[ "distutils.core.setup" ]
[((66, 264), 'distutils.core.setup', 'setup', ([], {'name': '"""neat-python"""', 'version': '"""0.1"""', 'description': '"""A NEAT (NeuroEvolution of Augmenting Topologies) implementation"""', 'packages': "['neat', 'neat/iznn', 'neat/nn', 'neat/ctrnn', 'neat/ifnn']"}), "(name='neat-python', version='0.1', description=\...
from unittest import TestCase from src.stack import StackWithMaxValue class TestStackWithMaxValue(TestCase): def test_push(self): stack = StackWithMaxValue() stack.push(1) stack.push(2) stack.push(3) self.assertEqual([1, 2, 3], stack.as_list()) def test_pop(self): ...
[ "src.stack.StackWithMaxValue" ]
[((154, 173), 'src.stack.StackWithMaxValue', 'StackWithMaxValue', ([], {}), '()\n', (171, 173), False, 'from src.stack import StackWithMaxValue\n'), ((334, 353), 'src.stack.StackWithMaxValue', 'StackWithMaxValue', ([], {}), '()\n', (351, 353), False, 'from src.stack import StackWithMaxValue\n'), ((573, 592), 'src.stack...
# -*- coding: utf-8 -*- from keras.models import load_model import numpy as np import os import cv2 from FaceQNet import load_Qnet_model, face_quality # Loading the pretrained model model = load_Qnet_model() IMG_PATH = '/home/sai/YANG/image/video/nanning/haha' dir = os.listdir(IMG_PATH) count = len(dir) print('count:...
[ "os.listdir", "FaceQNet.load_Qnet_model", "os.path.join", "cv2.imread", "FaceQNet.face_quality" ]
[((191, 208), 'FaceQNet.load_Qnet_model', 'load_Qnet_model', ([], {}), '()\n', (206, 208), False, 'from FaceQNet import load_Qnet_model, face_quality\n'), ((269, 289), 'os.listdir', 'os.listdir', (['IMG_PATH'], {}), '(IMG_PATH)\n', (279, 289), False, 'import os\n'), ((427, 452), 'os.path.join', 'os.path.join', (['IMG_P...
# Generated by Django 3.2.2 on 2021-05-10 04:54 from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('imgrepo', '0002_auto_20210509_2317'), ] operations = [ migrations.AlterField( model_name='image', name='image_src', ...
[ "django.db.models.ImageField" ]
[((338, 389), 'django.db.models.ImageField', 'models.ImageField', ([], {'upload_to': '"""static/media/images/"""'}), "(upload_to='static/media/images/')\n", (355, 389), False, 'from django.db import migrations, models\n')]
# Generated by Django 3.1.3 on 2022-02-28 16:55 from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('home', '0003_auto_20220228_1112'), ] operations = [ migrations.CreateModel( name='Results', fields=[ ...
[ "django.db.models.AutoField", "django.db.models.IntegerField" ]
[((328, 421), 'django.db.models.AutoField', 'models.AutoField', ([], {'auto_created': '(True)', 'primary_key': '(True)', 'serialize': '(False)', 'verbose_name': '"""ID"""'}), "(auto_created=True, primary_key=True, serialize=False,\n verbose_name='ID')\n", (344, 421), False, 'from django.db import migrations, models\...
from flask import Flask, request, render_template, flash, redirect, url_for from wtforms import Form, TextField, TextAreaField, validators, StringField, SubmitField from wtforms.validators import DataRequired from app.DineCision import yelprequest from flask_wtf import FlaskForm import random import json import os API...
[ "flask.render_template", "random.choice", "flask.flash", "flask.Flask", "wtforms.validators.required", "os.environ.get", "wtforms.SubmitField", "flask.url_for", "flask.request.form.get", "app.DineCision.yelprequest" ]
[((693, 708), 'flask.Flask', 'Flask', (['__name__'], {}), '(__name__)\n', (698, 708), False, 'from flask import Flask, request, render_template, flash, redirect, url_for\n'), ((327, 363), 'os.environ.get', 'os.environ.get', (['"""DINECISION_API_KEY"""'], {}), "('DINECISION_API_KEY')\n", (341, 363), False, 'import os\n'...
import unittest import numpy as np import torch from torch import optim from spn.structure.Base import Product, Sum from spn.structure.Base import assign_ids, rebuild_scopes_bottom_up from spn.structure.leaves.parametric.Parametric import Gaussian, Categorical from spn.gpu.TensorFlow import spn_to_tf_graph, optimize_...
[ "pytorch.SumNode.from_spn", "spn.structure.Base.Sum", "numpy.log", "spn.algorithms.Inference.log_likelihood", "spn.structure.leaves.parametric.Parametric.Gaussian", "numpy.array", "unittest.main", "pytorch.GaussianNode.from_spn", "spn.structure.leaves.parametric.Parametric.Categorical", "tensorflo...
[((7532, 7547), 'unittest.main', 'unittest.main', ([], {}), '()\n', (7545, 7547), False, 'import unittest\n'), ((790, 817), 'torch.log', 'torch.log', (['x[:, self.scope]'], {}), '(x[:, self.scope])\n', (799, 817), False, 'import torch\n'), ((1172, 1197), 'numpy.array', 'np.array', (['[0.1, 0.4, 0.5]'], {}), '([0.1, 0.4...
# -*- coding: utf-8 -*- """ Profile: http://hl7.org/fhir/StructureDefinition/TerminologyCapabilities Release: R4 Version: 4.0.1 Build ID: 9346c8cc45 Last updated: 2019-11-01T09:29:23.356+11:00 """ import io import json import os import unittest import pytest from .. import terminologycapabilities from ..fhirdate imp...
[ "json.load", "os.environ.get", "pytest.mark.usefixtures", "os.path.join" ]
[((370, 410), 'pytest.mark.usefixtures', 'pytest.mark.usefixtures', (['"""base_settings"""'], {}), "('base_settings')\n", (393, 410), False, 'import pytest\n'), ((526, 565), 'os.environ.get', 'os.environ.get', (['"""FHIR_UNITTEST_DATADIR"""'], {}), "('FHIR_UNITTEST_DATADIR')\n", (540, 565), False, 'import os\n'), ((677...
#! /usr/bin/env python import pandas as pd import click ''' gene expression matrix, with gene id in first column, gene expression level of each sample in othre columns. ''' @click.group(chain=True, invoke_without_command=True) @click.argument('exp_table', type=click.STRING, required=True) @click.pass_con...
[ "click.argument", "click.group", "click.option", "pandas.merge", "pandas.read_table" ]
[((187, 239), 'click.group', 'click.group', ([], {'chain': '(True)', 'invoke_without_command': '(True)'}), '(chain=True, invoke_without_command=True)\n', (198, 239), False, 'import click\n'), ((242, 303), 'click.argument', 'click.argument', (['"""exp_table"""'], {'type': 'click.STRING', 'required': '(True)'}), "('exp_t...
# Copyright (C) 2017 <NAME> <<EMAIL>> # This file is subject to the terms and conditions defined in # file 'LICENSE', which is part of this source code package. '''yaml checker''' import yaml from omnilint.error import Error from omnilint.checkers import Checker class Yaml(Checker): extensions = ['yaml', 'yml'...
[ "yaml.load", "omnilint.error.Error" ]
[((494, 507), 'yaml.load', 'yaml.load', (['fd'], {}), '(fd)\n', (503, 507), False, 'import yaml\n'), ((644, 766), 'omnilint.error.Error', 'Error', ([], {'msg': "(exc.context + ' ' + exc.problem)", 'file': 'origname', 'line': 'exc.problem_mark.line', 'column': 'exc.problem_mark.column'}), "(msg=exc.context + ' ' + exc.p...
from netcad.device.l2_interfaces import InterfaceL2Access, InterfaceL2Trunk from netcad.device import PeerInterfaceId from netcad_demo_meraki1.vlans import vlan_native_1 from .physical import port_UTP_1G class AccessVlan1(InterfaceL2Access): port_profile = port_UTP_1G vlan = vlan_native_1 desc = PeerInte...
[ "netcad.device.PeerInterfaceId" ]
[((312, 329), 'netcad.device.PeerInterfaceId', 'PeerInterfaceId', ([], {}), '()\n', (327, 329), False, 'from netcad.device import PeerInterfaceId\n')]
""" Ruuvi Counter (LoRaWAN) decoder See: https://bitbucket.org/iotpetri/hki_kuva_iot/src/master/ESP32/LORA/ESP32_RuuviTagGW_Lora_v2/ """ import json import pytz from dateutil.parser import parse from broker.providers.decoder import DecoderProvider from broker.utils import create_dataline from fvhexperiments.parsers.r...
[ "dateutil.parser.parse", "fvhexperiments.parsers.ruuvicounter.parse_ruuvicounter", "broker.utils.create_dataline" ]
[((652, 682), 'fvhexperiments.parsers.ruuvicounter.parse_ruuvicounter', 'parse_ruuvicounter', (['data', 'port'], {}), '(data, port)\n', (670, 682), False, 'from fvhexperiments.parsers.ruuvicounter import parse_ruuvicounter\n'), ((812, 867), 'broker.utils.create_dataline', 'create_dataline', (['timestamp', "ruuvicounter...
from bot import __version__ from setuptools import setup, find_packages REQUIREMENTS = [line.strip() for line in open("requirements.txt").readlines()] setup(name='guldai-telegram-bot', version=__version__, description='Telegram interface for the guldai bot', author='isysd', author_email='<EMAI...
[ "setuptools.setup" ]
[((153, 698), 'setuptools.setup', 'setup', ([], {'name': '"""guldai-telegram-bot"""', 'version': '__version__', 'description': '"""Telegram interface for the guldai bot"""', 'author': '"""isysd"""', 'author_email': '"""<EMAIL>"""', 'license': '"""MIT"""', 'url': '"""https://guld.io/"""', 'py_modules': "['bot']", 'insta...
import dash_mantine_components as dmc component = dmc.Spoiler( showLabel="Show more", hideLabel="Hide", maxHeight=50, children=[ dmc.Text( """We Butter the Bread with Butter was founded in 2007 by <NAME>, who was originally guitarist for <NAME>'s band, and <NAME>. The b...
[ "dash_mantine_components.Text" ]
[((154, 833), 'dash_mantine_components.Text', 'dmc.Text', (['"""We Butter the Bread with Butter was founded in 2007 by <NAME>, who was originally guitarist \n for <NAME>\'s band, and <NAME>. The band was originally meant as a joke, but progressed \n into being a more serious musical duo. The name ...
import json from mock import patch import jenkins from tests.base import JenkinsTestBase class JenkinsCredentialTestBase(JenkinsTestBase): config_xml = """<com.cloudbees.plugins.credentials.impl.UsernamePasswordCredentialsImpl> <scope>GLOBAL</scope> <id>Test Credential</id> <username>Test...
[ "mock.patch.object", "json.dumps", "jenkins.NotFoundException" ]
[((1505, 1550), 'mock.patch.object', 'patch.object', (['jenkins.Jenkins', '"""jenkins_open"""'], {}), "(jenkins.Jenkins, 'jenkins_open')\n", (1517, 1550), False, 'from mock import patch\n'), ((1787, 1832), 'mock.patch.object', 'patch.object', (['jenkins.Jenkins', '"""jenkins_open"""'], {}), "(jenkins.Jenkins, 'jenkins_...
import logging import pytest from pathlib import Path from leaf_focus.pdf.images.component import Component from tests.base_test import BaseTest class TestPdfImagesComponent(BaseTest): def test_no_exe(self): with pytest.raises(ValueError, match="Must supply exe file."): Component(logging.getLo...
[ "logging.getLogger", "pytest.raises", "pathlib.Path" ]
[((919, 942), 'pathlib.Path', 'Path', (['tmp_path', '"""image"""'], {}), "(tmp_path, 'image')\n", (923, 942), False, 'from pathlib import Path\n'), ((227, 283), 'pytest.raises', 'pytest.raises', (['ValueError'], {'match': '"""Must supply exe file."""'}), "(ValueError, match='Must supply exe file.')\n", (240, 283), Fals...
import pynput from pynput.keyboard import Key, Listener def on_press(key): print(key + " was pressed") def on_release(key): if key == key.esc: return False with Listener(on_press=on_press, on_release=on_release) as listener: listener.join()
[ "pynput.keyboard.Listener" ]
[((186, 236), 'pynput.keyboard.Listener', 'Listener', ([], {'on_press': 'on_press', 'on_release': 'on_release'}), '(on_press=on_press, on_release=on_release)\n', (194, 236), False, 'from pynput.keyboard import Key, Listener\n')]
import pstats p = pstats.Stats('output.prof') p.sort_stats('cumulative').print_stats(15)
[ "pstats.Stats" ]
[((23, 50), 'pstats.Stats', 'pstats.Stats', (['"""output.prof"""'], {}), "('output.prof')\n", (35, 50), False, 'import pstats\n')]
from django.db import models class Restaurant(models.Model): """ Restaurants' Model. """ id = models.CharField(primary_key = True, editable = True, max_length = 255, verbose_name = u'Id') rating = models.IntegerField(verbose_name = u'Rating') name = models.CharField(max_length = 255, verbose_na...
[ "django.db.models.FloatField", "django.db.models.CharField", "django.db.models.IntegerField" ]
[((111, 200), 'django.db.models.CharField', 'models.CharField', ([], {'primary_key': '(True)', 'editable': '(True)', 'max_length': '(255)', 'verbose_name': 'u"""Id"""'}), "(primary_key=True, editable=True, max_length=255,\n verbose_name=u'Id')\n", (127, 200), False, 'from django.db import models\n'), ((218, 261), 'd...
""" Useful functions for the admin panel of ImageLabeller. In particular: * Download labels from database to json or csv * Upload images to the database, from json (catalogue of image locations) or zip archive NOTE: When uploading files, or an archive full of files, we will attempt to match the filename to a regex wi...
[ "re.search", "image_labeller.db.session.commit", "os.listdir", "os.makedirs", "image_labeller.schema.Image", "os.path.join", "image_labeller.schema.Label.query.all", "datetime.datetime.now", "json.dump", "image_labeller.db.session.add" ]
[((835, 852), 'image_labeller.schema.Label.query.all', 'Label.query.all', ([], {}), '()\n', (850, 852), False, 'from image_labeller.schema import Label, User, Image, Category\n'), ((1937, 1971), 'os.makedirs', 'os.makedirs', (['tmpdir'], {'exist_ok': '(True)'}), '(tmpdir, exist_ok=True)\n', (1948, 1971), False, 'import...
# Licensed under a 3-clause BSD style license - see LICENSE.rst """Command line tool to run tests. This file is called `check` instead of `test` to prevent confusion for developers and the test runner from including it in test collection. """ from __future__ import (absolute_import, division, print_function, ...
[ "logging.basicConfig", "gammapy.test" ]
[((714, 792), 'logging.basicConfig', 'logging.basicConfig', ([], {'level': 'logging.DEBUG', 'format': '"""%(levelname)s - %(message)s"""'}), "(level=logging.DEBUG, format='%(levelname)s - %(message)s')\n", (733, 792), False, 'import logging\n'), ((817, 852), 'gammapy.test', 'gammapy.test', (['package'], {'verbose': '(T...
import json from abc import ABC, abstractmethod from construct import Struct, Container class StateCore(ABC): layout: Struct = None @classmethod @abstractmethod def from_container(cls, container: Container): pass @classmethod def parse(cls, bytes_data: bytes, factor: int): co...
[ "json.dumps" ]
[((696, 742), 'json.dumps', 'json.dumps', (['my_dict'], {'sort_keys': '(False)', 'indent': '(4)'}), '(my_dict, sort_keys=False, indent=4)\n', (706, 742), False, 'import json\n')]
from conans import ConanFile, tools import os class ClaraConan(ConanFile): name = "clara" version = "1.1.5" description = "A simple to use, composable, command line parser for C++ 11 and beyond" url = "https://github.com/bincrafters/conan-clara" homepage = "https://github.com/catchorg/Clara" t...
[ "os.path.join" ]
[((782, 836), 'os.path.join', 'os.path.join', (['self._source_subfolder', '"""single_include"""'], {}), "(self._source_subfolder, 'single_include')\n", (794, 836), False, 'import os\n')]
# -*- coding: utf-8 -*- # @Time : 2019-09-01 17:49 # @Author : EchoShoot # @Email : <EMAIL> # @URL : https://github.com/EchoShoot # @File : test_others.py # @Explain : from sheen import Str import pytest class TestOthers(object): raw = 'xxooAß西xoox' obj = Str.red(raw) obj[2:-2] = Str.green ...
[ "sheen.Str.green", "sheen.Str", "sheen.Str.red", "pytest.raises" ]
[((280, 292), 'sheen.Str.red', 'Str.red', (['raw'], {}), '(raw)\n', (287, 292), False, 'from sheen import Str\n'), ((2331, 2364), 'pytest.raises', 'pytest.raises', (['UnicodeEncodeError'], {}), '(UnicodeEncodeError)\n', (2344, 2364), False, 'import pytest\n'), ((2440, 2473), 'pytest.raises', 'pytest.raises', (['Unicode...
""" Validate CAL DAC settings XML files. The command line is: valDACsettings [-V] [-r] [-R <root_file>] [-L <log_file>] FLE|FHE|LAC|ULD <MeV | margin> <dac_slopes_file> <dac_xml_file> where: -r = generate ROOT output with default name -R <root_file> = output validation diagnostics in ROOT...
[ "logging.basicConfig", "logging.getLogger", "os.path.exists", "calDacXML.calSettingsXML", "getopt.getopt", "ROOT.TH1F", "logging.Formatter", "os.path.splitext", "numarray.logical_not", "logging.FileHandler", "calCalibXML.calDacSlopesCalibXML", "numarray.where", "sys.exit", "ROOT.TFile", ...
[((1320, 1360), 'ROOT.TCanvas', 'ROOT.TCanvas', (['"""c_Summary"""', '"""Summary"""', '(-1)'], {}), "('c_Summary', 'Summary', -1)\n", (1332, 1360), False, 'import ROOT\n'), ((1435, 1541), 'ROOT.TH1F', 'ROOT.TH1F', (['hName', "('DAC_Val_%s: %s' % (dacType, fileName))", '(100)', '(MeV - errLimit * 2)', '(MeV + errLimit *...
"""! @brief Test templates for ROCK clustering module. @authors <NAME> (<EMAIL>) @date 2014-2020 @copyright BSD-3-Clause """ from pyclustering.cluster.rock import rock; from pyclustering.utils import read_sample; from random import random; class RockTestTemplates: @staticmethod def ...
[ "pyclustering.utils.read_sample", "random.random", "pyclustering.cluster.rock.rock" ]
[((447, 472), 'pyclustering.utils.read_sample', 'read_sample', (['path_to_file'], {}), '(path_to_file)\n', (458, 472), False, 'from pyclustering.utils import read_sample\n'), ((509, 564), 'pyclustering.cluster.rock.rock', 'rock', (['sample', 'radius', 'cluster_numbers', 'threshold', 'ccore'], {}), '(sample, radius, clu...
from ryu.app.wsgi import WSGIApplication from ryu.base import app_manager from ryu.lib import hub from ryu.lib.dpid import dpid_to_str from ryu.controller.handler import MAIN_DISPATCHER, DEAD_DISPATCHER, CONFIG_DISPATCHER from ryu.controller.handler import set_ev_cls from ryu.controller import ofp_event from ryu.lib.p...
[ "process_stats_flow.avg_rates", "ryu.lib.packet.packet.Packet", "process_stats_port.avg_rates", "ryu.base.app_manager.require_app", "process_stats_port.stats_event", "ryu.lib.hub.spawn", "ryu.controller.handler.set_ev_cls", "process_stats_flow.stats_event", "ryu.lib.dpid.dpid_to_str", "ryu.lib.hub...
[((9680, 9736), 'ryu.base.app_manager.require_app', 'app_manager.require_app', (['"""ryu.app.simple_switch_13_lldp"""'], {}), "('ryu.app.simple_switch_13_lldp')\n", (9703, 9736), False, 'from ryu.base import app_manager\n'), ((9737, 9785), 'ryu.base.app_manager.require_app', 'app_manager.require_app', (['"""ryu.app.res...
import painter import sys from utils import distance from collections import deque #returns last node (destination_node) def dfs(origin_node, dest_node, board, screen): #put origin node on open_nodes_list (first to be expanded) open_nodes=deque() closed_nodes=list() open_nodes.append(origin_node) ...
[ "collections.deque", "painter.paint_search" ]
[((249, 256), 'collections.deque', 'deque', ([], {}), '()\n', (254, 256), False, 'from collections import deque\n'), ((476, 526), 'painter.paint_search', 'painter.paint_search', (['screen', 'curr_node.pos', 'board'], {}), '(screen, curr_node.pos, board)\n', (496, 526), False, 'import painter\n')]
#!/bin/python import roomai.common from roomai.sevenking import SevenKingPublicState from roomai.sevenking import SevenKingPrivateState from roomai.sevenking import SevenKingPersonState from roomai.sevenking import SevenKingAction from roomai.sevenking import SevenKingPokerCard from roomai.sevenking import AllSevenKing...
[ "random.shuffle", "roomai.sevenking.SevenKingPokerCard.compare", "roomai.sevenking.SevenKingAction", "roomai.sevenking.SevenKingPersonState", "roomai.sevenking.SevenKingPrivateState", "roomai.sevenking.AllSevenKingPokerCards.values", "roomai.sevenking.SevenKingAction.lookup", "roomai.sevenking.SevenKi...
[((1590, 1612), 'roomai.sevenking.SevenKingPublicState', 'SevenKingPublicState', ([], {}), '()\n', (1610, 1612), False, 'from roomai.sevenking import SevenKingPublicState\n'), ((1642, 1665), 'roomai.sevenking.SevenKingPrivateState', 'SevenKingPrivateState', ([], {}), '()\n', (1663, 1665), False, 'from roomai.sevenking ...
from __future__ import unicode_literals from django.contrib.auth.models import AnonymousUser, User from django.http import HttpRequest from django.template import RequestContext, Template from djblets.avatars.services.gravatar import GravatarService from reviewboard.testing import TestCase from reviewboard.avatars im...
[ "reviewboard.avatars.tests.DummyAvatarService", "django.template.Template", "django.contrib.auth.models.AnonymousUser", "reviewboard.avatars.avatar_services.set_default_service", "django.template.RequestContext", "reviewboard.avatars.avatar_services.reset", "reviewboard.avatars.avatar_services.get_avata...
[((580, 603), 'reviewboard.avatars.avatar_services.reset', 'avatar_services.reset', ([], {}), '()\n', (601, 603), False, 'from reviewboard.avatars import avatar_services\n'), ((631, 700), 'reviewboard.avatars.avatar_services.get_avatar_service', 'avatar_services.get_avatar_service', (['GravatarService.avatar_service_id...
import jwt import requests from cryptography.hazmat.backends import default_backend from cryptography.hazmat.primitives.asymmetric.rsa import RSAPublicNumbers from oauth2 import Oauth2, ResponseObject, RemoteUserProfile, OAuthDiscoveryError class OpenID(Oauth2): state_token: str or None = None well_know_pat...
[ "jwt.decode", "requests.get", "jwt.get_unverified_header", "jwt.api_jws.base64url_decode", "oauth2.OAuthDiscoveryError", "cryptography.hazmat.backends.default_backend", "cryptography.hazmat.primitives.asymmetric.rsa.RSAPublicNumbers" ]
[((1388, 1420), 'jwt.get_unverified_header', 'jwt.get_unverified_header', (['token'], {}), '(token)\n', (1413, 1420), False, 'import jwt\n'), ((2101, 2132), 'jwt.api_jws.base64url_decode', 'jwt.api_jws.base64url_decode', (['e'], {}), '(e)\n', (2129, 2132), False, 'import jwt\n'), ((2162, 2193), 'jwt.api_jws.base64url_d...
from dvc_preprocessing import plot, preprocessing, constants from skimage.filters import threshold_otsu import numpy as np def auto_processing(filename, dirpath='./', data_type=np.int16, init_slice=0, final_slice="last", outname="output", ret="True"): ''' TODO: add outpath ''' stack = preprocessing.r...
[ "dvc_preprocessing.constants.INT8MINVAL", "dvc_preprocessing.preprocessing.save_3d_tiff", "skimage.filters.threshold_otsu", "dvc_preprocessing.preprocessing.volume_CoM", "dvc_preprocessing.preprocessing.read_images_from_h5", "dvc_preprocessing.preprocessing.crop_around_CoM", "dvc_preprocessing.preproces...
[((305, 368), 'dvc_preprocessing.preprocessing.read_images_from_h5', 'preprocessing.read_images_from_h5', (['filename', 'data_type', 'dirpath'], {}), '(filename, data_type, dirpath)\n', (338, 368), False, 'from dvc_preprocessing import plot, preprocessing, constants\n'), ((392, 413), 'skimage.filters.threshold_otsu', '...
from __future__ import absolute_import, division, print_function import pytest from ..spparser import Scanner scanner = Scanner() # Test of a single instance of each token. Does not test them in # context, but at least it tests that each one is recognized. tokens = [ # bug: the original pysynphot could not reco...
[ "pytest.mark.xfail", "pytest.mark.parametrize" ]
[((2241, 6212), 'pytest.mark.parametrize', 'pytest.mark.parametrize', (["('text', 'result')", "[('spec($PYSYN_CDBS//calspec/gd71_mod_005.fits)', [('IDENTIFIER', 'spec'),\n ('LPAREN', None), ('IDENTIFIER',\n '$PYSYN_CDBS//calspec/gd71_mod_005.fits'), ('RPAREN', None)]), (\n 'spec(earthshine.fits)*0.5+rn(spec(Zo...
# # This program is free software: you can redistribute it and/or modify # it under the terms of the GNU Lesser General Public License as published by # the Free Software Foundation, either version 3 of the License, or # (at your option) any later version. # # This program is distributed in the hope that it will be use...
[ "os.path.realpath", "proton.template.get_template" ]
[((1270, 1313), 'proton.template.get_template', 'template.get_template', (['"""twotemplates.xhtml"""'], {}), "('twotemplates.xhtml')\n", (1291, 1313), False, 'from proton import template\n'), ((1686, 1727), 'proton.template.get_template', 'template.get_template', (['"""twotemplates.xml"""'], {}), "('twotemplates.xml')\...
# -*- coding: utf-8 -*- # PLEASE DO NOT EDIT THIS FILE, IT IS GENERATED AND WILL BE OVERWRITTEN: # https://github.com/ccxt/ccxt/blob/master/CONTRIBUTING.md#how-to-contribute-code from ccxt.base.exchange import Exchange from ccxt.base.errors import ExchangeError from ccxt.base.errors import BadRequest from ccxt.base.e...
[ "datetime.datetime.fromtimestamp", "json.dumps", "sys.exc_info", "datetime.datetime.now", "ccxt.base.errors.ExchangeError" ]
[((5652, 5700), 'datetime.datetime.fromtimestamp', 'datetime.fromtimestamp', (['(currentTimestamp * 0.001)'], {}), '(currentTimestamp * 0.001)\n', (5674, 5700), False, 'from datetime import datetime\n'), ((11228, 11276), 'datetime.datetime.fromtimestamp', 'datetime.fromtimestamp', (['(currentTimestamp * 0.001)'], {}), ...
from c6 import People # Student类继承People class Student(People): def __init__(self,school,name,age): self.school = school # People.__init__(self,name,age) super(Student,self).__init__(name,age) Student.sum += 1 def do_homework(self): print('do homework') student1 = ...
[ "c6.People.do_homework" ]
[((425, 460), 'c6.People.do_homework', 'People.do_homework', (['""""""', '"""xiaolinzi"""'], {}), "('', 'xiaolinzi')\n", (443, 460), False, 'from c6 import People\n')]
from django.http import HttpResponseRedirect from thedaily.models import OAuthState from thedaily.views import get_or_create_user_profile def get_phone_number(backend, uid, user=None, social=None, *args, **kwargs): subscriber = get_or_create_user_profile(user) if not subscriber.phone: state = kwargs[...
[ "thedaily.views.get_or_create_user_profile", "thedaily.models.OAuthState.objects.get", "django.http.HttpResponseRedirect" ]
[((235, 267), 'thedaily.views.get_or_create_user_profile', 'get_or_create_user_profile', (['user'], {}), '(user)\n', (261, 267), False, 'from thedaily.views import get_or_create_user_profile\n'), ((737, 804), 'django.http.HttpResponseRedirect', 'HttpResponseRedirect', (["('/usuarios/registrate/google/' + query_params)"...
from django.core.management.base import BaseCommand import numpy as np import pandas as pd from django.conf import settings from baseball.models import Player, PlayerStats import sys import requests import datetime HITTING_BASE_URL = 'http://lookup-service-prod.mlb.com/json/named.sport_hitting_tm.bam' PITCHING_BASE_...
[ "baseball.models.Player.objects.all", "baseball.models.PlayerStats.objects.create", "requests.get", "datetime.datetime.now", "baseball.models.PlayerStats.objects.get" ]
[((528, 548), 'baseball.models.Player.objects.all', 'Player.objects.all', ([], {}), '()\n', (546, 548), False, 'from baseball.models import Player, PlayerStats\n'), ((480, 503), 'datetime.datetime.now', 'datetime.datetime.now', ([], {}), '()\n', (501, 503), False, 'import datetime\n'), ((799, 855), 'baseball.models.Pla...
from typing import Optional from uuid import UUID import attr from airflow_monitor.shared.base_monitor_config import BaseMonitorConfig @attr.s class BaseServerConfig(object): source_name: str = attr.ib() source_type: str = attr.ib() tracking_source_uid: UUID = attr.ib() sync_interval: int = attr.ib...
[ "attr.ib" ]
[((202, 211), 'attr.ib', 'attr.ib', ([], {}), '()\n', (209, 211), False, 'import attr\n'), ((235, 244), 'attr.ib', 'attr.ib', ([], {}), '()\n', (242, 244), False, 'import attr\n'), ((277, 286), 'attr.ib', 'attr.ib', ([], {}), '()\n', (284, 286), False, 'import attr\n'), ((313, 332), 'attr.ib', 'attr.ib', ([], {'default...
from unittest import TestCase, mock import requests from waterdata.commands.lookup_generation.wqp_lookups import ( get_lookup_by_json, is_us_county, get_nwis_state_lookup, get_nwis_county_lookup) @mock.patch('waterdata.commands.lookup_generation.wqp_lookups.execute_get_request') class GetLookupByJsonTestCase(T...
[ "waterdata.commands.lookup_generation.wqp_lookups.get_nwis_county_lookup", "requests.Response", "waterdata.commands.lookup_generation.wqp_lookups.get_lookup_by_json", "waterdata.commands.lookup_generation.wqp_lookups.get_nwis_state_lookup", "waterdata.commands.lookup_generation.wqp_lookups.is_us_county", ...
[((206, 293), 'unittest.mock.patch', 'mock.patch', (['"""waterdata.commands.lookup_generation.wqp_lookups.execute_get_request"""'], {}), "(\n 'waterdata.commands.lookup_generation.wqp_lookups.execute_get_request')\n", (216, 293), False, 'from unittest import TestCase, mock\n'), ((421, 440), 'requests.Response', 'req...
import csv # # Calculate the proportion of Four types of datasets - energy, piezo, elasticity, diel # # Energy # energy = [] # with open('training/energy/energy.csv', 'r', encoding='utf-8') as en: # reader = csv.reader(en) # for row in reader: # energy.append(row[0]) # print(len(energy)) # # # elas...
[ "csv.reader" ]
[((2693, 2707), 'csv.reader', 'csv.reader', (['en'], {}), '(en)\n', (2703, 2707), False, 'import csv\n'), ((3174, 3188), 'csv.reader', 'csv.reader', (['en'], {}), '(en)\n', (3184, 3188), False, 'import csv\n'), ((3645, 3659), 'csv.reader', 'csv.reader', (['en'], {}), '(en)\n', (3655, 3659), False, 'import csv\n')]
import numpy as np import pandas as pd from welib.tools.clean_exceptions import * from welib.FEM.graph import Node as GraphNode from welib.FEM.graph import Element as GraphElement from welib.FEM.graph import NodeProperty from welib.FEM.graph import GraphModel class MaterialProperty(NodeProperty): def __init__(se...
[ "welib.FEM.graph.Element.__init__", "welib.FEM.reduction.CraigBampton", "numpy.zeros", "welib.FEM.graph.Node.__init__", "welib.FEM.graph.GraphModel.__init__", "numpy.set_printoptions" ]
[((2329, 2363), 'numpy.set_printoptions', 'np.set_printoptions', ([], {'linewidth': '(500)'}), '(linewidth=500)\n', (2348, 2363), True, 'import numpy as np\n'), ((464, 501), 'welib.FEM.graph.Node.__init__', 'GraphNode.__init__', (['self', 'ID', 'x', 'y', 'z'], {}), '(self, ID, x, y, z)\n', (482, 501), True, 'from welib...
from django.contrib import admin from .models import Label, Choice, Quiz, Question, Answer, Submit # Register your models here. admin.site.register(Quiz) admin.site.register(Question) admin.site.register(Answer) admin.site.register(Label) admin.site.register(Submit) admin.site.register(Choice)
[ "django.contrib.admin.site.register" ]
[((129, 154), 'django.contrib.admin.site.register', 'admin.site.register', (['Quiz'], {}), '(Quiz)\n', (148, 154), False, 'from django.contrib import admin\n'), ((155, 184), 'django.contrib.admin.site.register', 'admin.site.register', (['Question'], {}), '(Question)\n', (174, 184), False, 'from django.contrib import ad...