code stringlengths 22 1.05M | apis listlengths 1 3.31k | extract_api stringlengths 75 3.25M |
|---|---|---|
# Generated by Django 3.1.12 on 2021-08-18 21:24
from django.db import migrations, models
class Migration(migrations.Migration):
dependencies = [
('reo', '0115_auto_20210810_1550'),
]
operations = [
migrations.AddField(
model_name='chpmodel',
name='supplementary_... | [
"django.db.models.FloatField"
] | [((358, 398), 'django.db.models.FloatField', 'models.FloatField', ([], {'blank': '(True)', 'null': '(True)'}), '(blank=True, null=True)\n', (375, 398), False, 'from django.db import migrations, models\n'), ((550, 590), 'django.db.models.FloatField', 'models.FloatField', ([], {'blank': '(True)', 'null': '(True)'}), '(bl... |
import factory
class ServiceCategoryFactory(factory.django.DjangoModelFactory):
class Meta:
model = "core.ServiceCategory"
name = factory.Sequence(lambda n: f"Service Category {n}")
slug = factory.Sequence(lambda n: f"service-category-{n}")
description = factory.Faker("sentence")
icon = "... | [
"factory.Faker",
"factory.Sequence",
"factory.SubFactory"
] | [((149, 200), 'factory.Sequence', 'factory.Sequence', (["(lambda n: f'Service Category {n}')"], {}), "(lambda n: f'Service Category {n}')\n", (165, 200), False, 'import factory\n'), ((212, 263), 'factory.Sequence', 'factory.Sequence', (["(lambda n: f'service-category-{n}')"], {}), "(lambda n: f'service-category-{n}')\n... |
# Generated by Django 2.1.7 on 2019-03-22 01:36
from django.db import migrations, models
class Migration(migrations.Migration):
dependencies = [
('nxos_config_import', '0009_auto_20190322_1408'),
]
operations = [
migrations.AddField(
model_name='objectconfigurationstatus',
... | [
"django.db.models.URLField"
] | [((366, 397), 'django.db.models.URLField', 'models.URLField', ([], {'default': '"""none"""'}), "(default='none')\n", (381, 397), False, 'from django.db import migrations, models\n')] |
import logging
import psycopg2
import psycopg2.extras
import socket
import sys
import time
from cluster_under_test import *
class DbRetriable:
"""
Wrapper around psycopg2, which offers convenient retry functionality.
If connection to postgres is lost during query execution or between
queries, retry wi... | [
"psycopg2.connect",
"sys.stdout.flush",
"time.sleep",
"sys.stdout.write"
] | [((1299, 1360), 'psycopg2.connect', 'psycopg2.connect', ([], {'host': 'resolved'}), '(host=resolved, **self.other_connection_args)\n', (1315, 1360), False, 'import psycopg2\n'), ((2585, 2602), 'time.sleep', 'time.sleep', (['delay'], {}), '(delay)\n', (2595, 2602), False, 'import time\n'), ((2413, 2434), 'sys.stdout.wri... |
#!/usr/bin/python3
import json
from argparse import ArgumentParser
def get_args():
p = ArgumentParser(description='Merge CLOSURE xdconf.ini files')
p.add_argument('-f', '--files', required=True, type=str, help='Input files')
p.add_argument('-o', '--outfile', required=False, type=str, default='xdconf.ini',... | [
"json.load",
"json.dump",
"argparse.ArgumentParser"
] | [((97, 157), 'argparse.ArgumentParser', 'ArgumentParser', ([], {'description': '"""Merge CLOSURE xdconf.ini files"""'}), "(description='Merge CLOSURE xdconf.ini files')\n", (111, 157), False, 'from argparse import ArgumentParser\n'), ((1406, 1437), 'json.dump', 'json.dump', (['data', 'outf'], {'indent': '(2)'}), '(data... |
'''
Copyright (c) The Dojo Foundation 2011. All Rights Reserved.
Copyright (c) IBM Corporation 2008, 2011. All Rights Reserved.
'''
# std lib
import ConfigParser
# tornado
import tornado.web
from base import AuthBase
class IniAuth(AuthBase):
cookieName = 'coweb.auth.ini.username'
def __init__(self, container, ... | [
"ConfigParser.ConfigParser"
] | [((1083, 1110), 'ConfigParser.ConfigParser', 'ConfigParser.ConfigParser', ([], {}), '()\n', (1108, 1110), False, 'import ConfigParser\n')] |
from django.conf import settings
from django_filters.rest_framework import DjangoFilterBackend
from django_filters.rest_framework import UUIDFilter
from rest_framework.permissions import IsAuthenticated
from contentcuration.celery import app
from contentcuration.models import Channel
from contentcuration.models import... | [
"contentcuration.celery.app.control.revoke",
"django_filters.rest_framework.UUIDFilter",
"contentcuration.celery.app.AsyncResult",
"contentcuration.models.Task.objects.all",
"contentcuration.models.Task.objects.filter",
"contentcuration.models.Channel.objects.all"
] | [((563, 598), 'django_filters.rest_framework.UUIDFilter', 'UUIDFilter', ([], {'method': '"""filter_channel"""'}), "(method='filter_channel')\n", (573, 598), False, 'from django_filters.rest_framework import UUIDFilter\n'), ((1036, 1054), 'contentcuration.models.Task.objects.all', 'Task.objects.all', ([], {}), '()\n', (... |
from django.shortcuts import render
from django.http import HttpResponse, JsonResponse
from django.core import serializers
from .models import Event
from .serializers import EventSerializer
from django.db.models import Q
import datetime
def get_events(request):
startDate = request.GET.get('start')
endDate = r... | [
"django.http.JsonResponse",
"datetime.datetime.strptime",
"datetime.date.today",
"django.db.models.Q",
"datetime.timedelta"
] | [((352, 373), 'datetime.date.today', 'datetime.date.today', ([], {}), '()\n', (371, 373), False, 'import datetime\n'), ((382, 403), 'datetime.date.today', 'datetime.date.today', ([], {}), '()\n', (401, 403), False, 'import datetime\n'), ((796, 799), 'django.db.models.Q', 'Q', ([], {}), '()\n', (797, 799), False, 'from ... |
# ------------------------------------
# Copyright (c) Microsoft Corporation.
# Licensed under the MIT License.
# ------------------------------------
from azure.keyvault.certificates._shared import HttpChallengeCache
from azure.keyvault.certificates._shared.client_base import DEFAULT_VERSION
from devtools_testutils im... | [
"azure.keyvault.certificates._shared.HttpChallengeCache.clear",
"pytest.skip"
] | [((640, 666), 'azure.keyvault.certificates._shared.HttpChallengeCache.clear', 'HttpChallengeCache.clear', ([], {}), '()\n', (664, 666), False, 'from azure.keyvault.certificates._shared import HttpChallengeCache\n'), ((1445, 1518), 'pytest.skip', 'pytest.skip', (['"""This test only uses the default API version for live ... |
import logging
from pyramid.httpexceptions import HTTPNotImplemented
from pyramid.renderers import render, render_to_response
log = logging.getLogger(__name__)
class RestView(object):
renderers = {}
def __init__(self, request):
self.request = request
self.params = request.params
se... | [
"logging.getLogger",
"pyramid.renderers.render",
"pyramid.httpexceptions.HTTPNotImplemented",
"pyramid.renderers.render_to_response"
] | [((134, 161), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (151, 161), False, 'import logging\n'), ((527, 550), 'pyramid.renderers.render', 'render', (['*args'], {}), '(*args, **kwargs)\n', (533, 550), False, 'from pyramid.renderers import render, render_to_response\n'), ((647, 682), 'p... |
# Let's make an database
#
# this is like the worst code ever
#
# just to make an test DB for burn the subs
import time
from datetime import datetime
from bts.dataBaseClass import Sub
def main():
fileName = open("subscriberListTest.txt")
print("making:")
for entry in fileName:
en... | [
"bts.dataBaseClass.Sub.create",
"time.sleep",
"datetime.datetime.utcnow"
] | [((360, 377), 'datetime.datetime.utcnow', 'datetime.utcnow', ([], {}), '()\n', (375, 377), False, 'from datetime import datetime\n'), ((407, 512), 'bts.dataBaseClass.Sub.create', 'Sub.create', ([], {'userName': 'entry', 'entryTime': 'dateTime', 'status': '(2)', 'fontSize': '(72)', 'positionX': '(1000)', 'positionY': '(... |
from getpass import getpass
from nxapi_plumbing import Device
from lxml import etree
from pprint import pprint as pp
# Disable Self-signed Certificate Warnings
import requests
from requests.packages.urllib3.exceptions import InsecureRequestWarning
requests.packages.urllib3.disable_warnings(InsecureRequestWarning)
'''... | [
"requests.packages.urllib3.disable_warnings",
"getpass.getpass",
"lxml.etree.tostring"
] | [((249, 315), 'requests.packages.urllib3.disable_warnings', 'requests.packages.urllib3.disable_warnings', (['InsecureRequestWarning'], {}), '(InsecureRequestWarning)\n', (291, 315), False, 'import requests\n'), ((1227, 1236), 'getpass.getpass', 'getpass', ([], {}), '()\n', (1234, 1236), False, 'from getpass import getp... |
import os
import os.path
import webapp2
import logging
from webapp2 import WSGIApplication, Route
from google.appengine.api import users
# hack until we can make this public
cache = dict()
class Content(webapp2.RequestHandler):
def get(self, *args, **kwargs):
urlPath = args[0]
root = os.path.split(__file__... | [
"webapp2.Route",
"os.path.join",
"logging.exception",
"os.path.isfile",
"os.path.split"
] | [((341, 380), 'os.path.join', 'os.path.join', (['root', '"""404"""', '"""index.html"""'], {}), "(root, '404', 'index.html')\n", (353, 380), False, 'import os\n'), ((1781, 1813), 'webapp2.Route', 'Route', (['"""/<:.*>"""'], {'handler': 'Content'}), "('/<:.*>', handler=Content)\n", (1786, 1813), False, 'from webapp2 impo... |
# This file is used to define an AML pipeline for training the teacher on new labeled data
import json
import shutil
import os
from azureml.core import Workspace, Run, Experiment, Datastore
from azureml.data.data_reference import DataReference
from azureml.pipeline.core.schedule import ScheduleRecurrence, Schedule
fr... | [
"azureml.core.runconfig.CondaDependencies.create",
"azureml.train.hyperdrive.loguniform",
"azureml.core.compute.AmlCompute.provisioning_configuration",
"azureml.core.compute.AmlCompute",
"azureml.pipeline.core.PipelineData",
"utils.azure.disable_pipeline",
"azureml.pipeline.steps.HyperDriveStep",
"azu... | [((1277, 1329), 'azureml.core.Workspace.from_config', 'Workspace.from_config', ([], {'path': 'config_json', 'auth': 'svc_pr'}), '(path=config_json, auth=svc_pr)\n', (1298, 1329), False, 'from azureml.core import Workspace, Run, Experiment, Datastore\n'), ((1677, 1718), 'os.makedirs', 'os.makedirs', (['script_folder'], ... |
"""
-------------------------------------------------------------------------
Library of RTL queues
-------------------------------------------------------------------------
Author : <NAME>
Date : Mar 23, 2019
"""
from pymtl3 import *
from pymtl3.stdlib.ifcs import DeqIfcRTL, EnqIfcRTL
from pymtl3.stdlib.rtl impor... | [
"pymtl3.stdlib.ifcs.DeqIfcRTL",
"pymtl3.stdlib.rtl.Mux",
"pymtl3.stdlib.ifcs.EnqIfcRTL",
"pymtl3.stdlib.rtl.RegisterFile"
] | [((3348, 3368), 'pymtl3.stdlib.ifcs.EnqIfcRTL', 'EnqIfcRTL', (['EntryType'], {}), '(EntryType)\n', (3357, 3368), False, 'from pymtl3.stdlib.ifcs import DeqIfcRTL, EnqIfcRTL\n'), ((3385, 3405), 'pymtl3.stdlib.ifcs.DeqIfcRTL', 'DeqIfcRTL', (['EntryType'], {}), '(EntryType)\n', (3394, 3405), False, 'from pymtl3.stdlib.ifc... |
# coding=utf-8
#
# Copyright Zucker
#
# Licensed under the Apache License, Version 2.0 (the "License"); you may
# not use this file except in compliance with the License. You may obtain
# a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in w... | [
"zipfile.ZipFile",
"re.compile",
"copy.deepcopy",
"os.path.islink",
"os.walk",
"os.path.exists",
"os.listdir",
"os.readlink",
"pathlib.Path",
"sys.argv.append",
"os.path.isdir",
"subprocess.call",
"os.mkdir",
"shutil.Error",
"shutil.copystat",
"os.path.getsize",
"subprocess.check_cal... | [((30736, 30765), 'sys.argv.append', 'sys.argv.append', (['"""ZuckerDemo"""'], {}), "('ZuckerDemo')\n", (30751, 30765), False, 'import sys\n'), ((30839, 30850), 'os.getcwd', 'os.getcwd', ([], {}), '()\n', (30848, 30850), False, 'import os\n'), ((30879, 30914), 'os.path.join', 'os.path.join', (['currentPath', '"""output... |
# Licensed to the Apache Software Foundation (ASF) under one
# or more contributor license agreements. See the NOTICE file
# distributed with this work for additional information
# regarding copyright ownership. The ASF licenses this file
# to you under the Apache License, Version 2.0 (the
# "License"); you may not u... | [
"pytest.mark.parametrize",
"unittest.mock.MagicMock",
"tests.test_utils.mock_operators.MockOperator"
] | [((1135, 1650), 'pytest.mark.parametrize', 'pytest.mark.parametrize', (['"""region_name, aws_partition,keywords,expected_value"""', "[('eu-central-1', 'aws', {}, {'region_name': 'eu-central-1', 'aws_domain':\n 'aws.amazon.com'}), ('cn-north-1', 'aws-cn', {}, {'region_name':\n 'cn-north-1', 'aws_domain': 'amazonaw... |
from django.db import models
from articles.models import Article
from users.models import User
class Comment(models.Model):
id = models.IntegerField(primary_key=True,
editable=False,
auto_created=True)
author = models.ForeignKey(User,
... | [
"django.db.models.ForeignKey",
"django.db.models.DateField",
"django.db.models.TextField",
"django.db.models.IntegerField"
] | [((136, 208), 'django.db.models.IntegerField', 'models.IntegerField', ([], {'primary_key': '(True)', 'editable': '(False)', 'auto_created': '(True)'}), '(primary_key=True, editable=False, auto_created=True)\n', (155, 208), False, 'from django.db import models\n'), ((280, 354), 'django.db.models.ForeignKey', 'models.For... |
# -*- coding: utf-8 -*-
import requests_mock
from chaoslib.run import EventHandlerRegistry
from logzero import logger
from chaosgrafana.controls.loki import configure_control
def test_sending_to_loki():
with requests_mock.Mocker() as m:
m.post("http://localhost.test:3100/loki/api/v1/push", status_code=20... | [
"chaoslib.run.EventHandlerRegistry",
"requests_mock.Mocker",
"logzero.logger.error",
"chaosgrafana.controls.loki.configure_control"
] | [((215, 237), 'requests_mock.Mocker', 'requests_mock.Mocker', ([], {}), '()\n', (235, 237), False, 'import requests_mock\n'), ((343, 365), 'chaoslib.run.EventHandlerRegistry', 'EventHandlerRegistry', ([], {}), '()\n', (363, 365), False, 'from chaoslib.run import EventHandlerRegistry\n'), ((374, 586), 'chaosgrafana.cont... |
#!/usr/bin/env python3
# -*- coding: utf-8 -*-
"""数据模型"""
__author__ = 'stone'
import re
import LianJiaWangConfig
class LianJiaWangModel(object):
def __init__(self):
self.title = ''
self.des = ''
self.time = ''
"""命名 bottom 是为了跟 HTML 代码中的名字保持一致并方便批量赋值;但其实的意义是 features"""
... | [
"re.findall",
"re.compile"
] | [((968, 994), 're.compile', 're.compile', (['"""href="(.+?)\\""""'], {}), '(\'href="(.+?)"\')\n', (978, 994), False, 'import re\n'), ((1016, 1044), 're.findall', 're.findall', (['pattern', 'striped'], {}), '(pattern, striped)\n', (1026, 1044), False, 'import re\n')] |
#!/usr/bin/env python3
from xmlrpc.client import Boolean
from sqlalchemy import false, true
import yaml
import os
import rospy
from ros_pepper_pkg.srv import *
from std_msgs.msg import String
from std_msgs.msg import Bool
def callback(message, tts_service, publisher):
"""
Topic callback.
Parameters
--... | [
"yaml.full_load",
"rospy.Publisher",
"rospy.init_node",
"rospy.get_param",
"rospy.ServiceProxy",
"os.path.join",
"rospy.spin",
"os.path.abspath",
"rospy.logdebug",
"rospy.wait_for_service"
] | [((749, 787), 'rospy.logdebug', 'rospy.logdebug', (['"""Tts interface READY."""'], {}), "('Tts interface READY.')\n", (763, 787), False, 'import rospy\n'), ((792, 804), 'rospy.spin', 'rospy.spin', ([], {}), '()\n', (802, 804), False, 'import rospy\n'), ((1114, 1140), 'rospy.init_node', 'rospy.init_node', (['node_name']... |
import sys
import random
from test_base import *
class TestBlockLD(TestBase):
def generate(self):
self.clear_tag()
for n in range(50000):
store_not_load = random.randint(0,1)
tag = random.randint(0, 15)
index = random.randint(0,self.sets_p-1)
taddr = self.get_addr(tag,index)
... | [
"random.randint"
] | [((176, 196), 'random.randint', 'random.randint', (['(0)', '(1)'], {}), '(0, 1)\n', (190, 196), False, 'import random\n'), ((208, 229), 'random.randint', 'random.randint', (['(0)', '(15)'], {}), '(0, 15)\n', (222, 229), False, 'import random\n'), ((244, 278), 'random.randint', 'random.randint', (['(0)', '(self.sets_p -... |
import sys
import json
import nltk
import nltk
from nltk.corpus import stopwords
from nltk.tokenize import word_tokenize, sent_tokenize
class Partsofspeech():
def pos(txt):
tokens = nltk.word_tokenize(txt)
return (nltk.pos_tag(tokens))
# stop_words = set(stopwords.words('english')) | [
"nltk.pos_tag",
"nltk.word_tokenize"
] | [((202, 225), 'nltk.word_tokenize', 'nltk.word_tokenize', (['txt'], {}), '(txt)\n', (220, 225), False, 'import nltk\n'), ((242, 262), 'nltk.pos_tag', 'nltk.pos_tag', (['tokens'], {}), '(tokens)\n', (254, 262), False, 'import nltk\n')] |
#!/bin/python3
# author: <NAME>
import sys
import time
basic_format = '{desc} {n:02d}/{total:02d} [{elapsed}/{remaining}] [{item_duration}/item]'
reverse_format = '{n:02d}/{total:02d} {desc} [{elapsed}/{remaining}] [{item_duration}/item]'
class Line(object):
def __init__(self, iterable=None, total=None, format=... | [
"time.time"
] | [((902, 913), 'time.time', 'time.time', ([], {}), '()\n', (911, 913), False, 'import time\n'), ((995, 1006), 'time.time', 'time.time', ([], {}), '()\n', (1004, 1006), False, 'import time\n'), ((1263, 1274), 'time.time', 'time.time', ([], {}), '()\n', (1272, 1274), False, 'import time\n')] |
import game_framework
import pico2d
import start_state
pico2d.open_canvas()
game_framework.run(start_state)
pico2d.close_canvas()
| [
"game_framework.run",
"pico2d.close_canvas",
"pico2d.open_canvas"
] | [((57, 77), 'pico2d.open_canvas', 'pico2d.open_canvas', ([], {}), '()\n', (75, 77), False, 'import pico2d\n'), ((78, 109), 'game_framework.run', 'game_framework.run', (['start_state'], {}), '(start_state)\n', (96, 109), False, 'import game_framework\n'), ((110, 131), 'pico2d.close_canvas', 'pico2d.close_canvas', ([], {... |
import os
import subprocess
from pathlib import Path
from ..views.viewhelper import delay_refresh_detail
from ..helper.config import config
def edit(filepath: Path, loop):
if isinstance(filepath, str):
filepath = Path(filepath)
editor = os.environ.get('EDITOR', 'vi').lower()
# vim
if editor ==... | [
"subprocess.check_output",
"pathlib.Path",
"pathlib.Path.cwd",
"subprocess.check_call",
"os.environ.get",
"os.chdir",
"subprocess.call",
"os.system"
] | [((1274, 1288), 'os.system', 'os.system', (['cmd'], {}), '(cmd)\n', (1283, 1288), False, 'import os\n'), ((1349, 1363), 'os.system', 'os.system', (['cmd'], {}), '(cmd)\n', (1358, 1363), False, 'import os\n'), ((227, 241), 'pathlib.Path', 'Path', (['filepath'], {}), '(filepath)\n', (231, 241), False, 'from pathlib impor... |
import json
import sys
import os
from algoliasearch.search_client import SearchClient
# This script updates the product list in products/data/product_list.json
# This is done by downloading the data from algolia index and replacing the product_list.json with the new content.
def main():
product_list_filepa... | [
"algoliasearch.search_client.SearchClient.create",
"json.dump",
"os.getcwd"
] | [((458, 528), 'algoliasearch.search_client.SearchClient.create', 'SearchClient.create', (['"""9SXIDIVU1E"""', "os.environ['ALGOLIA_ADMIN_API_KEY']"], {}), "('9SXIDIVU1E', os.environ['ALGOLIA_ADMIN_API_KEY'])\n", (477, 528), False, 'from algoliasearch.search_client import SearchClient\n'), ((325, 336), 'os.getcwd', 'os.... |
#!/usr/bin/env python
import asyncio
import io
import http.server
import json
import numpy
import os
import queue
import re
import sys
import threading
import websockets
from obspy.io.quakeml.core import Pickler
from obspy.core.stream import Stream
from obspy.core.utcdatetime import UTCDateTime
from obspy.core.event.... | [
"logging.getLogger",
"threading.Thread.__init__",
"obspy.core.stream.Stream",
"logging.StreamHandler",
"obspy.core.event.catalog.Catalog",
"json.JSONEncoder.default",
"re.compile",
"asyncio.get_event_loop",
"asyncio.new_event_loop",
"json.dumps",
"asyncio.Queue",
"re.match",
"io.BytesIO",
... | [((471, 501), 'logging.getLogger', 'logging.getLogger', (['"""viewobspy"""'], {}), "('viewobspy')\n", (488, 501), False, 'import logging\n'), ((550, 573), 'logging.StreamHandler', 'logging.StreamHandler', ([], {}), '()\n', (571, 573), False, 'import logging\n'), ((13750, 13783), 're.compile', 're.compile', (['"""eventi... |
#!/usr/bin/env python
# Created by "Thieu" at 00:16, 15/03/2022 ----------%
# Email: <EMAIL> %
# Github: https://github.com/thieu1995 % ... | [
"mealpy.bio_based.BBO.OriginalBBO",
"pytest.mark.parametrize",
"numpy.sum",
"pytest.raises",
"pytest.fixture",
"mealpy.bio_based.BBO.BaseBBO"
] | [((485, 515), 'pytest.fixture', 'pytest.fixture', ([], {'scope': '"""module"""'}), "(scope='module')\n", (499, 515), False, 'import pytest\n'), ((3242, 3489), 'pytest.mark.parametrize', 'pytest.mark.parametrize', (['"""problem, p_m, system_code"""', "[(problem, None, 0), (problem, 'hello', 0), (problem, -1.0, 0), (prob... |
from io import StringIO
from hypothesis import given, strategies
from intentionally_blank import api
def test_format_empty_with_empty_string():
with StringIO() as in_file, StringIO() as out_file:
api.format_from_file_to_file(in_file, out_file, format_names=["identity"])
assert len(out_file.getva... | [
"hypothesis.strategies.text",
"io.StringIO",
"intentionally_blank.api.format_from_file_to_file"
] | [((157, 167), 'io.StringIO', 'StringIO', ([], {}), '()\n', (165, 167), False, 'from io import StringIO\n'), ((180, 190), 'io.StringIO', 'StringIO', ([], {}), '()\n', (188, 190), False, 'from io import StringIO\n'), ((212, 286), 'intentionally_blank.api.format_from_file_to_file', 'api.format_from_file_to_file', (['in_fi... |
"""
Mask R-CNN
Train on the toy Balloon dataset and implement color splash effect.
Copyright (c) 2018 Matterport, Inc.
Licensed under the MIT License (see LICENSE for details)
Written by <NAME>
------------------------------------------------------------"""
import os
import sys
import json
import numpy as... | [
"mrcnn.model.MaskRCNN",
"os.path.join",
"numpy.array",
"mrcnn.visualize.display_instances",
"os.path.abspath",
"sys.path.append",
"matplotlib.pyplot.subplots"
] | [((541, 566), 'os.path.abspath', 'os.path.abspath', (['"""../../"""'], {}), "('../../')\n", (556, 566), False, 'import os\n'), ((590, 615), 'sys.path.append', 'sys.path.append', (['ROOT_DIR'], {}), '(ROOT_DIR)\n', (605, 615), False, 'import sys\n'), ((1185, 1215), 'os.path.join', 'os.path.join', (['ROOT_DIR', '"""logs"... |
import os, csv
path = 'F:\Movies-TV'
with open('C:\wsl\local-movies\db\movies.csv', 'w', newline='') as csvfile:
writer = csv.writer(csvfile)
for root,dirs, files in os.walk(path):
for folders in dirs:
if folders == "Subs" or folders == "Subtitles" or folders == "Other" or folders == "subtitles":
... | [
"csv.writer",
"os.walk"
] | [((126, 145), 'csv.writer', 'csv.writer', (['csvfile'], {}), '(csvfile)\n', (136, 145), False, 'import os, csv\n'), ((172, 185), 'os.walk', 'os.walk', (['path'], {}), '(path)\n', (179, 185), False, 'import os, csv\n')] |
import io
import logging
from functools import cached_property
from .base_experiment import BaseExperiment
from .utils import insert_suffix
class TqdmToLogger(io.StringIO):
buffer = ''
def __init__(self, logger, level=logging.DEBUG):
super().__init__()
self.logger = logger
self.level = ... | [
"logging.Formatter",
"logging.FileHandler",
"logging.info"
] | [((1041, 1076), 'logging.FileHandler', 'logging.FileHandler', (['name'], {'mode': '"""w"""'}), "(name, mode='w')\n", (1060, 1076), False, 'import logging\n'), ((1278, 1313), 'logging.info', 'logging.info', (['"""Logging in %s"""', 'name'], {}), "('Logging in %s', name)\n", (1290, 1313), False, 'import logging\n'), ((11... |
#!/usr/bin/python3
# Entry-point for running from the CLI when not installed via Pip, Pip will handle the console_scripts entry_points's from setup.py
# It's recommended to use `pip3 install changedetection.io` and start with `changedetection.py` instead, it will be linkd to your global path.
# or Docker.
# Read more ... | [
"mb_changedetectionio.mb_changedetection.main"
] | [((457, 482), 'mb_changedetectionio.mb_changedetection.main', 'mb_changedetection.main', ([], {}), '()\n', (480, 482), False, 'from mb_changedetectionio import mb_changedetection\n')] |
'''Autogenerated by get_gl_extensions script, do not edit!'''
from OpenGL import platform as _p
from OpenGL.GL import glget
EXTENSION_NAME = 'GL_INGR_color_clamp'
_p.unpack_constants( """GL_RED_MIN_CLAMP_INGR 0x8560
GL_GREEN_MIN_CLAMP_INGR 0x8561
GL_BLUE_MIN_CLAMP_INGR 0x8562
GL_ALPHA_MIN_CLAMP_INGR 0x8563
GL_RED_MAX_C... | [
"OpenGL.extensions.hasGLExtension",
"OpenGL.GL.glget.addGLGetConstant"
] | [((444, 495), 'OpenGL.GL.glget.addGLGetConstant', 'glget.addGLGetConstant', (['GL_RED_MIN_CLAMP_INGR', '(1,)'], {}), '(GL_RED_MIN_CLAMP_INGR, (1,))\n', (466, 495), False, 'from OpenGL.GL import glget\n'), ((498, 551), 'OpenGL.GL.glget.addGLGetConstant', 'glget.addGLGetConstant', (['GL_GREEN_MIN_CLAMP_INGR', '(1,)'], {}... |
# -*- coding: utf-8 -*-
"""
Created on Tue Nov 10 22:12:12 2020
@author: vxr131730
"""
import glob
import os
import sys
import random
import time
import numpy as np
import cv2
from test import *
from casadi import *
from numpy import random as npr
from casadi.tools import *
try:
sys.path.appe... | [
"time.sleep",
"carla.Client",
"carla.Color",
"numpy.transpose",
"glob.glob"
] | [((633, 664), 'carla.Client', 'carla.Client', (['"""localhost"""', '(2000)'], {}), "('localhost', 2000)\n", (645, 664), False, 'import carla\n'), ((4756, 4773), 'numpy.transpose', 'np.transpose', (['lbx'], {}), '(lbx)\n', (4768, 4773), True, 'import numpy as np\n'), ((4785, 4802), 'numpy.transpose', 'np.transpose', (['... |
"""
Collection of functions to calculate lag correlations
and significance following Ebisuzaki 97 JCLIM
"""
def phaseran(recblk, nsurr,ax):
""" Phaseran by <NAME>: http://www.mathworks.nl/matlabcentral/fileexchange/32621-phase-randomization/content/phaseran.m
Args:
recblk (2D array): Row: time sample.... | [
"numpy.tile",
"numpy.mean",
"numpy.flip",
"numpy.random.rand",
"numpy.flipud",
"numpy.fft.fft",
"numpy.ndim",
"numpy.exp",
"numpy.zeros",
"ds21grl.misc.AxRoll",
"numpy.concatenate",
"numpy.std",
"numpy.moveaxis",
"numpy.percentile",
"numpy.fft.ifft",
"numpy.arange"
] | [((1166, 1184), 'ds21grl.misc.AxRoll', 'AxRoll', (['recblk', 'ax'], {}), '(recblk, ax)\n', (1172, 1184), False, 'from ds21grl.misc import AxRoll\n'), ((1453, 1478), 'numpy.arange', 'np.arange', (['(1)', '(len_ser + 1)'], {}), '(1, len_ser + 1)\n', (1462, 1478), True, 'import numpy as np\n'), ((1491, 1520), 'numpy.arang... |
from django.conf.urls import url
from . import views
urlpatterns = [
url(r'^$', views.index, name="home"),
url(r'success/(?P<id>\d+)', views.success, name="success"),
url(r'register$', views.register, name="register"),
url(r'login$', views.login, name="login"),
url(r'users$', views.users, name="use... | [
"django.conf.urls.url"
] | [((74, 109), 'django.conf.urls.url', 'url', (['"""^$"""', 'views.index'], {'name': '"""home"""'}), "('^$', views.index, name='home')\n", (77, 109), False, 'from django.conf.urls import url\n'), ((116, 174), 'django.conf.urls.url', 'url', (['"""success/(?P<id>\\\\d+)"""', 'views.success'], {'name': '"""success"""'}), "(... |
import argparse
import os
import mlflow
import numpy as np
import pandas as pd
import torch
import torch.optim as optim
from matplotlib import pyplot as plt
from matplotlib.backends.backend_agg import FigureCanvasAgg
from mlflow import log_metric, log_param, get_artifact_uri
from skimage.io import imsave
from sklearn.m... | [
"mlflow.create_experiment",
"matplotlib.pyplot.grid",
"numpy.sqrt",
"matplotlib.pyplot.ylabel",
"mlflow.log_param",
"torch.cuda.is_available",
"argparse.Namespace",
"numpy.arange",
"sklearn.model_selection.ParameterGrid",
"transform.transforms",
"torch.set_grad_enabled",
"argparse.ArgumentPars... | [((6499, 6585), 'sampler.TomoBatchSampler', 'TomoBatchSampler', ([], {'batch_size': 'args.batch_size', 'data_frame': 'dataset_train.data_frame'}), '(batch_size=args.batch_size, data_frame=dataset_train.\n data_frame)\n', (6515, 6585), False, 'from sampler import TomoBatchSampler\n'), ((6687, 6800), 'torch.utils.data... |
from server_commands.argument_helpers import TunableInstanceParam, get_tunable_instance
import services
import sims4.commands
ZONE_MODIFIER_CAP = 3
@sims4.commands.Command('zone_modifier.add_zone_modifier', command_type=sims4.commands.CommandType.DebugOnly)
def add_zone_modifier(zone_modifier:TunableInstanceParam(sims... | [
"services.current_zone_id",
"server_commands.argument_helpers.get_tunable_instance",
"services.get_zone_modifier_service",
"services.get_persistence_service",
"server_commands.argument_helpers.TunableInstanceParam"
] | [((507, 541), 'services.get_persistence_service', 'services.get_persistence_service', ([], {}), '()\n', (539, 541), False, 'import services\n'), ((1593, 1627), 'services.get_persistence_service', 'services.get_persistence_service', ([], {}), '()\n', (1625, 1627), False, 'import services\n'), ((2411, 2445), 'services.ge... |
# This file is part of Ansible
#
# Ansible is free software: you can redistribute it and/or modify
# it under the terms of the GNU General Public License as published by
# the Free Software Foundation, either version 3 of the License, or
# (at your option) any later version.
#
# Ansible is distributed in the hope that ... | [
"os.path.dirname"
] | [((13231, 13256), 'os.path.dirname', 'os.path.dirname', (['__file__'], {}), '(__file__)\n', (13246, 13256), False, 'import os\n'), ((13760, 13785), 'os.path.dirname', 'os.path.dirname', (['__file__'], {}), '(__file__)\n', (13775, 13785), False, 'import os\n'), ((14466, 14491), 'os.path.dirname', 'os.path.dirname', (['_... |
import numpy as np
def point_to_seg(x1, x2) -> np.ndarray:
'''
Method:
-------
Transform 2 points into a parametrized segment. Implicitely phi is in
[-pi/2; pi/2], it is the oriented angle the segment makes with the
horizontal line passing through its middle c.
'''
c = (x1[:2] + x2[:2]... | [
"numpy.abs",
"numpy.hstack",
"numpy.sum",
"numpy.cos",
"numpy.sign",
"numpy.sin"
] | [((793, 815), 'numpy.hstack', 'np.hstack', (['[c, r, phi]'], {}), '([c, r, phi])\n', (802, 815), True, 'import numpy as np\n'), ((395, 417), 'numpy.sum', 'np.sum', (['((x2 - x1) ** 2)'], {}), '((x2 - x1) ** 2)\n', (401, 417), True, 'import numpy as np\n'), ((472, 494), 'numpy.sum', 'np.sum', (['((x2 - x1) ** 2)'], {}),... |
#!/usr/bin/python
from __future__ import absolute_import
from __future__ import print_function
import sys, os
import syslog
try:
import time
import daemon
import pwd
from . import pid
djangopath = os.path.join(os.path.dirname(sys.argv[0]), "../../")
sys.path.append(djangopath)
os.environ... | [
"main.models.ProjectUpload.objects.filter",
"pwd.getpwnam",
"os.getuid",
"time.sleep",
"syslog.syslog",
"os.path.dirname",
"sys.exc_info",
"sys.path.append"
] | [((278, 305), 'sys.path.append', 'sys.path.append', (['djangopath'], {}), '(djangopath)\n', (293, 305), False, 'import sys, os\n'), ((878, 922), 'main.models.ProjectUpload.objects.filter', 'ProjectUpload.objects.filter', ([], {'complete': '(False)'}), '(complete=False)\n', (906, 922), False, 'from main.models import Pr... |
# -*- coding: utf-8 -*-
# Copyright (C) 2010-2016 <NAME> All rights reserved
# Langstrasse 4, A--2244 Spannberg, Austria. <EMAIL>
# ****************************************************************************
# This module is part of the package GTW.__test__.
#
# This module is licensed under the terms of the BSD 3-Cla... | [
"_TFL.Regexp.Re_Replacer"
] | [((1502, 1576), '_TFL.Regexp.Re_Replacer', 'Re_Replacer', (['"""\'start\', \'\\\\d{4}-\\\\d{2}-\\\\d{2}\'"""', '"""\'start\', <date instance>"""'], {}), '("\'start\', \'\\\\d{4}-\\\\d{2}-\\\\d{2}\'", "\'start\', <date instance>")\n', (1513, 1576), False, 'from _TFL.Regexp import Re_Replacer, re\n')] |
#!/usr/bin/env python
# -*- coding: utf-8 -*-
"""
sfftk.unittests.test_readers
This testing module should have no side-effects because it only reads.
"""
from __future__ import division, print_function
import glob
import os
import struct
import sys
import unittest
import numpy
import random_words
import __init__ a... | [
"random_words.RandomWords",
"os.path.join",
"os.path.basename",
"unittest.main",
"glob.glob",
"os.remove"
] | [((544, 570), 'random_words.RandomWords', 'random_words.RandomWords', ([], {}), '()\n', (568, 570), False, 'import random_words\n'), ((17550, 17565), 'unittest.main', 'unittest.main', ([], {}), '()\n', (17563, 17565), False, 'import unittest\n'), ((693, 760), 'os.path.join', 'os.path.join', (['tests.TEST_DATA_PATH', '"... |
#!/usr/bin/env python3
# _*_ coding: utf-8 _*_
###
# Project : SubLime
# FileName : util.py
# -----------------------------------------------------------------------------
# Author : sham
# E-Mail : <EMAIL>
# -------------------------------------------------------------------------... | [
"logging.getLogger",
"os.path.splitext",
"sublime.util.get_exe_dir",
"csv.reader"
] | [((439, 466), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (456, 466), False, 'import logging\n'), ((3886, 3912), 'os.path.splitext', 'os.path.splitext', (['filepath'], {}), '(filepath)\n', (3902, 3912), False, 'import os\n'), ((2500, 2518), 'sublime.util.get_exe_dir', 'util.get_exe_dir... |
# -*- coding: utf-8 -*-
"""
@author:XuMing(<EMAIL>), <NAME>(<EMAIL>)
@description: Graph classify
"""
import numpy
from sklearn.metrics import f1_score, accuracy_score
from sklearn.multiclass import OneVsRestClassifier
from sklearn.preprocessing import MultiLabelBinarizer
class TopKRanker(OneVsRestClassifier):
d... | [
"sklearn.metrics.f1_score",
"numpy.asarray",
"numpy.random.seed",
"sklearn.preprocessing.MultiLabelBinarizer",
"sklearn.metrics.accuracy_score"
] | [((702, 727), 'numpy.asarray', 'numpy.asarray', (['all_labels'], {}), '(all_labels)\n', (715, 727), False, 'import numpy\n'), ((894, 933), 'sklearn.preprocessing.MultiLabelBinarizer', 'MultiLabelBinarizer', ([], {'sparse_output': '(True)'}), '(sparse_output=True)\n', (913, 933), False, 'from sklearn.preprocessing impor... |
###############################################################################
# Module: ddl_statement
# Purpose: Parent class for DDL (Data Definition Language) statements
#
# Notes:
#
###############################################################################
import data_pipeline.constants.const as const
... | [
"data_pipeline.constants.const.COMMASPACE.join"
] | [((889, 918), 'data_pipeline.constants.const.COMMASPACE.join', 'const.COMMASPACE.join', (['params'], {}), '(params)\n', (910, 918), True, 'import data_pipeline.constants.const as const\n')] |
# Copyright 2016 Google Inc.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, ... | [
"google.cloud.logging.entries.ProtobufEntry.from_api_repr",
"google.cloud.logging.entries.StructEntry.from_api_repr",
"google.cloud.logging.entries.TextEntry.from_api_repr"
] | [((1481, 1531), 'google.cloud.logging.entries.TextEntry.from_api_repr', 'TextEntry.from_api_repr', (['resource', 'client', 'loggers'], {}), '(resource, client, loggers)\n', (1504, 1531), False, 'from google.cloud.logging.entries import TextEntry\n'), ((1583, 1635), 'google.cloud.logging.entries.StructEntry.from_api_rep... |
import os
import pandas as pd
import numpy as np
def inp_mortality_tot():
df=pd.read_csv('total_country.csv')
mortal_rate=[]
cured_rate=[]
i=0
while(i<len(df)):
res=df.iloc[i]['Deaths']/df.iloc[i]['Confirmed']
res_2=df.iloc[i]['Cured']/df.iloc[i]['Confirmed']
mortal_rate.append(res)
cured_rate.append... | [
"os.listdir",
"pandas.read_csv"
] | [((80, 112), 'pandas.read_csv', 'pd.read_csv', (['"""total_country.csv"""'], {}), "('total_country.csv')\n", (91, 112), True, 'import pandas as pd\n'), ((482, 509), 'os.listdir', 'os.listdir', (['"""details_state"""'], {}), "('details_state')\n", (492, 509), False, 'import os\n'), ((966, 1002), 'pandas.read_csv', 'pd.r... |
# encoding: utf-8
from collections import OrderedDict
import string
from pydatacube.pydatacube import _DataCube
import px_reader
# A bit scandinavian specific
default_translate = dict(zip(
u"äöä -",
u"aoa__"
))
class Sluger(object):
def __init__(self, translate=default_translate):
self.given_out = {}
self.tra... | [
"pydatacube.pydatacube._DataCube",
"px_reader.Px",
"collections.OrderedDict"
] | [((834, 859), 'px_reader.Px', 'px_reader.Px', (['pcaxis_data'], {}), '(pcaxis_data)\n', (846, 859), False, 'import px_reader\n'), ((868, 881), 'collections.OrderedDict', 'OrderedDict', ([], {}), '()\n', (879, 881), False, 'from collections import OrderedDict\n'), ((894, 907), 'collections.OrderedDict', 'OrderedDict', (... |
# Licensed to the Apache Software Foundation (ASF) under one
# or more contributor license agreements. See the NOTICE file
# distributed with this work for additional information
# regarding copyright ownership. The ASF licenses this file
# to you under the Apache License, Version 2.0 (the
# "License"); you may not u... | [
"numpy.ones",
"tvm.relay.backend.contrib.ethosu.codegen._create_cascader",
"tvm.contrib.ethosu.cascader.EthosuDeviceConfig",
"pytest.importorskip",
"tvm.contrib.ethosu.cascader.MemoryRegion",
"tvm.IRModule",
"tvm.relay.transform.InferType",
"tvm.relay.analysis.free_vars",
"tvm.relay.var"
] | [((906, 940), 'pytest.importorskip', 'pytest.importorskip', (['"""ethosu.vela"""'], {}), "('ethosu.vela')\n", (925, 940), False, 'import pytest\n'), ((1329, 1461), 'tvm.contrib.ethosu.cascader.MemoryRegion', 'MemoryRegion', ([], {'name': '"""SRAM"""', 'size': '(10 ** 6)', 'read_bandwidth': '(16)', 'write_bandwidth': '(... |
# encoding: utf-8
'''
Created on Dec 18, 2018
@author: <NAME>
'''
import time
from array import *
from ctypes import *
from sys import exit
from multiprocessing import Process
from multiprocessing import Queue
import numpy as np
class EmotivDeviceReader(object):
'''
classdocs
This class is used to read ... | [
"multiprocessing.Process",
"time.sleep",
"numpy.array",
"sys.exit",
"multiprocessing.Queue"
] | [((9695, 9708), 'time.sleep', 'time.sleep', (['(5)'], {}), '(5)\n', (9705, 9708), False, 'import time\n'), ((494, 511), 'multiprocessing.Queue', 'Queue', ([], {'maxsize': '(-1)'}), '(maxsize=-1)\n', (499, 511), False, 'from multiprocessing import Queue\n'), ((8523, 8548), 'multiprocessing.Process', 'Process', ([], {'ta... |
import scipy
import matplotlib.pyplot as plt
import numpy as np
x = [
0.001, 0.019, 0.039, 0.058, 0.080, 0.098, 0.119, 0.139,
0.159, 0.180, 0.198, 0.249, 0.298, 0.349, 0.398, 0.419,
0.439, 0.460, 0.479, 0.499, 0.519, 0.540, 0.558, 0.578,
0.598, 0.649, 0.698, 0.749, 0.798, 0.819, 0.839, 0.859,
0.879... | [
"numpy.polyfit",
"numpy.poly1d",
"matplotlib.pyplot.subplots",
"matplotlib.pyplot.show"
] | [((724, 738), 'matplotlib.pyplot.subplots', 'plt.subplots', ([], {}), '()\n', (736, 738), True, 'import matplotlib.pyplot as plt\n'), ((1037, 1047), 'matplotlib.pyplot.show', 'plt.show', ([], {}), '()\n', (1045, 1047), True, 'import matplotlib.pyplot as plt\n'), ((827, 862), 'numpy.polyfit', 'np.polyfit', (['x', 'y', '... |
#!/usr/bin/env python
from setuptools import setup
import backtracepython
setup(
name='backtracepython',
version=backtracepython.version_string,
description='Backtrace error reporting tool for Python',
author='<NAME>',
author_email='<EMAIL>',
packages=['backtracepython'],
test_suite="test... | [
"setuptools.setup"
] | [((77, 370), 'setuptools.setup', 'setup', ([], {'name': '"""backtracepython"""', 'version': 'backtracepython.version_string', 'description': '"""Backtrace error reporting tool for Python"""', 'author': '"""<NAME>"""', 'author_email': '"""<EMAIL>"""', 'packages': "['backtracepython']", 'test_suite': '"""tests"""', 'url'... |
#
# MIT License
#
# Copyright (c) 2020 Airbyte
#
# Permission is hereby granted, free of charge, to any person obtaining a copy
# of this software and associated documentation files (the "Software"), to deal
# in the Software without restriction, including without limitation the rights
# to use, copy, modify, merge, pu... | [
"boto3.client",
"airbyte_cdk.logger.AirbyteLogger",
"time.sleep",
"boto3.resource",
"json.load"
] | [((1435, 1450), 'airbyte_cdk.logger.AirbyteLogger', 'AirbyteLogger', ([], {}), '()\n', (1448, 1450), False, 'from airbyte_cdk.logger import AirbyteLogger\n'), ((2201, 2360), 'boto3.client', 'boto3.client', (['"""s3"""'], {'aws_access_key_id': "credentials['aws_access_key_id']", 'aws_secret_access_key': "credentials['aw... |
from serverless_crud.model import BaseModel
try:
from troposphere import Sub
except ImportError:
class Sub:
def __init__(self, name):
self.name = name
def __str__(self):
return self.name
class PolicyBuilder:
def __init__(self, statements=None):
statements... | [
"troposphere.Sub"
] | [((1630, 1727), 'troposphere.Sub', 'Sub', (['f"""arn:aws:dynamodb:${{AWS::Region}}:${{AWS::AccountId}}:table/{model._meta.table_name}"""'], {}), "(f'arn:aws:dynamodb:${{AWS::Region}}:${{AWS::AccountId}}:table/{model._meta.table_name}'\n )\n", (1633, 1727), False, 'from troposphere import Sub\n'), ((1814, 1919), 'tro... |
# Installation script
from distutils.core import setup, Extension
setup(
name='neat-python',
version='0.1',
description='A NEAT (NeuroEvolution of Augmenting Topologies) implementation',
packages=['neat', 'neat/iznn', 'neat/nn', 'neat/ctrnn', 'neat/ifnn'],
#ext_modules=[
# Ex... | [
"distutils.core.setup"
] | [((66, 264), 'distutils.core.setup', 'setup', ([], {'name': '"""neat-python"""', 'version': '"""0.1"""', 'description': '"""A NEAT (NeuroEvolution of Augmenting Topologies) implementation"""', 'packages': "['neat', 'neat/iznn', 'neat/nn', 'neat/ctrnn', 'neat/ifnn']"}), "(name='neat-python', version='0.1', description=\... |
from unittest import TestCase
from src.stack import StackWithMaxValue
class TestStackWithMaxValue(TestCase):
def test_push(self):
stack = StackWithMaxValue()
stack.push(1)
stack.push(2)
stack.push(3)
self.assertEqual([1, 2, 3], stack.as_list())
def test_pop(self):
... | [
"src.stack.StackWithMaxValue"
] | [((154, 173), 'src.stack.StackWithMaxValue', 'StackWithMaxValue', ([], {}), '()\n', (171, 173), False, 'from src.stack import StackWithMaxValue\n'), ((334, 353), 'src.stack.StackWithMaxValue', 'StackWithMaxValue', ([], {}), '()\n', (351, 353), False, 'from src.stack import StackWithMaxValue\n'), ((573, 592), 'src.stack... |
# -*- coding: utf-8 -*-
from keras.models import load_model
import numpy as np
import os
import cv2
from FaceQNet import load_Qnet_model, face_quality
# Loading the pretrained model
model = load_Qnet_model()
IMG_PATH = '/home/sai/YANG/image/video/nanning/haha'
dir = os.listdir(IMG_PATH)
count = len(dir)
print('count:... | [
"os.listdir",
"FaceQNet.load_Qnet_model",
"os.path.join",
"cv2.imread",
"FaceQNet.face_quality"
] | [((191, 208), 'FaceQNet.load_Qnet_model', 'load_Qnet_model', ([], {}), '()\n', (206, 208), False, 'from FaceQNet import load_Qnet_model, face_quality\n'), ((269, 289), 'os.listdir', 'os.listdir', (['IMG_PATH'], {}), '(IMG_PATH)\n', (279, 289), False, 'import os\n'), ((427, 452), 'os.path.join', 'os.path.join', (['IMG_P... |
# Generated by Django 3.2.2 on 2021-05-10 04:54
from django.db import migrations, models
class Migration(migrations.Migration):
dependencies = [
('imgrepo', '0002_auto_20210509_2317'),
]
operations = [
migrations.AlterField(
model_name='image',
name='image_src',
... | [
"django.db.models.ImageField"
] | [((338, 389), 'django.db.models.ImageField', 'models.ImageField', ([], {'upload_to': '"""static/media/images/"""'}), "(upload_to='static/media/images/')\n", (355, 389), False, 'from django.db import migrations, models\n')] |
# Generated by Django 3.1.3 on 2022-02-28 16:55
from django.db import migrations, models
class Migration(migrations.Migration):
dependencies = [
('home', '0003_auto_20220228_1112'),
]
operations = [
migrations.CreateModel(
name='Results',
fields=[
... | [
"django.db.models.AutoField",
"django.db.models.IntegerField"
] | [((328, 421), 'django.db.models.AutoField', 'models.AutoField', ([], {'auto_created': '(True)', 'primary_key': '(True)', 'serialize': '(False)', 'verbose_name': '"""ID"""'}), "(auto_created=True, primary_key=True, serialize=False,\n verbose_name='ID')\n", (344, 421), False, 'from django.db import migrations, models\... |
from flask import Flask, request, render_template, flash, redirect, url_for
from wtforms import Form, TextField, TextAreaField, validators, StringField, SubmitField
from wtforms.validators import DataRequired
from app.DineCision import yelprequest
from flask_wtf import FlaskForm
import random
import json
import os
API... | [
"flask.render_template",
"random.choice",
"flask.flash",
"flask.Flask",
"wtforms.validators.required",
"os.environ.get",
"wtforms.SubmitField",
"flask.url_for",
"flask.request.form.get",
"app.DineCision.yelprequest"
] | [((693, 708), 'flask.Flask', 'Flask', (['__name__'], {}), '(__name__)\n', (698, 708), False, 'from flask import Flask, request, render_template, flash, redirect, url_for\n'), ((327, 363), 'os.environ.get', 'os.environ.get', (['"""DINECISION_API_KEY"""'], {}), "('DINECISION_API_KEY')\n", (341, 363), False, 'import os\n'... |
import unittest
import numpy as np
import torch
from torch import optim
from spn.structure.Base import Product, Sum
from spn.structure.Base import assign_ids, rebuild_scopes_bottom_up
from spn.structure.leaves.parametric.Parametric import Gaussian, Categorical
from spn.gpu.TensorFlow import spn_to_tf_graph, optimize_... | [
"pytorch.SumNode.from_spn",
"spn.structure.Base.Sum",
"numpy.log",
"spn.algorithms.Inference.log_likelihood",
"spn.structure.leaves.parametric.Parametric.Gaussian",
"numpy.array",
"unittest.main",
"pytorch.GaussianNode.from_spn",
"spn.structure.leaves.parametric.Parametric.Categorical",
"tensorflo... | [((7532, 7547), 'unittest.main', 'unittest.main', ([], {}), '()\n', (7545, 7547), False, 'import unittest\n'), ((790, 817), 'torch.log', 'torch.log', (['x[:, self.scope]'], {}), '(x[:, self.scope])\n', (799, 817), False, 'import torch\n'), ((1172, 1197), 'numpy.array', 'np.array', (['[0.1, 0.4, 0.5]'], {}), '([0.1, 0.4... |
# -*- coding: utf-8 -*-
"""
Profile: http://hl7.org/fhir/StructureDefinition/TerminologyCapabilities
Release: R4
Version: 4.0.1
Build ID: 9346c8cc45
Last updated: 2019-11-01T09:29:23.356+11:00
"""
import io
import json
import os
import unittest
import pytest
from .. import terminologycapabilities
from ..fhirdate imp... | [
"json.load",
"os.environ.get",
"pytest.mark.usefixtures",
"os.path.join"
] | [((370, 410), 'pytest.mark.usefixtures', 'pytest.mark.usefixtures', (['"""base_settings"""'], {}), "('base_settings')\n", (393, 410), False, 'import pytest\n'), ((526, 565), 'os.environ.get', 'os.environ.get', (['"""FHIR_UNITTEST_DATADIR"""'], {}), "('FHIR_UNITTEST_DATADIR')\n", (540, 565), False, 'import os\n'), ((677... |
#! /usr/bin/env python
import pandas as pd
import click
'''
gene expression matrix, with gene id in first column,
gene expression level of each sample in othre columns.
'''
@click.group(chain=True, invoke_without_command=True)
@click.argument('exp_table', type=click.STRING, required=True)
@click.pass_con... | [
"click.argument",
"click.group",
"click.option",
"pandas.merge",
"pandas.read_table"
] | [((187, 239), 'click.group', 'click.group', ([], {'chain': '(True)', 'invoke_without_command': '(True)'}), '(chain=True, invoke_without_command=True)\n', (198, 239), False, 'import click\n'), ((242, 303), 'click.argument', 'click.argument', (['"""exp_table"""'], {'type': 'click.STRING', 'required': '(True)'}), "('exp_t... |
# Copyright (C) 2017 <NAME> <<EMAIL>>
# This file is subject to the terms and conditions defined in
# file 'LICENSE', which is part of this source code package.
'''yaml checker'''
import yaml
from omnilint.error import Error
from omnilint.checkers import Checker
class Yaml(Checker):
extensions = ['yaml', 'yml'... | [
"yaml.load",
"omnilint.error.Error"
] | [((494, 507), 'yaml.load', 'yaml.load', (['fd'], {}), '(fd)\n', (503, 507), False, 'import yaml\n'), ((644, 766), 'omnilint.error.Error', 'Error', ([], {'msg': "(exc.context + ' ' + exc.problem)", 'file': 'origname', 'line': 'exc.problem_mark.line', 'column': 'exc.problem_mark.column'}), "(msg=exc.context + ' ' + exc.p... |
from netcad.device.l2_interfaces import InterfaceL2Access, InterfaceL2Trunk
from netcad.device import PeerInterfaceId
from netcad_demo_meraki1.vlans import vlan_native_1
from .physical import port_UTP_1G
class AccessVlan1(InterfaceL2Access):
port_profile = port_UTP_1G
vlan = vlan_native_1
desc = PeerInte... | [
"netcad.device.PeerInterfaceId"
] | [((312, 329), 'netcad.device.PeerInterfaceId', 'PeerInterfaceId', ([], {}), '()\n', (327, 329), False, 'from netcad.device import PeerInterfaceId\n')] |
"""
Ruuvi Counter (LoRaWAN) decoder
See: https://bitbucket.org/iotpetri/hki_kuva_iot/src/master/ESP32/LORA/ESP32_RuuviTagGW_Lora_v2/
"""
import json
import pytz
from dateutil.parser import parse
from broker.providers.decoder import DecoderProvider
from broker.utils import create_dataline
from fvhexperiments.parsers.r... | [
"dateutil.parser.parse",
"fvhexperiments.parsers.ruuvicounter.parse_ruuvicounter",
"broker.utils.create_dataline"
] | [((652, 682), 'fvhexperiments.parsers.ruuvicounter.parse_ruuvicounter', 'parse_ruuvicounter', (['data', 'port'], {}), '(data, port)\n', (670, 682), False, 'from fvhexperiments.parsers.ruuvicounter import parse_ruuvicounter\n'), ((812, 867), 'broker.utils.create_dataline', 'create_dataline', (['timestamp', "ruuvicounter... |
from bot import __version__
from setuptools import setup, find_packages
REQUIREMENTS = [line.strip() for line in open("requirements.txt").readlines()]
setup(name='guldai-telegram-bot',
version=__version__,
description='Telegram interface for the guldai bot',
author='isysd',
author_email='<EMAI... | [
"setuptools.setup"
] | [((153, 698), 'setuptools.setup', 'setup', ([], {'name': '"""guldai-telegram-bot"""', 'version': '__version__', 'description': '"""Telegram interface for the guldai bot"""', 'author': '"""isysd"""', 'author_email': '"""<EMAIL>"""', 'license': '"""MIT"""', 'url': '"""https://guld.io/"""', 'py_modules': "['bot']", 'insta... |
import dash_mantine_components as dmc
component = dmc.Spoiler(
showLabel="Show more",
hideLabel="Hide",
maxHeight=50,
children=[
dmc.Text(
"""We Butter the Bread with Butter was founded in 2007 by <NAME>, who was originally guitarist
for <NAME>'s band, and <NAME>. The b... | [
"dash_mantine_components.Text"
] | [((154, 833), 'dash_mantine_components.Text', 'dmc.Text', (['"""We Butter the Bread with Butter was founded in 2007 by <NAME>, who was originally guitarist \n for <NAME>\'s band, and <NAME>. The band was originally meant as a joke, but progressed \n into being a more serious musical duo. The name ... |
import json
from mock import patch
import jenkins
from tests.base import JenkinsTestBase
class JenkinsCredentialTestBase(JenkinsTestBase):
config_xml = """<com.cloudbees.plugins.credentials.impl.UsernamePasswordCredentialsImpl>
<scope>GLOBAL</scope>
<id>Test Credential</id>
<username>Test... | [
"mock.patch.object",
"json.dumps",
"jenkins.NotFoundException"
] | [((1505, 1550), 'mock.patch.object', 'patch.object', (['jenkins.Jenkins', '"""jenkins_open"""'], {}), "(jenkins.Jenkins, 'jenkins_open')\n", (1517, 1550), False, 'from mock import patch\n'), ((1787, 1832), 'mock.patch.object', 'patch.object', (['jenkins.Jenkins', '"""jenkins_open"""'], {}), "(jenkins.Jenkins, 'jenkins_... |
import logging
import pytest
from pathlib import Path
from leaf_focus.pdf.images.component import Component
from tests.base_test import BaseTest
class TestPdfImagesComponent(BaseTest):
def test_no_exe(self):
with pytest.raises(ValueError, match="Must supply exe file."):
Component(logging.getLo... | [
"logging.getLogger",
"pytest.raises",
"pathlib.Path"
] | [((919, 942), 'pathlib.Path', 'Path', (['tmp_path', '"""image"""'], {}), "(tmp_path, 'image')\n", (923, 942), False, 'from pathlib import Path\n'), ((227, 283), 'pytest.raises', 'pytest.raises', (['ValueError'], {'match': '"""Must supply exe file."""'}), "(ValueError, match='Must supply exe file.')\n", (240, 283), Fals... |
import pynput
from pynput.keyboard import Key, Listener
def on_press(key):
print(key + " was pressed")
def on_release(key):
if key == key.esc:
return False
with Listener(on_press=on_press, on_release=on_release) as listener:
listener.join() | [
"pynput.keyboard.Listener"
] | [((186, 236), 'pynput.keyboard.Listener', 'Listener', ([], {'on_press': 'on_press', 'on_release': 'on_release'}), '(on_press=on_press, on_release=on_release)\n', (194, 236), False, 'from pynput.keyboard import Key, Listener\n')] |
import pstats
p = pstats.Stats('output.prof')
p.sort_stats('cumulative').print_stats(15)
| [
"pstats.Stats"
] | [((23, 50), 'pstats.Stats', 'pstats.Stats', (['"""output.prof"""'], {}), "('output.prof')\n", (35, 50), False, 'import pstats\n')] |
from django.db import models
class Restaurant(models.Model):
"""
Restaurants' Model.
"""
id = models.CharField(primary_key = True, editable = True, max_length = 255, verbose_name = u'Id')
rating = models.IntegerField(verbose_name = u'Rating')
name = models.CharField(max_length = 255, verbose_na... | [
"django.db.models.FloatField",
"django.db.models.CharField",
"django.db.models.IntegerField"
] | [((111, 200), 'django.db.models.CharField', 'models.CharField', ([], {'primary_key': '(True)', 'editable': '(True)', 'max_length': '(255)', 'verbose_name': 'u"""Id"""'}), "(primary_key=True, editable=True, max_length=255,\n verbose_name=u'Id')\n", (127, 200), False, 'from django.db import models\n'), ((218, 261), 'd... |
"""
Useful functions for the admin panel of ImageLabeller. In particular:
* Download labels from database to json or csv
* Upload images to the database, from json (catalogue of image locations) or zip archive
NOTE:
When uploading files, or an archive full of files, we will attempt to
match the filename to a regex wi... | [
"re.search",
"image_labeller.db.session.commit",
"os.listdir",
"os.makedirs",
"image_labeller.schema.Image",
"os.path.join",
"image_labeller.schema.Label.query.all",
"datetime.datetime.now",
"json.dump",
"image_labeller.db.session.add"
] | [((835, 852), 'image_labeller.schema.Label.query.all', 'Label.query.all', ([], {}), '()\n', (850, 852), False, 'from image_labeller.schema import Label, User, Image, Category\n'), ((1937, 1971), 'os.makedirs', 'os.makedirs', (['tmpdir'], {'exist_ok': '(True)'}), '(tmpdir, exist_ok=True)\n', (1948, 1971), False, 'import... |
# Licensed under a 3-clause BSD style license - see LICENSE.rst
"""Command line tool to run tests.
This file is called `check` instead of `test` to prevent confusion
for developers and the test runner from including it in test collection.
"""
from __future__ import (absolute_import, division, print_function,
... | [
"logging.basicConfig",
"gammapy.test"
] | [((714, 792), 'logging.basicConfig', 'logging.basicConfig', ([], {'level': 'logging.DEBUG', 'format': '"""%(levelname)s - %(message)s"""'}), "(level=logging.DEBUG, format='%(levelname)s - %(message)s')\n", (733, 792), False, 'import logging\n'), ((817, 852), 'gammapy.test', 'gammapy.test', (['package'], {'verbose': '(T... |
import json
from abc import ABC, abstractmethod
from construct import Struct, Container
class StateCore(ABC):
layout: Struct = None
@classmethod
@abstractmethod
def from_container(cls, container: Container):
pass
@classmethod
def parse(cls, bytes_data: bytes, factor: int):
co... | [
"json.dumps"
] | [((696, 742), 'json.dumps', 'json.dumps', (['my_dict'], {'sort_keys': '(False)', 'indent': '(4)'}), '(my_dict, sort_keys=False, indent=4)\n', (706, 742), False, 'import json\n')] |
from conans import ConanFile, tools
import os
class ClaraConan(ConanFile):
name = "clara"
version = "1.1.5"
description = "A simple to use, composable, command line parser for C++ 11 and beyond"
url = "https://github.com/bincrafters/conan-clara"
homepage = "https://github.com/catchorg/Clara"
t... | [
"os.path.join"
] | [((782, 836), 'os.path.join', 'os.path.join', (['self._source_subfolder', '"""single_include"""'], {}), "(self._source_subfolder, 'single_include')\n", (794, 836), False, 'import os\n')] |
# -*- coding: utf-8 -*-
# @Time : 2019-09-01 17:49
# @Author : EchoShoot
# @Email : <EMAIL>
# @URL : https://github.com/EchoShoot
# @File : test_others.py
# @Explain :
from sheen import Str
import pytest
class TestOthers(object):
raw = 'xxooAß西xoox'
obj = Str.red(raw)
obj[2:-2] = Str.green
... | [
"sheen.Str.green",
"sheen.Str",
"sheen.Str.red",
"pytest.raises"
] | [((280, 292), 'sheen.Str.red', 'Str.red', (['raw'], {}), '(raw)\n', (287, 292), False, 'from sheen import Str\n'), ((2331, 2364), 'pytest.raises', 'pytest.raises', (['UnicodeEncodeError'], {}), '(UnicodeEncodeError)\n', (2344, 2364), False, 'import pytest\n'), ((2440, 2473), 'pytest.raises', 'pytest.raises', (['Unicode... |
"""
Validate CAL DAC settings XML files. The command line is:
valDACsettings [-V] [-r] [-R <root_file>] [-L <log_file>] FLE|FHE|LAC|ULD <MeV | margin> <dac_slopes_file> <dac_xml_file>
where:
-r = generate ROOT output with default name
-R <root_file> = output validation diagnostics in ROOT... | [
"logging.basicConfig",
"logging.getLogger",
"os.path.exists",
"calDacXML.calSettingsXML",
"getopt.getopt",
"ROOT.TH1F",
"logging.Formatter",
"os.path.splitext",
"numarray.logical_not",
"logging.FileHandler",
"calCalibXML.calDacSlopesCalibXML",
"numarray.where",
"sys.exit",
"ROOT.TFile",
... | [((1320, 1360), 'ROOT.TCanvas', 'ROOT.TCanvas', (['"""c_Summary"""', '"""Summary"""', '(-1)'], {}), "('c_Summary', 'Summary', -1)\n", (1332, 1360), False, 'import ROOT\n'), ((1435, 1541), 'ROOT.TH1F', 'ROOT.TH1F', (['hName', "('DAC_Val_%s: %s' % (dacType, fileName))", '(100)', '(MeV - errLimit * 2)', '(MeV + errLimit *... |
"""!
@brief Test templates for ROCK clustering module.
@authors <NAME> (<EMAIL>)
@date 2014-2020
@copyright BSD-3-Clause
"""
from pyclustering.cluster.rock import rock;
from pyclustering.utils import read_sample;
from random import random;
class RockTestTemplates:
@staticmethod
def ... | [
"pyclustering.utils.read_sample",
"random.random",
"pyclustering.cluster.rock.rock"
] | [((447, 472), 'pyclustering.utils.read_sample', 'read_sample', (['path_to_file'], {}), '(path_to_file)\n', (458, 472), False, 'from pyclustering.utils import read_sample\n'), ((509, 564), 'pyclustering.cluster.rock.rock', 'rock', (['sample', 'radius', 'cluster_numbers', 'threshold', 'ccore'], {}), '(sample, radius, clu... |
from ryu.app.wsgi import WSGIApplication
from ryu.base import app_manager
from ryu.lib import hub
from ryu.lib.dpid import dpid_to_str
from ryu.controller.handler import MAIN_DISPATCHER, DEAD_DISPATCHER, CONFIG_DISPATCHER
from ryu.controller.handler import set_ev_cls
from ryu.controller import ofp_event
from ryu.lib.p... | [
"process_stats_flow.avg_rates",
"ryu.lib.packet.packet.Packet",
"process_stats_port.avg_rates",
"ryu.base.app_manager.require_app",
"process_stats_port.stats_event",
"ryu.lib.hub.spawn",
"ryu.controller.handler.set_ev_cls",
"process_stats_flow.stats_event",
"ryu.lib.dpid.dpid_to_str",
"ryu.lib.hub... | [((9680, 9736), 'ryu.base.app_manager.require_app', 'app_manager.require_app', (['"""ryu.app.simple_switch_13_lldp"""'], {}), "('ryu.app.simple_switch_13_lldp')\n", (9703, 9736), False, 'from ryu.base import app_manager\n'), ((9737, 9785), 'ryu.base.app_manager.require_app', 'app_manager.require_app', (['"""ryu.app.res... |
import painter
import sys
from utils import distance
from collections import deque
#returns last node (destination_node)
def dfs(origin_node, dest_node, board, screen):
#put origin node on open_nodes_list (first to be expanded)
open_nodes=deque()
closed_nodes=list()
open_nodes.append(origin_node)
... | [
"collections.deque",
"painter.paint_search"
] | [((249, 256), 'collections.deque', 'deque', ([], {}), '()\n', (254, 256), False, 'from collections import deque\n'), ((476, 526), 'painter.paint_search', 'painter.paint_search', (['screen', 'curr_node.pos', 'board'], {}), '(screen, curr_node.pos, board)\n', (496, 526), False, 'import painter\n')] |
#!/bin/python
import roomai.common
from roomai.sevenking import SevenKingPublicState
from roomai.sevenking import SevenKingPrivateState
from roomai.sevenking import SevenKingPersonState
from roomai.sevenking import SevenKingAction
from roomai.sevenking import SevenKingPokerCard
from roomai.sevenking import AllSevenKing... | [
"random.shuffle",
"roomai.sevenking.SevenKingPokerCard.compare",
"roomai.sevenking.SevenKingAction",
"roomai.sevenking.SevenKingPersonState",
"roomai.sevenking.SevenKingPrivateState",
"roomai.sevenking.AllSevenKingPokerCards.values",
"roomai.sevenking.SevenKingAction.lookup",
"roomai.sevenking.SevenKi... | [((1590, 1612), 'roomai.sevenking.SevenKingPublicState', 'SevenKingPublicState', ([], {}), '()\n', (1610, 1612), False, 'from roomai.sevenking import SevenKingPublicState\n'), ((1642, 1665), 'roomai.sevenking.SevenKingPrivateState', 'SevenKingPrivateState', ([], {}), '()\n', (1663, 1665), False, 'from roomai.sevenking ... |
from __future__ import unicode_literals
from django.contrib.auth.models import AnonymousUser, User
from django.http import HttpRequest
from django.template import RequestContext, Template
from djblets.avatars.services.gravatar import GravatarService
from reviewboard.testing import TestCase
from reviewboard.avatars im... | [
"reviewboard.avatars.tests.DummyAvatarService",
"django.template.Template",
"django.contrib.auth.models.AnonymousUser",
"reviewboard.avatars.avatar_services.set_default_service",
"django.template.RequestContext",
"reviewboard.avatars.avatar_services.reset",
"reviewboard.avatars.avatar_services.get_avata... | [((580, 603), 'reviewboard.avatars.avatar_services.reset', 'avatar_services.reset', ([], {}), '()\n', (601, 603), False, 'from reviewboard.avatars import avatar_services\n'), ((631, 700), 'reviewboard.avatars.avatar_services.get_avatar_service', 'avatar_services.get_avatar_service', (['GravatarService.avatar_service_id... |
import jwt
import requests
from cryptography.hazmat.backends import default_backend
from cryptography.hazmat.primitives.asymmetric.rsa import RSAPublicNumbers
from oauth2 import Oauth2, ResponseObject, RemoteUserProfile, OAuthDiscoveryError
class OpenID(Oauth2):
state_token: str or None = None
well_know_pat... | [
"jwt.decode",
"requests.get",
"jwt.get_unverified_header",
"jwt.api_jws.base64url_decode",
"oauth2.OAuthDiscoveryError",
"cryptography.hazmat.backends.default_backend",
"cryptography.hazmat.primitives.asymmetric.rsa.RSAPublicNumbers"
] | [((1388, 1420), 'jwt.get_unverified_header', 'jwt.get_unverified_header', (['token'], {}), '(token)\n', (1413, 1420), False, 'import jwt\n'), ((2101, 2132), 'jwt.api_jws.base64url_decode', 'jwt.api_jws.base64url_decode', (['e'], {}), '(e)\n', (2129, 2132), False, 'import jwt\n'), ((2162, 2193), 'jwt.api_jws.base64url_d... |
from dvc_preprocessing import plot, preprocessing, constants
from skimage.filters import threshold_otsu
import numpy as np
def auto_processing(filename, dirpath='./', data_type=np.int16, init_slice=0, final_slice="last", outname="output", ret="True"):
'''
TODO: add outpath
'''
stack = preprocessing.r... | [
"dvc_preprocessing.constants.INT8MINVAL",
"dvc_preprocessing.preprocessing.save_3d_tiff",
"skimage.filters.threshold_otsu",
"dvc_preprocessing.preprocessing.volume_CoM",
"dvc_preprocessing.preprocessing.read_images_from_h5",
"dvc_preprocessing.preprocessing.crop_around_CoM",
"dvc_preprocessing.preproces... | [((305, 368), 'dvc_preprocessing.preprocessing.read_images_from_h5', 'preprocessing.read_images_from_h5', (['filename', 'data_type', 'dirpath'], {}), '(filename, data_type, dirpath)\n', (338, 368), False, 'from dvc_preprocessing import plot, preprocessing, constants\n'), ((392, 413), 'skimage.filters.threshold_otsu', '... |
from __future__ import absolute_import, division, print_function
import pytest
from ..spparser import Scanner
scanner = Scanner()
# Test of a single instance of each token. Does not test them in
# context, but at least it tests that each one is recognized.
tokens = [
# bug: the original pysynphot could not reco... | [
"pytest.mark.xfail",
"pytest.mark.parametrize"
] | [((2241, 6212), 'pytest.mark.parametrize', 'pytest.mark.parametrize', (["('text', 'result')", "[('spec($PYSYN_CDBS//calspec/gd71_mod_005.fits)', [('IDENTIFIER', 'spec'),\n ('LPAREN', None), ('IDENTIFIER',\n '$PYSYN_CDBS//calspec/gd71_mod_005.fits'), ('RPAREN', None)]), (\n 'spec(earthshine.fits)*0.5+rn(spec(Zo... |
#
# This program is free software: you can redistribute it and/or modify
# it under the terms of the GNU Lesser General Public License as published by
# the Free Software Foundation, either version 3 of the License, or
# (at your option) any later version.
#
# This program is distributed in the hope that it will be use... | [
"os.path.realpath",
"proton.template.get_template"
] | [((1270, 1313), 'proton.template.get_template', 'template.get_template', (['"""twotemplates.xhtml"""'], {}), "('twotemplates.xhtml')\n", (1291, 1313), False, 'from proton import template\n'), ((1686, 1727), 'proton.template.get_template', 'template.get_template', (['"""twotemplates.xml"""'], {}), "('twotemplates.xml')\... |
# -*- coding: utf-8 -*-
# PLEASE DO NOT EDIT THIS FILE, IT IS GENERATED AND WILL BE OVERWRITTEN:
# https://github.com/ccxt/ccxt/blob/master/CONTRIBUTING.md#how-to-contribute-code
from ccxt.base.exchange import Exchange
from ccxt.base.errors import ExchangeError
from ccxt.base.errors import BadRequest
from ccxt.base.e... | [
"datetime.datetime.fromtimestamp",
"json.dumps",
"sys.exc_info",
"datetime.datetime.now",
"ccxt.base.errors.ExchangeError"
] | [((5652, 5700), 'datetime.datetime.fromtimestamp', 'datetime.fromtimestamp', (['(currentTimestamp * 0.001)'], {}), '(currentTimestamp * 0.001)\n', (5674, 5700), False, 'from datetime import datetime\n'), ((11228, 11276), 'datetime.datetime.fromtimestamp', 'datetime.fromtimestamp', (['(currentTimestamp * 0.001)'], {}), ... |
from c6 import People
# Student类继承People
class Student(People):
def __init__(self,school,name,age):
self.school = school
# People.__init__(self,name,age)
super(Student,self).__init__(name,age)
Student.sum += 1
def do_homework(self):
print('do homework')
student1 = ... | [
"c6.People.do_homework"
] | [((425, 460), 'c6.People.do_homework', 'People.do_homework', (['""""""', '"""xiaolinzi"""'], {}), "('', 'xiaolinzi')\n", (443, 460), False, 'from c6 import People\n')] |
from django.http import HttpResponseRedirect
from thedaily.models import OAuthState
from thedaily.views import get_or_create_user_profile
def get_phone_number(backend, uid, user=None, social=None, *args, **kwargs):
subscriber = get_or_create_user_profile(user)
if not subscriber.phone:
state = kwargs[... | [
"thedaily.views.get_or_create_user_profile",
"thedaily.models.OAuthState.objects.get",
"django.http.HttpResponseRedirect"
] | [((235, 267), 'thedaily.views.get_or_create_user_profile', 'get_or_create_user_profile', (['user'], {}), '(user)\n', (261, 267), False, 'from thedaily.views import get_or_create_user_profile\n'), ((737, 804), 'django.http.HttpResponseRedirect', 'HttpResponseRedirect', (["('/usuarios/registrate/google/' + query_params)"... |
from django.core.management.base import BaseCommand
import numpy as np
import pandas as pd
from django.conf import settings
from baseball.models import Player, PlayerStats
import sys
import requests
import datetime
HITTING_BASE_URL = 'http://lookup-service-prod.mlb.com/json/named.sport_hitting_tm.bam'
PITCHING_BASE_... | [
"baseball.models.Player.objects.all",
"baseball.models.PlayerStats.objects.create",
"requests.get",
"datetime.datetime.now",
"baseball.models.PlayerStats.objects.get"
] | [((528, 548), 'baseball.models.Player.objects.all', 'Player.objects.all', ([], {}), '()\n', (546, 548), False, 'from baseball.models import Player, PlayerStats\n'), ((480, 503), 'datetime.datetime.now', 'datetime.datetime.now', ([], {}), '()\n', (501, 503), False, 'import datetime\n'), ((799, 855), 'baseball.models.Pla... |
from typing import Optional
from uuid import UUID
import attr
from airflow_monitor.shared.base_monitor_config import BaseMonitorConfig
@attr.s
class BaseServerConfig(object):
source_name: str = attr.ib()
source_type: str = attr.ib()
tracking_source_uid: UUID = attr.ib()
sync_interval: int = attr.ib... | [
"attr.ib"
] | [((202, 211), 'attr.ib', 'attr.ib', ([], {}), '()\n', (209, 211), False, 'import attr\n'), ((235, 244), 'attr.ib', 'attr.ib', ([], {}), '()\n', (242, 244), False, 'import attr\n'), ((277, 286), 'attr.ib', 'attr.ib', ([], {}), '()\n', (284, 286), False, 'import attr\n'), ((313, 332), 'attr.ib', 'attr.ib', ([], {'default... |
from unittest import TestCase, mock
import requests
from waterdata.commands.lookup_generation.wqp_lookups import (
get_lookup_by_json, is_us_county, get_nwis_state_lookup, get_nwis_county_lookup)
@mock.patch('waterdata.commands.lookup_generation.wqp_lookups.execute_get_request')
class GetLookupByJsonTestCase(T... | [
"waterdata.commands.lookup_generation.wqp_lookups.get_nwis_county_lookup",
"requests.Response",
"waterdata.commands.lookup_generation.wqp_lookups.get_lookup_by_json",
"waterdata.commands.lookup_generation.wqp_lookups.get_nwis_state_lookup",
"waterdata.commands.lookup_generation.wqp_lookups.is_us_county",
... | [((206, 293), 'unittest.mock.patch', 'mock.patch', (['"""waterdata.commands.lookup_generation.wqp_lookups.execute_get_request"""'], {}), "(\n 'waterdata.commands.lookup_generation.wqp_lookups.execute_get_request')\n", (216, 293), False, 'from unittest import TestCase, mock\n'), ((421, 440), 'requests.Response', 'req... |
import csv
# # Calculate the proportion of Four types of datasets - energy, piezo, elasticity, diel
# # Energy
# energy = []
# with open('training/energy/energy.csv', 'r', encoding='utf-8') as en:
# reader = csv.reader(en)
# for row in reader:
# energy.append(row[0])
# print(len(energy))
#
# # elas... | [
"csv.reader"
] | [((2693, 2707), 'csv.reader', 'csv.reader', (['en'], {}), '(en)\n', (2703, 2707), False, 'import csv\n'), ((3174, 3188), 'csv.reader', 'csv.reader', (['en'], {}), '(en)\n', (3184, 3188), False, 'import csv\n'), ((3645, 3659), 'csv.reader', 'csv.reader', (['en'], {}), '(en)\n', (3655, 3659), False, 'import csv\n')] |
import numpy as np
import pandas as pd
from welib.tools.clean_exceptions import *
from welib.FEM.graph import Node as GraphNode
from welib.FEM.graph import Element as GraphElement
from welib.FEM.graph import NodeProperty
from welib.FEM.graph import GraphModel
class MaterialProperty(NodeProperty):
def __init__(se... | [
"welib.FEM.graph.Element.__init__",
"welib.FEM.reduction.CraigBampton",
"numpy.zeros",
"welib.FEM.graph.Node.__init__",
"welib.FEM.graph.GraphModel.__init__",
"numpy.set_printoptions"
] | [((2329, 2363), 'numpy.set_printoptions', 'np.set_printoptions', ([], {'linewidth': '(500)'}), '(linewidth=500)\n', (2348, 2363), True, 'import numpy as np\n'), ((464, 501), 'welib.FEM.graph.Node.__init__', 'GraphNode.__init__', (['self', 'ID', 'x', 'y', 'z'], {}), '(self, ID, x, y, z)\n', (482, 501), True, 'from welib... |
from django.contrib import admin
from .models import Label, Choice, Quiz, Question, Answer, Submit
# Register your models here.
admin.site.register(Quiz)
admin.site.register(Question)
admin.site.register(Answer)
admin.site.register(Label)
admin.site.register(Submit)
admin.site.register(Choice) | [
"django.contrib.admin.site.register"
] | [((129, 154), 'django.contrib.admin.site.register', 'admin.site.register', (['Quiz'], {}), '(Quiz)\n', (148, 154), False, 'from django.contrib import admin\n'), ((155, 184), 'django.contrib.admin.site.register', 'admin.site.register', (['Question'], {}), '(Question)\n', (174, 184), False, 'from django.contrib import ad... |