code stringlengths 22 1.05M | apis listlengths 1 3.31k | extract_api stringlengths 75 3.25M |
|---|---|---|
# pip install psycopg
import psycopg
# Connect to database
client = psycopg.connect("postgres://username:password@127.0.0.1")
cursor = client.cursor()
# Create table
cursor.execute(
"""
CREATE TABLE table_name(
sku integer,
name varchar(255),
description text,
category varchar(... | [
"psycopg.connect"
] | [((69, 126), 'psycopg.connect', 'psycopg.connect', (['"""postgres://username:password@127.0.0.1"""'], {}), "('postgres://username:password@127.0.0.1')\n", (84, 126), False, 'import psycopg\n')] |
import autocomplete_light
from party.functions import portfolio_render_reference
from party.models import Portfolio
class PortfolioAutocomplete(autocomplete_light.AutocompleteModelBase):
choices = Portfolio.objects.filter().order_by('-ordering')
search_fields = ['title']
display_edit_link = True
fie... | [
"party.functions.portfolio_render_reference",
"autocomplete_light.register",
"party.models.Portfolio.objects.filter"
] | [((631, 692), 'autocomplete_light.register', 'autocomplete_light.register', (['Portfolio', 'PortfolioAutocomplete'], {}), '(Portfolio, PortfolioAutocomplete)\n', (658, 692), False, 'import autocomplete_light\n'), ((396, 471), 'party.functions.portfolio_render_reference', 'portfolio_render_reference', (['choice', 'self.... |
'''
Reinforcement learning Enviroment Definition for Network anomaly detection
'''
import logging.config
from gym import spaces
import gym
from helpers_data_preprocessing import data_cls
class NetworkClassificationEnv(gym.Env,data_cls):
'''
Environment class definition. It represents the dynamics of the A... | [
"helpers_data_preprocessing.data_cls.__init__",
"gym.spaces.Discrete"
] | [((1371, 1428), 'helpers_data_preprocessing.data_cls.__init__', 'data_cls.__init__', (['self', 'train_test', 'attack_map'], {}), '(self, train_test, attack_map, **kwargs)\n', (1388, 1428), False, 'from helpers_data_preprocessing import data_cls\n'), ((1747, 1782), 'gym.spaces.Discrete', 'spaces.Discrete', (['self.data_... |
"""Quirks for Develco Products A/S."""
from zigpy import types as t
from zigpy.quirks import CustomCluster
from zigpy.zcl.clusters.security import IasZone
from zhaquirks import PowerConfigurationCluster
FRIENT = "frient A/S"
DEVELCO = "Develco Products A/S"
class DevelcoPowerConfiguration(PowerConfigurationCluster... | [
"zigpy.types.Optional"
] | [((673, 694), 'zigpy.types.Optional', 't.Optional', (['t.bitmap8'], {}), '(t.bitmap8)\n', (683, 694), True, 'from zigpy import types as t\n'), ((712, 733), 'zigpy.types.Optional', 't.Optional', (['t.uint8_t'], {}), '(t.uint8_t)\n', (722, 733), True, 'from zigpy import types as t\n'), ((751, 773), 'zigpy.types.Optional'... |
from PIL import Image, ImageDraw, ImageFont
import numpy as np
import random
from phi.fluidformat import *
def text_to_pixels(text, size=10, binary=False, as_numpy_array=True):
image = Image.new("1" if binary else "L", (len(text)*size*3//4, size), 0)
draw = ImageDraw.Draw(image)
try:
font = ImageF... | [
"random.choice",
"PIL.ImageFont.truetype",
"numpy.sum",
"numpy.array",
"PIL.ImageDraw.Draw",
"numpy.zeros",
"numpy.linspace",
"numpy.all",
"random.randint"
] | [((268, 289), 'PIL.ImageDraw.Draw', 'ImageDraw.Draw', (['image'], {}), '(image)\n', (282, 289), False, 'from PIL import Image, ImageDraw, ImageFont\n'), ((861, 876), 'numpy.zeros', 'np.zeros', (['shape'], {}), '(shape)\n', (869, 876), True, 'import numpy as np\n'), ((1620, 1635), 'numpy.zeros', 'np.zeros', (['shape'], ... |
from schematics import Model
from schematics.types import ListType, StringType, PolyModelType, DictType
from .base import BaseResponse
class RegionResource(Model):
name = StringType(default="")
region_code = StringType()
provider = StringType(default="azure")
tags = DictType(StringType)
class Region... | [
"schematics.types.PolyModelType",
"schematics.types.StringType",
"schematics.types.DictType",
"schematics.types.ListType"
] | [((177, 199), 'schematics.types.StringType', 'StringType', ([], {'default': '""""""'}), "(default='')\n", (187, 199), False, 'from schematics.types import ListType, StringType, PolyModelType, DictType\n'), ((218, 230), 'schematics.types.StringType', 'StringType', ([], {}), '()\n', (228, 230), False, 'from schematics.ty... |
# Copyright (C) 2015-2021 by Vd.
# This file is part of Rocketgram, the modern Telegram bot framework.
# Rocketgram is released under the MIT License (see LICENSE).
from dataclasses import dataclass
from typing import Optional
from .input_message_content import InputMessageContent
@dataclass(frozen=True)
class Inp... | [
"dataclasses.dataclass"
] | [((288, 310), 'dataclasses.dataclass', 'dataclass', ([], {'frozen': '(True)'}), '(frozen=True)\n', (297, 310), False, 'from dataclasses import dataclass\n')] |
import sys
from typing import Callable, Dict, Optional, Tuple, Union
import numpy as np
import scipy.sparse
import scipy.sparse.linalg
from datafold.utils.general import is_symmetric_matrix, sort_eigenpairs
class NumericalMathError(Exception):
"""Use for numerical problems/issues, such as singular matrices or t... | [
"numpy.ones",
"datafold.utils.general.is_symmetric_matrix",
"numpy.any",
"numpy.real",
"datafold.utils.general.sort_eigenpairs",
"numpy.isfinite",
"numpy.linalg.norm"
] | [((8006, 8040), 'datafold.utils.general.sort_eigenpairs', 'sort_eigenpairs', (['eigvals', 'eigvects'], {}), '(eigvals, eigvects)\n', (8021, 8040), False, 'from datafold.utils.general import is_symmetric_matrix, sort_eigenpairs\n'), ((7519, 7572), 'numpy.any', 'np.any', (['(eigvals.imag > 100.0 * sys.float_info.epsilon)... |
# Licensed Materials - Property of IBM
# Copyright IBM Corp. 2016
import unittest
import sys
import dill
import pickle
import datetime
import time
import random
from streamsx.spl.types import Timestamp
from streamsx.topology import schema
class TestTypes(unittest.TestCase):
def test_Timestamp(self):
s = rand... | [
"pickle.dumps",
"streamsx.spl.types.Timestamp.from_datetime",
"datetime.datetime.now",
"streamsx.spl.types.Timestamp.now",
"streamsx.spl.types.Timestamp.from_time",
"streamsx.spl.types.Timestamp",
"dill.dumps",
"time.time",
"random.randint"
] | [((316, 347), 'random.randint', 'random.randint', (['(0)', '(999999999999)'], {}), '(0, 999999999999)\n', (330, 347), False, 'import random\n'), ((359, 388), 'random.randint', 'random.randint', (['(0)', '(1000000000)'], {}), '(0, 1000000000)\n', (373, 388), False, 'import random\n'), ((401, 426), 'random.randint', 'ran... |
# Zope Imports
from zope.interface import implementer
# Twisted Imports
from twisted.internet import reactor, defer, serialport
from twisted.internet.endpoints import TCP4ClientEndpoint
from twisted.internet.interfaces import IAddress
from twisted.python.util import FancyEqMixin
#
# Transports have a connect() functi... | [
"twisted.internet.endpoints.TCP4ClientEndpoint",
"zope.interface.implementer"
] | [((713, 734), 'zope.interface.implementer', 'implementer', (['IAddress'], {}), '(IAddress)\n', (724, 734), False, 'from zope.interface import implementer\n'), ((535, 574), 'twisted.internet.endpoints.TCP4ClientEndpoint', 'TCP4ClientEndpoint', (['reactor', 'host', 'port'], {}), '(reactor, host, port)\n', (553, 574), Fal... |
# Copyright 2018 Google LLC
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, ... | [
"sys.path.insert",
"tensorflow.compat.v1.gfile.GFile",
"re.compile",
"sgf_wrapper.get_sgf_root_node",
"coords.from_flat",
"sys.exit",
"json.dumps",
"absl.flags.DEFINE_boolean",
"absl.app.run",
"coords.from_sgf",
"tensorflow.compat.v1.gfile.ListDirectory",
"collections.namedtuple",
"os.path.s... | [((1213, 1236), 'sys.path.insert', 'sys.path.insert', (['(0)', '"""."""'], {}), "(0, '.')\n", (1228, 1236), False, 'import sys\n'), ((1467, 1593), 'collections.namedtuple', 'collections.namedtuple', (['"""DebugRow"""', "['move', 'action', 'Q', 'U', 'prior', 'orig_prior', 'N', 'soft_N',\n 'p_delta', 'p_rel']"], {}), ... |
########################################
__author__ = "<NAME>"
__license__ = "GNU GPLv3"
__version__ = "0.1"
__maintainer__ = "<NAME>"
__email__ = "<EMAIL>"
########################################
import torch
import torch.nn as nn
import torch.nn.functional as F
from torch.nn.parameter import Parameter
from torch.nn... | [
"torch.nn.functional.conv2d",
"torch.nn.parameter.Parameter",
"torch.rand",
"numpy.arange",
"torch.nn.init.xavier_uniform_",
"torch.unsqueeze",
"torch.Tensor",
"torch.exp",
"torch.nn.init.kaiming_uniform_",
"torch.nn.functional.sigmoid",
"torch.nn.functional.softplus",
"torch.sum",
"scipy.st... | [((1536, 1581), 'torch.nn.functional.max_pool2d', 'F.max_pool2d', (['c1', 'ds', 'ds'], {'return_indices': '(True)'}), '(c1, ds, ds, return_indices=True)\n', (1548, 1581), True, 'import torch.nn.functional as F\n'), ((2012, 2060), 'torch.nn.functional.max_pool2d', 'F.max_pool2d', (['c2_ds', 'ds', 'ds'], {'return_indices... |
import os
import tempfile
def load_model(path, compile=True):
import train
return train._load_model(path, compile=compile)
def copy_weights(from_model, to_model, by_name=True):
with tempfile.TemporaryDirectory(dir='.') as d:
path = os.path.join(d, 'weights.h5')
from_model.save_weights(pa... | [
"tempfile.TemporaryDirectory",
"train._load_model",
"os.path.join"
] | [((92, 132), 'train._load_model', 'train._load_model', (['path'], {'compile': 'compile'}), '(path, compile=compile)\n', (109, 132), False, 'import train\n'), ((198, 234), 'tempfile.TemporaryDirectory', 'tempfile.TemporaryDirectory', ([], {'dir': '"""."""'}), "(dir='.')\n", (225, 234), False, 'import tempfile\n'), ((256... |
import logging
import praw
import redis
import threading
from rusertracker.stream import Stream
class SubredditWorker(threading.Thread):
def __init__(self, evt: threading.Event, subreddit: str, user_agent: str, user: str, redis_kwargs: dict,
ttl: int=7 * 86400):
self._logger = logging.get... | [
"logging.getLogger",
"rusertracker.stream.Stream",
"praw.Reddit",
"redis.Redis"
] | [((309, 362), 'logging.getLogger', 'logging.getLogger', (['f"""rusertracker.worker.{subreddit}"""'], {}), "(f'rusertracker.worker.{subreddit}')\n", (326, 362), False, 'import logging\n'), ((598, 625), 'redis.Redis', 'redis.Redis', ([], {}), '(**redis_kwargs)\n', (609, 625), False, 'import redis\n'), ((821, 871), 'praw.... |
from django.conf import settings
from django.db import migrations, models
class Migration(migrations.Migration):
initial = True
dependencies = [
migrations.swappable_dependency(settings.AUTH_USER_MODEL),
]
operations = [
migrations.CreateModel(
name='Bookmark',
... | [
"django.db.migrations.AlterUniqueTogether",
"django.db.migrations.swappable_dependency",
"django.db.models.AutoField",
"django.db.models.ForeignKey"
] | [((165, 222), 'django.db.migrations.swappable_dependency', 'migrations.swappable_dependency', (['settings.AUTH_USER_MODEL'], {}), '(settings.AUTH_USER_MODEL)\n', (196, 222), False, 'from django.db import migrations, models\n'), ((889, 978), 'django.db.migrations.AlterUniqueTogether', 'migrations.AlterUniqueTogether', (... |
"""Trace Logger Class"""
# standard library
import logging
from inspect import getframeinfo, stack
# Create trace logging level
logging.TRACE = logging.DEBUG - 5
logging.addLevelName(logging.TRACE, 'TRACE')
class TraceLogger(logging.Logger):
"""Add trace level to logging"""
# supports updated call for Pytho... | [
"inspect.stack",
"logging.addLevelName"
] | [((163, 207), 'logging.addLevelName', 'logging.addLevelName', (['logging.TRACE', '"""TRACE"""'], {}), "(logging.TRACE, 'TRACE')\n", (183, 207), False, 'import logging\n'), ((804, 811), 'inspect.stack', 'stack', ([], {}), '()\n', (809, 811), False, 'from inspect import getframeinfo, stack\n')] |
import pendulum
import pytest
from toggl.api import TimeEntry
from toggl import exceptions
class TestTimeEntries:
def test_ls(self, cmd, fake, factories):
midpoint = pendulum.now(tz='utc') - pendulum.duration(days=2)
start = fake.date_time_between(start_date="-5d", end_date=midpoint,)
... | [
"toggl.api.TimeEntry.objects.current",
"toggl.api.TimeEntry.objects.filter",
"pendulum.duration",
"toggl.api.TimeEntry.start_and_save",
"pendulum.now"
] | [((5499, 5539), 'toggl.api.TimeEntry.objects.current', 'TimeEntry.objects.current', ([], {'config': 'config'}), '(config=config)\n', (5524, 5539), False, 'from toggl.api import TimeEntry\n'), ((5759, 5799), 'toggl.api.TimeEntry.objects.current', 'TimeEntry.objects.current', ([], {'config': 'config'}), '(config=config)\... |
import time, os, django, pathlib, pandas as pd
os.environ.setdefault("DJANGO_SETTINGS_MODULE", "xna2.settings")
django.setup()
from noma.models import NomaSet, NomaFunc, NomaStrMap, NomaStrMapSet, NomaSetAct as sa
from noma.utils import nomaMain, nomaCreateTbl, nomaExecT
from django.db import connection, connections
i... | [
"os.environ.setdefault",
"noma.models.NomaSetAct",
"django.setup",
"pathlib.Path",
"noma.models.NomaFunc.objects.get",
"noma.utils.nomaCreateTbl",
"pandas.ExcelFile",
"noma.models.NomaSet",
"noma.utils.nomaMain",
"time.time",
"noma.models.NomaStrMap.objects.all"
] | [((47, 111), 'os.environ.setdefault', 'os.environ.setdefault', (['"""DJANGO_SETTINGS_MODULE"""', '"""xna2.settings"""'], {}), "('DJANGO_SETTINGS_MODULE', 'xna2.settings')\n", (68, 111), False, 'import time, os, django, pathlib, pandas as pd\n'), ((112, 126), 'django.setup', 'django.setup', ([], {}), '()\n', (124, 126),... |
# Computing player vs player statistics by parsing over all the IPL yaml files downloaded from the cricsheet.org
import yaml
import re
import csv
import os
playerStatistics = {} # Player statistics is the ditionary which will have all the player vs player statistics.
playerList = []
ok = 1
def probabilityFunction(fi... | [
"os.listdir",
"re.compile",
"csv.writer",
"os.path.join",
"yaml.load"
] | [((7121, 7144), 'csv.writer', 'csv.writer', (['filePointer'], {}), '(filePointer)\n', (7131, 7144), False, 'import csv\n'), ((7664, 7687), 'csv.writer', 'csv.writer', (['filePointer'], {}), '(filePointer)\n', (7674, 7687), False, 'import csv\n'), ((569, 593), 're.compile', 're.compile', (['"""dates:(.*)"""'], {}), "('d... |
import os
from askapdev.rbuild.builders import Autotools as Builder
import askapdev.rbuild.utils as utils
builder = Builder(buildsubdir="build_unix",
confcommand='../dist/configure')
builder.remote_archive = "db-5.3.21.NC.tar.gz"
if os.uname()[4] == 'x86_64':
builder.add_option('--with-pic')
#... | [
"os.environ.get",
"askapdev.rbuild.builders.Autotools",
"askapdev.rbuild.utils.get_platform",
"os.environ.has_key",
"os.uname"
] | [((118, 184), 'askapdev.rbuild.builders.Autotools', 'Builder', ([], {'buildsubdir': '"""build_unix"""', 'confcommand': '"""../dist/configure"""'}), "(buildsubdir='build_unix', confcommand='../dist/configure')\n", (125, 184), True, 'from askapdev.rbuild.builders import Autotools as Builder\n'), ((746, 766), 'askapdev.rb... |
import unittest
from pychess.core.gamer import Game
from pychess.element.squarer import Square
from pychess.element.piecer import Piece
from pychess import constant as c
from pychess.core.mover import Move
def _check_win(game):
if game.is_game_over:
winner = str(game.winner.name).capitalize()
pr... | [
"pychess.element.piecer.Piece",
"pychess.core.gamer.Game.parse_move_spec",
"pychess.core.gamer.Game",
"unittest.main",
"pychess.element.squarer.Square"
] | [((11835, 11850), 'unittest.main', 'unittest.main', ([], {}), '()\n', (11848, 11850), False, 'import unittest\n'), ((801, 807), 'pychess.core.gamer.Game', 'Game', ([], {}), '()\n', (805, 807), False, 'from pychess.core.gamer import Game\n'), ((4687, 4734), 'pychess.element.piecer.Piece', 'Piece', (['c.PieceType.pawn', ... |
# Copyright (c) Meta Platforms, Inc. and affiliates.
#
# This source code is licensed under the MIT license found in the
# LICENSE file in the root directory of this source tree.
# pyre-unsafe
from typing import Any, Dict, List, Optional, Sequence, Tuple, Type
from libcst._exceptions import PartialParserSyntaxError
f... | [
"libcst._nodes.expression.Name",
"libcst._parser.types.partials.ImportPartial",
"libcst._parser.whitespace_parser.parse_parenthesizable_whitespace",
"libcst._parser.types.partials.SimpleStatementPartial",
"libcst._parser.types.partials.WithLeadingWhitespace",
"libcst._parser.production_decorator.with_prod... | [((2513, 2560), 'libcst._parser.production_decorator.with_production', 'with_production', (['"""stmt_input"""', '"""stmt ENDMARKER"""'], {}), "('stmt_input', 'stmt ENDMARKER')\n", (2528, 2560), False, 'from libcst._parser.production_decorator import with_production\n'), ((2693, 2752), 'libcst._parser.production_decorat... |
import scrapy
from locations.items import GeojsonPointItem
import json
class FarmersInsuranceSpider(scrapy.Spider):
name = "farmers-insurance"
start_urls = ["https://agents.farmers.com/"]
def get_hours(self, hours_json):
days = json.loads(hours_json)
out_hours = []
for day in da... | [
"json.loads",
"scrapy.Request",
"locations.items.GeojsonPointItem"
] | [((252, 274), 'json.loads', 'json.loads', (['hours_json'], {}), '(hours_json)\n', (262, 274), False, 'import json\n'), ((910, 962), 'scrapy.Request', 'scrapy.Request', ([], {'url': 'path', 'callback': 'self.parse_cities'}), '(url=path, callback=self.parse_cities)\n', (924, 962), False, 'import scrapy\n'), ((1194, 1246)... |
import discord
from discord.ext import commands
class Example(commands.Cog):
def __init__(self , Client):
self.Client = Client
@commands.command()
async def helloworld(self , ctx):
ctx.send('Hello World!')
def setup(Client):
Client.add_cog(Example(Client))
| [
"discord.ext.commands.command"
] | [((147, 165), 'discord.ext.commands.command', 'commands.command', ([], {}), '()\n', (163, 165), False, 'from discord.ext import commands\n')] |
# Copyright 2018, <NAME> LLC
# License: Apache License Version 2.0
# -------------------------------------------------------------------------
# logger.py - basic wrapper around Python standard logging just so in
# case we need to change this behavior it is all in one place
# -------------------------------------... | [
"logging.getLogger",
"logging.Formatter",
"logging.StreamHandler"
] | [((830, 858), 'logging.getLogger', 'logging.getLogger', (['"""vespene"""'], {}), "('vespene')\n", (847, 858), False, 'import logging\n'), ((919, 942), 'logging.StreamHandler', 'logging.StreamHandler', ([], {}), '()\n', (940, 942), False, 'import logging\n'), ((967, 1040), 'logging.Formatter', 'logging.Formatter', (['""... |
from collections import Counter
from pathlib import Path
from typing import List, Dict
def calculate_triangular_value(count: int) -> int:
result: int = 0
for i in range(count):
result += i
return result
class CrabCollection:
def __init__(self, initial_lineup: List[int]):
self.lineup ... | [
"pathlib.Path"
] | [((1046, 1076), 'pathlib.Path', 'Path', (['"""../input/day7_test.txt"""'], {}), "('../input/day7_test.txt')\n", (1050, 1076), False, 'from pathlib import Path\n'), ((1098, 1125), 'pathlib.Path', 'Path', (['"""../input/day7_1.txt"""'], {}), "('../input/day7_1.txt')\n", (1102, 1125), False, 'from pathlib import Path\n')] |
# -*- coding: utf-8 -*-
import six
from abc import ABCMeta, abstractmethod
@six.add_metaclass(ABCMeta)
class SettingBase(object):
@abstractmethod
def generate(self, args):
"""
Generate setting file
:param dict args: setting parameter
:raises ParseError:
""... | [
"six.add_metaclass"
] | [((79, 105), 'six.add_metaclass', 'six.add_metaclass', (['ABCMeta'], {}), '(ABCMeta)\n', (96, 105), False, 'import six\n')] |
# --------------------------------------------------------------------------
# Source file provided under Apache License, Version 2.0, January 2004,
# http://www.apache.org/licenses/
# (c) Copyright IBM Corp. 2015, 2016
# --------------------------------------------------------------------------
# gendoc: ignore
from... | [
"docplex.mp.utils.is_string",
"docplex.mp.internal.mloader.import_class",
"docplex.mp.error_handler.docplex_fatal",
"docplex.mp.docloud_engine.DOcloudEngine",
"inspect.isclass",
"docplex.mp.context.has_credentials"
] | [((4201, 4240), 'docplex.mp.context.has_credentials', 'has_credentials', (['context.solver.docloud'], {}), '(context.solver.docloud)\n', (4216, 4240), False, 'from docplex.mp.context import has_credentials\n'), ((4997, 5027), 'docplex.mp.docloud_engine.DOcloudEngine', 'DOcloudEngine', (['model'], {}), '(model, **kwargs... |
import threading
from unittest import TestCase
import multiprocessing
import multiprocessing_utils
class LocalTestCaseMixin(object):
def test_local(self):
local = multiprocessing_utils.local()
self.assertFalse(hasattr(local, 'value'))
self.assertRaises(AttributeError, getattr, local, 'va... | [
"multiprocessing_utils.local"
] | [((179, 208), 'multiprocessing_utils.local', 'multiprocessing_utils.local', ([], {}), '()\n', (206, 208), False, 'import multiprocessing_utils\n')] |
# GitHub URL: https://github.com/giswqs/qgis-earthengine-examples/tree/master/Visualization/random_color_visualizer.py
import ee
from ee_plugin import Map
dataset = ee.Image('USGS/NLCD/NLCD2016')
landcover = ee.Image(dataset.select('landcover'))
Map.setCenter(-95, 38, 5)
Map.addLayer(landcover.randomVisualizer(), {}... | [
"ee.Image",
"ee_plugin.Map.setCenter"
] | [((167, 197), 'ee.Image', 'ee.Image', (['"""USGS/NLCD/NLCD2016"""'], {}), "('USGS/NLCD/NLCD2016')\n", (175, 197), False, 'import ee\n'), ((249, 274), 'ee_plugin.Map.setCenter', 'Map.setCenter', (['(-95)', '(38)', '(5)'], {}), '(-95, 38, 5)\n', (262, 274), False, 'from ee_plugin import Map\n')] |
from collections import deque
class Process:
def __init__(self):
self.burst_times = deque([])
self.wait_times = deque([])
def create(self, filename):
inputfile = open(filename)
lines = inputfile.readlines()
self.pid = int(lines[0].rstrip('\n'))
self.arrival_time... | [
"collections.deque"
] | [((97, 106), 'collections.deque', 'deque', (['[]'], {}), '([])\n', (102, 106), False, 'from collections import deque\n'), ((133, 142), 'collections.deque', 'deque', (['[]'], {}), '([])\n', (138, 142), False, 'from collections import deque\n')] |
import json
import logging
from django.http import HttpResponse
from pollaris.app.models import SearchLog
def log_failed_search(request):
"""API route to log searches that failed on FE to the Pollaris DB"""
body = json.loads(request.body)
logging.info(f"Logging failed search: {body}")
address_json =... | [
"django.http.HttpResponse",
"json.loads",
"logging.info",
"logging.error"
] | [((226, 250), 'json.loads', 'json.loads', (['request.body'], {}), '(request.body)\n', (236, 250), False, 'import json\n'), ((255, 301), 'logging.info', 'logging.info', (['f"""Logging failed search: {body}"""'], {}), "(f'Logging failed search: {body}')\n", (267, 301), False, 'import logging\n'), ((792, 816), 'django.htt... |
# Copyright (C) 2013 Google Inc. All rights reserved.
# coding=utf-8
#
# Redistribution and use in source and binary forms, with or without
# modification, are permitted provided that the following conditions are
# met:
#
# * Redistributions of source code must retain the above copyright
# notice, this list of cond... | [
"v8_utilities.runtime_enabled_feature_name",
"v8_methods.is_secure_context",
"v8_globals.includes.add",
"v8_utilities.cpp_name_or_partial",
"v8_globals.includes.update",
"v8_utilities.v8_class_name_or_partial",
"idl_types.inherits_interface",
"v8_attributes.filter_accessors",
"v8_attributes.filter_d... | [((9837, 9853), 'v8_globals.includes.clear', 'includes.clear', ([], {}), '()\n', (9851, 9853), False, 'from v8_globals import includes\n'), ((9858, 9897), 'v8_globals.includes.update', 'includes.update', (['INTERFACE_CPP_INCLUDES'], {}), '(INTERFACE_CPP_INCLUDES)\n', (9873, 9897), False, 'from v8_globals import include... |
# -*- coding: utf-8 -*-
import pandas as pd
import numpy as np
from sklearn.naive_bayes import MultinomialNB
from sklearn.feature_extraction.text import CountVectorizer
"""
Ler
"""
df_train = pd.read_csv('./data_train_test/df_train_imbd.csv')
df_test = pd.read_csv('./data_train_test/df_test_imbd.csv')
"""
Vetoriz... | [
"sklearn.feature_extraction.text.CountVectorizer",
"sklearn.naive_bayes.MultinomialNB",
"pandas.read_csv"
] | [((196, 246), 'pandas.read_csv', 'pd.read_csv', (['"""./data_train_test/df_train_imbd.csv"""'], {}), "('./data_train_test/df_train_imbd.csv')\n", (207, 246), True, 'import pandas as pd\n'), ((257, 306), 'pandas.read_csv', 'pd.read_csv', (['"""./data_train_test/df_test_imbd.csv"""'], {}), "('./data_train_test/df_test_im... |
# Licensed under the Apache License, Version 2.0 (the "License"); you may
# not use this file except in compliance with the License. You may obtain
# a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distribut... | [
"inspect.currentframe"
] | [((1733, 1755), 'inspect.currentframe', 'inspect.currentframe', ([], {}), '()\n', (1753, 1755), False, 'import inspect\n'), ((2224, 2246), 'inspect.currentframe', 'inspect.currentframe', ([], {}), '()\n', (2244, 2246), False, 'import inspect\n'), ((2727, 2749), 'inspect.currentframe', 'inspect.currentframe', ([], {}), ... |
import setuptools
setuptools.setup(
name="ngboost",
version="0.1.3",
author="<NAME>",
author_email="<EMAIL>",
description="Library for probabilistic predictions via gradient boosting.",
long_description="Please see Github for full description.",
long_description_content_type="text/markdown"... | [
"setuptools.find_packages"
] | [((423, 449), 'setuptools.find_packages', 'setuptools.find_packages', ([], {}), '()\n', (447, 449), False, 'import setuptools\n')] |
'''
Solve stokes equation for a moving sphere
- In practical applications the grid must be denser
'''
import mesh
import FBEM
output_folder = 'data/sphere/'
position = (0, 0, 0)
radius = 1
velocity = (2, 0, 0)
angular = (0, 0, 3)
sphere_grid = 4 # Mesh consisting of 4x4 nodes
#---
sphere = mesh.sphere_create('sphere'... | [
"mesh.sphere_create",
"FBEM.read_all_triangulation_input",
"matplotlib.pyplot.axes",
"FBEM.run",
"matplotlib.pyplot.show"
] | [((293, 371), 'mesh.sphere_create', 'mesh.sphere_create', (['"""sphere"""', 'position', 'radius', 'velocity', 'angular', 'sphere_grid'], {}), "('sphere', position, radius, velocity, angular, sphere_grid)\n", (311, 371), False, 'import mesh\n'), ((437, 468), 'FBEM.run', 'FBEM.run', (['sphere', 'output_folder'], {}), '(s... |
from telegram import ReplyKeyboardMarkup, KeyboardButton
def main_keyboard():
# http://unicode.org/emoji/charts/full-emoji-list.html
return ReplyKeyboardMarkup([
[KeyboardButton('/balance (beta)')],
[KeyboardButton('/markets 📈'), KeyboardButton('/volume 💰'), KeyboardButton('/exchange 🚀')],
... | [
"telegram.KeyboardButton"
] | [((181, 214), 'telegram.KeyboardButton', 'KeyboardButton', (['"""/balance (beta)"""'], {}), "('/balance (beta)')\n", (195, 214), False, 'from telegram import ReplyKeyboardMarkup, KeyboardButton\n'), ((226, 254), 'telegram.KeyboardButton', 'KeyboardButton', (['"""/markets 📈"""'], {}), "('/markets 📈')\n", (240, 254), F... |
#!/usr/bin/env python3
#-*- coding: utf-8 -*-
#comment from wenmei:
'''
nal_type:7和nal_type:8代表sps 和pps \
get length 8461, complete 1, type 0,
type 0 表示keyframe
'''
"""
Description: This app is used to parse the main log file and get the information of VoLTE video call
Version-0.3
Author: yi.qiu
E-m... | [
"tkinter.Menu",
"os.path.exists",
"tkinter.messagebox.showinfo",
"tkinter.messagebox.showwarning",
"time.strptime",
"tkinter.messagebox.askokcancel",
"tkinter.messagebox.showerror",
"threading.Lock",
"csv.writer",
"tkinter.filedialog.asksaveasfilename",
"tkinter.Tk",
"threading.Thread",
"tki... | [((10924, 10949), 'os.path.exists', 'os.path.exists', (['file_path'], {}), '(file_path)\n', (10938, 10949), False, 'import os\n'), ((11901, 11926), 'os.path.exists', 'os.path.exists', (['file_path'], {}), '(file_path)\n', (11915, 11926), False, 'import os\n'), ((16499, 16511), 'tkinter.Tk', 'tkinter.Tk', ([], {}), '()\... |
from conans import ConanFile, CMake, tools
import os
class LibmikmodConan(ConanFile):
name = "libmikmod"
description = "Module player and library supporting many formats, including mod, s3m, it, and xm."
topics = ("libmikmod", "audio")
url = "https://github.com/conan-io/conan-center-index"
homepag... | [
"os.rename",
"conans.CMake",
"conans.tools.patch",
"os.path.join",
"conans.tools.collect_libs",
"conans.tools.get",
"conans.tools.is_apple_os"
] | [((1659, 1694), 'conans.tools.is_apple_os', 'tools.is_apple_os', (['self.settings.os'], {}), '(self.settings.os)\n', (1676, 1694), False, 'from conans import ConanFile, CMake, tools\n'), ((2191, 2244), 'conans.tools.get', 'tools.get', ([], {}), "(**self.conan_data['sources'][self.version])\n", (2200, 2244), False, 'fro... |
# Licensed under the Apache License, Version 2.0 (the "License"); you may
# not use this file except in compliance with the License. You may obtain
# a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# d... | [
"stevedore.sphinxext._detailed_list",
"mock.Mock",
"stevedore.extension.Extension",
"stevedore.extension.ExtensionManager.make_test_instance",
"stevedore.sphinxext._simple_list"
] | [((888, 928), 'mock.Mock', 'mock.Mock', ([], {'spec': 'pkg_resources.EntryPoint'}), '(spec=pkg_resources.EntryPoint)\n', (897, 928), False, 'import mock\n'), ((977, 1025), 'mock.Mock', 'mock.Mock', ([], {'return_value': "('ENTRY_POINT(%s)' % name)"}), "(return_value='ENTRY_POINT(%s)' % name)\n", (986, 1025), False, 'im... |
# Copyright (C) 2016, A10 Networks Inc. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License"); you may
# not use this file except in compliance with the License. You may obtain
# a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless requ... | [
"logging.getLogger",
"neutron_lbaas.services.loadbalancer.plugin.LoadBalancerPluginv2"
] | [((964, 991), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (981, 991), False, 'import logging\n'), ((1293, 1322), 'neutron_lbaas.services.loadbalancer.plugin.LoadBalancerPluginv2', 'plugin.LoadBalancerPluginv2', ([], {}), '()\n', (1320, 1322), False, 'from neutron_lbaas.services.loadbal... |
#!/usr/bin/env python
# -*- coding: utf-8 -*-
# Copyright © 2017, CNRS-LAAS
# All rights reserved.
#
# Redistribution and use in source and binary forms, with or without
# modification, are permitted provided that the following conditions are met:
# * Redistributions of source code must retain the above copyright
... | [
"sys.path.append",
"ivy_msg_interface.PprzMessage",
"os.getenv"
] | [((1611, 1636), 'os.getenv', 'os.getenv', (['"""PPRZLINK_LIB"""'], {}), "('PPRZLINK_LIB')\n", (1620, 1636), False, 'import os\n'), ((1637, 1687), 'sys.path.append', 'sys.path.append', (["(PPRZLINK_LIB + '/lib/v1.0/python')"], {}), "(PPRZLINK_LIB + '/lib/v1.0/python')\n", (1652, 1687), False, 'import sys\n'), ((3335, 33... |
import math
import torch
import torch.nn as nn
import torch.nn.functional as F
from ops import linear, conv1d
class MLP(nn.Module):
def __init__(self, num_classes=1000):
super(MLP, self).__init__()
self.fc1 = nn.Linear(512, 512)
self.fc2 = nn.Linear(512, num_classes)
# when you ... | [
"torch.nn.MaxPool1d",
"torch.nn.ReLU",
"torch.nn.Dropout",
"torch.nn.functional.softmax",
"torch.mean",
"math.sqrt",
"torch.nn.BatchNorm1d",
"ops.linear",
"torch.nn.functional.relu",
"torch.nn.Linear",
"ops.conv1d",
"torch.nn.Conv1d"
] | [((233, 252), 'torch.nn.Linear', 'nn.Linear', (['(512)', '(512)'], {}), '(512, 512)\n', (242, 252), True, 'import torch.nn as nn\n'), ((272, 299), 'torch.nn.Linear', 'nn.Linear', (['(512)', 'num_classes'], {}), '(512, num_classes)\n', (281, 299), True, 'import torch.nn as nn\n'), ((805, 951), 'ops.linear', 'linear', ([... |
# Copyright 2020 DataStax, Inc
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing,... | [
"pysper.humanize.pluralize",
"pysper.diag.find_logs",
"pysper.parser.read_output_log",
"collections.defaultdict"
] | [((1053, 1069), 'collections.defaultdict', 'defaultdict', (['int'], {}), '(int)\n', (1064, 1069), False, 'from collections import defaultdict\n'), ((1505, 1532), 'pysper.parser.read_output_log', 'parser.read_output_log', (['log'], {}), '(log)\n', (1527, 1532), False, 'from pysper import parser\n'), ((3253, 3291), 'pysp... |
## -*- coding: utf-8 -*
from PyQt5.QtCore import *
from PyQt5.QtGui import *
from PyQt5.QtWidgets import *
import numpy as np
import math
class Draw(QGraphicsItem):
def __init__(self,width=180, height=180, size=90):
super(Draw,self).__init__()
self.offsetx = 10
self.offsety = 10
se... | [
"numpy.sin",
"numpy.cos",
"numpy.savetxt",
"math.radians"
] | [((7293, 7372), 'numpy.savetxt', 'np.savetxt', (['csv_filename', 'self.angvel_list'], {'delimiter': '""","""', 'header': '""" """', 'fmt': '"""%f"""'}), "(csv_filename, self.angvel_list, delimiter=',', header=' ', fmt='%f')\n", (7303, 7372), True, 'import numpy as np\n'), ((4987, 5032), 'numpy.sin', 'np.sin', (['((90.0... |
# Demo Python Datetime - The strftime() Method
'''
The strftime() Method
The datetime object has a method for formatting date objects into readable strings.
The method is called strftime(), and takes one parameter, format, to specify the format of the returned string.
Directive Description ... | [
"datetime.datetime.now"
] | [((2351, 2374), 'datetime.datetime.now', 'datetime.datetime.now', ([], {}), '()\n', (2372, 2374), False, 'import datetime\n')] |
import os
import logging
import discord
from operator import mul
from functools import reduce
from discord.ext import commands
from datetime import datetime as dt
logging.basicConfig(level=logging.INFO)
bot = commands.Bot(
command_prefix=commands.when_mentioned_or('/'),
help_command=None,
intents=discord.I... | [
"logging.basicConfig",
"discord.ext.commands.when_mentioned_or",
"functools.reduce",
"discord.Intents.all",
"datetime.datetime.now"
] | [((164, 203), 'logging.basicConfig', 'logging.basicConfig', ([], {'level': 'logging.INFO'}), '(level=logging.INFO)\n', (183, 203), False, 'import logging\n'), ((842, 885), 'functools.reduce', 'reduce', (['mul', '(3, 3, 80071, 152837, 4318511)'], {}), '(mul, (3, 3, 80071, 152837, 4318511))\n', (848, 885), False, 'from f... |
#!/usr/bin/python
import os
import json
import string
def path_to_dict(path):
file_name = os.path.basename(path);
title_name = string.replace(os.path.splitext(file_name)[0], '_', ' ').title()
d = {'title': title_name}
if os.path.isdir(path):
d['url'] = file_name
d['menu'] = [path_to_dict(os.path.join(path,x))... | [
"os.listdir",
"json.dumps",
"os.path.join",
"os.path.splitext",
"os.path.isdir",
"os.path.basename"
] | [((93, 115), 'os.path.basename', 'os.path.basename', (['path'], {}), '(path)\n', (109, 115), False, 'import os\n'), ((227, 246), 'os.path.isdir', 'os.path.isdir', (['path'], {}), '(path)\n', (240, 246), False, 'import os\n'), ((801, 817), 'json.dumps', 'json.dumps', (['menu'], {}), '(menu)\n', (811, 817), False, 'impor... |
from setuptools import setup, find_packages
classifiers = [
'Development Status :: 5 - Production/Stable',
'Intended Audience :: Education',
'Operating System :: Microsoft :: Windows :: Windows 10',
'License :: OSI Approved :: MIT License',
'Programming Language :: Python :: 3'
]
setup(
name='mariocard.... | [
"setuptools.find_packages"
] | [((651, 666), 'setuptools.find_packages', 'find_packages', ([], {}), '()\n', (664, 666), False, 'from setuptools import setup, find_packages\n')] |
import json
import warnings
from typing import Type
from urllib.parse import urlparse
from django.conf import settings
from django.core.signals import setting_changed
from django.db import models
from django.db.models.base import ModelBase
from django.http.request import validate_host
from django.utils.functional impo... | [
"django.core.signals.setting_changed.connect",
"urllib.parse.urlparse",
"requests.get",
"django.utils.module_loading.import_string",
"django.http.request.validate_host",
"warnings.warn"
] | [((2974, 3000), 'django.utils.module_loading.import_string', 'import_string', (['import_path'], {}), '(import_path)\n', (2987, 3000), False, 'from django.utils.module_loading import import_string\n'), ((1448, 1461), 'urllib.parse.urlparse', 'urlparse', (['url'], {}), '(url)\n', (1456, 1461), False, 'from urllib.parse i... |
from collections import OrderedDict
class MyObj(object):
b = 1
a = 2
def __init__(self):
object.__setattr__(self, '_attrs', OrderedDict())
self.c = 1
self.d = 2
def __setattr__(self, key, value):
assert key != '_attrs'
self._attrs[key] = value
def __getat... | [
"collections.OrderedDict"
] | [((147, 160), 'collections.OrderedDict', 'OrderedDict', ([], {}), '()\n', (158, 160), False, 'from collections import OrderedDict\n')] |
"""
Copyright (C) 2019 NVIDIA Corporation. All rights reserved.
Licensed under the CC BY-NC-SA 4.0 license
(https://creativecommons.org/licenses/by-nc-sa/4.0/legalcode).
"""
import torch
from torch.utils.data import Dataset
import numpy as np
import time
import os
import cv2
import sys
import utils
from datasets.sca... | [
"numpy.log",
"utils.resize_mask",
"numpy.array",
"utils.compute_overlaps",
"numpy.arctan2",
"numpy.linalg.norm",
"utils.writePointCloud",
"utils.compose_image_meta",
"numpy.save",
"numpy.arange",
"os.path.exists",
"numpy.where",
"numpy.stack",
"utils.extract_bboxes",
"numpy.random.seed",... | [((20837, 20957), 'utils.resize_image', 'utils.resize_image', (['image'], {'min_dim': 'config.IMAGE_MAX_DIM', 'max_dim': 'config.IMAGE_MAX_DIM', 'padding': 'config.IMAGE_PADDING'}), '(image, min_dim=config.IMAGE_MAX_DIM, max_dim=config.\n IMAGE_MAX_DIM, padding=config.IMAGE_PADDING)\n', (20855, 20957), False, 'impor... |
#!/usr/bin/env python
# -*- coding: utf-8 -*-
import socket
import struct
import marshal
from ComssServiceDevelopment.connectors.base import BaseConnector
class OutputMulticastConnector(BaseConnector):
TIMEOUT = 0.2
TTL = 2
def __init__(self, service_instance):
self._service_instance = service_i... | [
"socket.socket",
"struct.pack",
"marshal.dumps",
"socket.inet_aton",
"marshal.loads"
] | [((1288, 1306), 'marshal.dumps', 'marshal.dumps', (['msg'], {}), '(msg)\n', (1301, 1306), False, 'import marshal\n'), ((3279, 3298), 'marshal.loads', 'marshal.loads', (['data'], {}), '(data)\n', (3292, 3298), False, 'import marshal\n'), ((840, 888), 'socket.socket', 'socket.socket', (['socket.AF_INET', 'socket.SOCK_DGR... |
#!/usr/bin/env python3
#
# Copyright 2021 Graviti. Licensed under MIT License.
#
"""Implementation of gas tag."""
import click
from ..client.gas import DatasetClientType
from .tbrn import TBRN, TBRNType
from .utility import ContextInfo, error, exception_handler, get_dataset_client, get_gas
@exception_handler
def _... | [
"click.echo"
] | [((1059, 1108), 'click.echo', 'click.echo', (['f"""{tag_tbrn} is deleted successfully"""'], {}), "(f'{tag_tbrn} is deleted successfully')\n", (1069, 1108), False, 'import click\n'), ((1550, 1599), 'click.echo', 'click.echo', (['f"""{tag_tbrn} is created successfully"""'], {}), "(f'{tag_tbrn} is created successfully')\n... |
#!/usr/bin/env python3
import os
import configparser
from .app_version import *
# Obligatorily (on every OS):
# - config files go in $HOME/.config/IPyNotes
#
# By default (on every OS):
# - note files go in $HOME/IPyNotes/notes
# - change files go in $HOME/IPyNotes/changes
# TODO: Add a way to save settings, e.g.... | [
"os.path.exists",
"configparser.ConfigParser",
"os.makedirs",
"os.path.join",
"os.path.expanduser"
] | [((498, 521), 'os.path.expanduser', 'os.path.expanduser', (['"""~"""'], {}), "('~')\n", (516, 521), False, 'import os\n'), ((537, 576), 'os.path.join', 'os.path.join', (['home', '""".config"""', 'APP_NAME'], {}), "(home, '.config', APP_NAME)\n", (549, 576), False, 'import os\n'), ((585, 617), 'os.makedirs', 'os.makedir... |
from __future__ import absolute_import
from __future__ import division
from __future__ import print_function
import json
import requests
from plasticity.utils import utils
class Endpoint(object):
"""An Endpoint is a specific API action within an API service.
Attributes:
plasticity: a Plasticity ins... | [
"json.loads",
"json.dumps"
] | [((2533, 2563), 'json.loads', 'json.loads', (['self._request.body'], {}), '(self._request.body)\n', (2543, 2563), False, 'import json\n'), ((2335, 2360), 'json.loads', 'json.loads', (['response.text'], {}), '(response.text)\n', (2345, 2360), False, 'import json\n'), ((1464, 1483), 'json.dumps', 'json.dumps', (['payload... |
import os
import matplotlib.pyplot as plt
import numpy as np
import pandas as pd
import math
import re
from datetime import datetime
data_location = os.path.join(os.path.dirname(__file__), "..", "data")
from os import listdir
from os.path import isfile, join
onlyfiles = [f for f in listdir(data_location) if isfile(jo... | [
"re.split",
"os.listdir",
"datetime.datetime.strptime",
"os.path.join",
"os.path.dirname"
] | [((712, 740), 're.split', 're.split', (['"""-|\\\\."""', 'file_name'], {}), "('-|\\\\.', file_name)\n", (720, 740), False, 'import re\n'), ((859, 894), 'datetime.datetime.strptime', 'datetime.strptime', (['str_date', '"""%Y%m"""'], {}), "(str_date, '%Y%m')\n", (876, 894), False, 'from datetime import datetime\n'), ((16... |
#!/usr/bin/env python3
import csv
import datetime
import requests
# def get_file_lines(url):
def download_file(url):
"""Returns the lines contained in the file at the given URL"""
# Download the file over the internet
response = requests.get(url, stream=True)
lines = []
for line in response.iter... | [
"datetime.datetime",
"datetime.datetime.strptime",
"requests.get",
"datetime.datetime.today",
"datetime.timedelta",
"csv.reader"
] | [((244, 274), 'requests.get', 'requests.get', (['url'], {'stream': '(True)'}), '(url, stream=True)\n', (256, 274), False, 'import requests\n'), ((794, 829), 'datetime.datetime', 'datetime.datetime', (['year', 'month', 'day'], {}), '(year, month, day)\n', (811, 829), False, 'import datetime\n'), ((1227, 1252), 'datetime... |
from __future__ import absolute_import, unicode_literals
import sys
import pytest
from flaky import flaky
from virtualenv.activation import XonshActivator
from virtualenv.info import IS_PYPY, PY3
@pytest.mark.slow
@pytest.mark.skipif(
(sys.platform == "win32" and IS_PYPY and PY3) or sys.version_info[0:2] == (3... | [
"flaky.flaky",
"pytest.mark.skipif"
] | [((220, 404), 'pytest.mark.skipif', 'pytest.mark.skipif', (["(sys.platform == 'win32' and IS_PYPY and PY3 or sys.version_info[0:2] == (3, 9)\n )"], {'reason': '"""xonsh on Windows blocks indefinitely and is not stable yet on 3.9"""'}), "(sys.platform == 'win32' and IS_PYPY and PY3 or sys.\n version_info[0:2] == (... |
import pytest
from plenum.common.exceptions import MissingNodeOp, InvalidNodeOp
from plenum.common.messages.fields import NonNegativeNumberField, AnyValueField, HexField, BooleanField, Base58Field
from plenum.common.messages.message_base import MessageBase
from plenum.common.messages.node_message_factory import Messag... | [
"plenum.common.messages.node_message_factory.NodeMessageFactory",
"pytest.raises",
"plenum.common.messages.fields.NonNegativeNumberField",
"plenum.common.messages.node_message_factory.MessageFactory"
] | [((487, 547), 'plenum.common.messages.node_message_factory.MessageFactory', 'MessageFactory', (['"""plenum.test.input_validation.stub_messages"""'], {}), "('plenum.test.input_validation.stub_messages')\n", (501, 547), False, 'from plenum.common.messages.node_message_factory import MessageFactory, NodeMessageFactory\n')... |
import tensorflow as tf
from tensorflow import keras
# Open with keras first (direct conversion from SavedModel does not work for some reason)
model = keras.models.load_model("model")
model.compile(optimizer="adam",
loss=tf.keras.losses.SparseCategoricalCrossentropy(from_logits=True),
metrics=["accuracy"])
... | [
"tensorflow.keras.losses.SparseCategoricalCrossentropy",
"tensorflow.keras.models.load_model",
"tensorflow.lite.TFLiteConverter.from_keras_model"
] | [((156, 188), 'tensorflow.keras.models.load_model', 'keras.models.load_model', (['"""model"""'], {}), "('model')\n", (179, 188), False, 'from tensorflow import keras\n'), ((371, 418), 'tensorflow.lite.TFLiteConverter.from_keras_model', 'tf.lite.TFLiteConverter.from_keras_model', (['model'], {}), '(model)\n', (411, 418)... |
from ...accounts import Account, account_factory
import tempfile
import os
import pickle
from os import listdir
from os.path import isfile, join
class LocalFileSystemAccountAdapter():
def __init__(self, root=None):
if root is None: root = tempfile.gettempdir()
if not os.path.exists(root+"/accounts... | [
"os.path.exists",
"os.listdir",
"pickle.dump",
"os.makedirs",
"pickle.load",
"os.path.join",
"tempfile.gettempdir"
] | [((253, 274), 'tempfile.gettempdir', 'tempfile.gettempdir', ([], {}), '()\n', (272, 274), False, 'import tempfile\n'), ((290, 325), 'os.path.exists', 'os.path.exists', (["(root + '/accounts/')"], {}), "(root + '/accounts/')\n", (304, 325), False, 'import os\n'), ((337, 369), 'os.makedirs', 'os.makedirs', (["(root + '/a... |
import logging
from typing import Callable
from django.conf import settings
from django.http import HttpRequest, HttpResponse
from django.contrib.auth import get_user_model, logout
from django.contrib.messages import info
from .utils import now, seconds_until_idle_time_end, seconds_until_session_end
UserModel = get_u... | [
"logging.getLogger",
"django.contrib.auth.get_user_model",
"django.contrib.messages.info",
"django.contrib.auth.logout"
] | [((315, 331), 'django.contrib.auth.get_user_model', 'get_user_model', ([], {}), '()\n', (329, 331), False, 'from django.contrib.auth import get_user_model, logout\n'), ((341, 368), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (358, 368), False, 'import logging\n'), ((1303, 1318), 'djang... |
# Custom log levels for Python's logging module.
#
# Author: <NAME> <<EMAIL>>
# Last Change: August 7, 2017
# URL: https://verboselogs.readthedocs.io
"""
Custom log levels for Python's :mod:`logging` module.
The :mod:`verboselogs` module defines the :data:`NOTICE`, :data:`SPAM`,
:data:`SUCCESS` and :data:`VERBOSE` co... | [
"logging.Logger.__init__",
"logging.setLoggerClass",
"logging.getLogger",
"logging.addLevelName"
] | [((2673, 2710), 'logging.setLoggerClass', 'logging.setLoggerClass', (['VerboseLogger'], {}), '(VerboseLogger)\n', (2695, 2710), False, 'import logging\n'), ((2931, 2964), 'logging.addLevelName', 'logging.addLevelName', (['value', 'name'], {}), '(value, name)\n', (2951, 2964), False, 'import logging\n'), ((4367, 4409), ... |
from setuptools import setup, find_packages
from dexy.version import DEXY_VERSION
import platform
is_windows = platform.system() == 'Windows'
if is_windows:
os_specific_requires = []
else:
os_specific_requires = ['pexpect']
setup(
author='<NAME>',
author_email='<EMAIL>',
classifiers=[... | [
"platform.system",
"setuptools.find_packages"
] | [((112, 129), 'platform.system', 'platform.system', ([], {}), '()\n', (127, 129), False, 'import platform\n'), ((2008, 2023), 'setuptools.find_packages', 'find_packages', ([], {}), '()\n', (2021, 2023), False, 'from setuptools import setup, find_packages\n')] |
""" # NOQA
:mod:`aioopenssl` --- A transport for asyncio using :mod:`OpenSSL`
##################################################################
This package provides a socket-based :class:`asyncio.Transport` which uses
:mod:`OpenSSL` to create a TLS connection. Optionally, the TLS handshake can be
deferred and perfo... | [
"logging.getLogger",
"socket.socket",
"asyncio.Future"
] | [((1456, 1483), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (1473, 1483), False, 'import logging\n'), ((33610, 33635), 'asyncio.Future', 'asyncio.Future', ([], {'loop': 'loop'}), '(loop=loop)\n', (33624, 33635), False, 'import asyncio\n'), ((27355, 27371), 'asyncio.Future', 'asyncio.Fu... |
# coding: utf-8
import six
import pytest
from mock import sentinel, patch
from hamcrest.core.base_description import BaseDescription
from hamcrest.core.selfdescribing import SelfDescribing
from hamcrest.core.helpers.ismock import MOCKTYPES
__author__ = "<NAME>"
__copyright__ = "Copyright 2015 hamcrest.org"
__license_... | [
"pytest.mark.parametrize",
"six.u",
"six.b",
"pytest.mark.skipif"
] | [((1518, 1617), 'pytest.mark.parametrize', 'pytest.mark.parametrize', (['"""char, rep"""', '(("\'", "\'"), (\'\\n\', \'\\\\n\'), (\'\\r\', \'\\\\r\'), (\'\\t\', \'\\\\t\'))'], {}), '(\'char, rep\', (("\'", "\'"), (\'\\n\', \'\\\\n\'), (\'\\r\',\n \'\\\\r\'), (\'\\t\', \'\\\\t\')))\n', (1541, 1617), False, 'import py... |
from collections import defaultdict
from typing import Dict, List, Optional
import re
import sys
try:
# TODO: Rewrite this module to use ldap3 which gets installed in rootless installs too.
import ldap
except ModuleNotFoundError:
print("""\
*** ldap module unavailable; this comes from "python-ldap" which i... | [
"logging.getLogger",
"ldap.set_option",
"re.compile",
"ldap.initialize",
"collections.defaultdict",
"asn1.Encoder",
"hashlib.sha1"
] | [((577, 614), 'ldap.set_option', 'ldap.set_option', (['ldap.OPT_TIMEOUT', '(10)'], {}), '(ldap.OPT_TIMEOUT, 10)\n', (592, 614), False, 'import ldap\n'), ((615, 660), 'ldap.set_option', 'ldap.set_option', (['ldap.OPT_NETWORK_TIMEOUT', '(10)'], {}), '(ldap.OPT_NETWORK_TIMEOUT, 10)\n', (630, 660), False, 'import ldap\n'),... |
import json
from django.conf import settings
from django.contrib.auth.models import User
from django.utils.translation import ugettext_lazy as _
from cms.models import Placeholder
from cms.api import create_page, add_plugin, publish_page
placeholder = {}
with open('starting_page.json') as data_file:
content = ... | [
"django.utils.translation.ugettext_lazy",
"django.contrib.auth.models.User.objects.all",
"django.contrib.auth.models.User.objects.count",
"json.load",
"cms.api.add_plugin"
] | [((1158, 1216), 'cms.api.add_plugin', 'add_plugin', (["placeholder['main']", '"""MultiColumnPlugin"""', 'lang'], {}), "(placeholder['main'], 'MultiColumnPlugin', lang)\n", (1168, 1216), False, 'from cms.api import create_page, add_plugin, publish_page\n'), ((320, 340), 'json.load', 'json.load', (['data_file'], {}), '(d... |
# -*- coding: utf-8 -*-
from functools import partial
s_open = partial(open, mode='r')
| [
"functools.partial"
] | [((65, 88), 'functools.partial', 'partial', (['open'], {'mode': '"""r"""'}), "(open, mode='r')\n", (72, 88), False, 'from functools import partial\n')] |
import json
import logging.config
import os
import re
from string import Template
from jans.pycloudlib import get_manager
from jans.pycloudlib.persistence import render_couchbase_properties
from jans.pycloudlib.persistence import render_base_properties
from jans.pycloudlib.persistence import render_hybrid_properties
f... | [
"jans.pycloudlib.persistence.sync_ldap_truststore",
"jans.pycloudlib.get_manager",
"string.Template",
"jans.pycloudlib.persistence.render_hybrid_properties",
"jans.pycloudlib.persistence.render_sql_properties",
"jans.pycloudlib.persistence.sync_couchbase_truststore",
"jans.pycloudlib.persistence.render_... | [((938, 951), 'jans.pycloudlib.get_manager', 'get_manager', ([], {}), '()\n', (949, 951), False, 'from jans.pycloudlib import get_manager\n'), ((1099, 1395), 're.sub', 're.sub', (['"""<New id="DefaultHandler" class="org.eclipse.jetty.server.handler.DefaultHandler"/>"""', '"""<New id="DefaultHandler" class="org.eclipse.... |
import dash
import dash_bootstrap_components as dbc
from flask import Flask
server = Flask(__name__)
app = dash.Dash(__name__, server=server, external_stylesheets=[dbc.themes.BOOTSTRAP], suppress_callback_exceptions=True,
prevent_initial_callbacks=True)
app.server.config["SQLALCHEMY_DATABASE_URI"] = ... | [
"dash.Dash",
"flask.Flask"
] | [((86, 101), 'flask.Flask', 'Flask', (['__name__'], {}), '(__name__)\n', (91, 101), False, 'from flask import Flask\n'), ((108, 263), 'dash.Dash', 'dash.Dash', (['__name__'], {'server': 'server', 'external_stylesheets': '[dbc.themes.BOOTSTRAP]', 'suppress_callback_exceptions': '(True)', 'prevent_initial_callbacks': '(T... |
# -*- coding: utf-8 -*-
# Generated by Django 1.10.2 on 2017-01-18 11:25
from __future__ import unicode_literals
from django.db import migrations
class Migration(migrations.Migration):
dependencies = [
('meas_models', '0005_auto_20170118_1923'),
]
operations = [
migrations.RemoveField(
... | [
"django.db.migrations.RemoveField"
] | [((296, 357), 'django.db.migrations.RemoveField', 'migrations.RemoveField', ([], {'model_name': '"""formulaindex"""', 'name': '"""idf"""'}), "(model_name='formulaindex', name='idf')\n", (318, 357), False, 'from django.db import migrations\n')] |
import importlib
import sys
import tensorflow as tf
LSTM_SIZE = 2048
def resnet_rnn_model(features, model_params, example_description, training):
# Get hyperparameters
dropout_rate = model_params['resnet_rnn'].get('dropout_rate', 0.5)
# Reshape inputs into proper dimensions
for (name, f), d in zip(... | [
"tensorflow.contrib.framework.init_from_checkpoint",
"tensorflow.layers.flatten",
"importlib.import_module",
"tensorflow.transpose",
"tensorflow.variable_scope",
"tensorflow.reduce_sum",
"tensorflow.layers.max_pooling2d",
"tensorflow.contrib.layers.l2_regularizer",
"tensorflow.nn.rnn_cell.LSTMCell",... | [((749, 799), 'tensorflow.transpose', 'tf.transpose', (['telescope_data'], {'perm': '[1, 0, 2, 3, 4]'}), '(telescope_data, perm=[1, 0, 2, 3, 4])\n', (761, 799), True, 'import tensorflow as tf\n'), ((1539, 1587), 'sys.path.append', 'sys.path.append', (["model_params['model_directory']"], {}), "(model_params['model_direc... |
from molml.Datastructures.molecule import Molecule
import pandas as pd
def mols_from_csv(filename: str, smiles_col: str, label_col: str = None, id_col: str = None,
remove_empty_rows: bool = True):
# Read the data
df = pd.read_csv(filename, low_memory=True, usecols=list(filter(None, [smiles_c... | [
"molml.Datastructures.molecule.Molecule"
] | [((921, 944), 'molml.Datastructures.molecule.Molecule', 'Molecule', (['smiles', 'y', 'id'], {}), '(smiles, y, id)\n', (929, 944), False, 'from molml.Datastructures.molecule import Molecule\n')] |
import os
from functools import partial
from pathlib import Path
from typing import Dict, Optional, Union
from bw_tools.common.bw_api_tool import BWAPITool
from bw_tools.common.bw_node_selection import remove_dot_nodes
from bw_tools.modules.bw_settings.bw_settings import BWModuleSettings
from bw_tools.modules.bw_strai... | [
"bw_tools.common.bw_node_selection.remove_dot_nodes",
"pathlib.Path",
"bw_tools.modules.bw_straighten_connection.straighten_behavior.BWBreakAtSource",
"os.access",
"PySide2.QtWidgets.QMessageBox.question",
"functools.partial",
"bw_tools.modules.bw_straighten_connection.straighten_behavior.BWBreakAtTarge... | [((5573, 5582), 'PySide2.QtWidgets.QAction', 'QAction', ([], {}), '()\n', (5580, 5582), False, 'from PySide2.QtWidgets import QAction, QMessageBox\n'), ((3459, 3498), 'sd.tools.graphlayout.snapSDNodes', 'snapSDNodes', (['api.current_node_selection'], {}), '(api.current_node_selection)\n', (3470, 3498), False, 'from sd.... |
r"""
Cyclic sieving phenomenon
Implementation of the Cyclic Sieving Phenomenon as described in
Reiner, Stanton, White - The cyclic sieving phenomenon, Journal of Combinatorial Theory A 108 (2004)
We define the CyclicSievingPolynomial of a finite set S together with cyclic action cyc_act (of order n) to be the unique
... | [
"sage.rings.all.ZZ",
"sage.rings.arith.lcm"
] | [((3004, 3013), 'sage.rings.arith.lcm', 'lcm', (['keys'], {}), '(keys)\n', (3007, 3013), False, 'from sage.rings.arith import lcm\n'), ((3436, 3449), 'sage.rings.all.ZZ', 'ZZ', (['(order / n)'], {}), '(order / n)\n', (3438, 3449), False, 'from sage.rings.all import ZZ, QQ\n'), ((3380, 3385), 'sage.rings.all.ZZ', 'ZZ', ... |
# Copyright (c) 2016 Mirantis Inc.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writ... | [
"sahara.utils.files.get_file_text",
"sahara.plugins.provisioning.Config"
] | [((2995, 3211), 'sahara.plugins.provisioning.Config', 'p.Config', (['"""Hadoop OpenStack library URL"""', '"""general"""', '"""cluster"""'], {'priority': '(1)', 'default_value': 'DEFAULT_SWIFT_LIB_URL', 'description': '"""Library that adds Swift support to CDH. The file will be downloaded by VMs."""'}), "('Hadoop OpenS... |
import os, sys
import numpy as np
import pybullet as p
class Util:
def __init__(self, pid, np_random):
self.id = pid
self.ik_lower_limits = {}
self.ik_upper_limits = {}
self.ik_joint_ranges = {}
self.ik_rest_poses = {}
self.np_random = np_random
def enable_gpu(... | [
"pkgutil.get_loader",
"numpy.mean",
"numpy.abs",
"numpy.cross",
"GPUtil.getGPUs",
"numpy.any",
"numpy.array",
"numpy.dot",
"GPUtil.showUtilization",
"numpy.cos",
"numpy.concatenate",
"numpy.linalg.norm",
"numpy.sin",
"pybullet.loadPlugin"
] | [((904, 925), 'GPUtil.showUtilization', 'GPU.showUtilization', ([], {}), '()\n', (923, 925), True, 'import GPUtil as GPU\n'), ((988, 1017), 'GPUtil.showUtilization', 'GPU.showUtilization', ([], {'all': '(True)'}), '(all=True)\n', (1007, 1017), True, 'import GPUtil as GPU\n'), ((1219, 1232), 'GPUtil.getGPUs', 'GPU.getGP... |
# Copyright (c) 2010-2014 <NAME>.
#
# Permission is hereby granted, free of charge, to any person obtaining a copy
# of this software and associated documentation files (the "Software"), to deal
# in the Software without restriction, including without limitation the rights
# to use, copy, modify, merge, publish, distri... | [
"pyrimaa.util.TimeControl"
] | [((1295, 1316), 'pyrimaa.util.TimeControl', 'TimeControl', (['"""30s/1s"""'], {}), "('30s/1s')\n", (1306, 1316), False, 'from pyrimaa.util import TimeControl\n'), ((1368, 1388), 'pyrimaa.util.TimeControl', 'TimeControl', (['"""5m/1s"""'], {}), "('5m/1s')\n", (1379, 1388), False, 'from pyrimaa.util import TimeControl\n'... |
import bpy
import numpy
import mathutils
from enum import Enum
from mathutils import Matrix, Quaternion, Vector
import xml.etree.ElementTree as ET
import os
os.system('cls')
mesh_targets = {}
controller_targets = {}
images = {}
class SourceType(Enum):
Name_array = 0
float_array = 1
class DataType(Enum):
... | [
"mathutils.Matrix.Identity",
"numpy.asarray",
"xml.etree.ElementTree.Element",
"xml.etree.ElementTree.ElementTree",
"mathutils.Quaternion",
"mathutils.Matrix.Translation",
"os.system",
"xml.etree.ElementTree.SubElement"
] | [((158, 174), 'os.system', 'os.system', (['"""cls"""'], {}), "('cls')\n", (167, 174), False, 'import os\n'), ((695, 726), 'xml.etree.ElementTree.SubElement', 'ET.SubElement', (['domNode', '"""input"""'], {}), "(domNode, 'input')\n", (708, 726), True, 'import xml.etree.ElementTree as ET\n'), ((975, 1007), 'xml.etree.Ele... |
"""Time-related utilities."""
from __future__ import absolute_import
from __future__ import division
from __future__ import generator_stop
from __future__ import print_function
import datetime
def datetime_range(start, end, fmt='%Y%m%d'):
"""Change `start` and `end` string to datetime object.
Pa... | [
"datetime.datetime.strptime"
] | [((790, 828), 'datetime.datetime.strptime', 'datetime.datetime.strptime', (['start', 'fmt'], {}), '(start, fmt)\n', (816, 828), False, 'import datetime\n'), ((844, 880), 'datetime.datetime.strptime', 'datetime.datetime.strptime', (['end', 'fmt'], {}), '(end, fmt)\n', (870, 880), False, 'import datetime\n'), ((943, 981)... |
# Generated by Django 3.2.7 on 2021-10-26 08:55
import ckeditor_uploader.fields
from django.conf import settings
from django.db import migrations, models
import django.db.models.deletion
class Migration(migrations.Migration):
initial = True
dependencies = [
migrations.swappable_dependency(settings.... | [
"django.db.models.EmailField",
"django.db.models.TextField",
"django.db.models.ForeignKey",
"django.db.models.ManyToManyField",
"django.db.models.ImageField",
"django.db.models.BigAutoField",
"django.db.models.DateTimeField",
"django.db.migrations.swappable_dependency",
"django.db.models.CharField"
... | [((279, 336), 'django.db.migrations.swappable_dependency', 'migrations.swappable_dependency', (['settings.AUTH_USER_MODEL'], {}), '(settings.AUTH_USER_MODEL)\n', (310, 336), False, 'from django.db import migrations, models\n'), ((4009, 4123), 'django.db.models.ForeignKey', 'models.ForeignKey', ([], {'on_delete': 'djang... |
#!/usr/bin/python
import os
import sys
import glob
import argparse
import tempfile
import numpy as np
import matplotlib.pyplot as plt
import pickle
import scipy.stats as stats
from copy import deepcopy
from subprocess import Popen, PIPE
from get_qdec_info import get_qdec_info
from fs_load_stats import fs_load_stats
f... | [
"numpy.array",
"scipy.stats.ttest_ind",
"numpy.mean",
"argparse.ArgumentParser",
"subprocess.Popen",
"os.path.split",
"scipy.stats.spearmanr",
"get_qdec_info.get_qdec_info",
"pickle.load",
"numpy.nonzero",
"numpy.std",
"os.path.abspath",
"fs_load_stats.fs_load_stats",
"pickle.dump",
"os.... | [((770, 884), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {'description': '"""Analyze aparc12 surface annotation: Surface area and average thickness"""'}), "(description=\n 'Analyze aparc12 surface annotation: Surface area and average thickness')\n", (793, 884), False, 'import argparse\n'), ((1810, 18... |
# INPUTS: BrainProtonDensitySlice.png
# OUTPUTS: FastMarchingImageFilterOutput5.png
# 81 114 1.0 -0.5 3.0 100 100
#
# INPUTS: BrainProtonDensitySlice.png
# OUTPUTS: FastMarchingImageFilterOutput6.png
# 99 114 1.0 -0.5 3.0 100 100
#
# INPUTS: BrainProtonDensitySlice.png
# OUTPU... | [
"itk.auto_progress",
"sys.exit"
] | [((926, 946), 'itk.auto_progress', 'itk.auto_progress', (['(2)'], {}), '(2)\n', (943, 946), False, 'import itk\n'), ((915, 922), 'sys.exit', 'exit', (['(1)'], {}), '(1)\n', (919, 922), False, 'from sys import argv, stderr, exit\n')] |
import pandas
import os
import math
import numpy
from scipy import stats
pandas.set_option('display.max_rows', 200, 'display.max_columns', 200) # change it to see more or less rows and/or columns
# Ask inputs to read the TSV file
dirPath = input('Enter path to TSV file: ')
inputName = input('Enter TSV name (input fil... | [
"pandas.isnull",
"numpy.mean",
"pandas.merge",
"numpy.log",
"math.sqrt",
"os.path.join",
"os.path.splitext",
"pandas.set_option",
"math.isnan"
] | [((74, 144), 'pandas.set_option', 'pandas.set_option', (['"""display.max_rows"""', '(200)', '"""display.max_columns"""', '(200)'], {}), "('display.max_rows', 200, 'display.max_columns', 200)\n", (91, 144), False, 'import pandas\n'), ((5100, 5199), 'pandas.merge', 'pandas.merge', (['meanAN_DF_gnomAD', 'meanAN_DF_CSVS'],... |
import bitwise as bw
class TestBitwiseNOT4:
def test_BitwiseNOT4(self):
a_bus_1 = bw.wire.Wire()
a_bus_2 = bw.wire.Wire()
a_bus_3 = bw.wire.Wire()
a_bus_4 = bw.wire.Wire()
a_bus = bw.wire.Bus4(a_bus_1, a_bus_2, a_bus_3, a_bus_4)
o_bus_1 = bw.wire.Wire()
o_b... | [
"bitwise.wire.Bus4",
"bitwise.logic.BitwiseNOT4",
"bitwise.wire.Wire"
] | [((96, 110), 'bitwise.wire.Wire', 'bw.wire.Wire', ([], {}), '()\n', (108, 110), True, 'import bitwise as bw\n'), ((129, 143), 'bitwise.wire.Wire', 'bw.wire.Wire', ([], {}), '()\n', (141, 143), True, 'import bitwise as bw\n'), ((162, 176), 'bitwise.wire.Wire', 'bw.wire.Wire', ([], {}), '()\n', (174, 176), True, 'import ... |
import numpy as np
import pandas as pd
import qrcode
import os
import sys
import time
#图像类别:emoji,动图,商品图片,文字图片,二维码,小程序码
#图像分类
from cv2 import cv2
from PIL import Image,ImageDraw
from datetime import datetime
import time
from pytesseract import image_to_string
class Detect():
def __init__(self):
pass
... | [
"PIL.Image.open",
"os.listdir",
"cv2.cv2.imread",
"cv2.cv2.CascadeClassifier",
"datetime.datetime.now",
"numpy.array",
"PIL.ImageDraw.Draw",
"pytesseract.image_to_string",
"cv2.cv2.cvtColor"
] | [((4069, 4083), 'datetime.datetime.now', 'datetime.now', ([], {}), '()\n', (4081, 4083), False, 'from datetime import datetime\n'), ((545, 567), 'cv2.cv2.imread', 'cv2.imread', (['image_name'], {}), '(image_name)\n', (555, 567), False, 'from cv2 import cv2\n'), ((591, 676), 'cv2.cv2.CascadeClassifier', 'cv2.CascadeClas... |
"""
LEGACY DFU HAS NOT BEEN IMPLEMENTED YET!
We will implement it when we will need it.
"""
import math
import pexpect
import time
from array import array
from util import *
from nrf_ble_dfu_controller import NrfBleDfuController
verbose = False
class Procedures:
START_DFU = 1
INITIALIZE_D... | [
"time.time",
"time.sleep"
] | [((5186, 5197), 'time.time', 'time.time', ([], {}), '()\n', (5195, 5197), False, 'import time\n'), ((5223, 5234), 'time.time', 'time.time', ([], {}), '()\n', (5232, 5234), False, 'import time\n'), ((6906, 6919), 'time.sleep', 'time.sleep', (['(1)'], {}), '(1)\n', (6916, 6919), False, 'import time\n'), ((8414, 8429), 't... |
# Add custom plugins here.
# If you want to make git not to track this file anymore,
# use `git update-index --skip-worktree app/plugin/__init__.py`
import flask
import app.plugin.ddc_docker as ddc_plugin_docker
def init_app(app: flask.Flask):
ddc_plugin_docker.init_app(app)
| [
"app.plugin.ddc_docker.init_app"
] | [((251, 282), 'app.plugin.ddc_docker.init_app', 'ddc_plugin_docker.init_app', (['app'], {}), '(app)\n', (277, 282), True, 'import app.plugin.ddc_docker as ddc_plugin_docker\n')] |
# Copyright 2016 VMware, Inc.
#
# Licensed under the Apache License, Version 2.0 (the "License"); you may
# not use this file except in compliance with the License. You may obtain
# a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agree... | [
"sqlalchemy.String",
"sqlalchemy.PrimaryKeyConstraint",
"sqlalchemy.DateTime"
] | [((1310, 1344), 'sqlalchemy.PrimaryKeyConstraint', 'sa.PrimaryKeyConstraint', (['"""purpose"""'], {}), "('purpose')\n", (1333, 1344), True, 'import sqlalchemy as sa\n'), ((989, 1009), 'sqlalchemy.String', 'sa.String', ([], {'length': '(32)'}), '(length=32)\n', (998, 1009), True, 'import sqlalchemy as sa\n'), ((1061, 10... |
import os, sys
import numpy as np
from math import sqrt
# testing without install
#sys.path.insert(0, '../build/lib.macosx-10.9-x86_64-3.8')
import poppunk_refine
# Original PopPUNK function (with some improvements)
def withinBoundary(dists, x_max, y_max, slope=2):
boundary_test = np.ones((dists.shape[0]))
fo... | [
"poppunk_refine.assignThreshold",
"numpy.ones",
"math.sqrt",
"poppunk_refine.edgeThreshold",
"poppunk_refine.thresholdIterate2D",
"numpy.array",
"poppunk_refine.thresholdIterate1D",
"numpy.finfo",
"numpy.meshgrid",
"numpy.all",
"numpy.arange"
] | [((1345, 1383), 'numpy.arange', 'np.arange', (['(0)', '(1)', '(0.1)'], {'dtype': 'np.float32'}), '(0, 1, 0.1, dtype=np.float32)\n', (1354, 1383), True, 'import numpy as np\n'), ((1388, 1426), 'numpy.arange', 'np.arange', (['(0)', '(1)', '(0.1)'], {'dtype': 'np.float32'}), '(0, 1, 0.1, dtype=np.float32)\n', (1397, 1426)... |
import cv2
import numpy as np
from keras.datasets import cifar10
from keras import backend as K
from keras.utils import np_utils
nb_train_samples = 50000 # 3000 training samples
nb_valid_samples = 10000 # 100 validation samples
num_classes = 10
def load_cifar10_data(img_rows, img_cols, start=None, end=None, what_da... | [
"cv2.resize",
"keras.utils.np_utils.to_categorical",
"keras.datasets.cifar10.load_data",
"keras.backend.image_dim_ordering"
] | [((424, 443), 'keras.datasets.cifar10.load_data', 'cifar10.load_data', ([], {}), '()\n', (441, 443), False, 'from keras.datasets import cifar10\n'), ((1263, 1327), 'keras.utils.np_utils.to_categorical', 'np_utils.to_categorical', (['Y_train[:nb_train_samples]', 'num_classes'], {}), '(Y_train[:nb_train_samples], num_cla... |
from dagster import (
DependencyDefinition,
InputDefinition,
Int,
Output,
OutputDefinition,
PipelineDefinition,
RunConfig,
lambda_solid,
solid,
)
from dagster.core.execution.api import create_execution_plan, execute_plan
from dagster.core.instance import DagsterInstance
def define_... | [
"dagster.RunConfig",
"dagster.core.instance.DagsterInstance.ephemeral",
"dagster.core.execution.api.create_execution_plan",
"dagster.InputDefinition",
"dagster.Output",
"dagster.PipelineDefinition",
"dagster.OutputDefinition",
"dagster.DependencyDefinition"
] | [((799, 834), 'dagster.core.execution.api.create_execution_plan', 'create_execution_plan', (['pipeline_def'], {}), '(pipeline_def)\n', (820, 834), False, 'from dagster.core.execution.api import create_execution_plan, execute_plan\n'), ((2013, 2092), 'dagster.PipelineDefinition', 'PipelineDefinition', ([], {'name': '"""... |
from Raspi_MotorHAT.Raspi_PWM_Servo_Driver import PWM
class Servos(object):
def __init__(self, addr=0x6f, deflect_90_in_ms = 0.9):
"""addr: The i2c address of the PWM chip.
deflect_90_in_ms: set this to calibrate the servo motors.
it is what a deflection of 90 degrees is
... | [
"Raspi_MotorHAT.Raspi_PWM_Servo_Driver.PWM"
] | [((413, 422), 'Raspi_MotorHAT.Raspi_PWM_Servo_Driver.PWM', 'PWM', (['addr'], {}), '(addr)\n', (416, 422), False, 'from Raspi_MotorHAT.Raspi_PWM_Servo_Driver import PWM\n')] |
import swapper
from rest_framework.generics import ListAPIView, ListCreateAPIView
from rest_framework.response import Response
from rest_framework.views import APIView
from openwisp_users.api.authentication import BearerAuthentication
from openwisp_users.api.mixins import (
FilterByOrganizationManaged,
FilterB... | [
"rest_framework.response.Response",
"swapper.load_model"
] | [((856, 908), 'swapper.load_model', 'swapper.load_model', (['"""openwisp_users"""', '"""Organization"""'], {}), "('openwisp_users', 'Organization')\n", (874, 908), False, 'import swapper\n'), ((1497, 1509), 'rest_framework.response.Response', 'Response', (['{}'], {}), '({})\n', (1505, 1509), False, 'from rest_framework... |
from torch.utils.data import ConcatDataset
from ssd.config.path_catlog import DatasetCatalog
from .voc_dataset import VOCDataset
from .coco_dataset import COCODataset
def build_dataset(dataset_list, transform=None, target_transform=None, is_test=False):
assert len(dataset_list) > 0
datasets = []
for data... | [
"torch.utils.data.ConcatDataset",
"ssd.config.path_catlog.DatasetCatalog.get"
] | [((361, 393), 'ssd.config.path_catlog.DatasetCatalog.get', 'DatasetCatalog.get', (['dataset_name'], {}), '(dataset_name)\n', (379, 393), False, 'from ssd.config.path_catlog import DatasetCatalog\n'), ((753, 776), 'torch.utils.data.ConcatDataset', 'ConcatDataset', (['datasets'], {}), '(datasets)\n', (766, 776), False, '... |
import pandas as pd
def read_clean_csv(path, sep='|'):
"""will read and clean a csv file'"""
df = pd.read_csv(path, sep=sep)
if('Unnamed: 0' in df.columns.values):
df = remove_unnamed(df)
return(df)
def remove_unnamed(df):
df = df.drop('Unnamed: 0', 1)
return df
| [
"pandas.read_csv"
] | [((104, 130), 'pandas.read_csv', 'pd.read_csv', (['path'], {'sep': 'sep'}), '(path, sep=sep)\n', (115, 130), True, 'import pandas as pd\n')] |
"""Read a GO Association File (GAF) and store the data in a Python object.
Annotations available from the Gene Ontology Consortium:
http://geneontology.org/page/download-annotations
GAF format:
http://geneontology.org/page/go-annotation-file-formats
"""
import sys
import re
import collections... | [
"re.split",
"goatools.base.nopen",
"re.compile",
"collections.Counter",
"goatools.evidence_codes.EvidenceCodes"
] | [((8166, 8201), 're.compile', 're.compile', (['"""^!(\\\\w[\\\\w\\\\s-]+:.*)$"""'], {}), "('^!(\\\\w[\\\\w\\\\s-]+:.*)$')\n", (8176, 8201), False, 'import re\n'), ((2562, 2577), 'goatools.evidence_codes.EvidenceCodes', 'EvidenceCodes', ([], {}), '()\n', (2575, 2577), False, 'from goatools.evidence_codes import Evidence... |