code
stringlengths
22
1.05M
apis
listlengths
1
3.31k
extract_api
stringlengths
75
3.25M
#!/usr/bin/env python3 import argparse import hashlib import json import re import signal import subprocess import sys import time import urllib.request DEFAULT_SERVER_HOSTNAME = "127.0.0.1" DEFAULT_SERVER_PORT = 6878 SERVER_POLL_TIME = 2 SERVER_STATUS_STREAM_ACTIVE = "dl" def exit_error(message): print(f"Erro...
[ "signal.signal", "argparse.ArgumentParser", "time.sleep", "sys.exit", "json.load", "re.search" ]
[((356, 367), 'sys.exit', 'sys.exit', (['(1)'], {}), '(1)\n', (364, 367), False, 'import sys\n'), ((819, 988), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {'description': '"""Instructs server to commence a given program ID. Will optionally execute a local media player once playback has started."""'}), "(...
# -*- coding: utf-8 -*- """ Created on Wed Jun 13 13:10:41 2018 @author: crius """ import Hamiltonians as H import numpy as np import tools as t import spintensor as st import spinops as so import time import Expand as ex from matplotlib import pyplot as plt exp = np.exp N = 8 nlegs = 4 S = 0.5 c = np.sqrt(2) Jcu...
[ "tools.Statelist", "numpy.sqrt", "Expand.append", "spinops.SziOp", "matplotlib.pyplot.plot", "numpy.asarray", "numpy.kron", "numpy.linspace", "Expand.Expand", "tools.exval", "Hamiltonians.nlegHeisenberg.blockH", "numpy.linalg.eigh", "numpy.cos", "time.time", "spinops.sz" ]
[((306, 316), 'numpy.sqrt', 'np.sqrt', (['(2)'], {}), '(2)\n', (313, 316), True, 'import numpy as np\n'), ((343, 429), 'Hamiltonians.nlegHeisenberg.blockH', 'H.nlegHeisenberg.blockH', (['N', 'S', 'nlegs', 'c'], {'Js': '[1, 1]', 'gamma': '[0.0, 4.0]', 'full': '"""True"""'}), "(N, S, nlegs, c, Js=[1, 1], gamma=[0.0, 4.0]...
import pandas as pd from pandas import Timestamp import hillmaker file_stopdata = '../data/unit_stop_log_Experiment1_Scenario138_Rep17.csv' scenario_name = 'log_unitocc_test_5' in_fld_name = 'EnteredTS' out_fld_name = 'ExitedTS' cat_fld_name = 'Unit' start_analysis = '12/12/2015 00:00' end_analysis = '12/19/2021 00:...
[ "pandas.Timestamp", "hillmaker.make_hills", "pandas.DateOffset", "pandas.read_csv" ]
[((412, 438), 'pandas.read_csv', 'pd.read_csv', (['file_stopdata'], {}), '(file_stopdata)\n', (423, 438), True, 'import pandas as pd\n'), ((450, 480), 'pandas.Timestamp', 'Timestamp', (['"""20150215 00:00:00"""'], {}), "('20150215 00:00:00')\n", (459, 480), False, 'from pandas import Timestamp\n'), ((891, 1153), 'hillm...
import pandas as pd import numpy as np def augment(data): data['MalePercent'] = (data['MaleLowerQuartile'] + data['MaleLowerMiddleQuartile'] + data['MaleUpperMiddleQuartile'] + data['MaleTopQuartile']) * .25 data['FemalePercent'] = (data['FemaleLowerQuarti...
[ "pandas.read_csv" ]
[((2170, 2222), 'pandas.read_csv', 'pd.read_csv', (['"""data/ukgov-gpg-full-section-split.csv"""'], {}), "('data/ukgov-gpg-full-section-split.csv')\n", (2181, 2222), True, 'import pandas as pd\n')]
#!/usr/bin/env python import os from typing import List, Union _PROJECT_DIRECTORY = os.path.realpath(os.path.curdir) def _remove_files(files: Union[List[str], str]) -> None: """ Removes the list of files provided. :param files: list of filepath to remove :type files: Union[List[str], str] """ ...
[ "os.path.realpath", "os.path.join" ]
[((86, 118), 'os.path.realpath', 'os.path.realpath', (['os.path.curdir'], {}), '(os.path.curdir)\n', (102, 118), False, 'import os\n'), ((424, 466), 'os.path.join', 'os.path.join', (['_PROJECT_DIRECTORY', 'filepath'], {}), '(_PROJECT_DIRECTORY, filepath)\n', (436, 466), False, 'import os\n')]
import json import channels.layers from asgiref.sync import async_to_sync from django.conf import settings from django.core.paginator import Paginator from django.http import HttpResponseRedirect from django.shortcuts import render, get_object_or_404 from django.urls import reverse, reverse_lazy from django.views.gene...
[ "django.shortcuts.render", "django.urls.reverse" ]
[((526, 579), 'django.shortcuts.render', 'render', (['request', '"""user/profile.html"""'], {'context': 'context'}), "(request, 'user/profile.html', context=context)\n", (532, 579), False, 'from django.shortcuts import render, get_object_or_404\n'), ((1024, 1042), 'django.urls.reverse', 'reverse', (['"""profile"""'], {...
# /usr/bin/python3 import logging import sched import signal import sys import time from functools import partial import watchdog.events import watchdog.observers from watchdog.observers.polling import PollingObserver as PollingObserver from GlobusTransfer import GlobusTransfer from .args import Args from .handler ...
[ "sched.scheduler", "GlobusTransfer.GlobusTransfer", "functools.partial", "watchdog.observers.polling.PollingObserver" ]
[((720, 797), 'GlobusTransfer.GlobusTransfer', 'GlobusTransfer', (['args.source', 'args.destination', 'args.destination_dir', 'src_path'], {}), '(args.source, args.destination, args.destination_dir, src_path)\n', (734, 797), False, 'from GlobusTransfer import GlobusTransfer\n'), ((958, 975), 'watchdog.observers.polling...
import argparse import os import sys import torch from torch import nn, optim from torch.optim import optimizer from torchvision import datasets, models, transforms parser = argparse.ArgumentParser(description="Trains a neural network") parser.add_argument('data_dir', metavar='dir', type=str, help...
[ "torch.nn.ReLU", "torch.nn.Dropout", "torchvision.models.vgg19", "torchvision.models.densenet161", "torch.optim.optimizer.step", "torchvision.models.resnet18", "torch.exp", "torch.cuda.is_available", "torchvision.models.densenet121", "sys.exit", "torchvision.models.vgg11", "argparse.ArgumentPa...
[((176, 238), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {'description': '"""Trains a neural network"""'}), "(description='Trains a neural network')\n", (199, 238), False, 'import argparse\n'), ((6130, 6156), 'os.path.join', 'os.path.join', (['dir', '"""train"""'], {}), "(dir, 'train')\n", (6142, 6156),...
# Ejercicio 04 import time import matplotlib.pyplot as plt # Algoritmo def binarySearch(numbers, n): numElements = len(numbers) if(numElements == 1): return (n == numbers[0]) mitad = numElements // 2 if(numbers[mitad] == n): return True elif (n < numbers[mitad]): return...
[ "matplotlib.pyplot.subplots", "time.perf_counter", "matplotlib.pyplot.show" ]
[((515, 534), 'time.perf_counter', 'time.perf_counter', ([], {}), '()\n', (532, 534), False, 'import time\n'), ((574, 593), 'time.perf_counter', 'time.perf_counter', ([], {}), '()\n', (591, 593), False, 'import time\n'), ((1146, 1160), 'matplotlib.pyplot.subplots', 'plt.subplots', ([], {}), '()\n', (1158, 1160), True, ...
# Generated by Django 2.2.4 on 2020-06-29 10:53 from django.conf import settings from django.db import migrations, models import django.db.models.deletion import jsonfield.fields class Migration(migrations.Migration): initial = True dependencies = [ migrations.swappable_dependency(settings.AUTH_USE...
[ "django.db.models.ForeignKey", "django.db.models.BooleanField", "django.db.models.AutoField", "django.db.models.BigIntegerField", "django.db.models.DateTimeField", "django.db.migrations.swappable_dependency", "django.db.models.CharField" ]
[((271, 328), 'django.db.migrations.swappable_dependency', 'migrations.swappable_dependency', (['settings.AUTH_USER_MODEL'], {}), '(settings.AUTH_USER_MODEL)\n', (302, 328), False, 'from django.db import migrations, models\n'), ((462, 555), 'django.db.models.AutoField', 'models.AutoField', ([], {'auto_created': '(True)...
from os import getenv from dataclasses import dataclass from collections import ChainMap from dotenv import load_dotenv from bot.structs import CaseInsensitiveDict load_dotenv() MY_NAME = "LPUBeltbot" MY_REDDIT_PW = getenv("BELTBOT_REDDIT_PW") MY_REDDIT_CLIENT_ID = getenv("BELTBOT_REDDIT_CID") MY_REDDIT_SECRET...
[ "bot.structs.CaseInsensitiveDict", "collections.ChainMap", "os.getenv", "dotenv.load_dotenv" ]
[((168, 181), 'dotenv.load_dotenv', 'load_dotenv', ([], {}), '()\n', (179, 181), False, 'from dotenv import load_dotenv\n'), ((223, 250), 'os.getenv', 'getenv', (['"""BELTBOT_REDDIT_PW"""'], {}), "('BELTBOT_REDDIT_PW')\n", (229, 250), False, 'from os import getenv\n'), ((274, 302), 'os.getenv', 'getenv', (['"""BELTBOT_...
import mnist import numpy as np from PIL import Image from conv import Conv3x3 from maxpool import MaxPool2 from softmax import Softmax train_images = mnist.train_images()[:100] train_labels = mnist.train_labels()[:100] test_images = mnist.test_images()[:1000] test_labels = mnist.test_labels()[:1000] conv = Conv3x3(8...
[ "softmax.Softmax", "mnist.test_labels", "mnist.train_images", "numpy.log", "numpy.argmax", "mnist.test_images", "numpy.array", "numpy.zeros", "conv.Conv3x3", "maxpool.MaxPool2", "mnist.train_labels" ]
[((311, 321), 'conv.Conv3x3', 'Conv3x3', (['(8)'], {}), '(8)\n', (318, 321), False, 'from conv import Conv3x3\n'), ((329, 339), 'maxpool.MaxPool2', 'MaxPool2', ([], {}), '()\n', (337, 339), False, 'from maxpool import MaxPool2\n'), ((350, 374), 'softmax.Softmax', 'Softmax', (['(13 * 13 * 8)', '(10)'], {}), '(13 * 13 * ...
# -*- coding: utf-8 -*- """ :author: T8840 :tag: Thinking is a good thing! 纸上得来终觉浅,绝知此事要躬行! :description: 1.部署相关信息来自于nacos配置 { "server_info": { "host": "10.201.5.161", "port":22, "user" : "user", ...
[ "paramiko.SFTPClient.from_transport", "json.loads", "pathlib.Path", "pathlib.Path.cwd", "paramiko.AutoAddPolicy", "arrow.now", "paramiko.Transport", "requests.get", "platform.system", "paramiko.SSHClient" ]
[((1362, 1514), 'requests.get', 'requests.get', (['"""http://10.201.7.185:8848/nacos/v1/cs/configs?dataId=deploy_124&group=DEFAULT_GROUP&tenant=89362432-5255-497e-8e94-cb77d46cb1a9"""'], {}), "(\n 'http://10.201.7.185:8848/nacos/v1/cs/configs?dataId=deploy_124&group=DEFAULT_GROUP&tenant=89362432-5255-497e-8e94-cb77d...
# -*- coding: utf-8 -*- """ Functions for plotting reliability diagrams: smooths of simulated vs observed outcomes on the y-axis against predicted probabilities on the x-axis. """ from __future__ import absolute_import import matplotlib.pyplot as plt import numpy as np import seaborn as sbn from .plot_utils import _l...
[ "seaborn.set_style", "numpy.linspace", "matplotlib.pyplot.subplots", "past.builtins.range" ]
[((681, 706), 'seaborn.set_style', 'sbn.set_style', (['"""darkgrid"""'], {}), "('darkgrid')\n", (694, 706), True, 'import seaborn as sbn\n'), ((2708, 2754), 'numpy.linspace', 'np.linspace', (['min_ref_val', 'max_ref_val'], {'num': '(100)'}), '(min_ref_val, max_ref_val, num=100)\n', (2719, 2754), True, 'import numpy as ...
""" An AWS Lambda function used to run periodic background jobs on ECS. The complication with running these tasks is that we need to run them on the same version of the Docker image that the web servers are currently running on. """ import logging import boto3 from utils import env_list, env_param # Logging setup...
[ "logging.getLogger", "utils.env_param", "boto3.client", "utils.env_list" ]
[((348, 375), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (365, 375), False, 'import logging\n'), ((461, 480), 'boto3.client', 'boto3.client', (['"""ecs"""'], {}), "('ecs')\n", (473, 480), False, 'import boto3\n'), ((1198, 1218), 'utils.env_param', 'env_param', (['"""CLUSTER"""'], {}),...
#! python3 # -*- coding: utf-8 -*- import os from pypeapp import execute, Logger from pype.hosts.resolve.utils import get_resolve_module log = Logger().get_logger("Resolve") CURRENT_DIR = os.getenv("RESOLVE_UTILITY_SCRIPTS_DIR", "") python_dir = os.getenv("PYTHON36_RESOLVE") python_exe = os.path.normpath( os.path...
[ "os.getenv", "pype.hosts.resolve.utils.get_resolve_module", "os.path.join", "os.path.normpath", "pypeapp.Logger" ]
[((190, 234), 'os.getenv', 'os.getenv', (['"""RESOLVE_UTILITY_SCRIPTS_DIR"""', '""""""'], {}), "('RESOLVE_UTILITY_SCRIPTS_DIR', '')\n", (199, 234), False, 'import os\n'), ((248, 277), 'os.getenv', 'os.getenv', (['"""PYTHON36_RESOLVE"""'], {}), "('PYTHON36_RESOLVE')\n", (257, 277), False, 'import os\n'), ((365, 385), 'p...
from collections import OrderedDict import torch import torch.nn as nn import torch.nn.functional as F import logging from libs.modules.FuseBlock import MakeFB from .resnet_dilation import resnet50, resnet101, Bottleneck, conv1x1 BN_MOMENTUM = 0.1 logger = logging.getLogger(__name__) def conv3x3(in_planes, out_plan...
[ "logging.getLogger", "torch.nn.BatchNorm2d", "torch.nn.ReLU", "torch.nn.Sequential", "torch.nn.Dropout2d", "torch.nn.functional.sigmoid", "torch.nn.Conv2d", "torch.nn.Module", "torch.nn.functional.interpolate", "torch.nn.AdaptiveAvgPool2d", "libs.modules.FuseBlock.MakeFB", "torch.cat" ]
[((259, 286), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (276, 286), False, 'import logging\n'), ((385, 474), 'torch.nn.Conv2d', 'nn.Conv2d', (['in_planes', 'out_planes'], {'kernel_size': '(3)', 'stride': 'stride', 'padding': '(1)', 'bias': '(False)'}), '(in_planes, out_planes, kernel...
# -*- coding: utf-8 -*- """ @author : <NAME> @github : https://github.com/tianpangji @software : PyCharm @file : crud.py @create : 2020/12/9 20:44 """ from django.contrib.contenttypes.models import ContentType from easyaudit.models import CRUDEvent from rest_framework import serializers class CRUDSerializer...
[ "rest_framework.serializers.DateTimeField", "rest_framework.serializers.SerializerMethodField", "django.contrib.contenttypes.models.ContentType.objects.get" ]
[((376, 411), 'rest_framework.serializers.SerializerMethodField', 'serializers.SerializerMethodField', ([], {}), '()\n', (409, 411), False, 'from rest_framework import serializers\n'), ((427, 496), 'rest_framework.serializers.DateTimeField', 'serializers.DateTimeField', ([], {'format': '"""%Y-%m-%d %H:%M:%S"""', 'read_...
""" ICMP(Internet Control Message Protocol) - Echo Request: +-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+ |0|0|0|0|0|0|0|0|0|0|1|1|1|1|1|1|1|1|1|1|2|2|2|2|2|2|2|2|2|2|3|3| |0|1|2|3|4|5|6|7|8|9|0|1|2|3|4|5|6|7|8|9|0|1|2|3|4|5|6|7|8|9|0|1| +-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+...
[ "socket.gethostbyname", "socket.socket", "time.sleep", "struct.pack", "time.time" ]
[((1971, 2067), 'struct.pack', 'struct.pack', (['"""!BBHHHLL48s"""', 'Type', 'Code', '(0)', 'Identifier', 'SequenceNumber', 'TimeStamp', '(0)', 'Payload'], {}), "('!BBHHHLL48s', Type, Code, 0, Identifier, SequenceNumber,\n TimeStamp, 0, Payload)\n", (1982, 2067), False, 'import struct\n'), ((2257, 2360), 'struct.pac...
import random from model.contact import Contact from model.group import Group def test_to_delete_contact_from_group(app, db ,check_ui): app.group.open_groups_page() if len(db.get_group_list()) == 0: app.group.create() app.group.fill_group_form(Group(name='group for adding of contact')) ...
[ "model.group.Group", "random.choice", "model.contact.Contact" ]
[((443, 468), 'random.choice', 'random.choice', (['group_list'], {}), '(group_list)\n', (456, 468), False, 'import random\n'), ((1540, 1572), 'random.choice', 'random.choice', (['contacts_in_group'], {}), '(contacts_in_group)\n', (1553, 1572), False, 'import random\n'), ((1292, 1325), 'random.choice', 'random.choice', ...
import gzip import logging import logging.handlers import os from cStringIO import StringIO as IO from spreads.vendor.huey import SqliteHuey from spreads.vendor.huey.consumer import Consumer from spreads.vendor.pathlib import Path from flask import Flask, request from spreads.plugin import (HookPlugin, SubcommandHook...
[ "logging.getLogger", "os.path.exists", "spreads.util.add_argument_from_option", "cStringIO.StringIO", "flask.Flask", "logging.handlers.BufferingHandler", "spreads.vendor.huey.SqliteHuey", "spreads.vendor.huey.consumer.Consumer", "netifaces.ifaddresses", "gzip.GzipFile", "waitress.serve", "os.m...
[((439, 541), 'flask.Flask', 'Flask', (['"""spreadsplug.web"""'], {'static_url_path': '""""""', 'static_folder': '"""./client"""', 'template_folder': '"""./client"""'}), "('spreadsplug.web', static_url_path='', static_folder='./client',\n template_folder='./client')\n", (444, 541), False, 'from flask import Flask, r...
"""A tool to convert annotation files created with CVAT into ground-truth style images for machine learning. The initial code was copied from: https://gist.github.com/cheind/9850e35bb08cfe12500942fb8b55531f originally written for a similar purpose for the tool BeaverDam (which produces json), and was then adapted f...
[ "cv2.rectangle", "numpy.copy", "xml.etree.ElementTree.parse", "argparse.ArgumentParser", "cv2.VideoWriter", "cv2.imshow", "numpy.zeros", "cv2.destroyAllWindows", "cv2.VideoCapture", "cv2.VideoWriter_fourcc", "cv2.waitKey" ]
[((873, 891), 'xml.etree.ElementTree.parse', 'ET.parse', (['args.ann'], {}), '(args.ann)\n', (881, 891), True, 'import xml.etree.ElementTree as ET\n'), ((1101, 1129), 'cv2.VideoCapture', 'cv2.VideoCapture', (['args.video'], {}), '(args.video)\n', (1117, 1129), False, 'import cv2\n'), ((1464, 1495), 'cv2.VideoWriter_fou...
import time from selenium import webdriver from lxml import etree driver = webdriver.PhantomJS(executable_path='./phantomjs-2.1.1-macosx/bin/phantomjs') # 获取第一页的数据 def get_html(): url = "https://detail.tmall.com/item.htm?id=531993957001&skuId=3609796167425&user_id=268451883&cat_id=2&is_b=1&rn=71b9b0aeb233411c4f5...
[ "selenium.webdriver.PhantomJS", "lxml.etree.HTML", "time.sleep" ]
[((76, 153), 'selenium.webdriver.PhantomJS', 'webdriver.PhantomJS', ([], {'executable_path': '"""./phantomjs-2.1.1-macosx/bin/phantomjs"""'}), "(executable_path='./phantomjs-2.1.1-macosx/bin/phantomjs')\n", (95, 153), False, 'from selenium import webdriver\n'), ((359, 372), 'time.sleep', 'time.sleep', (['(5)'], {}), '(...
""" """ from flask import Flask import os import json from molder import db, site def create_app(instance_path=None, test_config=None): """ Create and configure the molder Flask application. Parameters ---------- instance_path : :class:`str`, optional Path to the instance directory of t...
[ "os.path.exists", "molder.db.init_app", "os.makedirs", "flask.Flask" ]
[((532, 607), 'flask.Flask', 'Flask', (['__name__'], {'instance_path': 'instance_path', 'instance_relative_config': '(True)'}), '(__name__, instance_path=instance_path, instance_relative_config=True)\n', (537, 607), False, 'from flask import Flask\n'), ((1460, 1476), 'molder.db.init_app', 'db.init_app', (['app'], {}), ...
from PIL import Image, ImageDraw, ImageFont from urllib.request import urlopen from textwrap import wrap import os BOLD_FONT_URL = "https://cdn.jsdelivr.net/gh/spoqa/spoqa-han-sans@latest/Subset/SpoqaHanSansNeo/SpoqaHanSansNeo-Bold.ttf" LIGHT_FONT_URL = "https://cdn.jsdelivr.net/gh/spoqa/spoqa-han-sans@latest/Subset/S...
[ "os.path.exists", "os.listdir", "PIL.Image.new", "PIL.ImageDraw.Draw", "os.mkdir", "textwrap.wrap", "urllib.request.urlopen" ]
[((780, 800), 'os.listdir', 'os.listdir', (['POSTPATH'], {}), '(POSTPATH)\n', (790, 800), False, 'import os\n'), ((2170, 2215), 'PIL.Image.new', 'Image.new', (['"""RGB"""', 'OPENGRAPH_SIZE'], {'color': 'WHITE'}), "('RGB', OPENGRAPH_SIZE, color=WHITE)\n", (2179, 2215), False, 'from PIL import Image, ImageDraw, ImageFont...
import os import yaml import pytest import testinfra.utils.ansible_runner testinfra_hosts = testinfra.utils.ansible_runner.AnsibleRunner( os.environ['MOLECULE_INVENTORY_FILE']).get_hosts('all') dir_path = os.path.dirname(os.path.abspath(__file__)) @pytest.fixture() def AnsibleDefaults(): with open(os.path.j...
[ "os.path.join", "yaml.load", "pytest.mark.parametrize", "pytest.fixture", "os.path.abspath" ]
[((257, 273), 'pytest.fixture', 'pytest.fixture', ([], {}), '()\n', (271, 273), False, 'import pytest\n'), ((418, 434), 'pytest.fixture', 'pytest.fixture', ([], {}), '()\n', (432, 434), False, 'import pytest\n'), ((569, 655), 'pytest.mark.parametrize', 'pytest.mark.parametrize', (['"""minio_bin_var"""', "['minio_server...
""" resourceview.py Contains administrative views for working with resources. """ from datetime import date from admin_helpers import * from sqlalchemy import or_, not_, func from flask import current_app, redirect, flash, request, url_for from flask.ext.admin import BaseView, expose from flask.ext.admin.actions im...
[ "sqlalchemy.func.count", "flask.ext.admin.expose", "flask.flash", "flask.request.form.getlist", "flask.ext.admin.actions.action", "flask.url_for", "flask.request.args.getlist", "remedy.rad.models.Category.name.asc", "wtforms.validators.Optional", "remedy.rad.models.Resource.name.asc", "datetime....
[((2137, 2259), 'flask.ext.admin.actions.action', 'action', (['"""togglevisible"""', '"""Toggle Visibility"""', '"""Are you sure you wish to toggle visibility for the selected resources?"""'], {}), "('togglevisible', 'Toggle Visibility',\n 'Are you sure you wish to toggle visibility for the selected resources?')\n",...
from typing import List import tensorflow as tf from tensorflow.keras.layers import Embedding, Layer, LSTM, Input from src.features.preprocessing import get_embedding, get_text_vectorization from src.models.embedding_model import EmbeddingModel class LSTMModel(EmbeddingModel): def __init__(self): super(...
[ "src.features.preprocessing.get_embedding", "src.features.preprocessing.get_text_vectorization", "tensorflow.keras.layers.LSTM", "tensorflow.keras.layers.Input" ]
[((360, 375), 'src.features.preprocessing.get_embedding', 'get_embedding', ([], {}), '()\n', (373, 375), False, 'from src.features.preprocessing import get_embedding, get_text_vectorization\n'), ((411, 435), 'src.features.preprocessing.get_text_vectorization', 'get_text_vectorization', ([], {}), '()\n', (433, 435), Fal...
# coding=utf-8 # *** WARNING: this file was generated by the Pulumi Terraform Bridge (tfgen) Tool. *** # *** Do not edit by hand unless you're certain you know what you are doing! *** import warnings import pulumi import pulumi.runtime from typing import Any, Mapping, Optional, Sequence, Union, overload from .. import...
[ "pulumi.get", "pulumi.getter", "pulumi.set", "pulumi.InvokeOptions", "pulumi.runtime.invoke" ]
[((4092, 4132), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""availabilityDomain"""'}), "(name='availabilityDomain')\n", (4105, 4132), False, 'import pulumi\n'), ((4382, 4417), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""compartmentId"""'}), "(name='compartmentId')\n", (4395, 4417), False, 'import pulumi\...
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Wed May 23 07:33:30 2018 @author: aaronpenne """ import os import pandas as pd import numpy as np code_dir = os.path.dirname(__file__) # Returns full path of this script data_dir = os.path.join(code_dir, 'data') output_dir = os.path.join(code_dir, 'outpu...
[ "os.path.dirname", "os.path.isdir", "os.path.join", "os.mkdir" ]
[((174, 199), 'os.path.dirname', 'os.path.dirname', (['__file__'], {}), '(__file__)\n', (189, 199), False, 'import os\n'), ((247, 277), 'os.path.join', 'os.path.join', (['code_dir', '"""data"""'], {}), "(code_dir, 'data')\n", (259, 277), False, 'import os\n'), ((291, 323), 'os.path.join', 'os.path.join', (['code_dir', ...
import os import tempfile import pytest import numpy as np try: import h5py except ImportError: h5py = None from msl.io import read, HDF5Writer, JSONWriter from msl.io.readers import HDF5Reader from helper import read_sample, roots_equal @pytest.mark.skipif(h5py is None, reason='h5py not installed') def te...
[ "helper.read_sample", "numpy.random.random", "msl.io.read", "msl.io.HDF5Writer", "numpy.array", "numpy.array_equal", "pytest.raises", "tempfile.gettempdir", "pytest.mark.skipif", "os.path.basename", "helper.roots_equal", "os.remove" ]
[((252, 313), 'pytest.mark.skipif', 'pytest.mark.skipif', (['(h5py is None)'], {'reason': '"""h5py not installed"""'}), "(h5py is None, reason='h5py not installed')\n", (270, 313), False, 'import pytest\n'), ((3668, 3729), 'pytest.mark.skipif', 'pytest.mark.skipif', (['(h5py is None)'], {'reason': '"""h5py not installe...
import logging import tkinter as tk class CalibrationMsgBox(): """ The message box class. """ MESSAGES = [ 'Turn the steering wheel fully LEFT and press x.', 'Turn the steering wheel fully RIGHT and press x.', 'Make sure the throttle is fully DEPRESSED and press x.', 'Ma...
[ "logging.getLogger", "tkinter.StringVar", "tkinter.Toplevel", "tkinter.Label" ]
[((811, 841), 'logging.getLogger', 'logging.getLogger', (['"""CALIB_MSG"""'], {}), "('CALIB_MSG')\n", (828, 841), False, 'import logging\n'), ((1039, 1056), 'tkinter.Toplevel', 'tk.Toplevel', (['root'], {}), '(root)\n', (1050, 1056), True, 'import tkinter as tk\n'), ((1230, 1253), 'tkinter.StringVar', 'tk.StringVar', (...
import numpy as np from ringity.classes.diagram import PersistenceDiagram def read_pdiagram(fname, **kwargs): """ Wrapper for numpy.genfromtxt. """ return PersistenceDiagram(np.genfromtxt(fname, **kwargs)) def write_pdiagram(dgm, fname, **kwargs): """ Wrapper for numpy.savetxt. """ ...
[ "numpy.array", "numpy.genfromtxt", "numpy.savetxt" ]
[((329, 342), 'numpy.array', 'np.array', (['dgm'], {}), '(dgm)\n', (337, 342), True, 'import numpy as np\n'), ((347, 381), 'numpy.savetxt', 'np.savetxt', (['fname', 'array'], {}), '(fname, array, **kwargs)\n', (357, 381), True, 'import numpy as np\n'), ((191, 221), 'numpy.genfromtxt', 'np.genfromtxt', (['fname'], {}), ...
import functools import os from argparse import ArgumentParser import networkx import numpy as np from visualize import heatmap class MatchingClustering(object): def __init__(self, n_clusters): self.n_clusters = n_clusters def fit_predict(self, X): total = len(X) grouping = [{i} for...
[ "networkx.algorithms.max_weight_matching", "argparse.ArgumentParser", "numpy.where", "os.path.join", "networkx.Graph", "numpy.max", "numpy.sum", "numpy.zeros" ]
[((1247, 1263), 'argparse.ArgumentParser', 'ArgumentParser', ([], {}), '()\n', (1261, 1263), False, 'from argparse import ArgumentParser\n'), ((1101, 1130), 'numpy.zeros', 'np.zeros', (['total'], {'dtype': 'np.int'}), '(total, dtype=np.int)\n', (1109, 1130), True, 'import numpy as np\n'), ((1428, 1470), 'os.path.join',...
#!/usr/bin/env python # -*- coding: utf-8 -*- import os from sfdc_cli.package_xml import PackageXml command_name = os.path.basename(__file__).split('.', 1)[0].replace("_", ":") def register(parser, subparsers, **kwargs): def handler(args): if args.scandir and args.savedir and args.name and args.apiversi...
[ "os.path.basename", "sfdc_cli.package_xml.PackageXml" ]
[((116, 142), 'os.path.basename', 'os.path.basename', (['__file__'], {}), '(__file__)\n', (132, 142), False, 'import os\n'), ((337, 364), 'sfdc_cli.package_xml.PackageXml', 'PackageXml', ([], {'project_dir': '"""."""'}), "(project_dir='.')\n", (347, 364), False, 'from sfdc_cli.package_xml import PackageXml\n')]
from django.conf.urls import url from basic_app import views # SET THE NAMESPACE! app_name = 'basic_app' urlpatterns=[ url(r'^register/$',views.register,name='register'), url(r'^user_login/$',views.user_login,name='user_login'), url(r'^add/$',views.add,name='add'), url(r'^bookadd/$',views.bookadd,name...
[ "django.conf.urls.url" ]
[((125, 176), 'django.conf.urls.url', 'url', (['"""^register/$"""', 'views.register'], {'name': '"""register"""'}), "('^register/$', views.register, name='register')\n", (128, 176), False, 'from django.conf.urls import url\n'), ((181, 238), 'django.conf.urls.url', 'url', (['"""^user_login/$"""', 'views.user_login'], {'...
# _*_ coding: utf-8 _*_ """ Implementation of quick sort algorithm. Reference: [1] https://runestone.academy/runestone/books/published/pythonds/SortSearch/TheQuickSort.html Author: <NAME> """ from typing import List def _quick_sort(nums: List[int], li: int, ri: int) -> None: if li >= ri: return sp...
[ "time.time" ]
[((1341, 1347), 'time.time', 'time', ([], {}), '()\n', (1345, 1347), False, 'from time import time\n'), ((1422, 1428), 'time.time', 'time', ([], {}), '()\n', (1426, 1428), False, 'from time import time\n')]
#!/usr/bin/env python """Command-line interface for skyviewbot""" # ASTERICS-OBELICS Good Coding Practices (skyviewbot.py) # <NAME> (<EMAIL>), with suggestions from <NAME> import sys from .functions import skyviewbot from argparse import ArgumentParser, RawTextHelpFormatter def main(*function_args): """Command-...
[ "argparse.ArgumentParser" ]
[((496, 548), 'argparse.ArgumentParser', 'ArgumentParser', ([], {'formatter_class': 'RawTextHelpFormatter'}), '(formatter_class=RawTextHelpFormatter)\n', (510, 548), False, 'from argparse import ArgumentParser, RawTextHelpFormatter\n')]
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Sun Dec 27 14:39:08 2020 @author: ravi """ import scipy.io as scio import scipy.io.wavfile as scwav import numpy as np import joblib import pyworld as pw import os import warnings warnings.filterwarnings('ignore') from tqdm import tqdm from concurrent.fut...
[ "feat_utils.preprocess_contour", "numpy.random.rand", "pyworld.code_spectral_envelope", "pyworld.cheaptrick", "tqdm.tqdm", "numpy.asarray", "extract_fold_data_hparams.Hparams", "os.path.join", "numpy.sum", "numpy.array", "numpy.random.randint", "scipy.io.wavfile.read", "functools.partial", ...
[((244, 277), 'warnings.filterwarnings', 'warnings.filterwarnings', (['"""ignore"""'], {}), "('ignore')\n", (267, 277), False, 'import warnings\n'), ((3408, 3474), 'joblib.load', 'joblib.load', (['"""/home/ravi/Downloads/Emo-Conv/speaker_file_info.pkl"""'], {}), "('/home/ravi/Downloads/Emo-Conv/speaker_file_info.pkl')\...
import numpy as np import cv2 import errno # set environment variable import os os.environ['OPENCV_IO_ENABLE_JASPER']= 'TRUE' # allows JPEG2000 format # path of this file det_path = os.path.split(os.path.abspath(__file__))[0] + '/' class DimensionError(Exception): """ raised when the image does not me...
[ "numpy.mean", "cv2.imwrite", "os.strerror", "cv2.dnn.readNetFromCaffe", "os.path.isfile", "numpy.array", "os.path.abspath", "cv2.resize", "cv2.imread" ]
[((1585, 1604), 'cv2.imread', 'cv2.imread', (['in_path'], {}), '(in_path)\n', (1595, 1604), False, 'import cv2\n'), ((1783, 1910), 'cv2.dnn.readNetFromCaffe', 'cv2.dnn.readNetFromCaffe', (["(det_path + 'models/deploy.prototxt')", "(det_path + 'models/res10_300x300_ssd_iter_140000.caffemodel')"], {}), "(det_path + 'mode...
# # Connects to SlideRule server at provided url and prints log messages # generated on server to local terminal # import sys import logging from sliderule import sliderule from sliderule import icesat2 ############################################################################### # GLOBAL CODE ####################...
[ "logging.basicConfig", "sliderule.icesat2.init", "sliderule.sliderule.source" ]
[((401, 440), 'logging.basicConfig', 'logging.basicConfig', ([], {'level': 'logging.INFO'}), '(level=logging.INFO)\n', (420, 440), False, 'import logging\n'), ((1300, 1323), 'sliderule.icesat2.init', 'icesat2.init', (['url', '(True)'], {}), '(url, True)\n', (1312, 1323), False, 'from sliderule import icesat2\n'), ((149...
import setuptools requirements = [ 'docopt', 'numpy', 'pyzmq' ] console_scripts = [ 'lambda_client=lambda_scope.zmq.client:main', 'lambda_forwarder=lambda_scope.zmq.forwarder:main', 'lambda_hub=lambda_scope.devices.hub_relay:main', 'lambda_publisher=lambda_scope.zmq.publisher:main', 'l...
[ "setuptools.setup" ]
[((1373, 2002), 'setuptools.setup', 'setuptools.setup', ([], {'name': '"""lambda_scope"""', 'version': '"""0.0.1"""', 'author': '"""<NAME>"""', 'author_email': '"""<EMAIL>"""', 'description': '"""Software to operate the customized imaging system, lambda."""', 'url': '"""https://github.com/venkatachalamlab/lambda"""', '...
""" NCL_coneff_16.py ================ This script illustrates the following concepts: - Showing features of the new color display model - Using a NCL colormap with levels to assign a color palette to contours - Drawing partially transparent filled contours See following URLs to see the reproduced NCL plot & s...
[ "geocat.datafiles.get", "geocat.viz.util.add_major_minor_ticks", "matplotlib.pyplot.colorbar", "cartopy.crs.PlateCarree", "matplotlib.pyplot.figure", "numpy.linspace", "matplotlib.pyplot.axes", "geocat.viz.util.add_lat_lon_ticklabels", "geocat.viz.util.set_titles_and_labels", "matplotlib.pyplot.sh...
[((1221, 1248), 'matplotlib.pyplot.figure', 'plt.figure', ([], {'figsize': '(14, 7)'}), '(figsize=(14, 7))\n', (1231, 1248), True, 'import matplotlib.pyplot as plt\n'), ((1294, 1312), 'cartopy.crs.PlateCarree', 'ccrs.PlateCarree', ([], {}), '()\n', (1310, 1312), True, 'import cartopy.crs as ccrs\n'), ((1318, 1349), 'ma...
from sqlalchemy.ext.declarative import declarative_base from sqlalchemy import Column, Integer, String from sqlalchemy import create_engine from sqlalchemy.ext.declarative import declarative_base from sqlalchemy.orm import sessionmaker engine = create_engine('sqlite:///gosts.db') Session = sessionmaker(bind=engine) s...
[ "sqlalchemy.orm.sessionmaker", "sqlalchemy.create_engine", "sqlalchemy.Column", "sqlalchemy.ext.declarative.declarative_base" ]
[((247, 282), 'sqlalchemy.create_engine', 'create_engine', (['"""sqlite:///gosts.db"""'], {}), "('sqlite:///gosts.db')\n", (260, 282), False, 'from sqlalchemy import create_engine\n'), ((293, 318), 'sqlalchemy.orm.sessionmaker', 'sessionmaker', ([], {'bind': 'engine'}), '(bind=engine)\n', (305, 318), False, 'from sqlal...
#!/usr/bin/env python # -*- coding: UTF-8 -*- __author__ = '<NAME>' import re from functools import lru_cache import numpy as np import pandas as pd from .reaction import Reaction from .metabolite import Metabolite from ..optim.optim import FBAOptimizer, TFBAOptimizer, ETFBAOptimizer from .pdict import PrettyDict ...
[ "re.match", "pandas.read_excel", "pandas.DataFrame", "functools.lru_cache", "re.search" ]
[((4031, 4042), 'functools.lru_cache', 'lru_cache', ([], {}), '()\n', (4040, 4042), False, 'from functools import lru_cache\n'), ((4854, 4865), 'functools.lru_cache', 'lru_cache', ([], {}), '()\n', (4863, 4865), False, 'from functools import lru_cache\n'), ((5988, 5999), 'functools.lru_cache', 'lru_cache', ([], {}), '(...
"""File with only WriteMessage class.""" import curses class WriteMessage: def __init__(self, history_file=False): """ Write data to the pad. Parameters ---------- n_line: int number of line of the pad n_col: int number of colu...
[ "curses.color_pair", "curses.start_color", "curses.init_pair", "curses.has_colors", "curses.newpad" ]
[((598, 618), 'curses.start_color', 'curses.start_color', ([], {}), '()\n', (616, 618), False, 'import curses\n'), ((631, 650), 'curses.has_colors', 'curses.has_colors', ([], {}), '()\n', (648, 650), False, 'import curses\n'), ((1436, 1478), 'curses.newpad', 'curses.newpad', (['self._max_line', 'self._n_col'], {}), '(s...
import matplotlib.pyplot as plt from string import ascii_uppercase def countSpecific(_path, _letter): _letter = _letter.strip().upper() file = open(_path, 'rb') text = str(file.read()) return text.count(_letter) + text.count(_letter.lower()) def countAll(_path): file = open(_path, "rb") text =...
[ "matplotlib.pyplot.show" ]
[((811, 821), 'matplotlib.pyplot.show', 'plt.show', ([], {}), '()\n', (819, 821), True, 'import matplotlib.pyplot as plt\n')]
import requests import pytest from apistar import TestClient from api.web.support import Status from tests.markers import smoke @pytest.fixture(scope="module") def response(client: TestClient) -> requests.Response: return client.get("/api") @smoke def test_get_vehicles_status(response: requests.Response) -> Non...
[ "pytest.fixture" ]
[((131, 161), 'pytest.fixture', 'pytest.fixture', ([], {'scope': '"""module"""'}), "(scope='module')\n", (145, 161), False, 'import pytest\n')]
import matplotlib.pyplot as plt import numpy as np from matplotlib.ticker import FuncFormatter filepath = '/Users/huangjiaming/Documents/developer/ETreeLearning/res/losses/delay_etree.txt' x = [] num = 0 with open(filepath) as fp: for line in fp: c = list(map(int, line.split())) x = c print(np....
[ "numpy.mean", "matplotlib.pyplot.savefig", "matplotlib.pyplot.subplots", "numpy.std" ]
[((360, 397), 'matplotlib.pyplot.subplots', 'plt.subplots', ([], {'nrows': '(2)', 'figsize': '(9, 6)'}), '(nrows=2, figsize=(9, 6))\n', (372, 397), True, 'import matplotlib.pyplot as plt\n'), ((853, 917), 'matplotlib.pyplot.savefig', 'plt.savefig', (['"""./reports/20200301/delay_etree_100_nodes"""'], {'dpi': '(600)'}),...
from flask_restx import reqparse from db_plugins.db.sql import models columns = [] for c in models.Object.__table__.columns: columns.append(str(c).split(".")[1]) for c in models.Probability.__table__.columns: columns.append(str(c).split(".")[1]) def str2bool(v): if isinstance(v, bool): return v ...
[ "flask_restx.reqparse.RequestParser", "flask_restx.reqparse.ArgumentTypeError" ]
[((619, 643), 'flask_restx.reqparse.RequestParser', 'reqparse.RequestParser', ([], {}), '()\n', (641, 643), False, 'from flask_restx import reqparse\n'), ((2341, 2365), 'flask_restx.reqparse.RequestParser', 'reqparse.RequestParser', ([], {}), '()\n', (2363, 2365), False, 'from flask_restx import reqparse\n'), ((2923, 2...
# -*- coding: utf-8 -*- """ Created on Wed Sep 25 15:06:45 2019 @author: garci """ import matplotlib.pyplot as plt import numpy as np import csv import xlwings as xw import pandas import os '''MAKE X-Y PLOTS WITH 2-COLUMN FILES <NAME>, 2019 ''' '''lastRow credit: answered Sep 14 '16 at 11:39 - Stefan https://st...
[ "matplotlib.pyplot.savefig", "matplotlib.pyplot.xlabel", "matplotlib.pyplot.plot", "os.walk", "matplotlib.pyplot.figure", "time.time", "matplotlib.pyplot.title", "xlwings.Book", "pandas.read_fwf", "matplotlib.pyplot.legend" ]
[((1217, 1237), 'xlwings.Book', 'xw.Book', (['(path + file)'], {}), '(path + file)\n', (1224, 1237), True, 'import xlwings as xw\n'), ((1480, 1492), 'matplotlib.pyplot.figure', 'plt.figure', ([], {}), '()\n', (1490, 1492), True, 'import matplotlib.pyplot as plt\n'), ((1497, 1511), 'matplotlib.pyplot.plot', 'plt.plot', ...
import urllib from bs4 import BeautifulSoup print ("Collecting data from IMDb charts....\n\n\n") print ("The current top 15 IMDB movies are the following: \n\n") response = urllib.request.urlopen("http://www.imdb.com/chart/top") html = response.read() soup = BeautifulSoup(html, 'html.parser') mytd = soup.findAll("td",...
[ "bs4.BeautifulSoup", "urllib.request.urlopen" ]
[((174, 229), 'urllib.request.urlopen', 'urllib.request.urlopen', (['"""http://www.imdb.com/chart/top"""'], {}), "('http://www.imdb.com/chart/top')\n", (196, 229), False, 'import urllib\n'), ((260, 294), 'bs4.BeautifulSoup', 'BeautifulSoup', (['html', '"""html.parser"""'], {}), "(html, 'html.parser')\n", (273, 294), Fa...
from django.urls import path import mainapp.views as mainapp app_name = "mainapp" urlpatterns = [ path("", mainapp.product, name="index"), path("<int:pk>/", mainapp.product, name="category"), path("<int:pk>/page/<int:page>/", mainapp.product, name="page"), path("product/<int:pk>/", mainapp.product_p...
[ "django.urls.path" ]
[((106, 145), 'django.urls.path', 'path', (['""""""', 'mainapp.product'], {'name': '"""index"""'}), "('', mainapp.product, name='index')\n", (110, 145), False, 'from django.urls import path\n'), ((151, 202), 'django.urls.path', 'path', (['"""<int:pk>/"""', 'mainapp.product'], {'name': '"""category"""'}), "('<int:pk>/',...
import datetime import os import random import time import requests from lxml import etree from selenium import webdriver # import config import threading # import numpy as np mUA_list = [ 'Mozilla/5.0 (iPhone; CPU iPhone OS 11_2_1 like Mac OS X) AppleWebKit/604.4.7 (KHTML, like Gecko) Version/11.0 Mobile/15C153...
[ "selenium.webdriver.Chrome", "selenium.webdriver.ChromeOptions", "time.sleep" ]
[((1617, 1630), 'time.sleep', 'time.sleep', (['(2)'], {}), '(2)\n', (1627, 1630), False, 'import time\n'), ((1661, 1675), 'time.sleep', 'time.sleep', (['(20)'], {}), '(20)\n', (1671, 1675), False, 'import time\n'), ((1037, 1062), 'selenium.webdriver.ChromeOptions', 'webdriver.ChromeOptions', ([], {}), '()\n', (1060, 10...
import os import random import string import json from django.core.management import BaseCommand __author__ = "<NAME>" __copyright__ = "Copyright 2018, <NAME>" __licence__ = "BSD 2-Clause Licence" __version__ = "1.0" __email__ = "<EMAIL>" class Command(BaseCommand): def is_valid_file(self, file): if no...
[ "os.path.isfile", "json.dumps", "random.SystemRandom", "os.path.expanduser" ]
[((322, 342), 'os.path.isfile', 'os.path.isfile', (['file'], {}), '(file)\n', (336, 342), False, 'import os\n'), ((1136, 1166), 'json.dumps', 'json.dumps', (['settings'], {'indent': '(2)'}), '(settings, indent=2)\n', (1146, 1166), False, 'import json\n'), ((581, 602), 'random.SystemRandom', 'random.SystemRandom', ([], ...
from flask_wtf import FlaskForm from wtforms import StringField,TextAreaField,SubmitField, SelectField from wtforms.validators import Required class PitchForm(FlaskForm): title = StringField('Pitch title',validators=[Required()]) category = SelectField('Pitch category', choices=[('Motivational', 'Motivational...
[ "wtforms.validators.Required", "wtforms.SubmitField" ]
[((487, 508), 'wtforms.SubmitField', 'SubmitField', (['"""Submit"""'], {}), "('Submit')\n", (498, 508), False, 'from wtforms import StringField, TextAreaField, SubmitField, SelectField\n'), ((627, 648), 'wtforms.SubmitField', 'SubmitField', (['"""Submit"""'], {}), "('Submit')\n", (638, 648), False, 'from wtforms import...
#!/usr/bin/env python # Copyright 2016 Amazon.com, Inc. or its # affiliates. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"). You may not # use this file except in compliance with the License. A copy of the License is # located at # # http://aws.amazon.com/apache2.0/ # # or i...
[ "os.path.exists", "json.loads", "PIL.Image.open", "boto3.client", "urllib.unquote_plus", "os.makedirs", "PIL.ExifTags.TAGS.get", "boto3.resource", "os.remove" ]
[((936, 978), 'boto3.client', 'boto3.client', (['"""s3"""'], {'region_name': 'aws_region'}), "('s3', region_name=aws_region)\n", (948, 978), False, 'import boto3\n'), ((985, 1030), 'boto3.resource', 'boto3.resource', (['"""sqs"""'], {'region_name': 'aws_region'}), "('sqs', region_name=aws_region)\n", (999, 1030), False...
from dash_extensions.enrich import Dash from _app.layout import serve_layout from _app.callback import register_callbacks external_stylesheets = ["https://codepen.io/chriddyp/pen/bWLwgP.css"] app = Dash(prevent_initial_callbacks=True, external_stylesheets=external_stylesheets ) register_c...
[ "_app.callback.register_callbacks", "_app.layout.serve_layout", "dash_extensions.enrich.Dash" ]
[((203, 282), 'dash_extensions.enrich.Dash', 'Dash', ([], {'prevent_initial_callbacks': '(True)', 'external_stylesheets': 'external_stylesheets'}), '(prevent_initial_callbacks=True, external_stylesheets=external_stylesheets)\n', (207, 282), False, 'from dash_extensions.enrich import Dash\n'), ((310, 333), '_app.callbac...
import TestCase from testcase_generation.Response.default import defaultTestCase import Constants def defaultMediaTypeTestCase(): testcase = defaultTestCase() # Request testcase.request.method = 'GET' # Response testcase.response.status_code = 200 testcase.response.expect_body = True return testcase def testC...
[ "testcase_generation.Response.default.defaultTestCase" ]
[((143, 160), 'testcase_generation.Response.default.defaultTestCase', 'defaultTestCase', ([], {}), '()\n', (158, 160), False, 'from testcase_generation.Response.default import defaultTestCase\n')]
import numpy as np from typing import List, Tuple class InterfaceSolver(): """ Informal interface for solving class needed to interact with rubiks environment. """ def __init__(self, depth:int, possible_moves: List[str]) -> None: """ Will be passed depth, i.e. number of backwa...
[ "numpy.sum", "numpy.array" ]
[((8033, 8074), 'numpy.sum', 'np.sum', (['(cube_state[i, :, :] == center_val)'], {}), '(cube_state[i, :, :] == center_val)\n', (8039, 8074), True, 'import numpy as np\n'), ((8674, 8706), 'numpy.array', 'np.array', (['self.cube_state_values'], {}), '(self.cube_state_values)\n', (8682, 8706), True, 'import numpy as np\n'...
#!/usr/bin/env python3 from selenium import webdriver from selenium.webdriver.common.by import By class TestLuckyApp: """E2E integration tests class.""" def setup_method(self, method): options = webdriver.FirefoxOptions() self.driver = webdriver.Remote('http://firefoxdriver:4444/wd/hub', opt...
[ "selenium.webdriver.FirefoxOptions", "selenium.webdriver.Remote" ]
[((215, 241), 'selenium.webdriver.FirefoxOptions', 'webdriver.FirefoxOptions', ([], {}), '()\n', (239, 241), False, 'from selenium import webdriver\n'), ((264, 333), 'selenium.webdriver.Remote', 'webdriver.Remote', (['"""http://firefoxdriver:4444/wd/hub"""'], {'options': 'options'}), "('http://firefoxdriver:4444/wd/hub...
from uuid import UUID from datetime import datetime def uuid_from_string(string): return UUID('{s}'.format(s=string)) def format_timestamp(string): if isinstance(string, str): return datetime.strptime(string, '%Y-%m-%dT%H:%M:%S.%fZ') if isinstance(string, datetime): return string
[ "datetime.datetime.strptime" ]
[((203, 253), 'datetime.datetime.strptime', 'datetime.strptime', (['string', '"""%Y-%m-%dT%H:%M:%S.%fZ"""'], {}), "(string, '%Y-%m-%dT%H:%M:%S.%fZ')\n", (220, 253), False, 'from datetime import datetime\n')]
import time def factorial(n): fact = 1 for x in range(2, n+1): fact = fact * x return fact # Timing function start = time.time() factorial(400000) end = time.time() print('Operation done in {} seconds'.format(end - start))
[ "time.time" ]
[((139, 150), 'time.time', 'time.time', ([], {}), '()\n', (148, 150), False, 'import time\n'), ((175, 186), 'time.time', 'time.time', ([], {}), '()\n', (184, 186), False, 'import time\n')]
""" ========================================================================== MeshNetworkCL_test.py ========================================================================== Test for NetworkCL Author : <NAME> Date : May 19, 2019 """ import pytest from pymtl3_net.meshnet.MeshNetworkCL import MeshNetworkCL from pym...
[ "pymtl3_net.ocnlib.utils.run_sim", "pymtl3_net.meshnet.MeshNetworkCL.MeshNetworkCL", "pymtl3_net.ocnlib.ifcs.positions.mk_mesh_pos", "pymtl3_net.ocnlib.ifcs.packets.mk_mesh_pkt", "pytest.mark.parametrize", "pymtl3.stdlib.test_utils.test_srcs.TestSrcCL", "pymtl3_net.ocnlib.test.net_sinks.TestNetSinkCL", ...
[((4006, 4237), 'pymtl3.stdlib.test_utils.mk_test_case_table', 'mk_test_case_table', (["['msg_list wid ht src_init src_intv sink_init sink_intv', ['simple2x2',\n simple_2x2, 2, 2, 0, 0, 0, 0], ['simple4x4', simple_4x4, 4, 4, 0, 0, 0,\n 0], ['simple8x8', simple_8x8, 8, 8, 0, 0, 0, 0]]"], {}), "([\n 'msg_li...
import os # Helpers def read_asset(*paths): dirname = os.path.dirname(__file__) return open(os.path.join(dirname, "assets", *paths)).read().strip() # General VERSION = read_asset("VERSION")
[ "os.path.dirname", "os.path.join" ]
[((62, 87), 'os.path.dirname', 'os.path.dirname', (['__file__'], {}), '(__file__)\n', (77, 87), False, 'import os\n'), ((104, 143), 'os.path.join', 'os.path.join', (['dirname', '"""assets"""', '*paths'], {}), "(dirname, 'assets', *paths)\n", (116, 143), False, 'import os\n')]
import sys from glob import glob from serial import Serial, SerialException import numpy as np BAUD_RATE = 9600 PORT = 'COM5' READ_TIMEOUT = 1 LOWER_BOUND = 0.01 UPPER_BOUND = 0.4 class SerialCommunication(): """ Manages the communication and sends the data to the Arduino """ def __init__(self): s...
[ "numpy.clip", "sys.platform.startswith", "serial.Serial", "numpy.isnan", "glob.glob" ]
[((342, 350), 'serial.Serial', 'Serial', ([], {}), '()\n', (348, 350), False, 'from serial import Serial, SerialException\n'), ((2715, 2745), 'sys.platform.startswith', 'sys.platform.startswith', (['"""win"""'], {}), "('win')\n", (2738, 2745), False, 'import sys\n'), ((3799, 3841), 'numpy.clip', 'np.clip', (['signals',...
"""add unique constraint to Source Revision ID: 4415298e147b Revises: 7392493a<PASSWORD> Create Date: 2020-01-02 16:41:03.424945 """ from alembic import op import sqlalchemy as sa # revision identifiers, used by Alembic. revision = '<KEY>' down_revision = '7392493a0768' branch_labels = None depends_on = None def ...
[ "alembic.op.drop_constraint", "alembic.op.create_unique_constraint" ]
[((335, 458), 'alembic.op.create_unique_constraint', 'op.create_unique_constraint', (['"""ori_id_canonical_fields"""', '"""source"""', "['resource_ori_id', 'canonical_id', 'canonical_iri']"], {}), "('ori_id_canonical_fields', 'source', [\n 'resource_ori_id', 'canonical_id', 'canonical_iri'])\n", (362, 458), False, '...
"""Helper methods for parsing EBS-related data from AWS SDK.""" import logging import myutils logger = myutils.get_logger(__name__, logging.DEBUG) @myutils.log_calls(level=logging.DEBUG) def parse(sdk_snapshots): """Process raw EBS snapshot data.""" snapshots = [] snapshots.extend( map(map_snaps...
[ "myutils.get_first", "myutils.log_calls", "myutils.get_logger" ]
[((105, 148), 'myutils.get_logger', 'myutils.get_logger', (['__name__', 'logging.DEBUG'], {}), '(__name__, logging.DEBUG)\n', (123, 148), False, 'import myutils\n'), ((152, 190), 'myutils.log_calls', 'myutils.log_calls', ([], {'level': 'logging.DEBUG'}), '(level=logging.DEBUG)\n', (169, 190), False, 'import myutils\n')...
import assets import webbrowser from PyQt5.Qt import QMessageBox from PyQt5.QtNetwork import QNetworkDiskCache from PyQt5.QtWebKitWidgets import QWebPage, QWebInspector class WebPage(QWebPage): def __init__(self): super(WebPage, self).__init__() self.inspector = QWebInspector() self.inspector.setPage(self) se...
[ "assets.dataJar.save", "assets.fs.dataPath", "PyQt5.QtWebKitWidgets.QWebPage.acceptNavigationRequest", "PyQt5.QtWebKitWidgets.QWebInspector", "PyQt5.QtNetwork.QNetworkDiskCache" ]
[((269, 284), 'PyQt5.QtWebKitWidgets.QWebInspector', 'QWebInspector', ([], {}), '()\n', (282, 284), False, 'from PyQt5.QtWebKitWidgets import QWebPage, QWebInspector\n'), ((365, 388), 'PyQt5.QtNetwork.QNetworkDiskCache', 'QNetworkDiskCache', (['self'], {}), '(self)\n', (382, 388), False, 'from PyQt5.QtNetwork import QN...
import azure.functions as func from .add_url_rule import app def main(req: func.HttpRequest, context: func.Context) -> func.HttpResponse: return func.WsgiMiddleware(app.wsgi_app).handle(req, context)
[ "azure.functions.WsgiMiddleware" ]
[((151, 184), 'azure.functions.WsgiMiddleware', 'func.WsgiMiddleware', (['app.wsgi_app'], {}), '(app.wsgi_app)\n', (170, 184), True, 'import azure.functions as func\n')]
from flask import Flask from app.services import model_manager from app.controllers.api_blueprint import api_router from app.controllers.main_blueprint import main_router def create_app(test_config=None): app = Flask(__name__) model_manager.init_app(app) app.register_blueprint(api_router) app.registe...
[ "app.services.model_manager.init_app", "flask.Flask" ]
[((216, 231), 'flask.Flask', 'Flask', (['__name__'], {}), '(__name__)\n', (221, 231), False, 'from flask import Flask\n'), ((237, 264), 'app.services.model_manager.init_app', 'model_manager.init_app', (['app'], {}), '(app)\n', (259, 264), False, 'from app.services import model_manager\n')]
import sys import time import threading import grpc import numpy import soundfile as sf import tensorflow as tf import _init_paths import audioset.vggish_input as vggish_input from tensorflow_serving.apis import predict_pb2 from tensorflow_serving.apis import prediction_service_pb2_grpc tf.app.flags.DEFINE_integer...
[ "sys.stdout.flush", "tensorflow.app.flags.DEFINE_integer", "tensorflow_serving.apis.predict_pb2.PredictRequest", "tensorflow_serving.apis.prediction_service_pb2_grpc.PredictionServiceStub", "grpc.insecure_channel", "tensorflow.app.flags.DEFINE_string", "sys.stdout.write", "numpy.pad", "audioset.vggi...
[((293, 381), 'tensorflow.app.flags.DEFINE_integer', 'tf.app.flags.DEFINE_integer', (['"""concurrency"""', '(1)', '"""concurrent inference requests limit"""'], {}), "('concurrency', 1,\n 'concurrent inference requests limit')\n", (320, 381), True, 'import tensorflow as tf\n'), ((378, 448), 'tensorflow.app.flags.DEFI...
import sqlite3 class Stats: con = sqlite3.connect("data.db") cur = con.cursor() stats_insert_com = "INSERT INTO Stats (Player, 'Win Count', 'Play Count', winPlayRatio) VALUES (?, ?, ?, ?);" matchhist_insert_com = "INSERT INTO MatchHistory (matchNum, playerList, 'Winner', numRounds) VALUES (?, ?, ?, ?);...
[ "sqlite3.connect" ]
[((39, 65), 'sqlite3.connect', 'sqlite3.connect', (['"""data.db"""'], {}), "('data.db')\n", (54, 65), False, 'import sqlite3\n')]
# -*- coding: utf-8 -*- """ Created on Wed Jun 24 21:46:56 2020 @author: adwait """ import numpy as np import cv2 import pims from tkinter import messagebox, Tk from PIL import ImageFont, ImageDraw, Image from PyQt5.QtGui import QIcon import logging class MainRecordFunctions: def recordVideo(...
[ "PyQt5.QtGui.QIcon", "logging.debug", "cv2.imshow", "pims.Video", "PIL.ImageDraw.Draw", "numpy.array", "cv2.resizeWindow", "cv2.line", "PIL.ImageFont.truetype", "cv2.VideoWriter", "numpy.empty", "cv2.VideoWriter_fourcc", "tkinter.messagebox.showinfo", "cv2.putText", "cv2.cvtColor", "cv...
[((352, 380), 'logging.debug', 'logging.debug', (['"""recordvideo"""'], {}), "('recordvideo')\n", (365, 380), False, 'import logging\n'), ((9583, 9612), 'logging.debug', 'logging.debug', (['"""record_frame"""'], {}), "('record_frame')\n", (9596, 9612), False, 'import logging\n'), ((11078, 11122), 'numpy.zeros', 'np.zer...
""" A simple class to allow quick testing of GPIB programs without instruments. All reads from an instrument return a semi-randomised number regardless of the specific command that may have been sent prior to reading. """ import stuff import time ## VisaIOError = False """ Old pyvisa calls. """ def ge...
[ "stuff.DataGen", "time.sleep" ]
[((596, 611), 'stuff.DataGen', 'stuff.DataGen', ([], {}), '()\n', (609, 611), False, 'import stuff\n'), ((666, 681), 'time.sleep', 'time.sleep', (['(0.1)'], {}), '(0.1)\n', (676, 681), False, 'import time\n'), ((749, 764), 'time.sleep', 'time.sleep', (['(0.2)'], {}), '(0.2)\n', (759, 764), False, 'import time\n'), ((15...
import argparse from core_extract_comments import * from core_utils import * def run(search, input_product_ids_filename): product_ids = list() if input_product_ids_filename is not None: with open(input_product_ids_filename, 'r') as r: for p in r.readlines(): pro_obj = p.st...
[ "argparse.ArgumentParser" ]
[((1171, 1196), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {}), '()\n', (1194, 1196), False, 'import argparse\n')]
''' Message Example b'620901.02908, 3, 0.242, 0.606, 9.527, 4, 0.020, -0.006, 0.001, 5, -24.762,-223.451,-98.491, 81, 173.805, -3.304, 1.315' ''' import time import string import socket, traceback host="192.168.0.23" # ip address of port=5555 s=socket.socket(socket.AF_INET, socket.SOCK_DGRAM) s.setsockopt(so...
[ "socket.socket" ]
[((256, 304), 'socket.socket', 'socket.socket', (['socket.AF_INET', 'socket.SOCK_DGRAM'], {}), '(socket.AF_INET, socket.SOCK_DGRAM)\n', (269, 304), False, 'import socket, traceback\n')]
import os from .local_config import * PROJECT_NAME = "Useful" SERVER_HOST = 'http://127.0.0.1:8000' # Secret key SECRET_KEY = b"<KEY>" BASE_DIR = os.path.dirname(os.path.dirname(os.path.abspath(__file__))) API_V1_STR = "/api/v1" # Token 60 minutes * 24 hours * 7 days = 7 days ACCESS_TOKEN_EXPIRE_MINUTES = 60 * 24...
[ "os.path.abspath", "os.environ.get" ]
[((845, 871), 'os.environ.get', 'os.environ.get', (['"""SMTP_TLS"""'], {}), "('SMTP_TLS')\n", (859, 871), False, 'import os\n'), ((884, 911), 'os.environ.get', 'os.environ.get', (['"""SMTP_PORT"""'], {}), "('SMTP_PORT')\n", (898, 911), False, 'import os\n'), ((924, 951), 'os.environ.get', 'os.environ.get', (['"""SMTP_H...
#%% # read full assignment # think algo before implementing # dont use a dict when you need a list # assignment is still = and not == # dont use itertools when you can use np.roll # check mathemathical functions if the parentheses are ok # networkx is awesome # %% import os import re import numpy as np try: os.chdir(...
[ "os.getcwd" ]
[((364, 375), 'os.getcwd', 'os.getcwd', ([], {}), '()\n', (373, 375), False, 'import os\n'), ((333, 344), 'os.getcwd', 'os.getcwd', ([], {}), '()\n', (342, 344), False, 'import os\n')]
from django.contrib import admin from .models import Succession,Succession_Seasons,Succession_Casts,Succession_Season_Episodes # Register your models here. # admin.site.register(Succession) # The model Succession is abstract so it can't be registered with admin admin.site.register(Succession_Seasons) ad...
[ "django.contrib.admin.site.register" ]
[((278, 317), 'django.contrib.admin.site.register', 'admin.site.register', (['Succession_Seasons'], {}), '(Succession_Seasons)\n', (297, 317), False, 'from django.contrib import admin\n'), ((318, 355), 'django.contrib.admin.site.register', 'admin.site.register', (['Succession_Casts'], {}), '(Succession_Casts)\n', (337,...
#!/usr/bin/env python3 import sys, os import arsdkparser #=============================================================================== class Writer(object): def __init__(self, fileobj): self.fileobj = fileobj def write(self, fmt, *args): if args: self.fileobj.write(fmt % (args)...
[ "os.path.exists", "os.path.join", "os.mkdirs" ]
[((3917, 3961), 'os.path.join', 'os.path.join', (['outdir', "(CONSTS_FILENAME + '.h')"], {}), "(outdir, CONSTS_FILENAME + '.h')\n", (3929, 3961), False, 'import sys, os\n'), ((4096, 4140), 'os.path.join', 'os.path.join', (['outdir', "(CONSTS_FILENAME + '.m')"], {}), "(outdir, CONSTS_FILENAME + '.m')\n", (4108, 4140), F...
""" **Project Name:** MakeHuman **Product Home Page:** http://www.makehuman.org/ **Code Home Page:** http://code.google.com/p/makehuman/ **Authors:** <NAME> **Copyright(c):** MakeHuman Team 2001-2009 **Licensing:** GPL3 (see also http://sites.google.com/site/makehumandocs...
[ "mhx_rig.writeDrivers", "mhx_rig.addPoseBone" ]
[((2763, 2953), 'mhx_rig.addPoseBone', 'addPoseBone', (['fp', '"""Jaw"""', '"""MHJaw"""', 'None', '(1, 1, 1)', '(0, 1, 0)', '(1, 1, 1)', '(1, 1, 1)', '(0)', "[('LimitRot', C_OW_LOCAL, 1, ['LimitRot', (-5 * D, 45 * D, 0, 0, -20 * D, \n 20 * D), (1, 1, 1)])]"], {}), "(fp, 'Jaw', 'MHJaw', None, (1, 1, 1), (0, 1, 0), (1...
import cv2 import numpy as np from matplotlib import pyplot as plt img = cv2.imread('home.jpg') hsv = cv2.cvtColor(img,cv2.COLOR_BGR2HSV) hist = cv2.calcHist( [hsv], [0, 1], None, [180, 256], [0, 180, 0, 256] ) plt.imshow(hist,interpolation = 'nearest') plt.show() # in numpy import cv2 import numpy as np from matplo...
[ "matplotlib.pyplot.imshow", "cv2.calcHist", "cv2.cvtColor", "cv2.imread", "matplotlib.pyplot.show" ]
[((74, 96), 'cv2.imread', 'cv2.imread', (['"""home.jpg"""'], {}), "('home.jpg')\n", (84, 96), False, 'import cv2\n'), ((103, 139), 'cv2.cvtColor', 'cv2.cvtColor', (['img', 'cv2.COLOR_BGR2HSV'], {}), '(img, cv2.COLOR_BGR2HSV)\n', (115, 139), False, 'import cv2\n'), ((146, 209), 'cv2.calcHist', 'cv2.calcHist', (['[hsv]',...
import network2 import logging import numpy as np logging.basicConfig(level=logging.DEBUG) # read in the data # logging.info("READING IN DATA...") # for reading in normal dataset # training, validation, test = network2.load_data_wrapper("data/mnist.pkl.gz") ### I WILL ADD AND COMMENT OUT SECTIONS OF CODE BASED ...
[ "logging.basicConfig", "network2.load_data_wrapper", "logging.info", "network2.Network" ]
[((51, 91), 'logging.basicConfig', 'logging.basicConfig', ([], {'level': 'logging.DEBUG'}), '(level=logging.DEBUG)\n', (70, 91), False, 'import logging\n'), ((7511, 7545), 'logging.info', 'logging.info', (['"""READING IN DATA..."""'], {}), "('READING IN DATA...')\n", (7523, 7545), False, 'import logging\n'), ((7608, 76...
from abc import ABC, abstractmethod import collections import statistics import numpy as np import sklearn.metrics import torch class Evaluator(ABC): """Class to evaluate model outputs and report the result. """ def __init__(self): self.reset() @abstractmethod def add_predictions(self, p...
[ "statistics.mean", "torch.mul", "torch.topk", "numpy.argmax", "numpy.array", "numpy.sum", "torch.add", "collections.defaultdict", "torch.zeros_like" ]
[((904, 930), 'torch.topk', 'torch.topk', (['predictions', '(1)'], {}), '(predictions, 1)\n', (914, 930), False, 'import torch\n'), ((1160, 1186), 'torch.topk', 'torch.topk', (['predictions', 'k'], {}), '(predictions, k)\n', (1170, 1186), False, 'import torch\n'), ((1366, 1414), 'torch.zeros_like', 'torch.zeros_like', ...
import unittest from lmctl.cli.format import TableFormat, Table, Column class DummyTable(Table): columns = [ Column('name', header='Name'), Column('status', header='Status', accessor=lambda x: 'OK' if x.get('status', None) in ['Excellent', 'Good'] else 'Unhealthy'), ] class DummyTableNoHeaders...
[ "lmctl.cli.format.Column" ]
[((122, 151), 'lmctl.cli.format.Column', 'Column', (['"""name"""'], {'header': '"""Name"""'}), "('name', header='Name')\n", (128, 151), False, 'from lmctl.cli.format import TableFormat, Table, Column\n'), ((353, 367), 'lmctl.cli.format.Column', 'Column', (['"""name"""'], {}), "('name')\n", (359, 367), False, 'from lmct...
import logging import azure.functions as func import numpy as np import json import requests from os import path def readjson_from_file(filename): try: fp = open(filename, "r") except: logging.info(f"WARNING: cant open file {filename} ") return {} try: ob...
[ "json.load", "json.dumps", "logging.info" ]
[((617, 651), 'logging.info', 'logging.info', (['f"""post url is {url}"""'], {}), "(f'post url is {url}')\n", (629, 651), False, 'import logging\n'), ((711, 740), 'logging.info', 'logging.info', (['"""payload added"""'], {}), "('payload added')\n", (723, 740), False, 'import logging\n'), ((848, 884), 'logging.info', 'l...
import django print(django.get_version()) print(f"boa {111 * 6}") print(f"{6*6}") input = input("Hey abuser, enter some stuff!\n") cmp = 1 > 0 print(type(cmp)) print(input) try: int(asdf) except Exception: print("Oh no, it failed!") while False: print("False!") listeL = [1,2,3,45,6,4,3,6,8,6,3...
[ "django.get_version" ]
[((20, 40), 'django.get_version', 'django.get_version', ([], {}), '()\n', (38, 40), False, 'import django\n')]
from start import client from modules import codeforces,delete,notes,hastebin,pin,pm,user,spam,rextester white=[] from telethon import TelegramClient,events import logging logging.basicConfig(format='[%(levelname) 5s/%(asctime)s] %(name)s: %(message)s', level=logging.WARNING) client.start().run_u...
[ "logging.basicConfig", "start.client.start" ]
[((172, 286), 'logging.basicConfig', 'logging.basicConfig', ([], {'format': '"""[%(levelname) 5s/%(asctime)s] %(name)s: %(message)s"""', 'level': 'logging.WARNING'}), "(format=\n '[%(levelname) 5s/%(asctime)s] %(name)s: %(message)s', level=logging.\n WARNING)\n", (191, 286), False, 'import logging\n'), ((300, 314...
# Let’s make a map! Using Geopandas, Pandas and Matplotlib to make a Choropleth map # https://towardsdatascience.com/lets-make-a-map-using-geopandas-pandas-and-matplotlib-to-make-a-chloropleth-map-dddc31c1983d import pandas as pd import matplotlib.pyplot as plt import geopandas as gpd from shapely.geometry import...
[ "matplotlib.pyplot.text", "pycountry.countries.search_fuzzy", "geopandas.read_file", "pandas.read_csv", "matplotlib.pyplot.title", "matplotlib.pyplot.subplots" ]
[((413, 601), 'geopandas.read_file', 'gpd.read_file', (['"""/Users/vivekparashar/OneDrive/OneDrive-GitHub/Challenges-and-Competitions/30DayMapChallenge/Maps/natural_earth_vector/110m_cultural/ne_110m_admin_0_countries.shp"""'], {}), "(\n '/Users/vivekparashar/OneDrive/OneDrive-GitHub/Challenges-and-Competitions/30Da...
""" This module exrtacts features from the data, saves the feauters from all measurements to global results file and creates one file for every sensor with all measurements. :copyright: (c) 2022 by <NAME>, Hochschule-Bonn-Rhein-Sieg :license: see LICENSE for more details. """ from pyexpat import features import pan...
[ "numpy.trapz", "numpy.ones", "pandas.read_csv", "matplotlib.pyplot.xticks", "pathlib.Path", "numpy.max", "matplotlib.pyplot.close", "scipy.signal.peak_widths", "scipy.signal.find_peaks", "pandas.DataFrame", "matplotlib.pyplot.subplots", "numpy.arange" ]
[((9673, 9752), 'scipy.signal.find_peaks', 'find_peaks', (['df[sensor]'], {'prominence': '(0)', 'width': '(1)', 'distance': '(20000)', 'height': 'threshold'}), '(df[sensor], prominence=0, width=1, distance=20000, height=threshold)\n', (9683, 9752), False, 'from scipy.signal import chirp, find_peaks, peak_widths\n'), ((...
from __future__ import annotations import json import os import shutil import subprocess import tempfile import uuid from abc import ABC, abstractmethod from typing import Any, Union from urllib.error import HTTPError from urllib.request import urlopen, urlretrieve import warnings import meerkat as mk import pandas a...
[ "pandas.read_csv", "os.remove", "os.path.exists", "meerkat.tools.lazy_loader.LazyLoader", "meerkat.DataPanel.read", "yaml.add_multi_representer", "urllib.request.urlretrieve", "yaml.add_constructor", "os.path.isdir", "subprocess.call", "warnings.warn", "urllib.request.urlopen", "json.loads",...
[((474, 508), 'meerkat.tools.lazy_loader.LazyLoader', 'LazyLoader', (['"""google.cloud.storage"""'], {}), "('google.cloud.storage')\n", (484, 508), False, 'from meerkat.tools.lazy_loader import LazyLoader\n'), ((517, 536), 'meerkat.tools.lazy_loader.LazyLoader', 'LazyLoader', (['"""torch"""'], {}), "('torch')\n", (527,...
import asyncio import json import zlib import aiohttp import errors API_BASE = 'https://discordapp.com/api/v6' CONFIG_FILE = json.load(open('data/config.json')) TOKEN = CONFIG_FILE['token'] HEADERS = {'Authorization': 'Bot ' + TOKEN, 'User-Agent': 'DiscordBot (https://www.github.com/fourjr/dapi-bot,\ ...
[ "json.loads", "errors.NotFound", "errors.Forbidden", "asyncio.get_event_loop", "zlib.decompress", "errors.RateLimit" ]
[((389, 413), 'asyncio.get_event_loop', 'asyncio.get_event_loop', ([], {}), '()\n', (411, 413), False, 'import asyncio\n'), ((670, 686), 'json.loads', 'json.loads', (['data'], {}), '(data)\n', (680, 686), False, 'import json\n'), ((1384, 1411), 'errors.Forbidden', 'errors.Forbidden', (['resp', 'obj'], {}), '(resp, obj)...
import sys import requests import argparse import logging import json import datetime import anticrlf from veracode_api_py import VeracodeAPI as vapi log = logging.getLogger(__name__) def setup_logger(): handler = logging.FileHandler('vcoffboard.log', encoding='utf8') handler.setFormatter(anticrlf.LogFormatt...
[ "logging.getLogger", "anticrlf.LogFormatter", "argparse.ArgumentParser", "datetime.datetime.strptime", "veracode_api_py.VeracodeAPI", "datetime.datetime.now", "logging.FileHandler" ]
[((158, 185), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (175, 185), False, 'import logging\n'), ((221, 275), 'logging.FileHandler', 'logging.FileHandler', (['"""vcoffboard.log"""'], {'encoding': '"""utf8"""'}), "('vcoffboard.log', encoding='utf8')\n", (240, 275), False, 'import loggi...
from expects import expect, contain, be_an class Bacon: ... sanduiche = 'sanduiche com queijo' expect(sanduiche).to(contain('queijo')) expect(sanduiche).to_not(be_an(Bacon))
[ "expects.be_an", "expects.contain", "expects.expect" ]
[((127, 144), 'expects.contain', 'contain', (['"""queijo"""'], {}), "('queijo')\n", (134, 144), False, 'from expects import expect, contain, be_an\n'), ((172, 184), 'expects.be_an', 'be_an', (['Bacon'], {}), '(Bacon)\n', (177, 184), False, 'from expects import expect, contain, be_an\n'), ((106, 123), 'expects.expect', ...
""" Main ============= Example """ # Import import numpy as np import pandas as pd # Specific from tpot import TPOTClassifier # Import own from pySML2.preprocessing.splitters import cvs_hos_split from pySML2.preprocessing.splitters import kfolds_split # --------------------------------------------- # Configuratio...
[ "tpot.TPOTClassifier", "pySML2.preprocessing.splitters.cvs_hos_split", "pandas.read_csv" ]
[((917, 938), 'pandas.read_csv', 'pd.read_csv', (['filepath'], {}), '(filepath)\n', (928, 938), True, 'import pandas as pd\n'), ((1211, 1249), 'pySML2.preprocessing.splitters.cvs_hos_split', 'cvs_hos_split', (['data'], {'selected_rows': 'cmp'}), '(data, selected_rows=cmp)\n', (1224, 1249), False, 'from pySML2.preproces...
#!/usr/bin/env python3.6 """Sherlock: Find Usernames Across Social Networks Module This module contains the main logic to search for usernames at social networks. """ import requests import csv import json import os import re from argparse import ArgumentParser, RawDescriptionHelpFormatter import platform module_nam...
[ "csv.DictWriter", "re.search", "argparse.ArgumentParser", "requests.get", "os.path.isfile", "json.load", "platform.python_version", "os.remove" ]
[((3910, 3934), 'os.path.isfile', 'os.path.isfile', (['filename'], {}), '(filename)\n', (3924, 3934), False, 'import os\n'), ((5626, 5647), 'os.path.isfile', 'os.path.isfile', (['fname'], {}), '(fname)\n', (5640, 5647), False, 'import os\n'), ((6915, 7033), 'argparse.ArgumentParser', 'ArgumentParser', ([], {'formatter_...
# 发现疑似实体,辅助训练 # 用ac自动机构建发现疑似实体的工具 import os from collections import defaultdict import json import re from .acmation import KeywordTree, add_to_ac, entity_files_folder, entity_folder from curLine_file import curLine, normal_transformer domain2entity_map = {} domain2entity_map["music"] = ["age", "singer", "song", "top...
[ "os.path.exists", "curLine_file.curLine", "re.compile", "os.path.join", "collections.defaultdict", "json.load" ]
[((549, 580), 're.compile', 're.compile', (['"""[0-9一二三四五六七八九十拾]+"""'], {}), "('[0-9一二三四五六七八九十拾]+')\n", (559, 580), False, 'import re\n'), ((3134, 3151), 'collections.defaultdict', 'defaultdict', (['list'], {}), '(list)\n', (3145, 3151), False, 'from collections import defaultdict\n'), ((746, 755), 'curLine_file.curLin...
#!python/bin/python3 """ Copyright (c) 2018 NSF Center for Space, High-performance, and Resilient Computing (SHREC) University of Pittsburgh. All rights reserved. Redistribution and use in source and binary forms, with or without modification, are permitted provided that the following conditions are met: 1. Redistribu...
[ "os.path.abspath", "collections.defaultdict", "importlib.import_module", "copy.deepcopy" ]
[((1658, 1686), 'importlib.import_module', 'import_module', (['"""src.targets"""'], {}), "('src.targets')\n", (1671, 1686), False, 'from importlib import import_module\n'), ((1784, 1801), 'collections.defaultdict', 'defaultdict', (['tree'], {}), '(tree)\n', (1795, 1801), False, 'from collections import defaultdict\n'),...
import logging import pickle import seaborn as sb import torch sb.set_style('whitegrid') # Class for efficiently handling configurations and parameters, enables to # easily set them and remember them when one config is reused # Read with config.key, set with config.update({'key': value}) or config[ # 'key'] = value ...
[ "logging.warning", "seaborn.set_style", "torch.cuda.is_available", "logging.info", "torch.cuda.current_device", "torch.linspace" ]
[((65, 90), 'seaborn.set_style', 'sb.set_style', (['"""whitegrid"""'], {}), "('whitegrid')\n", (77, 90), True, 'import seaborn as sb\n'), ((1620, 1669), 'torch.linspace', 'torch.linspace', (['self.t0', 'self.tf', 'self.nb_samples'], {}), '(self.t0, self.tf, self.nb_samples)\n', (1634, 1669), False, 'import torch\n'), (...