code stringlengths 22 1.05M | apis listlengths 1 3.31k | extract_api stringlengths 75 3.25M |
|---|---|---|
#!/usr/bin/env python3
import argparse
import hashlib
import json
import re
import signal
import subprocess
import sys
import time
import urllib.request
DEFAULT_SERVER_HOSTNAME = "127.0.0.1"
DEFAULT_SERVER_PORT = 6878
SERVER_POLL_TIME = 2
SERVER_STATUS_STREAM_ACTIVE = "dl"
def exit_error(message):
print(f"Erro... | [
"signal.signal",
"argparse.ArgumentParser",
"time.sleep",
"sys.exit",
"json.load",
"re.search"
] | [((356, 367), 'sys.exit', 'sys.exit', (['(1)'], {}), '(1)\n', (364, 367), False, 'import sys\n'), ((819, 988), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {'description': '"""Instructs server to commence a given program ID. Will optionally execute a local media player once playback has started."""'}), "(... |
# -*- coding: utf-8 -*-
"""
Created on Wed Jun 13 13:10:41 2018
@author: crius
"""
import Hamiltonians as H
import numpy as np
import tools as t
import spintensor as st
import spinops as so
import time
import Expand as ex
from matplotlib import pyplot as plt
exp = np.exp
N = 8
nlegs = 4
S = 0.5
c = np.sqrt(2)
Jcu... | [
"tools.Statelist",
"numpy.sqrt",
"Expand.append",
"spinops.SziOp",
"matplotlib.pyplot.plot",
"numpy.asarray",
"numpy.kron",
"numpy.linspace",
"Expand.Expand",
"tools.exval",
"Hamiltonians.nlegHeisenberg.blockH",
"numpy.linalg.eigh",
"numpy.cos",
"time.time",
"spinops.sz"
] | [((306, 316), 'numpy.sqrt', 'np.sqrt', (['(2)'], {}), '(2)\n', (313, 316), True, 'import numpy as np\n'), ((343, 429), 'Hamiltonians.nlegHeisenberg.blockH', 'H.nlegHeisenberg.blockH', (['N', 'S', 'nlegs', 'c'], {'Js': '[1, 1]', 'gamma': '[0.0, 4.0]', 'full': '"""True"""'}), "(N, S, nlegs, c, Js=[1, 1], gamma=[0.0, 4.0]... |
import pandas as pd
from pandas import Timestamp
import hillmaker
file_stopdata = '../data/unit_stop_log_Experiment1_Scenario138_Rep17.csv'
scenario_name = 'log_unitocc_test_5'
in_fld_name = 'EnteredTS'
out_fld_name = 'ExitedTS'
cat_fld_name = 'Unit'
start_analysis = '12/12/2015 00:00'
end_analysis = '12/19/2021 00:... | [
"pandas.Timestamp",
"hillmaker.make_hills",
"pandas.DateOffset",
"pandas.read_csv"
] | [((412, 438), 'pandas.read_csv', 'pd.read_csv', (['file_stopdata'], {}), '(file_stopdata)\n', (423, 438), True, 'import pandas as pd\n'), ((450, 480), 'pandas.Timestamp', 'Timestamp', (['"""20150215 00:00:00"""'], {}), "('20150215 00:00:00')\n", (459, 480), False, 'from pandas import Timestamp\n'), ((891, 1153), 'hillm... |
import pandas as pd
import numpy as np
def augment(data):
data['MalePercent'] = (data['MaleLowerQuartile'] +
data['MaleLowerMiddleQuartile'] + data['MaleUpperMiddleQuartile'] +
data['MaleTopQuartile']) * .25
data['FemalePercent'] = (data['FemaleLowerQuarti... | [
"pandas.read_csv"
] | [((2170, 2222), 'pandas.read_csv', 'pd.read_csv', (['"""data/ukgov-gpg-full-section-split.csv"""'], {}), "('data/ukgov-gpg-full-section-split.csv')\n", (2181, 2222), True, 'import pandas as pd\n')] |
#!/usr/bin/env python
import os
from typing import List, Union
_PROJECT_DIRECTORY = os.path.realpath(os.path.curdir)
def _remove_files(files: Union[List[str], str]) -> None:
"""
Removes the list of files provided.
:param files: list of filepath to remove
:type files: Union[List[str], str]
"""
... | [
"os.path.realpath",
"os.path.join"
] | [((86, 118), 'os.path.realpath', 'os.path.realpath', (['os.path.curdir'], {}), '(os.path.curdir)\n', (102, 118), False, 'import os\n'), ((424, 466), 'os.path.join', 'os.path.join', (['_PROJECT_DIRECTORY', 'filepath'], {}), '(_PROJECT_DIRECTORY, filepath)\n', (436, 466), False, 'import os\n')] |
import json
import channels.layers
from asgiref.sync import async_to_sync
from django.conf import settings
from django.core.paginator import Paginator
from django.http import HttpResponseRedirect
from django.shortcuts import render, get_object_or_404
from django.urls import reverse, reverse_lazy
from django.views.gene... | [
"django.shortcuts.render",
"django.urls.reverse"
] | [((526, 579), 'django.shortcuts.render', 'render', (['request', '"""user/profile.html"""'], {'context': 'context'}), "(request, 'user/profile.html', context=context)\n", (532, 579), False, 'from django.shortcuts import render, get_object_or_404\n'), ((1024, 1042), 'django.urls.reverse', 'reverse', (['"""profile"""'], {... |
# /usr/bin/python3
import logging
import sched
import signal
import sys
import time
from functools import partial
import watchdog.events
import watchdog.observers
from watchdog.observers.polling import PollingObserver as PollingObserver
from GlobusTransfer import GlobusTransfer
from .args import Args
from .handler ... | [
"sched.scheduler",
"GlobusTransfer.GlobusTransfer",
"functools.partial",
"watchdog.observers.polling.PollingObserver"
] | [((720, 797), 'GlobusTransfer.GlobusTransfer', 'GlobusTransfer', (['args.source', 'args.destination', 'args.destination_dir', 'src_path'], {}), '(args.source, args.destination, args.destination_dir, src_path)\n', (734, 797), False, 'from GlobusTransfer import GlobusTransfer\n'), ((958, 975), 'watchdog.observers.polling... |
import argparse
import os
import sys
import torch
from torch import nn, optim
from torch.optim import optimizer
from torchvision import datasets, models, transforms
parser = argparse.ArgumentParser(description="Trains a neural network")
parser.add_argument('data_dir', metavar='dir', type=str,
help... | [
"torch.nn.ReLU",
"torch.nn.Dropout",
"torchvision.models.vgg19",
"torchvision.models.densenet161",
"torch.optim.optimizer.step",
"torchvision.models.resnet18",
"torch.exp",
"torch.cuda.is_available",
"torchvision.models.densenet121",
"sys.exit",
"torchvision.models.vgg11",
"argparse.ArgumentPa... | [((176, 238), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {'description': '"""Trains a neural network"""'}), "(description='Trains a neural network')\n", (199, 238), False, 'import argparse\n'), ((6130, 6156), 'os.path.join', 'os.path.join', (['dir', '"""train"""'], {}), "(dir, 'train')\n", (6142, 6156),... |
# Ejercicio 04
import time
import matplotlib.pyplot as plt
# Algoritmo
def binarySearch(numbers, n):
numElements = len(numbers)
if(numElements == 1):
return (n == numbers[0])
mitad = numElements // 2
if(numbers[mitad] == n):
return True
elif (n < numbers[mitad]):
return... | [
"matplotlib.pyplot.subplots",
"time.perf_counter",
"matplotlib.pyplot.show"
] | [((515, 534), 'time.perf_counter', 'time.perf_counter', ([], {}), '()\n', (532, 534), False, 'import time\n'), ((574, 593), 'time.perf_counter', 'time.perf_counter', ([], {}), '()\n', (591, 593), False, 'import time\n'), ((1146, 1160), 'matplotlib.pyplot.subplots', 'plt.subplots', ([], {}), '()\n', (1158, 1160), True, ... |
# Generated by Django 2.2.4 on 2020-06-29 10:53
from django.conf import settings
from django.db import migrations, models
import django.db.models.deletion
import jsonfield.fields
class Migration(migrations.Migration):
initial = True
dependencies = [
migrations.swappable_dependency(settings.AUTH_USE... | [
"django.db.models.ForeignKey",
"django.db.models.BooleanField",
"django.db.models.AutoField",
"django.db.models.BigIntegerField",
"django.db.models.DateTimeField",
"django.db.migrations.swappable_dependency",
"django.db.models.CharField"
] | [((271, 328), 'django.db.migrations.swappable_dependency', 'migrations.swappable_dependency', (['settings.AUTH_USER_MODEL'], {}), '(settings.AUTH_USER_MODEL)\n', (302, 328), False, 'from django.db import migrations, models\n'), ((462, 555), 'django.db.models.AutoField', 'models.AutoField', ([], {'auto_created': '(True)... |
from os import getenv
from dataclasses import dataclass
from collections import ChainMap
from dotenv import load_dotenv
from bot.structs import CaseInsensitiveDict
load_dotenv()
MY_NAME = "LPUBeltbot"
MY_REDDIT_PW = getenv("BELTBOT_REDDIT_PW")
MY_REDDIT_CLIENT_ID = getenv("BELTBOT_REDDIT_CID")
MY_REDDIT_SECRET... | [
"bot.structs.CaseInsensitiveDict",
"collections.ChainMap",
"os.getenv",
"dotenv.load_dotenv"
] | [((168, 181), 'dotenv.load_dotenv', 'load_dotenv', ([], {}), '()\n', (179, 181), False, 'from dotenv import load_dotenv\n'), ((223, 250), 'os.getenv', 'getenv', (['"""BELTBOT_REDDIT_PW"""'], {}), "('BELTBOT_REDDIT_PW')\n", (229, 250), False, 'from os import getenv\n'), ((274, 302), 'os.getenv', 'getenv', (['"""BELTBOT_... |
import mnist
import numpy as np
from PIL import Image
from conv import Conv3x3
from maxpool import MaxPool2
from softmax import Softmax
train_images = mnist.train_images()[:100]
train_labels = mnist.train_labels()[:100]
test_images = mnist.test_images()[:1000]
test_labels = mnist.test_labels()[:1000]
conv = Conv3x3(8... | [
"softmax.Softmax",
"mnist.test_labels",
"mnist.train_images",
"numpy.log",
"numpy.argmax",
"mnist.test_images",
"numpy.array",
"numpy.zeros",
"conv.Conv3x3",
"maxpool.MaxPool2",
"mnist.train_labels"
] | [((311, 321), 'conv.Conv3x3', 'Conv3x3', (['(8)'], {}), '(8)\n', (318, 321), False, 'from conv import Conv3x3\n'), ((329, 339), 'maxpool.MaxPool2', 'MaxPool2', ([], {}), '()\n', (337, 339), False, 'from maxpool import MaxPool2\n'), ((350, 374), 'softmax.Softmax', 'Softmax', (['(13 * 13 * 8)', '(10)'], {}), '(13 * 13 * ... |
# -*- coding: utf-8 -*-
"""
:author: T8840
:tag: Thinking is a good thing!
纸上得来终觉浅,绝知此事要躬行!
:description:
1.部署相关信息来自于nacos配置
{
"server_info": {
"host": "10.201.5.161",
"port":22,
"user" : "user",
... | [
"paramiko.SFTPClient.from_transport",
"json.loads",
"pathlib.Path",
"pathlib.Path.cwd",
"paramiko.AutoAddPolicy",
"arrow.now",
"paramiko.Transport",
"requests.get",
"platform.system",
"paramiko.SSHClient"
] | [((1362, 1514), 'requests.get', 'requests.get', (['"""http://10.201.7.185:8848/nacos/v1/cs/configs?dataId=deploy_124&group=DEFAULT_GROUP&tenant=89362432-5255-497e-8e94-cb77d46cb1a9"""'], {}), "(\n 'http://10.201.7.185:8848/nacos/v1/cs/configs?dataId=deploy_124&group=DEFAULT_GROUP&tenant=89362432-5255-497e-8e94-cb77d... |
# -*- coding: utf-8 -*-
"""
Functions for plotting reliability diagrams: smooths of simulated vs observed
outcomes on the y-axis against predicted probabilities on the x-axis.
"""
from __future__ import absolute_import
import matplotlib.pyplot as plt
import numpy as np
import seaborn as sbn
from .plot_utils import _l... | [
"seaborn.set_style",
"numpy.linspace",
"matplotlib.pyplot.subplots",
"past.builtins.range"
] | [((681, 706), 'seaborn.set_style', 'sbn.set_style', (['"""darkgrid"""'], {}), "('darkgrid')\n", (694, 706), True, 'import seaborn as sbn\n'), ((2708, 2754), 'numpy.linspace', 'np.linspace', (['min_ref_val', 'max_ref_val'], {'num': '(100)'}), '(min_ref_val, max_ref_val, num=100)\n', (2719, 2754), True, 'import numpy as ... |
"""
An AWS Lambda function used to run periodic background jobs on ECS.
The complication with running these tasks is that we need to run them on
the same version of the Docker image that the web servers are currently
running on.
"""
import logging
import boto3
from utils import env_list, env_param
# Logging setup... | [
"logging.getLogger",
"utils.env_param",
"boto3.client",
"utils.env_list"
] | [((348, 375), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (365, 375), False, 'import logging\n'), ((461, 480), 'boto3.client', 'boto3.client', (['"""ecs"""'], {}), "('ecs')\n", (473, 480), False, 'import boto3\n'), ((1198, 1218), 'utils.env_param', 'env_param', (['"""CLUSTER"""'], {}),... |
#! python3
# -*- coding: utf-8 -*-
import os
from pypeapp import execute, Logger
from pype.hosts.resolve.utils import get_resolve_module
log = Logger().get_logger("Resolve")
CURRENT_DIR = os.getenv("RESOLVE_UTILITY_SCRIPTS_DIR", "")
python_dir = os.getenv("PYTHON36_RESOLVE")
python_exe = os.path.normpath(
os.path... | [
"os.getenv",
"pype.hosts.resolve.utils.get_resolve_module",
"os.path.join",
"os.path.normpath",
"pypeapp.Logger"
] | [((190, 234), 'os.getenv', 'os.getenv', (['"""RESOLVE_UTILITY_SCRIPTS_DIR"""', '""""""'], {}), "('RESOLVE_UTILITY_SCRIPTS_DIR', '')\n", (199, 234), False, 'import os\n'), ((248, 277), 'os.getenv', 'os.getenv', (['"""PYTHON36_RESOLVE"""'], {}), "('PYTHON36_RESOLVE')\n", (257, 277), False, 'import os\n'), ((365, 385), 'p... |
from collections import OrderedDict
import torch
import torch.nn as nn
import torch.nn.functional as F
import logging
from libs.modules.FuseBlock import MakeFB
from .resnet_dilation import resnet50, resnet101, Bottleneck, conv1x1
BN_MOMENTUM = 0.1
logger = logging.getLogger(__name__)
def conv3x3(in_planes, out_plan... | [
"logging.getLogger",
"torch.nn.BatchNorm2d",
"torch.nn.ReLU",
"torch.nn.Sequential",
"torch.nn.Dropout2d",
"torch.nn.functional.sigmoid",
"torch.nn.Conv2d",
"torch.nn.Module",
"torch.nn.functional.interpolate",
"torch.nn.AdaptiveAvgPool2d",
"libs.modules.FuseBlock.MakeFB",
"torch.cat"
] | [((259, 286), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (276, 286), False, 'import logging\n'), ((385, 474), 'torch.nn.Conv2d', 'nn.Conv2d', (['in_planes', 'out_planes'], {'kernel_size': '(3)', 'stride': 'stride', 'padding': '(1)', 'bias': '(False)'}), '(in_planes, out_planes, kernel... |
# -*- coding: utf-8 -*-
"""
@author : <NAME>
@github : https://github.com/tianpangji
@software : PyCharm
@file : crud.py
@create : 2020/12/9 20:44
"""
from django.contrib.contenttypes.models import ContentType
from easyaudit.models import CRUDEvent
from rest_framework import serializers
class CRUDSerializer... | [
"rest_framework.serializers.DateTimeField",
"rest_framework.serializers.SerializerMethodField",
"django.contrib.contenttypes.models.ContentType.objects.get"
] | [((376, 411), 'rest_framework.serializers.SerializerMethodField', 'serializers.SerializerMethodField', ([], {}), '()\n', (409, 411), False, 'from rest_framework import serializers\n'), ((427, 496), 'rest_framework.serializers.DateTimeField', 'serializers.DateTimeField', ([], {'format': '"""%Y-%m-%d %H:%M:%S"""', 'read_... |
"""
ICMP(Internet Control Message Protocol) - Echo Request:
+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+
|0|0|0|0|0|0|0|0|0|0|1|1|1|1|1|1|1|1|1|1|2|2|2|2|2|2|2|2|2|2|3|3|
|0|1|2|3|4|5|6|7|8|9|0|1|2|3|4|5|6|7|8|9|0|1|2|3|4|5|6|7|8|9|0|1|
+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+... | [
"socket.gethostbyname",
"socket.socket",
"time.sleep",
"struct.pack",
"time.time"
] | [((1971, 2067), 'struct.pack', 'struct.pack', (['"""!BBHHHLL48s"""', 'Type', 'Code', '(0)', 'Identifier', 'SequenceNumber', 'TimeStamp', '(0)', 'Payload'], {}), "('!BBHHHLL48s', Type, Code, 0, Identifier, SequenceNumber,\n TimeStamp, 0, Payload)\n", (1982, 2067), False, 'import struct\n'), ((2257, 2360), 'struct.pac... |
import random
from model.contact import Contact
from model.group import Group
def test_to_delete_contact_from_group(app, db ,check_ui):
app.group.open_groups_page()
if len(db.get_group_list()) == 0:
app.group.create()
app.group.fill_group_form(Group(name='group for adding of contact'))
... | [
"model.group.Group",
"random.choice",
"model.contact.Contact"
] | [((443, 468), 'random.choice', 'random.choice', (['group_list'], {}), '(group_list)\n', (456, 468), False, 'import random\n'), ((1540, 1572), 'random.choice', 'random.choice', (['contacts_in_group'], {}), '(contacts_in_group)\n', (1553, 1572), False, 'import random\n'), ((1292, 1325), 'random.choice', 'random.choice', ... |
import gzip
import logging
import logging.handlers
import os
from cStringIO import StringIO as IO
from spreads.vendor.huey import SqliteHuey
from spreads.vendor.huey.consumer import Consumer
from spreads.vendor.pathlib import Path
from flask import Flask, request
from spreads.plugin import (HookPlugin, SubcommandHook... | [
"logging.getLogger",
"os.path.exists",
"spreads.util.add_argument_from_option",
"cStringIO.StringIO",
"flask.Flask",
"logging.handlers.BufferingHandler",
"spreads.vendor.huey.SqliteHuey",
"spreads.vendor.huey.consumer.Consumer",
"netifaces.ifaddresses",
"gzip.GzipFile",
"waitress.serve",
"os.m... | [((439, 541), 'flask.Flask', 'Flask', (['"""spreadsplug.web"""'], {'static_url_path': '""""""', 'static_folder': '"""./client"""', 'template_folder': '"""./client"""'}), "('spreadsplug.web', static_url_path='', static_folder='./client',\n template_folder='./client')\n", (444, 541), False, 'from flask import Flask, r... |
"""A tool to convert annotation files created with CVAT into ground-truth style images
for machine learning. The initial code was copied from:
https://gist.github.com/cheind/9850e35bb08cfe12500942fb8b55531f
originally written for a similar purpose for the tool BeaverDam (which produces json),
and was then adapted f... | [
"cv2.rectangle",
"numpy.copy",
"xml.etree.ElementTree.parse",
"argparse.ArgumentParser",
"cv2.VideoWriter",
"cv2.imshow",
"numpy.zeros",
"cv2.destroyAllWindows",
"cv2.VideoCapture",
"cv2.VideoWriter_fourcc",
"cv2.waitKey"
] | [((873, 891), 'xml.etree.ElementTree.parse', 'ET.parse', (['args.ann'], {}), '(args.ann)\n', (881, 891), True, 'import xml.etree.ElementTree as ET\n'), ((1101, 1129), 'cv2.VideoCapture', 'cv2.VideoCapture', (['args.video'], {}), '(args.video)\n', (1117, 1129), False, 'import cv2\n'), ((1464, 1495), 'cv2.VideoWriter_fou... |
import time
from selenium import webdriver
from lxml import etree
driver = webdriver.PhantomJS(executable_path='./phantomjs-2.1.1-macosx/bin/phantomjs')
# 获取第一页的数据
def get_html():
url = "https://detail.tmall.com/item.htm?id=531993957001&skuId=3609796167425&user_id=268451883&cat_id=2&is_b=1&rn=71b9b0aeb233411c4f5... | [
"selenium.webdriver.PhantomJS",
"lxml.etree.HTML",
"time.sleep"
] | [((76, 153), 'selenium.webdriver.PhantomJS', 'webdriver.PhantomJS', ([], {'executable_path': '"""./phantomjs-2.1.1-macosx/bin/phantomjs"""'}), "(executable_path='./phantomjs-2.1.1-macosx/bin/phantomjs')\n", (95, 153), False, 'from selenium import webdriver\n'), ((359, 372), 'time.sleep', 'time.sleep', (['(5)'], {}), '(... |
"""
"""
from flask import Flask
import os
import json
from molder import db, site
def create_app(instance_path=None, test_config=None):
"""
Create and configure the molder Flask application.
Parameters
----------
instance_path : :class:`str`, optional
Path to the instance directory of t... | [
"os.path.exists",
"molder.db.init_app",
"os.makedirs",
"flask.Flask"
] | [((532, 607), 'flask.Flask', 'Flask', (['__name__'], {'instance_path': 'instance_path', 'instance_relative_config': '(True)'}), '(__name__, instance_path=instance_path, instance_relative_config=True)\n', (537, 607), False, 'from flask import Flask\n'), ((1460, 1476), 'molder.db.init_app', 'db.init_app', (['app'], {}), ... |
from PIL import Image, ImageDraw, ImageFont
from urllib.request import urlopen
from textwrap import wrap
import os
BOLD_FONT_URL = "https://cdn.jsdelivr.net/gh/spoqa/spoqa-han-sans@latest/Subset/SpoqaHanSansNeo/SpoqaHanSansNeo-Bold.ttf"
LIGHT_FONT_URL = "https://cdn.jsdelivr.net/gh/spoqa/spoqa-han-sans@latest/Subset/S... | [
"os.path.exists",
"os.listdir",
"PIL.Image.new",
"PIL.ImageDraw.Draw",
"os.mkdir",
"textwrap.wrap",
"urllib.request.urlopen"
] | [((780, 800), 'os.listdir', 'os.listdir', (['POSTPATH'], {}), '(POSTPATH)\n', (790, 800), False, 'import os\n'), ((2170, 2215), 'PIL.Image.new', 'Image.new', (['"""RGB"""', 'OPENGRAPH_SIZE'], {'color': 'WHITE'}), "('RGB', OPENGRAPH_SIZE, color=WHITE)\n", (2179, 2215), False, 'from PIL import Image, ImageDraw, ImageFont... |
import os
import yaml
import pytest
import testinfra.utils.ansible_runner
testinfra_hosts = testinfra.utils.ansible_runner.AnsibleRunner(
os.environ['MOLECULE_INVENTORY_FILE']).get_hosts('all')
dir_path = os.path.dirname(os.path.abspath(__file__))
@pytest.fixture()
def AnsibleDefaults():
with open(os.path.j... | [
"os.path.join",
"yaml.load",
"pytest.mark.parametrize",
"pytest.fixture",
"os.path.abspath"
] | [((257, 273), 'pytest.fixture', 'pytest.fixture', ([], {}), '()\n', (271, 273), False, 'import pytest\n'), ((418, 434), 'pytest.fixture', 'pytest.fixture', ([], {}), '()\n', (432, 434), False, 'import pytest\n'), ((569, 655), 'pytest.mark.parametrize', 'pytest.mark.parametrize', (['"""minio_bin_var"""', "['minio_server... |
"""
resourceview.py
Contains administrative views for working with resources.
"""
from datetime import date
from admin_helpers import *
from sqlalchemy import or_, not_, func
from flask import current_app, redirect, flash, request, url_for
from flask.ext.admin import BaseView, expose
from flask.ext.admin.actions im... | [
"sqlalchemy.func.count",
"flask.ext.admin.expose",
"flask.flash",
"flask.request.form.getlist",
"flask.ext.admin.actions.action",
"flask.url_for",
"flask.request.args.getlist",
"remedy.rad.models.Category.name.asc",
"wtforms.validators.Optional",
"remedy.rad.models.Resource.name.asc",
"datetime.... | [((2137, 2259), 'flask.ext.admin.actions.action', 'action', (['"""togglevisible"""', '"""Toggle Visibility"""', '"""Are you sure you wish to toggle visibility for the selected resources?"""'], {}), "('togglevisible', 'Toggle Visibility',\n 'Are you sure you wish to toggle visibility for the selected resources?')\n",... |
from typing import List
import tensorflow as tf
from tensorflow.keras.layers import Embedding, Layer, LSTM, Input
from src.features.preprocessing import get_embedding, get_text_vectorization
from src.models.embedding_model import EmbeddingModel
class LSTMModel(EmbeddingModel):
def __init__(self):
super(... | [
"src.features.preprocessing.get_embedding",
"src.features.preprocessing.get_text_vectorization",
"tensorflow.keras.layers.LSTM",
"tensorflow.keras.layers.Input"
] | [((360, 375), 'src.features.preprocessing.get_embedding', 'get_embedding', ([], {}), '()\n', (373, 375), False, 'from src.features.preprocessing import get_embedding, get_text_vectorization\n'), ((411, 435), 'src.features.preprocessing.get_text_vectorization', 'get_text_vectorization', ([], {}), '()\n', (433, 435), Fal... |
# coding=utf-8
# *** WARNING: this file was generated by the Pulumi Terraform Bridge (tfgen) Tool. ***
# *** Do not edit by hand unless you're certain you know what you are doing! ***
import warnings
import pulumi
import pulumi.runtime
from typing import Any, Mapping, Optional, Sequence, Union, overload
from .. import... | [
"pulumi.get",
"pulumi.getter",
"pulumi.set",
"pulumi.InvokeOptions",
"pulumi.runtime.invoke"
] | [((4092, 4132), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""availabilityDomain"""'}), "(name='availabilityDomain')\n", (4105, 4132), False, 'import pulumi\n'), ((4382, 4417), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""compartmentId"""'}), "(name='compartmentId')\n", (4395, 4417), False, 'import pulumi\... |
#!/usr/bin/env python3
# -*- coding: utf-8 -*-
"""
Created on Wed May 23 07:33:30 2018
@author: aaronpenne
"""
import os
import pandas as pd
import numpy as np
code_dir = os.path.dirname(__file__) # Returns full path of this script
data_dir = os.path.join(code_dir, 'data')
output_dir = os.path.join(code_dir, 'outpu... | [
"os.path.dirname",
"os.path.isdir",
"os.path.join",
"os.mkdir"
] | [((174, 199), 'os.path.dirname', 'os.path.dirname', (['__file__'], {}), '(__file__)\n', (189, 199), False, 'import os\n'), ((247, 277), 'os.path.join', 'os.path.join', (['code_dir', '"""data"""'], {}), "(code_dir, 'data')\n", (259, 277), False, 'import os\n'), ((291, 323), 'os.path.join', 'os.path.join', (['code_dir', ... |
import os
import tempfile
import pytest
import numpy as np
try:
import h5py
except ImportError:
h5py = None
from msl.io import read, HDF5Writer, JSONWriter
from msl.io.readers import HDF5Reader
from helper import read_sample, roots_equal
@pytest.mark.skipif(h5py is None, reason='h5py not installed')
def te... | [
"helper.read_sample",
"numpy.random.random",
"msl.io.read",
"msl.io.HDF5Writer",
"numpy.array",
"numpy.array_equal",
"pytest.raises",
"tempfile.gettempdir",
"pytest.mark.skipif",
"os.path.basename",
"helper.roots_equal",
"os.remove"
] | [((252, 313), 'pytest.mark.skipif', 'pytest.mark.skipif', (['(h5py is None)'], {'reason': '"""h5py not installed"""'}), "(h5py is None, reason='h5py not installed')\n", (270, 313), False, 'import pytest\n'), ((3668, 3729), 'pytest.mark.skipif', 'pytest.mark.skipif', (['(h5py is None)'], {'reason': '"""h5py not installe... |
import logging
import tkinter as tk
class CalibrationMsgBox():
"""
The message box class.
"""
MESSAGES = [
'Turn the steering wheel fully LEFT and press x.',
'Turn the steering wheel fully RIGHT and press x.',
'Make sure the throttle is fully DEPRESSED and press x.',
'Ma... | [
"logging.getLogger",
"tkinter.StringVar",
"tkinter.Toplevel",
"tkinter.Label"
] | [((811, 841), 'logging.getLogger', 'logging.getLogger', (['"""CALIB_MSG"""'], {}), "('CALIB_MSG')\n", (828, 841), False, 'import logging\n'), ((1039, 1056), 'tkinter.Toplevel', 'tk.Toplevel', (['root'], {}), '(root)\n', (1050, 1056), True, 'import tkinter as tk\n'), ((1230, 1253), 'tkinter.StringVar', 'tk.StringVar', (... |
import numpy as np
from ringity.classes.diagram import PersistenceDiagram
def read_pdiagram(fname, **kwargs):
"""
Wrapper for numpy.genfromtxt.
"""
return PersistenceDiagram(np.genfromtxt(fname, **kwargs))
def write_pdiagram(dgm, fname, **kwargs):
"""
Wrapper for numpy.savetxt.
"""
... | [
"numpy.array",
"numpy.genfromtxt",
"numpy.savetxt"
] | [((329, 342), 'numpy.array', 'np.array', (['dgm'], {}), '(dgm)\n', (337, 342), True, 'import numpy as np\n'), ((347, 381), 'numpy.savetxt', 'np.savetxt', (['fname', 'array'], {}), '(fname, array, **kwargs)\n', (357, 381), True, 'import numpy as np\n'), ((191, 221), 'numpy.genfromtxt', 'np.genfromtxt', (['fname'], {}), ... |
import functools
import os
from argparse import ArgumentParser
import networkx
import numpy as np
from visualize import heatmap
class MatchingClustering(object):
def __init__(self, n_clusters):
self.n_clusters = n_clusters
def fit_predict(self, X):
total = len(X)
grouping = [{i} for... | [
"networkx.algorithms.max_weight_matching",
"argparse.ArgumentParser",
"numpy.where",
"os.path.join",
"networkx.Graph",
"numpy.max",
"numpy.sum",
"numpy.zeros"
] | [((1247, 1263), 'argparse.ArgumentParser', 'ArgumentParser', ([], {}), '()\n', (1261, 1263), False, 'from argparse import ArgumentParser\n'), ((1101, 1130), 'numpy.zeros', 'np.zeros', (['total'], {'dtype': 'np.int'}), '(total, dtype=np.int)\n', (1109, 1130), True, 'import numpy as np\n'), ((1428, 1470), 'os.path.join',... |
#!/usr/bin/env python
# -*- coding: utf-8 -*-
import os
from sfdc_cli.package_xml import PackageXml
command_name = os.path.basename(__file__).split('.', 1)[0].replace("_", ":")
def register(parser, subparsers, **kwargs):
def handler(args):
if args.scandir and args.savedir and args.name and args.apiversi... | [
"os.path.basename",
"sfdc_cli.package_xml.PackageXml"
] | [((116, 142), 'os.path.basename', 'os.path.basename', (['__file__'], {}), '(__file__)\n', (132, 142), False, 'import os\n'), ((337, 364), 'sfdc_cli.package_xml.PackageXml', 'PackageXml', ([], {'project_dir': '"""."""'}), "(project_dir='.')\n", (347, 364), False, 'from sfdc_cli.package_xml import PackageXml\n')] |
from django.conf.urls import url
from basic_app import views
# SET THE NAMESPACE!
app_name = 'basic_app'
urlpatterns=[
url(r'^register/$',views.register,name='register'),
url(r'^user_login/$',views.user_login,name='user_login'),
url(r'^add/$',views.add,name='add'),
url(r'^bookadd/$',views.bookadd,name... | [
"django.conf.urls.url"
] | [((125, 176), 'django.conf.urls.url', 'url', (['"""^register/$"""', 'views.register'], {'name': '"""register"""'}), "('^register/$', views.register, name='register')\n", (128, 176), False, 'from django.conf.urls import url\n'), ((181, 238), 'django.conf.urls.url', 'url', (['"""^user_login/$"""', 'views.user_login'], {'... |
# _*_ coding: utf-8 _*_
"""
Implementation of quick sort algorithm.
Reference:
[1] https://runestone.academy/runestone/books/published/pythonds/SortSearch/TheQuickSort.html
Author: <NAME>
"""
from typing import List
def _quick_sort(nums: List[int], li: int, ri: int) -> None:
if li >= ri:
return
sp... | [
"time.time"
] | [((1341, 1347), 'time.time', 'time', ([], {}), '()\n', (1345, 1347), False, 'from time import time\n'), ((1422, 1428), 'time.time', 'time', ([], {}), '()\n', (1426, 1428), False, 'from time import time\n')] |
#!/usr/bin/env python
"""Command-line interface for skyviewbot"""
# ASTERICS-OBELICS Good Coding Practices (skyviewbot.py)
# <NAME> (<EMAIL>), with suggestions from <NAME>
import sys
from .functions import skyviewbot
from argparse import ArgumentParser, RawTextHelpFormatter
def main(*function_args):
"""Command-... | [
"argparse.ArgumentParser"
] | [((496, 548), 'argparse.ArgumentParser', 'ArgumentParser', ([], {'formatter_class': 'RawTextHelpFormatter'}), '(formatter_class=RawTextHelpFormatter)\n', (510, 548), False, 'from argparse import ArgumentParser, RawTextHelpFormatter\n')] |
#!/usr/bin/env python3
# -*- coding: utf-8 -*-
"""
Created on Sun Dec 27 14:39:08 2020
@author: ravi
"""
import scipy.io as scio
import scipy.io.wavfile as scwav
import numpy as np
import joblib
import pyworld as pw
import os
import warnings
warnings.filterwarnings('ignore')
from tqdm import tqdm
from concurrent.fut... | [
"feat_utils.preprocess_contour",
"numpy.random.rand",
"pyworld.code_spectral_envelope",
"pyworld.cheaptrick",
"tqdm.tqdm",
"numpy.asarray",
"extract_fold_data_hparams.Hparams",
"os.path.join",
"numpy.sum",
"numpy.array",
"numpy.random.randint",
"scipy.io.wavfile.read",
"functools.partial",
... | [((244, 277), 'warnings.filterwarnings', 'warnings.filterwarnings', (['"""ignore"""'], {}), "('ignore')\n", (267, 277), False, 'import warnings\n'), ((3408, 3474), 'joblib.load', 'joblib.load', (['"""/home/ravi/Downloads/Emo-Conv/speaker_file_info.pkl"""'], {}), "('/home/ravi/Downloads/Emo-Conv/speaker_file_info.pkl')\... |
import numpy as np
import cv2
import errno
# set environment variable
import os
os.environ['OPENCV_IO_ENABLE_JASPER']= 'TRUE' # allows JPEG2000 format
# path of this file
det_path = os.path.split(os.path.abspath(__file__))[0] + '/'
class DimensionError(Exception):
"""
raised when the image does not me... | [
"numpy.mean",
"cv2.imwrite",
"os.strerror",
"cv2.dnn.readNetFromCaffe",
"os.path.isfile",
"numpy.array",
"os.path.abspath",
"cv2.resize",
"cv2.imread"
] | [((1585, 1604), 'cv2.imread', 'cv2.imread', (['in_path'], {}), '(in_path)\n', (1595, 1604), False, 'import cv2\n'), ((1783, 1910), 'cv2.dnn.readNetFromCaffe', 'cv2.dnn.readNetFromCaffe', (["(det_path + 'models/deploy.prototxt')", "(det_path + 'models/res10_300x300_ssd_iter_140000.caffemodel')"], {}), "(det_path + 'mode... |
#
# Connects to SlideRule server at provided url and prints log messages
# generated on server to local terminal
#
import sys
import logging
from sliderule import sliderule
from sliderule import icesat2
###############################################################################
# GLOBAL CODE
####################... | [
"logging.basicConfig",
"sliderule.icesat2.init",
"sliderule.sliderule.source"
] | [((401, 440), 'logging.basicConfig', 'logging.basicConfig', ([], {'level': 'logging.INFO'}), '(level=logging.INFO)\n', (420, 440), False, 'import logging\n'), ((1300, 1323), 'sliderule.icesat2.init', 'icesat2.init', (['url', '(True)'], {}), '(url, True)\n', (1312, 1323), False, 'from sliderule import icesat2\n'), ((149... |
import setuptools
requirements = [
'docopt',
'numpy',
'pyzmq'
]
console_scripts = [
'lambda_client=lambda_scope.zmq.client:main',
'lambda_forwarder=lambda_scope.zmq.forwarder:main',
'lambda_hub=lambda_scope.devices.hub_relay:main',
'lambda_publisher=lambda_scope.zmq.publisher:main',
'l... | [
"setuptools.setup"
] | [((1373, 2002), 'setuptools.setup', 'setuptools.setup', ([], {'name': '"""lambda_scope"""', 'version': '"""0.0.1"""', 'author': '"""<NAME>"""', 'author_email': '"""<EMAIL>"""', 'description': '"""Software to operate the customized imaging system, lambda."""', 'url': '"""https://github.com/venkatachalamlab/lambda"""', '... |
"""
NCL_coneff_16.py
================
This script illustrates the following concepts:
- Showing features of the new color display model
- Using a NCL colormap with levels to assign a color palette to contours
- Drawing partially transparent filled contours
See following URLs to see the reproduced NCL plot & s... | [
"geocat.datafiles.get",
"geocat.viz.util.add_major_minor_ticks",
"matplotlib.pyplot.colorbar",
"cartopy.crs.PlateCarree",
"matplotlib.pyplot.figure",
"numpy.linspace",
"matplotlib.pyplot.axes",
"geocat.viz.util.add_lat_lon_ticklabels",
"geocat.viz.util.set_titles_and_labels",
"matplotlib.pyplot.sh... | [((1221, 1248), 'matplotlib.pyplot.figure', 'plt.figure', ([], {'figsize': '(14, 7)'}), '(figsize=(14, 7))\n', (1231, 1248), True, 'import matplotlib.pyplot as plt\n'), ((1294, 1312), 'cartopy.crs.PlateCarree', 'ccrs.PlateCarree', ([], {}), '()\n', (1310, 1312), True, 'import cartopy.crs as ccrs\n'), ((1318, 1349), 'ma... |
from sqlalchemy.ext.declarative import declarative_base
from sqlalchemy import Column, Integer, String
from sqlalchemy import create_engine
from sqlalchemy.ext.declarative import declarative_base
from sqlalchemy.orm import sessionmaker
engine = create_engine('sqlite:///gosts.db')
Session = sessionmaker(bind=engine)
s... | [
"sqlalchemy.orm.sessionmaker",
"sqlalchemy.create_engine",
"sqlalchemy.Column",
"sqlalchemy.ext.declarative.declarative_base"
] | [((247, 282), 'sqlalchemy.create_engine', 'create_engine', (['"""sqlite:///gosts.db"""'], {}), "('sqlite:///gosts.db')\n", (260, 282), False, 'from sqlalchemy import create_engine\n'), ((293, 318), 'sqlalchemy.orm.sessionmaker', 'sessionmaker', ([], {'bind': 'engine'}), '(bind=engine)\n', (305, 318), False, 'from sqlal... |
#!/usr/bin/env python
# -*- coding: UTF-8 -*-
__author__ = '<NAME>'
import re
from functools import lru_cache
import numpy as np
import pandas as pd
from .reaction import Reaction
from .metabolite import Metabolite
from ..optim.optim import FBAOptimizer, TFBAOptimizer, ETFBAOptimizer
from .pdict import PrettyDict
... | [
"re.match",
"pandas.read_excel",
"pandas.DataFrame",
"functools.lru_cache",
"re.search"
] | [((4031, 4042), 'functools.lru_cache', 'lru_cache', ([], {}), '()\n', (4040, 4042), False, 'from functools import lru_cache\n'), ((4854, 4865), 'functools.lru_cache', 'lru_cache', ([], {}), '()\n', (4863, 4865), False, 'from functools import lru_cache\n'), ((5988, 5999), 'functools.lru_cache', 'lru_cache', ([], {}), '(... |
"""File with only WriteMessage class."""
import curses
class WriteMessage:
def __init__(self, history_file=False):
"""
Write data to the pad.
Parameters
----------
n_line: int
number of line of the pad
n_col: int
number of colu... | [
"curses.color_pair",
"curses.start_color",
"curses.init_pair",
"curses.has_colors",
"curses.newpad"
] | [((598, 618), 'curses.start_color', 'curses.start_color', ([], {}), '()\n', (616, 618), False, 'import curses\n'), ((631, 650), 'curses.has_colors', 'curses.has_colors', ([], {}), '()\n', (648, 650), False, 'import curses\n'), ((1436, 1478), 'curses.newpad', 'curses.newpad', (['self._max_line', 'self._n_col'], {}), '(s... |
import matplotlib.pyplot as plt
from string import ascii_uppercase
def countSpecific(_path, _letter):
_letter = _letter.strip().upper()
file = open(_path, 'rb')
text = str(file.read())
return text.count(_letter) + text.count(_letter.lower())
def countAll(_path):
file = open(_path, "rb")
text =... | [
"matplotlib.pyplot.show"
] | [((811, 821), 'matplotlib.pyplot.show', 'plt.show', ([], {}), '()\n', (819, 821), True, 'import matplotlib.pyplot as plt\n')] |
import requests
import pytest
from apistar import TestClient
from api.web.support import Status
from tests.markers import smoke
@pytest.fixture(scope="module")
def response(client: TestClient) -> requests.Response:
return client.get("/api")
@smoke
def test_get_vehicles_status(response: requests.Response) -> Non... | [
"pytest.fixture"
] | [((131, 161), 'pytest.fixture', 'pytest.fixture', ([], {'scope': '"""module"""'}), "(scope='module')\n", (145, 161), False, 'import pytest\n')] |
import matplotlib.pyplot as plt
import numpy as np
from matplotlib.ticker import FuncFormatter
filepath = '/Users/huangjiaming/Documents/developer/ETreeLearning/res/losses/delay_etree.txt'
x = []
num = 0
with open(filepath) as fp:
for line in fp:
c = list(map(int, line.split()))
x = c
print(np.... | [
"numpy.mean",
"matplotlib.pyplot.savefig",
"matplotlib.pyplot.subplots",
"numpy.std"
] | [((360, 397), 'matplotlib.pyplot.subplots', 'plt.subplots', ([], {'nrows': '(2)', 'figsize': '(9, 6)'}), '(nrows=2, figsize=(9, 6))\n', (372, 397), True, 'import matplotlib.pyplot as plt\n'), ((853, 917), 'matplotlib.pyplot.savefig', 'plt.savefig', (['"""./reports/20200301/delay_etree_100_nodes"""'], {'dpi': '(600)'}),... |
from flask_restx import reqparse
from db_plugins.db.sql import models
columns = []
for c in models.Object.__table__.columns:
columns.append(str(c).split(".")[1])
for c in models.Probability.__table__.columns:
columns.append(str(c).split(".")[1])
def str2bool(v):
if isinstance(v, bool):
return v
... | [
"flask_restx.reqparse.RequestParser",
"flask_restx.reqparse.ArgumentTypeError"
] | [((619, 643), 'flask_restx.reqparse.RequestParser', 'reqparse.RequestParser', ([], {}), '()\n', (641, 643), False, 'from flask_restx import reqparse\n'), ((2341, 2365), 'flask_restx.reqparse.RequestParser', 'reqparse.RequestParser', ([], {}), '()\n', (2363, 2365), False, 'from flask_restx import reqparse\n'), ((2923, 2... |
# -*- coding: utf-8 -*-
"""
Created on Wed Sep 25 15:06:45 2019
@author: garci
"""
import matplotlib.pyplot as plt
import numpy as np
import csv
import xlwings as xw
import pandas
import os
'''MAKE X-Y PLOTS WITH 2-COLUMN FILES
<NAME>, 2019 '''
'''lastRow credit: answered Sep 14 '16 at 11:39 - Stefan
https://st... | [
"matplotlib.pyplot.savefig",
"matplotlib.pyplot.xlabel",
"matplotlib.pyplot.plot",
"os.walk",
"matplotlib.pyplot.figure",
"time.time",
"matplotlib.pyplot.title",
"xlwings.Book",
"pandas.read_fwf",
"matplotlib.pyplot.legend"
] | [((1217, 1237), 'xlwings.Book', 'xw.Book', (['(path + file)'], {}), '(path + file)\n', (1224, 1237), True, 'import xlwings as xw\n'), ((1480, 1492), 'matplotlib.pyplot.figure', 'plt.figure', ([], {}), '()\n', (1490, 1492), True, 'import matplotlib.pyplot as plt\n'), ((1497, 1511), 'matplotlib.pyplot.plot', 'plt.plot', ... |
import urllib
from bs4 import BeautifulSoup
print ("Collecting data from IMDb charts....\n\n\n")
print ("The current top 15 IMDB movies are the following: \n\n")
response = urllib.request.urlopen("http://www.imdb.com/chart/top")
html = response.read()
soup = BeautifulSoup(html, 'html.parser')
mytd = soup.findAll("td",... | [
"bs4.BeautifulSoup",
"urllib.request.urlopen"
] | [((174, 229), 'urllib.request.urlopen', 'urllib.request.urlopen', (['"""http://www.imdb.com/chart/top"""'], {}), "('http://www.imdb.com/chart/top')\n", (196, 229), False, 'import urllib\n'), ((260, 294), 'bs4.BeautifulSoup', 'BeautifulSoup', (['html', '"""html.parser"""'], {}), "(html, 'html.parser')\n", (273, 294), Fa... |
from django.urls import path
import mainapp.views as mainapp
app_name = "mainapp"
urlpatterns = [
path("", mainapp.product, name="index"),
path("<int:pk>/", mainapp.product, name="category"),
path("<int:pk>/page/<int:page>/", mainapp.product, name="page"),
path("product/<int:pk>/", mainapp.product_p... | [
"django.urls.path"
] | [((106, 145), 'django.urls.path', 'path', (['""""""', 'mainapp.product'], {'name': '"""index"""'}), "('', mainapp.product, name='index')\n", (110, 145), False, 'from django.urls import path\n'), ((151, 202), 'django.urls.path', 'path', (['"""<int:pk>/"""', 'mainapp.product'], {'name': '"""category"""'}), "('<int:pk>/',... |
import datetime
import os
import random
import time
import requests
from lxml import etree
from selenium import webdriver
# import config
import threading
# import numpy as np
mUA_list = [
'Mozilla/5.0 (iPhone; CPU iPhone OS 11_2_1 like Mac OS X) AppleWebKit/604.4.7 (KHTML, like Gecko) Version/11.0 Mobile/15C153... | [
"selenium.webdriver.Chrome",
"selenium.webdriver.ChromeOptions",
"time.sleep"
] | [((1617, 1630), 'time.sleep', 'time.sleep', (['(2)'], {}), '(2)\n', (1627, 1630), False, 'import time\n'), ((1661, 1675), 'time.sleep', 'time.sleep', (['(20)'], {}), '(20)\n', (1671, 1675), False, 'import time\n'), ((1037, 1062), 'selenium.webdriver.ChromeOptions', 'webdriver.ChromeOptions', ([], {}), '()\n', (1060, 10... |
import os
import random
import string
import json
from django.core.management import BaseCommand
__author__ = "<NAME>"
__copyright__ = "Copyright 2018, <NAME>"
__licence__ = "BSD 2-Clause Licence"
__version__ = "1.0"
__email__ = "<EMAIL>"
class Command(BaseCommand):
def is_valid_file(self, file):
if no... | [
"os.path.isfile",
"json.dumps",
"random.SystemRandom",
"os.path.expanduser"
] | [((322, 342), 'os.path.isfile', 'os.path.isfile', (['file'], {}), '(file)\n', (336, 342), False, 'import os\n'), ((1136, 1166), 'json.dumps', 'json.dumps', (['settings'], {'indent': '(2)'}), '(settings, indent=2)\n', (1146, 1166), False, 'import json\n'), ((581, 602), 'random.SystemRandom', 'random.SystemRandom', ([], ... |
from flask_wtf import FlaskForm
from wtforms import StringField,TextAreaField,SubmitField, SelectField
from wtforms.validators import Required
class PitchForm(FlaskForm):
title = StringField('Pitch title',validators=[Required()])
category = SelectField('Pitch category', choices=[('Motivational', 'Motivational... | [
"wtforms.validators.Required",
"wtforms.SubmitField"
] | [((487, 508), 'wtforms.SubmitField', 'SubmitField', (['"""Submit"""'], {}), "('Submit')\n", (498, 508), False, 'from wtforms import StringField, TextAreaField, SubmitField, SelectField\n'), ((627, 648), 'wtforms.SubmitField', 'SubmitField', (['"""Submit"""'], {}), "('Submit')\n", (638, 648), False, 'from wtforms import... |
#!/usr/bin/env python
# Copyright 2016 Amazon.com, Inc. or its
# affiliates. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License"). You may not
# use this file except in compliance with the License. A copy of the License is
# located at
#
# http://aws.amazon.com/apache2.0/
#
# or i... | [
"os.path.exists",
"json.loads",
"PIL.Image.open",
"boto3.client",
"urllib.unquote_plus",
"os.makedirs",
"PIL.ExifTags.TAGS.get",
"boto3.resource",
"os.remove"
] | [((936, 978), 'boto3.client', 'boto3.client', (['"""s3"""'], {'region_name': 'aws_region'}), "('s3', region_name=aws_region)\n", (948, 978), False, 'import boto3\n'), ((985, 1030), 'boto3.resource', 'boto3.resource', (['"""sqs"""'], {'region_name': 'aws_region'}), "('sqs', region_name=aws_region)\n", (999, 1030), False... |
from dash_extensions.enrich import Dash
from _app.layout import serve_layout
from _app.callback import register_callbacks
external_stylesheets = ["https://codepen.io/chriddyp/pen/bWLwgP.css"]
app = Dash(prevent_initial_callbacks=True,
external_stylesheets=external_stylesheets
)
register_c... | [
"_app.callback.register_callbacks",
"_app.layout.serve_layout",
"dash_extensions.enrich.Dash"
] | [((203, 282), 'dash_extensions.enrich.Dash', 'Dash', ([], {'prevent_initial_callbacks': '(True)', 'external_stylesheets': 'external_stylesheets'}), '(prevent_initial_callbacks=True, external_stylesheets=external_stylesheets)\n', (207, 282), False, 'from dash_extensions.enrich import Dash\n'), ((310, 333), '_app.callbac... |
import TestCase
from testcase_generation.Response.default import defaultTestCase
import Constants
def defaultMediaTypeTestCase():
testcase = defaultTestCase()
# Request
testcase.request.method = 'GET'
# Response
testcase.response.status_code = 200
testcase.response.expect_body = True
return testcase
def testC... | [
"testcase_generation.Response.default.defaultTestCase"
] | [((143, 160), 'testcase_generation.Response.default.defaultTestCase', 'defaultTestCase', ([], {}), '()\n', (158, 160), False, 'from testcase_generation.Response.default import defaultTestCase\n')] |
import numpy as np
from typing import List, Tuple
class InterfaceSolver():
"""
Informal interface for solving class needed to interact with rubiks
environment.
"""
def __init__(self, depth:int, possible_moves: List[str]) -> None:
"""
Will be passed depth, i.e. number of backwa... | [
"numpy.sum",
"numpy.array"
] | [((8033, 8074), 'numpy.sum', 'np.sum', (['(cube_state[i, :, :] == center_val)'], {}), '(cube_state[i, :, :] == center_val)\n', (8039, 8074), True, 'import numpy as np\n'), ((8674, 8706), 'numpy.array', 'np.array', (['self.cube_state_values'], {}), '(self.cube_state_values)\n', (8682, 8706), True, 'import numpy as np\n'... |
#!/usr/bin/env python3
from selenium import webdriver
from selenium.webdriver.common.by import By
class TestLuckyApp:
"""E2E integration tests class."""
def setup_method(self, method):
options = webdriver.FirefoxOptions()
self.driver = webdriver.Remote('http://firefoxdriver:4444/wd/hub', opt... | [
"selenium.webdriver.FirefoxOptions",
"selenium.webdriver.Remote"
] | [((215, 241), 'selenium.webdriver.FirefoxOptions', 'webdriver.FirefoxOptions', ([], {}), '()\n', (239, 241), False, 'from selenium import webdriver\n'), ((264, 333), 'selenium.webdriver.Remote', 'webdriver.Remote', (['"""http://firefoxdriver:4444/wd/hub"""'], {'options': 'options'}), "('http://firefoxdriver:4444/wd/hub... |
from uuid import UUID
from datetime import datetime
def uuid_from_string(string):
return UUID('{s}'.format(s=string))
def format_timestamp(string):
if isinstance(string, str):
return datetime.strptime(string, '%Y-%m-%dT%H:%M:%S.%fZ')
if isinstance(string, datetime):
return string
| [
"datetime.datetime.strptime"
] | [((203, 253), 'datetime.datetime.strptime', 'datetime.strptime', (['string', '"""%Y-%m-%dT%H:%M:%S.%fZ"""'], {}), "(string, '%Y-%m-%dT%H:%M:%S.%fZ')\n", (220, 253), False, 'from datetime import datetime\n')] |
import time
def factorial(n):
fact = 1
for x in range(2, n+1):
fact = fact * x
return fact
# Timing function
start = time.time()
factorial(400000)
end = time.time()
print('Operation done in {} seconds'.format(end - start))
| [
"time.time"
] | [((139, 150), 'time.time', 'time.time', ([], {}), '()\n', (148, 150), False, 'import time\n'), ((175, 186), 'time.time', 'time.time', ([], {}), '()\n', (184, 186), False, 'import time\n')] |
"""
==========================================================================
MeshNetworkCL_test.py
==========================================================================
Test for NetworkCL
Author : <NAME>
Date : May 19, 2019
"""
import pytest
from pymtl3_net.meshnet.MeshNetworkCL import MeshNetworkCL
from pym... | [
"pymtl3_net.ocnlib.utils.run_sim",
"pymtl3_net.meshnet.MeshNetworkCL.MeshNetworkCL",
"pymtl3_net.ocnlib.ifcs.positions.mk_mesh_pos",
"pymtl3_net.ocnlib.ifcs.packets.mk_mesh_pkt",
"pytest.mark.parametrize",
"pymtl3.stdlib.test_utils.test_srcs.TestSrcCL",
"pymtl3_net.ocnlib.test.net_sinks.TestNetSinkCL",
... | [((4006, 4237), 'pymtl3.stdlib.test_utils.mk_test_case_table', 'mk_test_case_table', (["['msg_list wid ht src_init src_intv sink_init sink_intv', ['simple2x2',\n simple_2x2, 2, 2, 0, 0, 0, 0], ['simple4x4', simple_4x4, 4, 4, 0, 0, 0,\n 0], ['simple8x8', simple_8x8, 8, 8, 0, 0, 0, 0]]"], {}), "([\n 'msg_li... |
import os
# Helpers
def read_asset(*paths):
dirname = os.path.dirname(__file__)
return open(os.path.join(dirname, "assets", *paths)).read().strip()
# General
VERSION = read_asset("VERSION")
| [
"os.path.dirname",
"os.path.join"
] | [((62, 87), 'os.path.dirname', 'os.path.dirname', (['__file__'], {}), '(__file__)\n', (77, 87), False, 'import os\n'), ((104, 143), 'os.path.join', 'os.path.join', (['dirname', '"""assets"""', '*paths'], {}), "(dirname, 'assets', *paths)\n", (116, 143), False, 'import os\n')] |
import sys
from glob import glob
from serial import Serial, SerialException
import numpy as np
BAUD_RATE = 9600
PORT = 'COM5'
READ_TIMEOUT = 1
LOWER_BOUND = 0.01
UPPER_BOUND = 0.4
class SerialCommunication():
""" Manages the communication and sends the data to the Arduino """
def __init__(self):
s... | [
"numpy.clip",
"sys.platform.startswith",
"serial.Serial",
"numpy.isnan",
"glob.glob"
] | [((342, 350), 'serial.Serial', 'Serial', ([], {}), '()\n', (348, 350), False, 'from serial import Serial, SerialException\n'), ((2715, 2745), 'sys.platform.startswith', 'sys.platform.startswith', (['"""win"""'], {}), "('win')\n", (2738, 2745), False, 'import sys\n'), ((3799, 3841), 'numpy.clip', 'np.clip', (['signals',... |
"""add unique constraint to Source
Revision ID: 4415298e147b
Revises: 7392493a<PASSWORD>
Create Date: 2020-01-02 16:41:03.424945
"""
from alembic import op
import sqlalchemy as sa
# revision identifiers, used by Alembic.
revision = '<KEY>'
down_revision = '7392493a0768'
branch_labels = None
depends_on = None
def ... | [
"alembic.op.drop_constraint",
"alembic.op.create_unique_constraint"
] | [((335, 458), 'alembic.op.create_unique_constraint', 'op.create_unique_constraint', (['"""ori_id_canonical_fields"""', '"""source"""', "['resource_ori_id', 'canonical_id', 'canonical_iri']"], {}), "('ori_id_canonical_fields', 'source', [\n 'resource_ori_id', 'canonical_id', 'canonical_iri'])\n", (362, 458), False, '... |
"""Helper methods for parsing EBS-related data from AWS SDK."""
import logging
import myutils
logger = myutils.get_logger(__name__, logging.DEBUG)
@myutils.log_calls(level=logging.DEBUG)
def parse(sdk_snapshots):
"""Process raw EBS snapshot data."""
snapshots = []
snapshots.extend(
map(map_snaps... | [
"myutils.get_first",
"myutils.log_calls",
"myutils.get_logger"
] | [((105, 148), 'myutils.get_logger', 'myutils.get_logger', (['__name__', 'logging.DEBUG'], {}), '(__name__, logging.DEBUG)\n', (123, 148), False, 'import myutils\n'), ((152, 190), 'myutils.log_calls', 'myutils.log_calls', ([], {'level': 'logging.DEBUG'}), '(level=logging.DEBUG)\n', (169, 190), False, 'import myutils\n')... |
import assets
import webbrowser
from PyQt5.Qt import QMessageBox
from PyQt5.QtNetwork import QNetworkDiskCache
from PyQt5.QtWebKitWidgets import QWebPage, QWebInspector
class WebPage(QWebPage):
def __init__(self):
super(WebPage, self).__init__()
self.inspector = QWebInspector()
self.inspector.setPage(self)
se... | [
"assets.dataJar.save",
"assets.fs.dataPath",
"PyQt5.QtWebKitWidgets.QWebPage.acceptNavigationRequest",
"PyQt5.QtWebKitWidgets.QWebInspector",
"PyQt5.QtNetwork.QNetworkDiskCache"
] | [((269, 284), 'PyQt5.QtWebKitWidgets.QWebInspector', 'QWebInspector', ([], {}), '()\n', (282, 284), False, 'from PyQt5.QtWebKitWidgets import QWebPage, QWebInspector\n'), ((365, 388), 'PyQt5.QtNetwork.QNetworkDiskCache', 'QNetworkDiskCache', (['self'], {}), '(self)\n', (382, 388), False, 'from PyQt5.QtNetwork import QN... |
import azure.functions as func
from .add_url_rule import app
def main(req: func.HttpRequest, context: func.Context) -> func.HttpResponse:
return func.WsgiMiddleware(app.wsgi_app).handle(req, context)
| [
"azure.functions.WsgiMiddleware"
] | [((151, 184), 'azure.functions.WsgiMiddleware', 'func.WsgiMiddleware', (['app.wsgi_app'], {}), '(app.wsgi_app)\n', (170, 184), True, 'import azure.functions as func\n')] |
from flask import Flask
from app.services import model_manager
from app.controllers.api_blueprint import api_router
from app.controllers.main_blueprint import main_router
def create_app(test_config=None):
app = Flask(__name__)
model_manager.init_app(app)
app.register_blueprint(api_router)
app.registe... | [
"app.services.model_manager.init_app",
"flask.Flask"
] | [((216, 231), 'flask.Flask', 'Flask', (['__name__'], {}), '(__name__)\n', (221, 231), False, 'from flask import Flask\n'), ((237, 264), 'app.services.model_manager.init_app', 'model_manager.init_app', (['app'], {}), '(app)\n', (259, 264), False, 'from app.services import model_manager\n')] |
import sys
import time
import threading
import grpc
import numpy
import soundfile as sf
import tensorflow as tf
import _init_paths
import audioset.vggish_input as vggish_input
from tensorflow_serving.apis import predict_pb2
from tensorflow_serving.apis import prediction_service_pb2_grpc
tf.app.flags.DEFINE_integer... | [
"sys.stdout.flush",
"tensorflow.app.flags.DEFINE_integer",
"tensorflow_serving.apis.predict_pb2.PredictRequest",
"tensorflow_serving.apis.prediction_service_pb2_grpc.PredictionServiceStub",
"grpc.insecure_channel",
"tensorflow.app.flags.DEFINE_string",
"sys.stdout.write",
"numpy.pad",
"audioset.vggi... | [((293, 381), 'tensorflow.app.flags.DEFINE_integer', 'tf.app.flags.DEFINE_integer', (['"""concurrency"""', '(1)', '"""concurrent inference requests limit"""'], {}), "('concurrency', 1,\n 'concurrent inference requests limit')\n", (320, 381), True, 'import tensorflow as tf\n'), ((378, 448), 'tensorflow.app.flags.DEFI... |
import sqlite3
class Stats:
con = sqlite3.connect("data.db")
cur = con.cursor()
stats_insert_com = "INSERT INTO Stats (Player, 'Win Count', 'Play Count', winPlayRatio) VALUES (?, ?, ?, ?);"
matchhist_insert_com = "INSERT INTO MatchHistory (matchNum, playerList, 'Winner', numRounds) VALUES (?, ?, ?, ?);... | [
"sqlite3.connect"
] | [((39, 65), 'sqlite3.connect', 'sqlite3.connect', (['"""data.db"""'], {}), "('data.db')\n", (54, 65), False, 'import sqlite3\n')] |
# -*- coding: utf-8 -*-
"""
Created on Wed Jun 24 21:46:56 2020
@author: adwait
"""
import numpy as np
import cv2
import pims
from tkinter import messagebox, Tk
from PIL import ImageFont, ImageDraw, Image
from PyQt5.QtGui import QIcon
import logging
class MainRecordFunctions:
def recordVideo(... | [
"PyQt5.QtGui.QIcon",
"logging.debug",
"cv2.imshow",
"pims.Video",
"PIL.ImageDraw.Draw",
"numpy.array",
"cv2.resizeWindow",
"cv2.line",
"PIL.ImageFont.truetype",
"cv2.VideoWriter",
"numpy.empty",
"cv2.VideoWriter_fourcc",
"tkinter.messagebox.showinfo",
"cv2.putText",
"cv2.cvtColor",
"cv... | [((352, 380), 'logging.debug', 'logging.debug', (['"""recordvideo"""'], {}), "('recordvideo')\n", (365, 380), False, 'import logging\n'), ((9583, 9612), 'logging.debug', 'logging.debug', (['"""record_frame"""'], {}), "('record_frame')\n", (9596, 9612), False, 'import logging\n'), ((11078, 11122), 'numpy.zeros', 'np.zer... |
"""
A simple class to allow quick testing of GPIB programs without instruments.
All reads from an instrument return a semi-randomised number regardless of
the specific command that may have been sent prior to reading.
"""
import stuff
import time
## VisaIOError = False
"""
Old pyvisa calls.
"""
def ge... | [
"stuff.DataGen",
"time.sleep"
] | [((596, 611), 'stuff.DataGen', 'stuff.DataGen', ([], {}), '()\n', (609, 611), False, 'import stuff\n'), ((666, 681), 'time.sleep', 'time.sleep', (['(0.1)'], {}), '(0.1)\n', (676, 681), False, 'import time\n'), ((749, 764), 'time.sleep', 'time.sleep', (['(0.2)'], {}), '(0.2)\n', (759, 764), False, 'import time\n'), ((15... |
import argparse
from core_extract_comments import *
from core_utils import *
def run(search, input_product_ids_filename):
product_ids = list()
if input_product_ids_filename is not None:
with open(input_product_ids_filename, 'r') as r:
for p in r.readlines():
pro_obj = p.st... | [
"argparse.ArgumentParser"
] | [((1171, 1196), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {}), '()\n', (1194, 1196), False, 'import argparse\n')] |
'''
Message Example
b'620901.02908, 3, 0.242, 0.606, 9.527, 4, 0.020, -0.006, 0.001, 5, -24.762,-223.451,-98.491, 81, 173.805, -3.304, 1.315'
'''
import time
import string
import socket, traceback
host="192.168.0.23" # ip address of
port=5555
s=socket.socket(socket.AF_INET, socket.SOCK_DGRAM)
s.setsockopt(so... | [
"socket.socket"
] | [((256, 304), 'socket.socket', 'socket.socket', (['socket.AF_INET', 'socket.SOCK_DGRAM'], {}), '(socket.AF_INET, socket.SOCK_DGRAM)\n', (269, 304), False, 'import socket, traceback\n')] |
import os
from .local_config import *
PROJECT_NAME = "Useful"
SERVER_HOST = 'http://127.0.0.1:8000'
# Secret key
SECRET_KEY = b"<KEY>"
BASE_DIR = os.path.dirname(os.path.dirname(os.path.abspath(__file__)))
API_V1_STR = "/api/v1"
# Token 60 minutes * 24 hours * 7 days = 7 days
ACCESS_TOKEN_EXPIRE_MINUTES = 60 * 24... | [
"os.path.abspath",
"os.environ.get"
] | [((845, 871), 'os.environ.get', 'os.environ.get', (['"""SMTP_TLS"""'], {}), "('SMTP_TLS')\n", (859, 871), False, 'import os\n'), ((884, 911), 'os.environ.get', 'os.environ.get', (['"""SMTP_PORT"""'], {}), "('SMTP_PORT')\n", (898, 911), False, 'import os\n'), ((924, 951), 'os.environ.get', 'os.environ.get', (['"""SMTP_H... |
#%%
# read full assignment
# think algo before implementing
# dont use a dict when you need a list
# assignment is still = and not ==
# dont use itertools when you can use np.roll
# check mathemathical functions if the parentheses are ok
# networkx is awesome
# %%
import os
import re
import numpy as np
try:
os.chdir(... | [
"os.getcwd"
] | [((364, 375), 'os.getcwd', 'os.getcwd', ([], {}), '()\n', (373, 375), False, 'import os\n'), ((333, 344), 'os.getcwd', 'os.getcwd', ([], {}), '()\n', (342, 344), False, 'import os\n')] |
from django.contrib import admin
from .models import Succession,Succession_Seasons,Succession_Casts,Succession_Season_Episodes
# Register your models here.
# admin.site.register(Succession) # The model Succession is abstract so it can't be registered with admin
admin.site.register(Succession_Seasons)
ad... | [
"django.contrib.admin.site.register"
] | [((278, 317), 'django.contrib.admin.site.register', 'admin.site.register', (['Succession_Seasons'], {}), '(Succession_Seasons)\n', (297, 317), False, 'from django.contrib import admin\n'), ((318, 355), 'django.contrib.admin.site.register', 'admin.site.register', (['Succession_Casts'], {}), '(Succession_Casts)\n', (337,... |
#!/usr/bin/env python3
import sys, os
import arsdkparser
#===============================================================================
class Writer(object):
def __init__(self, fileobj):
self.fileobj = fileobj
def write(self, fmt, *args):
if args:
self.fileobj.write(fmt % (args)... | [
"os.path.exists",
"os.path.join",
"os.mkdirs"
] | [((3917, 3961), 'os.path.join', 'os.path.join', (['outdir', "(CONSTS_FILENAME + '.h')"], {}), "(outdir, CONSTS_FILENAME + '.h')\n", (3929, 3961), False, 'import sys, os\n'), ((4096, 4140), 'os.path.join', 'os.path.join', (['outdir', "(CONSTS_FILENAME + '.m')"], {}), "(outdir, CONSTS_FILENAME + '.m')\n", (4108, 4140), F... |
"""
**Project Name:** MakeHuman
**Product Home Page:** http://www.makehuman.org/
**Code Home Page:** http://code.google.com/p/makehuman/
**Authors:** <NAME>
**Copyright(c):** MakeHuman Team 2001-2009
**Licensing:** GPL3 (see also http://sites.google.com/site/makehumandocs... | [
"mhx_rig.writeDrivers",
"mhx_rig.addPoseBone"
] | [((2763, 2953), 'mhx_rig.addPoseBone', 'addPoseBone', (['fp', '"""Jaw"""', '"""MHJaw"""', 'None', '(1, 1, 1)', '(0, 1, 0)', '(1, 1, 1)', '(1, 1, 1)', '(0)', "[('LimitRot', C_OW_LOCAL, 1, ['LimitRot', (-5 * D, 45 * D, 0, 0, -20 * D, \n 20 * D), (1, 1, 1)])]"], {}), "(fp, 'Jaw', 'MHJaw', None, (1, 1, 1), (0, 1, 0), (1... |
import cv2
import numpy as np
from matplotlib import pyplot as plt
img = cv2.imread('home.jpg')
hsv = cv2.cvtColor(img,cv2.COLOR_BGR2HSV)
hist = cv2.calcHist( [hsv], [0, 1], None, [180, 256], [0, 180, 0, 256] )
plt.imshow(hist,interpolation = 'nearest')
plt.show()
# in numpy
import cv2
import numpy as np
from matplo... | [
"matplotlib.pyplot.imshow",
"cv2.calcHist",
"cv2.cvtColor",
"cv2.imread",
"matplotlib.pyplot.show"
] | [((74, 96), 'cv2.imread', 'cv2.imread', (['"""home.jpg"""'], {}), "('home.jpg')\n", (84, 96), False, 'import cv2\n'), ((103, 139), 'cv2.cvtColor', 'cv2.cvtColor', (['img', 'cv2.COLOR_BGR2HSV'], {}), '(img, cv2.COLOR_BGR2HSV)\n', (115, 139), False, 'import cv2\n'), ((146, 209), 'cv2.calcHist', 'cv2.calcHist', (['[hsv]',... |
import network2
import logging
import numpy as np
logging.basicConfig(level=logging.DEBUG)
# read in the data
# logging.info("READING IN DATA...")
# for reading in normal dataset
# training, validation, test = network2.load_data_wrapper("data/mnist.pkl.gz")
### I WILL ADD AND COMMENT OUT SECTIONS OF CODE BASED ... | [
"logging.basicConfig",
"network2.load_data_wrapper",
"logging.info",
"network2.Network"
] | [((51, 91), 'logging.basicConfig', 'logging.basicConfig', ([], {'level': 'logging.DEBUG'}), '(level=logging.DEBUG)\n', (70, 91), False, 'import logging\n'), ((7511, 7545), 'logging.info', 'logging.info', (['"""READING IN DATA..."""'], {}), "('READING IN DATA...')\n", (7523, 7545), False, 'import logging\n'), ((7608, 76... |
from abc import ABC, abstractmethod
import collections
import statistics
import numpy as np
import sklearn.metrics
import torch
class Evaluator(ABC):
"""Class to evaluate model outputs and report the result.
"""
def __init__(self):
self.reset()
@abstractmethod
def add_predictions(self, p... | [
"statistics.mean",
"torch.mul",
"torch.topk",
"numpy.argmax",
"numpy.array",
"numpy.sum",
"torch.add",
"collections.defaultdict",
"torch.zeros_like"
] | [((904, 930), 'torch.topk', 'torch.topk', (['predictions', '(1)'], {}), '(predictions, 1)\n', (914, 930), False, 'import torch\n'), ((1160, 1186), 'torch.topk', 'torch.topk', (['predictions', 'k'], {}), '(predictions, k)\n', (1170, 1186), False, 'import torch\n'), ((1366, 1414), 'torch.zeros_like', 'torch.zeros_like', ... |
import unittest
from lmctl.cli.format import TableFormat, Table, Column
class DummyTable(Table):
columns = [
Column('name', header='Name'),
Column('status', header='Status', accessor=lambda x: 'OK' if x.get('status', None) in ['Excellent', 'Good'] else 'Unhealthy'),
]
class DummyTableNoHeaders... | [
"lmctl.cli.format.Column"
] | [((122, 151), 'lmctl.cli.format.Column', 'Column', (['"""name"""'], {'header': '"""Name"""'}), "('name', header='Name')\n", (128, 151), False, 'from lmctl.cli.format import TableFormat, Table, Column\n'), ((353, 367), 'lmctl.cli.format.Column', 'Column', (['"""name"""'], {}), "('name')\n", (359, 367), False, 'from lmct... |
import logging
import azure.functions as func
import numpy as np
import json
import requests
from os import path
def readjson_from_file(filename):
try:
fp = open(filename, "r")
except:
logging.info(f"WARNING: cant open file {filename} ")
return {}
try:
ob... | [
"json.load",
"json.dumps",
"logging.info"
] | [((617, 651), 'logging.info', 'logging.info', (['f"""post url is {url}"""'], {}), "(f'post url is {url}')\n", (629, 651), False, 'import logging\n'), ((711, 740), 'logging.info', 'logging.info', (['"""payload added"""'], {}), "('payload added')\n", (723, 740), False, 'import logging\n'), ((848, 884), 'logging.info', 'l... |
import django
print(django.get_version())
print(f"boa {111 * 6}")
print(f"{6*6}")
input = input("Hey abuser, enter some stuff!\n")
cmp = 1 > 0
print(type(cmp))
print(input)
try:
int(asdf)
except Exception:
print("Oh no, it failed!")
while False:
print("False!")
listeL = [1,2,3,45,6,4,3,6,8,6,3... | [
"django.get_version"
] | [((20, 40), 'django.get_version', 'django.get_version', ([], {}), '()\n', (38, 40), False, 'import django\n')] |
from start import client
from modules import codeforces,delete,notes,hastebin,pin,pm,user,spam,rextester
white=[]
from telethon import TelegramClient,events
import logging
logging.basicConfig(format='[%(levelname) 5s/%(asctime)s] %(name)s: %(message)s',
level=logging.WARNING)
client.start().run_u... | [
"logging.basicConfig",
"start.client.start"
] | [((172, 286), 'logging.basicConfig', 'logging.basicConfig', ([], {'format': '"""[%(levelname) 5s/%(asctime)s] %(name)s: %(message)s"""', 'level': 'logging.WARNING'}), "(format=\n '[%(levelname) 5s/%(asctime)s] %(name)s: %(message)s', level=logging.\n WARNING)\n", (191, 286), False, 'import logging\n'), ((300, 314... |
# Let’s make a map! Using Geopandas, Pandas and Matplotlib to make a Choropleth map
# https://towardsdatascience.com/lets-make-a-map-using-geopandas-pandas-and-matplotlib-to-make-a-chloropleth-map-dddc31c1983d
import pandas as pd
import matplotlib.pyplot as plt
import geopandas as gpd
from shapely.geometry import... | [
"matplotlib.pyplot.text",
"pycountry.countries.search_fuzzy",
"geopandas.read_file",
"pandas.read_csv",
"matplotlib.pyplot.title",
"matplotlib.pyplot.subplots"
] | [((413, 601), 'geopandas.read_file', 'gpd.read_file', (['"""/Users/vivekparashar/OneDrive/OneDrive-GitHub/Challenges-and-Competitions/30DayMapChallenge/Maps/natural_earth_vector/110m_cultural/ne_110m_admin_0_countries.shp"""'], {}), "(\n '/Users/vivekparashar/OneDrive/OneDrive-GitHub/Challenges-and-Competitions/30Da... |
"""
This module exrtacts features from the data, saves the feauters
from all measurements to global results file and creates
one file for every sensor with all measurements.
:copyright: (c) 2022 by <NAME>, Hochschule-Bonn-Rhein-Sieg
:license: see LICENSE for more details.
"""
from pyexpat import features
import pan... | [
"numpy.trapz",
"numpy.ones",
"pandas.read_csv",
"matplotlib.pyplot.xticks",
"pathlib.Path",
"numpy.max",
"matplotlib.pyplot.close",
"scipy.signal.peak_widths",
"scipy.signal.find_peaks",
"pandas.DataFrame",
"matplotlib.pyplot.subplots",
"numpy.arange"
] | [((9673, 9752), 'scipy.signal.find_peaks', 'find_peaks', (['df[sensor]'], {'prominence': '(0)', 'width': '(1)', 'distance': '(20000)', 'height': 'threshold'}), '(df[sensor], prominence=0, width=1, distance=20000, height=threshold)\n', (9683, 9752), False, 'from scipy.signal import chirp, find_peaks, peak_widths\n'), ((... |
from __future__ import annotations
import json
import os
import shutil
import subprocess
import tempfile
import uuid
from abc import ABC, abstractmethod
from typing import Any, Union
from urllib.error import HTTPError
from urllib.request import urlopen, urlretrieve
import warnings
import meerkat as mk
import pandas a... | [
"pandas.read_csv",
"os.remove",
"os.path.exists",
"meerkat.tools.lazy_loader.LazyLoader",
"meerkat.DataPanel.read",
"yaml.add_multi_representer",
"urllib.request.urlretrieve",
"yaml.add_constructor",
"os.path.isdir",
"subprocess.call",
"warnings.warn",
"urllib.request.urlopen",
"json.loads",... | [((474, 508), 'meerkat.tools.lazy_loader.LazyLoader', 'LazyLoader', (['"""google.cloud.storage"""'], {}), "('google.cloud.storage')\n", (484, 508), False, 'from meerkat.tools.lazy_loader import LazyLoader\n'), ((517, 536), 'meerkat.tools.lazy_loader.LazyLoader', 'LazyLoader', (['"""torch"""'], {}), "('torch')\n", (527,... |
import asyncio
import json
import zlib
import aiohttp
import errors
API_BASE = 'https://discordapp.com/api/v6'
CONFIG_FILE = json.load(open('data/config.json'))
TOKEN = CONFIG_FILE['token']
HEADERS = {'Authorization': 'Bot ' + TOKEN,
'User-Agent': 'DiscordBot (https://www.github.com/fourjr/dapi-bot,\
... | [
"json.loads",
"errors.NotFound",
"errors.Forbidden",
"asyncio.get_event_loop",
"zlib.decompress",
"errors.RateLimit"
] | [((389, 413), 'asyncio.get_event_loop', 'asyncio.get_event_loop', ([], {}), '()\n', (411, 413), False, 'import asyncio\n'), ((670, 686), 'json.loads', 'json.loads', (['data'], {}), '(data)\n', (680, 686), False, 'import json\n'), ((1384, 1411), 'errors.Forbidden', 'errors.Forbidden', (['resp', 'obj'], {}), '(resp, obj)... |
import sys
import requests
import argparse
import logging
import json
import datetime
import anticrlf
from veracode_api_py import VeracodeAPI as vapi
log = logging.getLogger(__name__)
def setup_logger():
handler = logging.FileHandler('vcoffboard.log', encoding='utf8')
handler.setFormatter(anticrlf.LogFormatt... | [
"logging.getLogger",
"anticrlf.LogFormatter",
"argparse.ArgumentParser",
"datetime.datetime.strptime",
"veracode_api_py.VeracodeAPI",
"datetime.datetime.now",
"logging.FileHandler"
] | [((158, 185), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (175, 185), False, 'import logging\n'), ((221, 275), 'logging.FileHandler', 'logging.FileHandler', (['"""vcoffboard.log"""'], {'encoding': '"""utf8"""'}), "('vcoffboard.log', encoding='utf8')\n", (240, 275), False, 'import loggi... |
from expects import expect, contain, be_an
class Bacon:
...
sanduiche = 'sanduiche com queijo'
expect(sanduiche).to(contain('queijo'))
expect(sanduiche).to_not(be_an(Bacon))
| [
"expects.be_an",
"expects.contain",
"expects.expect"
] | [((127, 144), 'expects.contain', 'contain', (['"""queijo"""'], {}), "('queijo')\n", (134, 144), False, 'from expects import expect, contain, be_an\n'), ((172, 184), 'expects.be_an', 'be_an', (['Bacon'], {}), '(Bacon)\n', (177, 184), False, 'from expects import expect, contain, be_an\n'), ((106, 123), 'expects.expect', ... |
"""
Main
=============
Example
"""
# Import
import numpy as np
import pandas as pd
# Specific
from tpot import TPOTClassifier
# Import own
from pySML2.preprocessing.splitters import cvs_hos_split
from pySML2.preprocessing.splitters import kfolds_split
# ---------------------------------------------
# Configuratio... | [
"tpot.TPOTClassifier",
"pySML2.preprocessing.splitters.cvs_hos_split",
"pandas.read_csv"
] | [((917, 938), 'pandas.read_csv', 'pd.read_csv', (['filepath'], {}), '(filepath)\n', (928, 938), True, 'import pandas as pd\n'), ((1211, 1249), 'pySML2.preprocessing.splitters.cvs_hos_split', 'cvs_hos_split', (['data'], {'selected_rows': 'cmp'}), '(data, selected_rows=cmp)\n', (1224, 1249), False, 'from pySML2.preproces... |
#!/usr/bin/env python3.6
"""Sherlock: Find Usernames Across Social Networks Module
This module contains the main logic to search for usernames at social
networks.
"""
import requests
import csv
import json
import os
import re
from argparse import ArgumentParser, RawDescriptionHelpFormatter
import platform
module_nam... | [
"csv.DictWriter",
"re.search",
"argparse.ArgumentParser",
"requests.get",
"os.path.isfile",
"json.load",
"platform.python_version",
"os.remove"
] | [((3910, 3934), 'os.path.isfile', 'os.path.isfile', (['filename'], {}), '(filename)\n', (3924, 3934), False, 'import os\n'), ((5626, 5647), 'os.path.isfile', 'os.path.isfile', (['fname'], {}), '(fname)\n', (5640, 5647), False, 'import os\n'), ((6915, 7033), 'argparse.ArgumentParser', 'ArgumentParser', ([], {'formatter_... |
# 发现疑似实体,辅助训练
# 用ac自动机构建发现疑似实体的工具
import os
from collections import defaultdict
import json
import re
from .acmation import KeywordTree, add_to_ac, entity_files_folder, entity_folder
from curLine_file import curLine, normal_transformer
domain2entity_map = {}
domain2entity_map["music"] = ["age", "singer", "song", "top... | [
"os.path.exists",
"curLine_file.curLine",
"re.compile",
"os.path.join",
"collections.defaultdict",
"json.load"
] | [((549, 580), 're.compile', 're.compile', (['"""[0-9一二三四五六七八九十拾]+"""'], {}), "('[0-9一二三四五六七八九十拾]+')\n", (559, 580), False, 'import re\n'), ((3134, 3151), 'collections.defaultdict', 'defaultdict', (['list'], {}), '(list)\n', (3145, 3151), False, 'from collections import defaultdict\n'), ((746, 755), 'curLine_file.curLin... |
#!python/bin/python3
"""
Copyright (c) 2018 NSF Center for Space, High-performance, and Resilient Computing (SHREC)
University of Pittsburgh. All rights reserved.
Redistribution and use in source and binary forms, with or without modification, are permitted provided
that the following conditions are met:
1. Redistribu... | [
"os.path.abspath",
"collections.defaultdict",
"importlib.import_module",
"copy.deepcopy"
] | [((1658, 1686), 'importlib.import_module', 'import_module', (['"""src.targets"""'], {}), "('src.targets')\n", (1671, 1686), False, 'from importlib import import_module\n'), ((1784, 1801), 'collections.defaultdict', 'defaultdict', (['tree'], {}), '(tree)\n', (1795, 1801), False, 'from collections import defaultdict\n'),... |
import logging
import pickle
import seaborn as sb
import torch
sb.set_style('whitegrid')
# Class for efficiently handling configurations and parameters, enables to
# easily set them and remember them when one config is reused
# Read with config.key, set with config.update({'key': value}) or config[
# 'key'] = value
... | [
"logging.warning",
"seaborn.set_style",
"torch.cuda.is_available",
"logging.info",
"torch.cuda.current_device",
"torch.linspace"
] | [((65, 90), 'seaborn.set_style', 'sb.set_style', (['"""whitegrid"""'], {}), "('whitegrid')\n", (77, 90), True, 'import seaborn as sb\n'), ((1620, 1669), 'torch.linspace', 'torch.linspace', (['self.t0', 'self.tf', 'self.nb_samples'], {}), '(self.t0, self.tf, self.nb_samples)\n', (1634, 1669), False, 'import torch\n'), (... |