code stringlengths 22 1.05M | apis listlengths 1 3.31k | extract_api stringlengths 75 3.25M |
|---|---|---|
text = """
ala ma kota
a kot ma ale
"""
# ------------------------------------------------------------------------------
# TODO as class
chars = list(sorted(set(text))) # stabilne indeksy
len_chars = len(chars)+1
c_to_i = {c:i+1 for i,c in enumerate(chars)}
i_to_c = {i+1:c for i,c in enumerate(chars)}
def text_to_i... | [
"keras.models.load_model",
"numpy.log",
"numpy.argmax",
"numpy.random.multinomial",
"numpy.exp",
"keras.models.Sequential",
"keras.layers.LSTM",
"numpy.sum",
"keras.layers.Dense",
"keras.optimizers.RMSprop"
] | [((1137, 1149), 'keras.models.Sequential', 'Sequential', ([], {}), '()\n', (1147, 1149), False, 'from keras.models import Sequential, load_model\n'), ((1338, 1364), 'keras.optimizers.RMSprop', 'RMSprop', ([], {'learning_rate': '(0.1)'}), '(learning_rate=0.1)\n', (1345, 1364), False, 'from keras.optimizers import RMSpro... |
import tensorflow as tf
from lime import lime_image
from skimage.segmentation import mark_boundaries
from xplainer.backend.tools.abstract_tool import AbstractTool, GeneralSettings
from xplainer.backend.utils.image import prepare_for_prediction, get_base64png
class Lime(AbstractTool):
def name(self):
retu... | [
"skimage.segmentation.mark_boundaries",
"tensorflow.image.convert_image_dtype",
"xplainer.backend.utils.image.prepare_for_prediction",
"lime.lime_image.LimeImageExplainer",
"xplainer.backend.utils.image.get_base64png",
"tensorflow.cast"
] | [((2660, 2701), 'xplainer.backend.utils.image.prepare_for_prediction', 'prepare_for_prediction', (['model', 'image_path'], {}), '(model, image_path)\n', (2682, 2701), False, 'from xplainer.backend.utils.image import prepare_for_prediction, get_base64png\n'), ((2793, 2825), 'tensorflow.cast', 'tf.cast', (['image'], {'dt... |
#!/usr/bin/env python
# encoding: utf-8
"""
@Author: yangwenhao
@Contact: <EMAIL>
@Software: PyCharm
@File: model.py
@Overview: The deep speaker model is not entirely the same as ResNet, as there are convolutional layers between blocks.
"""
import math
import torch
import torch.nn as nn
from torch.autograd import Fu... | [
"torch.nn.CrossEntropyLoss",
"torch.nn.Sequential",
"torch.sqrt",
"math.sqrt",
"torch.pow",
"torch.nn.init.xavier_normal",
"torch.numel",
"torch.nn.BatchNorm1d",
"torch.sum",
"torch.nn.BatchNorm2d",
"torch.mean",
"Define_Model.Loss.SoftmaxLoss.AngleLinear",
"torch.nn.AdaptiveAvgPool2d",
"t... | [((2456, 2545), 'torch.nn.Conv2d', 'nn.Conv2d', (['in_planes', 'out_planes'], {'kernel_size': '(3)', 'stride': 'stride', 'padding': '(1)', 'bias': '(False)'}), '(in_planes, out_planes, kernel_size=3, stride=stride, padding=1,\n bias=False)\n', (2465, 2545), True, 'import torch.nn as nn\n'), ((805, 823), 'torch.abs',... |
import requests
from datetime import datetime
CLIENT_ID = '2OWPT1X5RNXEE0DGOM5VWO2FBM5R5TNTVPPLH50NCSZAX3QD'
CLIENT_SECRET = '<KEY>'
DATE = datetime.today().strftime('%Y%m%d')
BASE_URL = 'https://api.foursquare.com/v2'
AUTH_URL_PART = 'client_id={}&client_secret={}&v={}'.format(CLIENT_ID, CLIENT_SECRET, DATE)
# Four... | [
"datetime.datetime.today",
"requests.get"
] | [((2138, 2155), 'requests.get', 'requests.get', (['url'], {}), '(url)\n', (2150, 2155), False, 'import requests\n'), ((141, 157), 'datetime.datetime.today', 'datetime.today', ([], {}), '()\n', (155, 157), False, 'from datetime import datetime\n'), ((1832, 1849), 'requests.get', 'requests.get', (['url'], {}), '(url)\n',... |
from django.conf.urls import include, url
from django.contrib import admin
from . import views
urlpatterns = [
url(r'^$', views.ViewView.as_view(), name='profile_own_view'),
url(r'^edit/', views.EditView.as_view(), name='profile_edit'),
url(r'^view/$', views.ViewView.as_view(), name='profile_own_view'),
... | [
"django.conf.urls.url"
] | [((420, 481), 'django.conf.urls.url', 'url', (['"""^delete/$"""', 'views.delete_account'], {'name': '"""delete_account"""'}), "('^delete/$', views.delete_account, name='delete_account')\n", (423, 481), False, 'from django.conf.urls import include, url\n'), ((488, 577), 'django.conf.urls.url', 'url', (['"""^follow/(?P<u... |
import os
import sys
import json
import logging
import requests
class ApiException(Exception):
""" API Exception class. """
pass
logger = logging.getLogger()
logger.setLevel(logging.INFO)
try:
api_url = os.environ['BUS_API_URL']
if api_url[-1] == '/':
api_url = api_url[0:-1]
except KeyError... | [
"logging.getLogger",
"requests.post",
"requests.get",
"sys.exit"
] | [((150, 169), 'logging.getLogger', 'logging.getLogger', ([], {}), '()\n', (167, 169), False, 'import logging\n'), ((464, 496), 'requests.get', 'requests.get', (["(api_url + '/lines')"], {}), "(api_url + '/lines')\n", (476, 496), False, 'import requests\n'), ((709, 745), 'requests.get', 'requests.get', (["(api_url + '/l... |
import os
import os.path as osp
import sys
import numpy as np
from sklearn.svm import LinearSVC
from tqdm import tqdm
sys.path.append(osp.dirname(osp.dirname(osp.abspath(__file__))))
import torch.utils.data as data
from dataset.modelnet40 import LatentCapsulesModelNet40, LatentVectorsModelNet40
from utils.utils impo... | [
"sklearn.svm.LinearSVC",
"os.path.join",
"dataset.modelnet40.LatentCapsulesModelNet40",
"numpy.zeros",
"dataset.modelnet40.LatentVectorsModelNet40",
"torch.utils.data.DataLoader",
"os.path.abspath",
"utils.utils.initialize_main",
"utils.utils.create_save_folder"
] | [((392, 409), 'utils.utils.initialize_main', 'initialize_main', ([], {}), '()\n', (407, 409), False, 'from utils.utils import create_save_folder, initialize_main\n'), ((559, 599), 'os.path.join', 'os.path.join', (['logdir', "args['train_root']"], {}), "(logdir, args['train_root'])\n", (571, 599), False, 'import os\n'),... |
from django.db import models, migrations
class Migration(migrations.Migration):
dependencies = [
('libretto', '0009_auto_20150423_2042'),
]
operations = [
migrations.AlterModelOptions(
name='pupitre',
options={'ordering': ('-soliste', 'partie'), 'verbose_name': 'p... | [
"django.db.migrations.AlterModelOptions",
"django.db.models.ManyToManyField",
"django.db.models.CharField",
"django.db.models.IntegerField"
] | [((187, 348), 'django.db.migrations.AlterModelOptions', 'migrations.AlterModelOptions', ([], {'name': '"""pupitre"""', 'options': "{'ordering': ('-soliste', 'partie'), 'verbose_name': 'pupitre',\n 'verbose_name_plural': 'pupitres'}"}), "(name='pupitre', options={'ordering': (\n '-soliste', 'partie'), 'verbose_nam... |
from unittest.mock import Mock
from pelican.tests.support import unittest
class Test_abbr_role(unittest.TestCase):
def call_it(self, text):
from pelican.rstdirectives import abbr_role
rawtext = text
lineno = 42
inliner = Mock(name='inliner')
nodes, system_messages = abbr_r... | [
"pelican.rstdirectives.abbr_role",
"unittest.mock.Mock"
] | [((260, 280), 'unittest.mock.Mock', 'Mock', ([], {'name': '"""inliner"""'}), "(name='inliner')\n", (264, 280), False, 'from unittest.mock import Mock\n'), ((314, 363), 'pelican.rstdirectives.abbr_role', 'abbr_role', (['"""abbr"""', 'rawtext', 'text', 'lineno', 'inliner'], {}), "('abbr', rawtext, text, lineno, inliner)\... |
# Copyright 2011 <NAME> (<EMAIL>)
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writ... | [
"tmapi.exceptions.ModelConstraintException",
"django.db.models.TextField",
"locator.Locator",
"django.db.models.ForeignKey"
] | [((1017, 1065), 'django.db.models.ForeignKey', 'models.ForeignKey', (['"""Topic"""'], {'related_name': '"""names"""'}), "('Topic', related_name='names')\n", (1034, 1065), False, 'from django.db import models\n'), ((1078, 1096), 'django.db.models.TextField', 'models.TextField', ([], {}), '()\n', (1094, 1096), False, 'fr... |
#!/usr/bin/env python
from __future__ import print_function
from platform import node
from sys import argv
from os import popen
from re import *
def matlabRange(threads):
limits = threads.split(":")
if len(limits)==1:
return range(int(limits[0]),int(limits[0])+1)
if len(limits)==2:
return... | [
"os.popen",
"platform.node"
] | [((1777, 1791), 'os.popen', 'popen', (['command'], {}), '(command)\n', (1782, 1791), False, 'from os import popen\n'), ((1617, 1623), 'platform.node', 'node', ([], {}), '()\n', (1621, 1623), False, 'from platform import node\n')] |
import os
from typing import Dict, Optional
import tomlkit
def _get_project_meta(pyproj_path: str = "./pyproject.toml") -> Dict[str, str]:
if os.path.exists(pyproj_path):
with open(pyproj_path, "r") as pyproject:
file_contents = pyproject.read()
return tomlkit.parse(file_contents)["to... | [
"tomlkit.parse",
"os.path.exists"
] | [((149, 176), 'os.path.exists', 'os.path.exists', (['pyproj_path'], {}), '(pyproj_path)\n', (163, 176), False, 'import os\n'), ((288, 316), 'tomlkit.parse', 'tomlkit.parse', (['file_contents'], {}), '(file_contents)\n', (301, 316), False, 'import tomlkit\n')] |
import datetime as dt
import logging
from collections import Counter, OrderedDict, defaultdict
from dataclasses import dataclass
from io import StringIO
from operator import itemgetter
input = """8
2017-01-03,16:18:50,AAPL,142.64
2017-01-03,16:25:22,AMD,13.86
2017-01-03,16:25:25,AAPL,141.64
2017-01-03,16:25:28,AMZN,84... | [
"logging.getLogger",
"datetime.time",
"collections.Counter",
"collections.defaultdict",
"operator.itemgetter",
"io.StringIO",
"datetime.time.fromisoformat",
"datetime.date.fromisoformat"
] | [((2052, 2104), 'collections.Counter', 'Counter', (['(feed.time.hour for feed in trading_day_feed)'], {}), '(feed.time.hour for feed in trading_day_feed)\n', (2059, 2104), False, 'from collections import Counter, OrderedDict, defaultdict\n'), ((2326, 2375), 'collections.Counter', 'Counter', (['(feed.symbol for feed in ... |
from wiki_scrape_db import get_headlines
headlines = []
for year in list(range(1995, 2020)):
try:
headlines.append(get_headlines(year))
except Exception as e:
print(e)
headlines.append(get_headlines(1994, start_month='July'))
headlines.append(get_headlines(2020, end_month='November'))
with o... | [
"wiki_scrape_db.get_headlines"
] | [((212, 251), 'wiki_scrape_db.get_headlines', 'get_headlines', (['(1994)'], {'start_month': '"""July"""'}), "(1994, start_month='July')\n", (225, 251), False, 'from wiki_scrape_db import get_headlines\n'), ((270, 311), 'wiki_scrape_db.get_headlines', 'get_headlines', (['(2020)'], {'end_month': '"""November"""'}), "(202... |
#!/usr/bin/env python
"""Utility and General purpose functions."""
import inspect
import warnings
import os
import re
import shatter.constants as cts
__author__ = '<NAME>'
def read_file(absolute_path):
"""
:param absolute_path: string path.
:return: list with lines of the file.
"""
return [line... | [
"os.path.exists",
"inspect.stack",
"shatter.constants.PARTICULAR_DEFINITION.pattern.format",
"re.match",
"os.remove",
"re.search"
] | [((463, 487), 'os.path.exists', 'os.path.exists', (['filename'], {}), '(filename)\n', (477, 487), False, 'import os\n'), ((497, 516), 'os.remove', 'os.remove', (['filename'], {}), '(filename)\n', (506, 516), False, 'import os\n'), ((3241, 3265), 're.search', 're.search', (['pattern', 'line'], {}), '(pattern, line)\n', ... |
#!/usr/bin/python
#
import argparse
import os
import pandas as pd
corr = {'pH': [1, 1.02, 1.04, 1.06, 1.08, 1.1, 1.12],
'cl': [1.000000, 1.017857, 1.035714, 1.053571, 1.071429, 1.089286,
1.107143, 1.160714, 1.196429]}
corr = {'pH': [1, 1.02, 1.04, 1.06, 1.08, 1.1],
'cl': [1.000000, 1.... | [
"os.makedirs",
"argparse.ArgumentParser",
"pandas.read_csv",
"os.path.join",
"os.path.isdir"
] | [((490, 538), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {'description': 'description'}), '(description=description)\n', (513, 538), False, 'import argparse\n'), ((1010, 1033), 'pandas.read_csv', 'pd.read_csv', (['args.ffile'], {}), '(args.ffile)\n', (1021, 1033), True, 'import pandas as pd\n'), ((1193,... |
from system import System
from src.basic.sessions.cmd_session import CmdSession
class CmdSystem(System):
def __init__(self):
super(CmdSystem, self).__init__()
@classmethod
def name(cls):
return 'cmd'
def new_session(self, agent, kb):
return CmdSession(agent, kb)
| [
"src.basic.sessions.cmd_session.CmdSession"
] | [((284, 305), 'src.basic.sessions.cmd_session.CmdSession', 'CmdSession', (['agent', 'kb'], {}), '(agent, kb)\n', (294, 305), False, 'from src.basic.sessions.cmd_session import CmdSession\n')] |
# Import necessary modules
from flask import render_template, request, redirect, url_for, flash, Blueprint
from flask_app import app, db
from flask_app.decorators import check_confirmed
from flask_app.models import *
from flask_login import login_required, current_user
from sqlalchemy import or_, and_, func
import rand... | [
"flask.render_template",
"flask.request.args.get",
"random.randint",
"flask.flash",
"flask_app.app.route",
"flask_app.db.session.query",
"requests.get",
"flask.url_for",
"flask.redirect",
"flask_app.app.errorhandler",
"flask_app.db.session.commit",
"flask_app.db.session.add",
"sqlalchemy.fun... | [((410, 437), 'flask.Blueprint', 'Blueprint', (['"""main"""', '__name__'], {}), "('main', __name__)\n", (419, 437), False, 'from flask import render_template, request, redirect, url_for, flash, Blueprint\n'), ((759, 773), 'flask_app.app.route', 'app.route', (['"""/"""'], {}), "('/')\n", (768, 773), False, 'from flask_a... |
from datetime import timedelta
import logging
from typing import Union
from pyschism.enums import (
IofHydroVariables,
IofDvdVariables,
IofWwmVariables,
IofGenVariables,
IofAgeVariables,
IofSedVariables,
IofEcoVariables,
IofIcmVariables,
IofCosVariables,
IofFibVariables,
Iof... | [
"logging.getLogger",
"datetime.timedelta"
] | [((430, 457), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (447, 457), False, 'import logging\n'), ((5495, 5522), 'datetime.timedelta', 'timedelta', ([], {'hours': 'nspool_sta'}), '(hours=nspool_sta)\n', (5504, 5522), False, 'from datetime import timedelta\n'), ((6589, 6616), 'datetime.... |
# -*- coding: utf-8 -*-
from abc import ABCMeta, abstractmethod, abstractproperty
from collections import Sequence
import numpy as np
class MatrixBase(object):
__metaclass__ = ABCMeta
_base_tags = set()
@abstractmethod
def __init__(self, backend, ioshape, iopacking, tags):
self.backend = ... | [
"numpy.unique",
"numpy.any"
] | [((3740, 3757), 'numpy.unique', 'np.unique', (['matmap'], {}), '(matmap)\n', (3749, 3757), True, 'import numpy as np\n'), ((3999, 4021), 'numpy.any', 'np.any', (['(stridemap == 0)'], {}), '(stridemap == 0)\n', (4005, 4021), True, 'import numpy as np\n')] |
import re
from typing import List
from .consts import *
# =================== #
# INTERNALS FUNCTIONS #
# =================== #
def my_re_escape(text):
escape_char = r"[]"
returned_text = ""
for c in text:
if c in escape_char:
returned_text += "\\"
returned_text += c
retur... | [
"re.compile"
] | [((2584, 2687), 're.compile', 're.compile', (['f"""\n {ESCAPED_CSI}\n \\\\d*\n (?:\n ;\\\\d*\n )*\n m\n"""', 're.VERBOSE'], {}), '(\n f"""\n {ESCAPED_CSI}\n \\\\d*\n (?:\n ;\\\\d*\n )*\n m\n""",\n re.VERBOSE)\n', (2594, 2687), False, 'import re\n')] |
from flask import Flask,render_template as render, request
from models import *
app = Flask(__name__,template_folder='./templates')
@app.route('/GETPage',methods = ['GET'])
def Gpage():
return f"You have landed on the page which allows the {request.method} method."
@app.route('/POSTPage', methods = ['GET','P... | [
"flask.render_template",
"json.loads",
"flask.Flask"
] | [((90, 136), 'flask.Flask', 'Flask', (['__name__'], {'template_folder': '"""./templates"""'}), "(__name__, template_folder='./templates')\n", (95, 136), False, 'from flask import Flask, render_template as render, request\n'), ((478, 504), 'flask.render_template', 'render', (['"""SuccessPage.html"""'], {}), "('SuccessPa... |
#!/usr/bin/env python
# -*- coding: utf-8 -*-
# pylint: disable=too-few-public-methods
"""
Web application endpoint
========================
Starts an http endpoint to serve requests
"""
import logging
import mimetypes
import os
import site
import sys
import falcon
from gunicorn.app.base import BaseApplication
logg... | [
"logging.getLogger",
"falcon.API",
"os.path.join",
"os.path.isfile",
"os.path.dirname",
"os.path.isdir",
"mimetypes.guess_type"
] | [((325, 352), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (342, 352), False, 'import logging\n'), ((3569, 3581), 'falcon.API', 'falcon.API', ([], {}), '()\n', (3579, 3581), False, 'import falcon\n'), ((873, 925), 'os.path.join', 'os.path.join', (['sys.prefix', '"""orion-dashboard"""', ... |
from __future__ import division
import argparse, logging, os, math, tqdm
import numpy as np
import mxnet as mx
from mxnet import gluon, nd, image
from mxnet.gluon.data.vision import transforms
import matplotlib.pyplot as plt
import gluoncv as gcv
from gluoncv import data
from gluoncv.data import mscoco
from gluoncv.... | [
"gluoncv.data.transforms.presets.yolo.load_test",
"mxnet.nd.sign",
"gluoncv.data.transforms.pose.get_final_preds",
"numpy.array",
"mxnet.nd.zeros_like",
"gluoncv.data.transforms.pose.get_max_pred",
"argparse.ArgumentParser",
"matplotlib.pyplot.scatter",
"mxnet.nd.array",
"matplotlib.pyplot.ylim",
... | [((415, 502), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {'description': '"""Predict ImageNet classes from a given image"""'}), "(description=\n 'Predict ImageNet classes from a given image')\n", (438, 502), False, 'import argparse, logging, os, math, tqdm\n'), ((2061, 2083), 'mxnet.nd.stack', 'nd.st... |
#! /usr/bin/env python
import sys
import os
import extargsparse
import re
import time
##importdebugstart
sys.path.append(os.path.abspath(os.path.dirname(__file__)))
from strparser import *
from filehdl import *
from fmthdl import *
from extract_ob import *
from obmaklib import *
##importdebugend
REPLACE_IMPORT_LIB=1... | [
"re.split",
"os.path.exists",
"extargsparse.ExtArgsParse",
"os.path.join",
"os.path.dirname",
"sys.stderr.write",
"sys.exit",
"extargsparse.ExtArgsOptions",
"time.time",
"sys.stdout.write"
] | [((1726, 1756), 'extargsparse.ExtArgsOptions', 'extargsparse.ExtArgsOptions', (['d'], {}), '(d)\n', (1753, 1756), False, 'import extargsparse\n'), ((1769, 1780), 'time.time', 'time.time', ([], {}), '()\n', (1778, 1780), False, 'import time\n'), ((1794, 1828), 'extargsparse.ExtArgsParse', 'extargsparse.ExtArgsParse', ([... |
from ..cross_box import CrossBox
from unittest.mock import Mock, patch
from pyglet.gl import GL_LINES
import unittest
class TestCrossBox(unittest.TestCase):
"""Test rendering of cross box graphics."""
def setUp(self):
"""Provides the following to all tests:
* ``self.rectangle``: Mock rectang... | [
"unittest.mock.patch",
"unittest.mock.Mock"
] | [((1298, 1344), 'unittest.mock.patch', 'patch', (['"""engine.graphics.cross_box.vertex_list"""'], {}), "('engine.graphics.cross_box.vertex_list')\n", (1303, 1344), False, 'from unittest.mock import Mock, patch\n'), ((1689, 1735), 'unittest.mock.patch', 'patch', (['"""engine.graphics.cross_box.vertex_list"""'], {}), "('... |
import pandas as pd
from tqdm import tqdm
from ..binarize import to_binary
from cana.boolean_node import BooleanNode
# set up variables
n_inputs = 2**2
n_rules = 2**(2**2)
df_dict = []
for rule in tqdm(range(n_rules)):
canal = {} # becomes row of dataframe
arr = to_binary(rule, digits=4)
print(arr)
... | [
"pandas.DataFrame",
"cana.boolean_node.BooleanNode.from_output_list"
] | [((857, 878), 'pandas.DataFrame', 'pd.DataFrame', (['df_dict'], {}), '(df_dict)\n', (869, 878), True, 'import pandas as pd\n'), ((367, 419), 'cana.boolean_node.BooleanNode.from_output_list', 'BooleanNode.from_output_list', ([], {'outputs': 'arr', 'name': 'rule'}), '(outputs=arr, name=rule)\n', (395, 419), False, 'from ... |
from django.shortcuts import render
from django.views.generic import ListView, CreateView, DetailView
from django.contrib.auth.mixins import LoginRequiredMixin
from posts.models import Post
from posts.forms import PostForm
from django.urls import reverse_lazy
# Create your views here.
class PostListView(LoginRequired... | [
"django.urls.reverse_lazy"
] | [((598, 624), 'django.urls.reverse_lazy', 'reverse_lazy', (['"""posts:list"""'], {}), "('posts:list')\n", (610, 624), False, 'from django.urls import reverse_lazy\n')] |
import cv2
import numpy as np
import random as rd
import os
from tensorflow.keras.models import Sequential, Model
from tensorflow.keras.layers import Dense, Dropout, Activation, Flatten
from tensorflow.keras.layers import Conv2D, MaxPooling2D, Input
from tensorflow.keras.optimizers import Adadelta
from tensorflow.keras... | [
"tensorflow.keras.layers.Conv2D",
"tensorflow.keras.layers.MaxPooling2D",
"os.rename",
"os.path.join",
"numpy.asarray",
"tensorflow.keras.optimizers.Adadelta",
"tensorflow.keras.layers.BatchNormalization",
"tensorflow.keras.callbacks.EarlyStopping",
"random.random",
"tensorflow.keras.layers.Dense"... | [((1236, 1272), 'tensorflow.keras.models.Model', 'Model', (['img_input', 'x'], {'name': '"""calvonet"""'}), "(img_input, x, name='calvonet')\n", (1241, 1272), False, 'from tensorflow.keras.models import Sequential, Model\n'), ((4349, 4388), 'os.path.join', 'os.path.join', (["(output_model_path + '.h5')"], {}), "(output... |
from django.views.generic import ListView, DetailView, TemplateView, CreateView, UpdateView, DeleteView
from board.models import Post
from django.contrib.auth.mixins import LoginRequiredMixin
from django.urls import reverse_lazy
from mysite.views import OwnerOnlyMixin
from django.conf import settings
#--- ListView
c... | [
"board.models.Post.objects.filter",
"django.urls.reverse_lazy"
] | [((702, 729), 'django.urls.reverse_lazy', 'reverse_lazy', (['"""board:index"""'], {}), "('board:index')\n", (714, 729), False, 'from django.urls import reverse_lazy\n'), ((1171, 1198), 'django.urls.reverse_lazy', 'reverse_lazy', (['"""board:index"""'], {}), "('board:index')\n", (1183, 1198), False, 'from django.urls im... |
from django.conf import settings
from django.db import models
from django.dispatch import Signal
from django.utils import timezone
from django.utils.translation import gettext_lazy as _
from froide.foirequest.models import FoiMessage
from .utils import inform_user_problem_resolved
class ProblemChoices(models.TextCh... | [
"django.db.models.TextField",
"django.db.models.ForeignKey",
"django.utils.translation.gettext_lazy",
"django.dispatch.Signal",
"django.db.models.BooleanField",
"django.utils.timezone.now",
"django.db.models.DateTimeField",
"django.db.models.CharField"
] | [((1854, 1862), 'django.dispatch.Signal', 'Signal', ([], {}), '()\n', (1860, 1862), False, 'from django.dispatch import Signal\n'), ((1873, 1881), 'django.dispatch.Signal', 'Signal', ([], {}), '()\n', (1879, 1881), False, 'from django.dispatch import Signal\n'), ((1893, 1901), 'django.dispatch.Signal', 'Signal', ([], {... |
# In The Name of God
# =======================================
# [] File Name : pipe.py
#
# [] Creation Date : 27-11-2019
#
# [] Created By : <NAME> <<EMAIL>>
# =======================================
import os
def child(n, w):
print('I am Child')
f = os.fdopen(w, 'w')
# the old way
f.write('hello %d... | [
"os.fdopen",
"os.fork",
"os.pipe"
] | [((263, 280), 'os.fdopen', 'os.fdopen', (['w', '"""w"""'], {}), "(w, 'w')\n", (272, 280), False, 'import os\n'), ((405, 414), 'os.pipe', 'os.pipe', ([], {}), '()\n', (412, 414), False, 'import os\n'), ((433, 442), 'os.fork', 'os.fork', ([], {}), '()\n', (440, 442), False, 'import os\n'), ((520, 537), 'os.fdopen', 'os.f... |
# -*- coding: utf-8 -*-
import logging
logger = logging.getLogger()
logger.basicConfig = logging.basicConfig(level=logging.DEBUG)
import numpy as np
import matplotlib.pyplot as plt
import logictensornetworks_wrapper as ltnw
nr_samples=500
data=np.random.uniform([0,0],[1.,1.],(nr_samples,2)).astype(np.float32)
data_A... | [
"logging.getLogger",
"logging.basicConfig",
"logictensornetworks_wrapper.constant",
"logictensornetworks_wrapper.variable",
"matplotlib.pyplot.colorbar",
"logictensornetworks_wrapper.train",
"numpy.square",
"logictensornetworks_wrapper.predicate",
"logictensornetworks_wrapper.axiom",
"logictensorn... | [((48, 67), 'logging.getLogger', 'logging.getLogger', ([], {}), '()\n', (65, 67), False, 'import logging\n'), ((89, 129), 'logging.basicConfig', 'logging.basicConfig', ([], {'level': 'logging.DEBUG'}), '(level=logging.DEBUG)\n', (108, 129), False, 'import logging\n'), ((448, 476), 'logictensornetworks_wrapper.variable'... |
from api.subgroups_names import ClassicalSubgroups
from graph_constructor import get_graph
from plotter.graph_plotter import GraphPlotter
from plotter.geodesic_plotter import GeodesicPlotter
from plotter.marker_plotter import MarkerPlotter
from special_polygon import SpecialPolygon
from fimath import Matrix, Field
from... | [
"graph_constructor.get_graph",
"fimath.Matrix.beautify",
"plotter.geodesic_plotter.GeodesicPlotter",
"fimath.Matrix.from_str",
"reduction.Decomposer",
"plotter.marker_plotter.MarkerPlotter",
"special_polygon.SpecialPolygon",
"fimath.Field",
"plotter.graph_plotter.GraphPlotter"
] | [((1674, 1692), 'plotter.graph_plotter.GraphPlotter', 'GraphPlotter', (['axes'], {}), '(axes)\n', (1686, 1692), False, 'from plotter.graph_plotter import GraphPlotter\n'), ((1858, 1885), 'plotter.graph_plotter.GraphPlotter', 'GraphPlotter', ([], {'bokeh_fig': 'fig'}), '(bokeh_fig=fig)\n', (1870, 1885), False, 'from plo... |
"""
Module description:
"""
__version__ = '0.3.1'
__author__ = '<NAME>, <NAME>'
__email__ = '<EMAIL>, <EMAIL>'
import tensorflow as tf
import numpy as np
import random
class Sampler():
def __init__(self, indexed_ratings=None, m=None, num_users=None, num_items=None, transactions=None, batch_size=512, random_seed=... | [
"tensorflow.data.Dataset.from_generator",
"numpy.random.seed",
"random.seed"
] | [((333, 360), 'numpy.random.seed', 'np.random.seed', (['random_seed'], {}), '(random_seed)\n', (347, 360), True, 'import numpy as np\n'), ((369, 393), 'random.seed', 'random.seed', (['random_seed'], {}), '(random_seed)\n', (380, 393), False, 'import random\n'), ((2386, 2498), 'tensorflow.data.Dataset.from_generator', '... |
#! /usr/bin/env python
import socket
def send_ping_data(HOST, PORT):
with socket.socket(socket.AF_INET, socket.SOCK_STREAM) as s:
s.connect((HOST,PORT))
s.sendall(b'Hello, world')
data=s.recv(1024)
return data
if __name__== "__main__" :
print('Received',repr(send_ping_data('127... | [
"socket.socket"
] | [((80, 129), 'socket.socket', 'socket.socket', (['socket.AF_INET', 'socket.SOCK_STREAM'], {}), '(socket.AF_INET, socket.SOCK_STREAM)\n', (93, 129), False, 'import socket\n')] |
# bot.py
import os
import random
import discord
from dotenv import load_dotenv
import commands as cm
import qrys
load_dotenv()
TOKEN = os.getenv('DISCORD_TOKEN')
GUILD = os.getenv('DISCORD_GUILD')
client = discord.Client()
prefix = '$'
@client.event
async def on_ready():
for guild in client.guilds:
if g... | [
"qrys.connect_db",
"os.getenv",
"commands.m_usage",
"dotenv.load_dotenv",
"discord.Client"
] | [((115, 128), 'dotenv.load_dotenv', 'load_dotenv', ([], {}), '()\n', (126, 128), False, 'from dotenv import load_dotenv\n'), ((137, 163), 'os.getenv', 'os.getenv', (['"""DISCORD_TOKEN"""'], {}), "('DISCORD_TOKEN')\n", (146, 163), False, 'import os\n'), ((172, 198), 'os.getenv', 'os.getenv', (['"""DISCORD_GUILD"""'], {}... |
# flake8: noqa
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distrib... | [
"sqlalchemy.true",
"alembic.op.drop_table",
"alembic.op.bulk_insert",
"sqlalchemy.String",
"sqlalchemy.CheckConstraint"
] | [((1195, 1239), 'alembic.op.bulk_insert', 'op.bulk_insert', (['table', "[{'worker_uuid': ''}]"], {}), "(table, [{'worker_uuid': ''}])\n", (1209, 1239), False, 'from alembic import op\n'), ((1277, 1306), 'alembic.op.drop_table', 'op.drop_table', (['RESOURCE_TABLE'], {}), '(RESOURCE_TABLE)\n', (1290, 1306), False, 'from ... |
import torch
import torch.nn.functional as F
from ..models.progressive import ProGANGenerator, ProGANDiscriminator
from ..modules.gan_loss import ImprovedWGANLoss
from ..modules.instance_refiner import InstanceRefiner
from tools.utils import to_cuda
from models import load_network, save_network, print_network
... | [
"torch.mul",
"torch.nn.functional.kl_div",
"torch.nn.functional.mse_loss",
"torch.log",
"torch.nn.functional.l1_loss",
"models.save_network",
"torch.mean",
"torch.Tensor",
"torch.empty_like",
"torch.tensor",
"torch.sum",
"models.load_network",
"torch.nn.functional.interpolate",
"tools.util... | [((2887, 2909), 'tools.utils.to_cuda', 'to_cuda', (['data', '"""z_seg"""'], {}), "(data, 'z_seg')\n", (2894, 2909), False, 'from tools.utils import to_cuda\n'), ((2937, 2961), 'tools.utils.to_cuda', 'to_cuda', (['data', '"""sem_seg"""'], {}), "(data, 'sem_seg')\n", (2944, 2961), False, 'from tools.utils import to_cuda\... |
import os
import sys
TOPIC = 'studio.schoolpower.SchoolPower'
PS_API = 'https://powerschool.mapleleaf.cn'
CACHE_DB_LOCATION = os.environ.get("CACHE_DB_LOCATION", None)
DB_LOCATION = os.environ.get('DB_LOCATION', 'users.db')
PEM_FILE_PATH = os.environ.get("APNS_CERT_FILE", None)
SECRET = os.environ.get("SECRET", "test"... | [
"os.environ.get"
] | [((127, 168), 'os.environ.get', 'os.environ.get', (['"""CACHE_DB_LOCATION"""', 'None'], {}), "('CACHE_DB_LOCATION', None)\n", (141, 168), False, 'import os\n'), ((183, 224), 'os.environ.get', 'os.environ.get', (['"""DB_LOCATION"""', '"""users.db"""'], {}), "('DB_LOCATION', 'users.db')\n", (197, 224), False, 'import os\... |
# Generated by Django 2.1.1 on 2018-10-01 17:34
import django.contrib.postgres.fields
from django.db import migrations, models
import uuid
class Migration(migrations.Migration):
dependencies = [
('custom', '0002_riskfield_risk_type'),
]
operations = [
migrations.AlterModelOptions(
... | [
"django.db.migrations.AlterModelOptions",
"django.db.models.UUIDField",
"django.db.models.CharField"
] | [((285, 434), 'django.db.migrations.AlterModelOptions', 'migrations.AlterModelOptions', ([], {'name': '"""riskfield"""', 'options': "{'ordering': ('id',), 'verbose_name': 'Risk Field', 'verbose_name_plural':\n 'Risk Fields'}"}), "(name='riskfield', options={'ordering': ('id',),\n 'verbose_name': 'Risk Field', 've... |
import os
import sys
import requests
import logging
import time
import json
import pandas as pd
import numpy as np
from concurrent.futures import ProcessPoolExecutor
from git import Git
class FPL_Gameweek:
""" Get the Gameweek state """
def __init__(self, logger, season_data):
"""
Args:
... | [
"logging.basicConfig",
"logging.getLogger",
"os.path.exists",
"pandas.DataFrame",
"os.makedirs",
"git.Git",
"os.path.join",
"requests.get",
"time.sleep",
"numpy.random.randint",
"concurrent.futures.ProcessPoolExecutor",
"json.load"
] | [((8774, 8849), 'logging.basicConfig', 'logging.basicConfig', ([], {'level': 'logging.INFO', 'format': '"""%(asctime)s - %(message)s"""'}), "(level=logging.INFO, format='%(asctime)s - %(message)s')\n", (8793, 8849), False, 'import logging\n'), ((8879, 8906), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}),... |
from chroma_core.lib.storage_plugin.api import attributes
from chroma_core.lib.storage_plugin.api.identifiers import GlobalId, ScopedId
from chroma_core.lib.storage_plugin.api.plugin import Plugin
from chroma_core.lib.storage_plugin.api import resources
from chroma_core.lib.storage_plugin.api import relations
version ... | [
"chroma_core.lib.storage_plugin.api.relations.Provide",
"chroma_core.lib.storage_plugin.api.identifiers.ScopedId",
"chroma_core.lib.storage_plugin.api.attributes.String",
"chroma_core.lib.storage_plugin.api.relations.Subscribe",
"chroma_core.lib.storage_plugin.api.attributes.Integer",
"chroma_core.lib.sto... | [((445, 464), 'chroma_core.lib.storage_plugin.api.attributes.String', 'attributes.String', ([], {}), '()\n', (462, 464), False, 'from chroma_core.lib.storage_plugin.api import attributes\n'), ((572, 591), 'chroma_core.lib.storage_plugin.api.attributes.String', 'attributes.String', ([], {}), '()\n', (589, 591), False, '... |
import json
import time
import logging
import requests
import websocket
from . import abc
from .. import events, sansio, methods, exceptions
LOG = logging.getLogger(__name__)
class SlackAPI(abc.SlackAPI):
"""
`requests` implementation of :class:`slack.io.abc.SlackAPI`
Args:
session: HTTP sessi... | [
"logging.getLogger",
"requests.session",
"json.loads",
"time.sleep",
"websocket.create_connection",
"time.time"
] | [((150, 177), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (167, 177), False, 'import logging\n'), ((720, 752), 'websocket.create_connection', 'websocket.create_connection', (['url'], {}), '(url)\n', (747, 752), False, 'import websocket\n'), ((942, 961), 'time.sleep', 'time.sleep', (['s... |
from django import forms
from student.models import Major, UserProfile
from django.contrib.auth.models import User
class MajorForm(forms.ModelForm):
code = forms.CharField(max_length=20)
viName = forms.CharField(max_length=128)
enName = forms.CharField(max_length=128)
class Meta:
model = Major
fields ... | [
"django.forms.CharField",
"django.forms.DateField",
"django.forms.ChoiceField",
"student.models.Major.objects.all",
"django.forms.IntegerField"
] | [((159, 189), 'django.forms.CharField', 'forms.CharField', ([], {'max_length': '(20)'}), '(max_length=20)\n', (174, 189), False, 'from django import forms\n'), ((200, 231), 'django.forms.CharField', 'forms.CharField', ([], {'max_length': '(128)'}), '(max_length=128)\n', (215, 231), False, 'from django import forms\n'),... |
# -*- coding: utf-8 -*-
# Standard library imports
import sys
# Third party imports
from Qt import QtWidgets
# Local imports
from .ui import Dialog
if __name__ == '__main__':
app = QtWidgets.QApplication(sys.argv)
d = Dialog()
sys.exit(d.exec_())
| [
"Qt.QtWidgets.QApplication"
] | [((190, 222), 'Qt.QtWidgets.QApplication', 'QtWidgets.QApplication', (['sys.argv'], {}), '(sys.argv)\n', (212, 222), False, 'from Qt import QtWidgets\n')] |
from flask import url_for
from tests.conftest import normalize_spaces
def test_set_inbound_sms_sets_a_number_for_service(
logged_in_client,
mock_add_sms_sender,
multiple_available_inbound_numbers,
service_one,
fake_uuid,
mock_no_inbound_number_for_service,
mocker
):
mocker.patch('app.... | [
"flask.url_for"
] | [((504, 576), 'flask.url_for', 'url_for', (['"""main.service_set_inbound_number"""'], {'service_id': "service_one['id']"}), "('main.service_set_inbound_number', service_id=service_one['id'])\n", (511, 576), False, 'from flask import url_for\n')] |
from ..function.node import *
from ..function.tree import *
import numpy as np
def generate_tree(name='test'):
p_branch = .2
p_infertile = .1
p_channel = 1 - p_branch - p_infertile
decay = .25
branch_nodes = [Max, Sum, Mean, Min, Product, Median]
infertile_nodes = [Constant, Input, Uniform, N... | [
"numpy.random.choice"
] | [((1244, 1338), 'numpy.random.choice', 'np.random.choice', (["['branch', 'infertile', 'channel']"], {'p': '[p_branch, p_infertile, p_channel]'}), "(['branch', 'infertile', 'channel'], p=[p_branch,\n p_infertile, p_channel])\n", (1260, 1338), True, 'import numpy as np\n'), ((1396, 1426), 'numpy.random.choice', 'np.ra... |
from grbl import *
from pyb import I2C, delay, millis
from pyb_i2c_lcd import I2cLcd
from lcd_v_minus import *
import time
"""
X
01234567890123456789
X=-xxx.xx F=xxxx* W
Y=-xxx.xx S=xxxx* FM
Z=-xxx.xx Idle. XYZP
message
---------------------
* override
F - feed
S - spindle
W - coordinate W M O
FM - flood mist
... | [
"time.time_ns",
"pyb_i2c_lcd.I2cLcd",
"pyb.I2C",
"time.sleep_ms"
] | [((1057, 1082), 'pyb.I2C', 'I2C', (['i2c_port', 'I2C.MASTER'], {}), '(i2c_port, I2C.MASTER)\n', (1060, 1082), False, 'from pyb import I2C, delay, millis\n'), ((1315, 1343), 'pyb_i2c_lcd.I2cLcd', 'I2cLcd', (['i2c', 'i2c_addr', '(4)', '(20)'], {}), '(i2c, i2c_addr, 4, 20)\n', (1321, 1343), False, 'from pyb_i2c_lcd import... |
import random
from django.core.management.base import BaseCommand, CommandError
from django.conf import settings
from django.core.cache import caches
class Command(BaseCommand):
help = "Loads cache with test objects"
def add_arguments(self, parser):
parser.add_argument("-c", "--cache", nargs="+", typ... | [
"random.choice",
"django.conf.settings.CACHES.keys"
] | [((955, 977), 'django.conf.settings.CACHES.keys', 'settings.CACHES.keys', ([], {}), '()\n', (975, 977), False, 'from django.conf import settings\n'), ((859, 881), 'django.conf.settings.CACHES.keys', 'settings.CACHES.keys', ([], {}), '()\n', (879, 881), False, 'from django.conf import settings\n'), ((1465, 1489), 'rando... |
import os
import tensorflow as tf
from .. import preprocessing
def tf_parse_line(line, data_dir):
line_split = tf.strings.split(line, '\t')
audio_fn = line_split[1]
transcription = line_split[2]
audio_filepath = tf.strings.join([data_dir, 'clips', audio_fn], '/')
wav_filepath = tf.strings.subst... | [
"tensorflow.strings.length",
"tensorflow.strings.split",
"tensorflow.data.TextLineDataset",
"tensorflow.strings.join"
] | [((118, 146), 'tensorflow.strings.split', 'tf.strings.split', (['line', '"""\t"""'], {}), "(line, '\\t')\n", (134, 146), True, 'import tensorflow as tf\n'), ((233, 284), 'tensorflow.strings.join', 'tf.strings.join', (["[data_dir, 'clips', audio_fn]", '"""/"""'], {}), "([data_dir, 'clips', audio_fn], '/')\n", (248, 284)... |
import numpy as np
import math
class virtual_factory(object):
def __init__(self, blade_specs , operation, gating_ct, non_gating_ct, options):
self.options = options
# Blade inputs
self.n_webs = blade_specs['n_webs']
... | [
"numpy.pmt"
] | [((26268, 26325), 'numpy.pmt', 'np.pmt', (['(self.crr / 100.0 / 12.0)', '(life * 12.0)', '(-investment)'], {}), '(self.crr / 100.0 / 12.0, life * 12.0, -investment)\n', (26274, 26325), True, 'import numpy as np\n'), ((24092, 24253), 'numpy.pmt', 'np.pmt', (['(self.crr / 100.0 / 12.0)', 'self.wcp', '(-(self.wcp / 12.0 *... |
from hikari import Permissions
from lightbulb import Context, Check, errors
from datetime import datetime, timedelta
from unicodedata import normalize
from aiohttp import ClientSession
from typing import Union
from os import environ
async def api_call(link: str, headers: dict = None, post: bool = False, json: bool =... | [
"aiohttp.ClientSession",
"lightbulb.errors.ConverterFailure",
"datetime.datetime.utcnow",
"datetime.datetime.now",
"lightbulb.errors.MissingRequiredPermission",
"unicodedata.normalize",
"datetime.timedelta",
"lightbulb.Check"
] | [((1776, 1793), 'lightbulb.Check', 'Check', (['_is_higher'], {}), '(_is_higher)\n', (1781, 1793), False, 'from lightbulb import Context, Check, errors\n'), ((1702, 1761), 'lightbulb.errors.MissingRequiredPermission', 'errors.MissingRequiredPermission', (['Permissions.ADMINISTRATOR'], {}), '(Permissions.ADMINISTRATOR)\n... |
import torchvision
__all__ = ["plot_compare"]
def plot_compare(sr, hr, baseline, filename):
"""
Plot Super-Resolution and High-Resolution image comparison
"""
sr, hr, baseline = sr.squeeze(), hr.squeeze(), baseline.squeeze()
grid = torchvision.utils.make_grid([hr, baseline, sr])
torchvision.... | [
"torchvision.utils.make_grid",
"torchvision.utils.save_image"
] | [((256, 303), 'torchvision.utils.make_grid', 'torchvision.utils.make_grid', (['[hr, baseline, sr]'], {}), '([hr, baseline, sr])\n', (283, 303), False, 'import torchvision\n'), ((308, 352), 'torchvision.utils.save_image', 'torchvision.utils.save_image', (['grid', 'filename'], {}), '(grid, filename)\n', (336, 352), False... |
import torch
from sklearn.model_selection import train_test_split
from torch.utils.data import TensorDataset, DataLoader
def data_loader(targets, labels):
batch_size = 10
train_samples, test_samples, train_labels, test_labels = train_test_split(targets, labels, test_size=0.2)
train_samples = torch.FloatTe... | [
"sklearn.model_selection.train_test_split",
"torch.LongTensor",
"torch.utils.data.TensorDataset",
"torch.utils.data.DataLoader",
"torch.FloatTensor"
] | [((238, 286), 'sklearn.model_selection.train_test_split', 'train_test_split', (['targets', 'labels'], {'test_size': '(0.2)'}), '(targets, labels, test_size=0.2)\n', (254, 286), False, 'from sklearn.model_selection import train_test_split\n'), ((307, 339), 'torch.FloatTensor', 'torch.FloatTensor', (['train_samples'], {}... |
import numbers
import xnmt.tensor_tools as tt
import xnmt.modelparts.decoders as decoders
import xnmt.transducers.recurrent as recurrent
import xnmt.transducers.base as transducers_base
import xnmt.expression_seqs as expr_seq
import xnmt.vocabs as vocabs
class SimultaneousState(decoders.AutoRegressiveDecoderState):
... | [
"xnmt.expression_seqs.ExpressionSequence",
"xnmt.transducers.base.FinalTransducerState"
] | [((1684, 1727), 'xnmt.transducers.base.FinalTransducerState', 'transducers_base.FinalTransducerState', (['h', 'c'], {}), '(h, c)\n', (1721, 1727), True, 'import xnmt.transducers.base as transducers_base\n'), ((2230, 2281), 'xnmt.expression_seqs.ExpressionSequence', 'expr_seq.ExpressionSequence', ([], {'expr_list': 'src... |
#!/usr/bin/env python
# -*- coding: utf-8 -*-
# @Time : 2019/12/27 下午6:04
# @Title : 83. 删除排序链表中的重复元素
# @Link : https://leetcode-cn.com/problems/remove-duplicates-from-sorted-list/
QUESTION = """
给定一个排序链表,删除所有重复的元素,使得每个元素只出现一次。
示例 1:
输入: 1->1->2
输出: 1->2
示例 2:
输入: 1->1->2->3->3
输出: 1->2->3
"""
THINKING = ... | [
"utils.linked_list.LinkedListGen.nodes_to_list",
"utils.linked_list.LinkedListGen.list_to_nodes"
] | [((1210, 1254), 'utils.linked_list.LinkedListGen.list_to_nodes', 'LinkedListGen.list_to_nodes', (['[1, 1, 2, 3, 3]'], {}), '([1, 1, 2, 3, 3])\n', (1237, 1254), False, 'from utils.linked_list import LinkedListGen\n'), ((1312, 1356), 'utils.linked_list.LinkedListGen.nodes_to_list', 'LinkedListGen.nodes_to_list', (['dupli... |
import numpy as np
import pickle as pkl
from envs.babyai.oracle.teacher import Teacher
class XYCorrections(Teacher):
def __init__(self, *args, **kwargs):
super(XYCorrections, self).__init__(*args, **kwargs)
self.next_state_coords = self.empty_feedback()
def empty_feedback(self):
"""
... | [
"pickle.dumps",
"numpy.zeros",
"numpy.concatenate",
"numpy.random.uniform"
] | [((541, 572), 'numpy.random.uniform', 'np.random.uniform', (['(0)', '(1)'], {'size': '(8)'}), '(0, 1, size=8)\n', (558, 572), True, 'import numpy as np\n'), ((907, 947), 'numpy.concatenate', 'np.concatenate', (['[self.next_state_coords]'], {}), '([self.next_state_coords])\n', (921, 947), True, 'import numpy as np\n'), ... |
TEST_TEMP_RAW = 529191
TEST_TEMP_CMP = 24.7894877676
TEST_PRES_RAW = 326816
TEST_PRES_CMP = 1006.61517564
TEST_ALT_CMP = 57.3174
def test_temperature():
from tools import SMBusFakeDevice
from bmp280 import BMP280
from calibration import BMP280Calibration
dev = SMBusFakeDevice(1)
# Load the fake ... | [
"calibration.BMP280Calibration",
"bmp280.BMP280",
"tools.SMBusFakeDevice"
] | [((280, 298), 'tools.SMBusFakeDevice', 'SMBusFakeDevice', (['(1)'], {}), '(1)\n', (295, 298), False, 'from tools import SMBusFakeDevice\n'), ((530, 549), 'bmp280.BMP280', 'BMP280', ([], {'i2c_dev': 'dev'}), '(i2c_dev=dev)\n', (536, 549), False, 'from bmp280 import BMP280\n'), ((654, 673), 'calibration.BMP280Calibration... |
# mnist example, Downloaded from PML github
import torch
import torch.nn as nn
import torch.nn.functional as F
import torch.optim as optim
### MNIST code originally from https://github.com/pytorch/examples/blob/master/mnist/main.py ###
from torchvision import datasets, transforms
from pytorch_metric_learning import d... | [
"torch.device",
"torch.nn.Dropout2d",
"torch.nn.Conv2d",
"pytorch_metric_learning.utils.accuracy_calculator.AccuracyCalculator",
"torchvision.datasets.MNIST",
"torch.utils.data.DataLoader",
"torch.nn.Linear",
"torch.nn.functional.relu",
"torch.nn.functional.max_pool2d",
"torchvision.transforms.Nor... | [((2839, 2859), 'torch.device', 'torch.device', (['"""cuda"""'], {}), "('cuda')\n", (2851, 2859), False, 'import torch\n'), ((2997, 3064), 'torchvision.datasets.MNIST', 'datasets.MNIST', (['"""."""'], {'train': '(True)', 'download': '(True)', 'transform': 'transform'}), "('.', train=True, download=True, transform=trans... |
import time
while True:
print("Test program")
time.sleep(10)
pass | [
"time.sleep"
] | [((54, 68), 'time.sleep', 'time.sleep', (['(10)'], {}), '(10)\n', (64, 68), False, 'import time\n')] |
#!/usr/bin/env python3.6
# -*- mode: python -*-
# =============================================================================
# @@-COPYRIGHT-START-@@
#
# Copyright (c) 2020 of Qualcomm Innovation Center, Inc. All rights reserved.
#
# @@-COPYRIGHT-END-@@
# ===========================================================... | [
"logging.getLogger",
"tensorflow.shape",
"pycocotools.cocoeval.COCOeval",
"tensorflow.transpose",
"tensorflow.data.Dataset.list_files",
"tensorflow.contrib.slim.tfexample_decoder.TFExampleDecoder",
"tensorflow.cast",
"aimet_tensorflow.quantsim.QuantizationSimModel",
"tensorflow.variables_initializer... | [((1760, 1787), 'logging.getLogger', 'logging.getLogger', (['__file__'], {}), '(__file__)\n', (1777, 1787), False, 'import logging\n'), ((2079, 2111), 'tensorflow.GPUOptions', 'tf.GPUOptions', ([], {'allow_growth': '(True)'}), '(allow_growth=True)\n', (2092, 2111), True, 'import tensorflow as tf\n'), ((2125, 2191), 'te... |
import torch
from torchvision import datasets
import shutil
import argparse
import os
import numpy as np
from tqdm import tqdm
########### Help ###########
'''
#size = (h,w)
python split_train_val.py \
--data_dir /Users/aman.gupta/Documents/self/datasets/blank_page_detection/letterbox_training_data/ \
--val... | [
"os.makedirs",
"argparse.ArgumentParser",
"tqdm.tqdm",
"numpy.floor",
"os.path.join",
"torchvision.datasets.ImageFolder",
"os.path.basename",
"shutil.copy",
"numpy.random.shuffle"
] | [((524, 662), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {'description': '"""this script splits classification data into train and val based on ratio provided by user"""'}), "(description=\n 'this script splits classification data into train and val based on ratio provided by user'\n )\n", (547, 6... |
""" Yahoo Finance Model """
__docformat__ = "numpy"
import logging
import pandas as pd
import yfinance as yf
from gamestonk_terminal.decorators import log_start_end
from gamestonk_terminal.rich_config import console
logger = logging.getLogger(__name__)
INDICES = {
"sp500": {"name": "S&P 500", "ticker": "^GSPC"... | [
"logging.getLogger",
"pandas.Series",
"gamestonk_terminal.decorators.log_start_end",
"gamestonk_terminal.rich_config.console.print",
"yfinance.download"
] | [((229, 256), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (246, 256), False, 'import logging\n'), ((1064, 1089), 'gamestonk_terminal.decorators.log_start_end', 'log_start_end', ([], {'log': 'logger'}), '(log=logger)\n', (1077, 1089), False, 'from gamestonk_terminal.decorators import lo... |
# Simple XML against XSD Validator for Python 2.7 - 3.2
# to run this script you need additionally: lxml (http://lxml.de)
# author: <NAME>, 2013
import sys
from lxml import etree
xsd_files = []
xml_files = []
def usage():
print("Usage: ")
print("python XSDValidator.py <list of xml files> <list of xsd file... | [
"lxml.etree.XMLSchema",
"lxml.etree.parse",
"sys.exit"
] | [((525, 553), 'lxml.etree.XMLSchema', 'etree.XMLSchema', ([], {'file': 'schema'}), '(file=schema)\n', (540, 553), False, 'from lxml import etree\n'), ((1019, 1029), 'sys.exit', 'sys.exit', ([], {}), '()\n', (1027, 1029), False, 'import sys\n'), ((1284, 1294), 'sys.exit', 'sys.exit', ([], {}), '()\n', (1292, 1294), Fals... |
from __future__ import unicode_literals
from django.db import models
from django.utils.encoding import python_2_unicode_compatible
from zshoes.stores.models import Store
@python_2_unicode_compatible
class Article(models.Model):
"""
Entity that represents the articles of the store
"""
#: Name of the ... | [
"django.db.models.FloatField",
"django.db.models.TextField",
"django.db.models.ForeignKey",
"django.db.models.PositiveIntegerField",
"django.db.models.CharField"
] | [((339, 370), 'django.db.models.CharField', 'models.CharField', ([], {'max_length': '(45)'}), '(max_length=45)\n', (355, 370), False, 'from django.db import models\n'), ((423, 462), 'django.db.models.TextField', 'models.TextField', ([], {'null': '(True)', 'blank': '(True)'}), '(null=True, blank=True)\n', (439, 462), Fa... |
#!/usr/bin/env python3
import sys, os, json
import random
# Check to make sure we are running the correct version of Python
assert sys.version_info >= (3,7), "This script requires at least Python 3.7"
# The game and item description files (in the same folder as this script)
game_file = 'game.json'
# Load the content... | [
"os.path.join",
"os.getcwd",
"os.path.dirname",
"os._exit",
"json.load"
] | [((693, 713), 'json.load', 'json.load', (['json_file'], {}), '(json_file)\n', (702, 713), False, 'import sys, os, json\n'), ((829, 840), 'os._exit', 'os._exit', (['(1)'], {}), '(1)\n', (837, 840), False, 'import sys, os, json\n'), ((574, 585), 'os.getcwd', 'os.getcwd', ([], {}), '()\n', (583, 585), False, 'import sys, ... |
"""
This module provides a sequence class which can be used for cyclic values
"""
from typing import Generic, TypeVar
T = TypeVar('T')
class Cycle(list,Generic[T]):
"""This class can be used to store cyclic values"""
def __getitem__(self,key:int) -> T:
return super().__getitem__(key%len(self))
... | [
"typing.TypeVar"
] | [((124, 136), 'typing.TypeVar', 'TypeVar', (['"""T"""'], {}), "('T')\n", (131, 136), False, 'from typing import Generic, TypeVar\n')] |
# -*- coding: utf-8 -*-
# Generated by Django 1.9.4 on 2016-03-17 12:32
from __future__ import unicode_literals
from django.db import migrations, models
class Migration(migrations.Migration):
initial = True
dependencies = [
]
operations = [
migrations.CreateModel(
name='orders'... | [
"django.db.models.DateField",
"django.db.models.AutoField",
"django.db.models.PositiveIntegerField",
"django.db.models.DecimalField",
"django.db.models.CharField"
] | [((366, 459), 'django.db.models.AutoField', 'models.AutoField', ([], {'auto_created': '(True)', 'primary_key': '(True)', 'serialize': '(False)', 'verbose_name': '"""ID"""'}), "(auto_created=True, primary_key=True, serialize=False,\n verbose_name='ID')\n", (382, 459), False, 'from django.db import migrations, models\... |
###############################################
# <NAME> - PG Applied AI - Programming
# Unit tests, for the graph algorithms
###############################################
import unittest # unit testing ftw
from board import Board
import numpy as np
import play
class TestMethods(unittest.TestCas... | [
"board.Board"
] | [((379, 387), 'board.Board', 'Board', (['(4)'], {}), '(4)\n', (384, 387), False, 'from board import Board\n'), ((686, 694), 'board.Board', 'Board', (['(4)'], {}), '(4)\n', (691, 694), False, 'from board import Board\n'), ((991, 999), 'board.Board', 'Board', (['(4)'], {}), '(4)\n', (996, 999), False, 'from board import ... |
import dash_bootstrap_components as dbc
import dash_core_components as dcc
import dash_html_components as html
from dash.dependencies import Input, Output
from . import dash_managers, dash_queue, dash_service
from .app import app
from .navbar import navbar
body = dbc.Container(
[
dbc.Row([
dbc... | [
"dash_bootstrap_components.Button",
"dash.dependencies.Output",
"dash_core_components.Location",
"dash.dependencies.Input",
"dash_html_components.H2",
"dash_core_components.Graph",
"dash_html_components.P",
"dash_html_components.Div"
] | [((1018, 1052), 'dash.dependencies.Output', 'Output', (['"""page-content"""', '"""children"""'], {}), "('page-content', 'children')\n", (1024, 1052), False, 'from dash.dependencies import Input, Output\n'), ((933, 970), 'dash_core_components.Location', 'dcc.Location', ([], {'id': '"""url"""', 'refresh': '(False)'}), "(... |
# -*- coding: utf-8 -*-
"""
Created on Mon Apr 8 11:09:33 2019
@author: 10365
"""
#CreateDataSet
import numpy as np
import sys
sys.path.append('../subway_system')
sys.path.append('../ato_agent')
import TrainAndRoadCharacter as trc
import trainRunningModel as trm
import pandas as pds
import matplotlib.pyplot as pl... | [
"numpy.mat",
"TrainAndRoadCharacter.getRoadGradinet",
"trainRunningModel.Train_model",
"TrainAndRoadCharacter.TrainAndRoadData",
"pandas.DataFrame",
"atoController.PidController",
"TrainAndRoadCharacter.getNextSpeedLimit",
"TrainAndRoadCharacter.plotSpeedLimitRoadGrad",
"matplotlib.pyplot.plot",
"... | [((133, 168), 'sys.path.append', 'sys.path.append', (['"""../subway_system"""'], {}), "('../subway_system')\n", (148, 168), False, 'import sys\n'), ((169, 200), 'sys.path.append', 'sys.path.append', (['"""../ato_agent"""'], {}), "('../ato_agent')\n", (184, 200), False, 'import sys\n'), ((1810, 1843), 'TrainAndRoadChara... |
#!/usr/bin/python3
import boto3
import json
if __name__ == '__main__':
ec2_client = boto3.client('ec2')
ec2_filter = [{'Name': 'tag:role', 'Values': ['ecs-cluster']}]
instances=ec2_client.describe_tags(Filters=ec2_filter)
#get only the instance_ids
instance_ids = []
for i in instances['Tags'... | [
"json.dumps",
"boto3.client"
] | [((91, 110), 'boto3.client', 'boto3.client', (['"""ec2"""'], {}), "('ec2')\n", (103, 110), False, 'import boto3\n'), ((882, 905), 'json.dumps', 'json.dumps', (['output_dict'], {}), '(output_dict)\n', (892, 905), False, 'import json\n')] |
from cstream import stdwar
from svgen import Point, Vector, SVG, Figure, Camera, Domain, Map, Surface, Animation
from math import radians, sin, cos, pi, hypot, sqrt
import sys
from tqdm import tqdm
from svgen.svglib.math import Transform
COLOR = "#3703b3"
if len(sys.argv) > 1 and sys.argv[1] == "-o":
proj = C... | [
"svgen.Domain",
"svgen.Surface",
"math.radians",
"svgen.Animation",
"math.cos",
"svgen.svglib.math.Transform.scale",
"math.sin",
"svgen.SVG"
] | [((557, 578), 'svgen.Animation', 'Animation', ([], {'delay': '(1000)'}), '(delay=1000)\n', (566, 578), False, 'from svgen import Point, Vector, SVG, Figure, Camera, Domain, Map, Surface, Animation\n'), ((584, 616), 'svgen.Domain', 'Domain', (['(-250, 250)', '(250, -250)'], {}), '((-250, 250), (250, -250))\n', (590, 616... |
import sys
sys.path.insert(0, '../')
from mocap.settings import get_amass_validation_files, get_amass_test_files
from mocap.math.amass_fk import rotmat2euclidean, exp2euclidean
from mocap.visualization.sequence import SequenceVisualizer
from mocap.math.mirror_smpl import mirror_p3d
from mocap.datasets.dataset import Li... | [
"mocap.datasets.combined.Combined",
"sys.path.insert",
"mocap.settings.get_amass_validation_files",
"mocap.datasets.h36m.H36M_FixedSkeleton",
"mocap.datasets.amass.AMASS_SMPL3d",
"numpy.array",
"mocap.settings.get_amass_test_files",
"mocap.visualization.sequence.SequenceVisualizer"
] | [((11, 36), 'sys.path.insert', 'sys.path.insert', (['(0)', '"""../"""'], {}), "(0, '../')\n", (26, 36), False, 'import sys\n'), ((617, 645), 'mocap.settings.get_amass_validation_files', 'get_amass_validation_files', ([], {}), '()\n', (643, 645), False, 'from mocap.settings import get_amass_validation_files, get_amass_t... |
from django.test import TestCase
from review.models import Review
class TestReviewModel(TestCase):
'''
Test suite for review modules.
'''
def setUp(self):
'''
Set up test data for the review model.
'''
Review.objects.create(
feedback='Test rev... | [
"review.models.Review.objects.all",
"review.models.Review.objects.create",
"review.models.Review.objects.get"
] | [((266, 351), 'review.models.Review.objects.create', 'Review.objects.create', ([], {'feedback': '"""Test review"""', 'riderReview': '"""Test review content"""'}), "(feedback='Test review', riderReview='Test review content'\n )\n", (287, 351), False, 'from review.models import Review\n'), ((651, 693), 'review.models.... |
"""Script to start webserving."""
from wembedder.app import create_app
app = create_app()
if __name__ == '__main__':
app.run(debug=True)
| [
"wembedder.app.create_app"
] | [((81, 93), 'wembedder.app.create_app', 'create_app', ([], {}), '()\n', (91, 93), False, 'from wembedder.app import create_app\n')] |
from openprocurement.tender.core.procedure.serializers.base import ListSerializer
from openprocurement.tender.core.procedure.serializers.document import ConfidentialDocumentSerializer
from openprocurement.tender.core.procedure.serializers.parameter import ParameterSerializer
from openprocurement.tender.esco.procedure.s... | [
"openprocurement.tender.core.procedure.serializers.base.ListSerializer"
] | [((667, 701), 'openprocurement.tender.core.procedure.serializers.base.ListSerializer', 'ListSerializer', (['LotValueSerializer'], {}), '(LotValueSerializer)\n', (681, 701), False, 'from openprocurement.tender.core.procedure.serializers.base import ListSerializer\n'), ((724, 770), 'openprocurement.tender.core.procedure.... |
from typing import List
import argparse
from detectron2.evaluation import COCOEvaluator, inference_on_dataset
from detectron2.config import get_cfg
from detectron2 import model_zoo
from detectron2.data.datasets import register_coco_instances
from detectron2.data import build_detection_test_loader
from trainers import... | [
"trainers.MyTrainer",
"detectron2.data.datasets.register_coco_instances",
"detectron2.config.get_cfg",
"argparse.ArgumentParser",
"detectron2.model_zoo.get_config_file",
"detectron2.evaluation.inference_on_dataset",
"detectron2.data.build_detection_test_loader",
"detectron2.evaluation.COCOEvaluator"
] | [((356, 435), 'detectron2.data.datasets.register_coco_instances', 'register_coco_instances', (['args.ds_test', '{}', 'args.ds_test_json', 'args.ds_test_imgs'], {}), '(args.ds_test, {}, args.ds_test_json, args.ds_test_imgs)\n', (379, 435), False, 'from detectron2.data.datasets import register_coco_instances\n'), ((446, ... |
# -*- coding: utf-8 -*-
# Form implementation generated from reading ui file 'npc/gui/uis/new_character.ui'
#
# Created by: PyQt5 UI code generator 5.7.1
#
# WARNING! All changes made in this file will be lost!
from PyQt5 import QtCore, QtGui, QtWidgets
class Ui_NewCharacterDialog(object):
def setupUi(self, NewC... | [
"PyQt5.QtWidgets.QDialogButtonBox",
"PyQt5.QtWidgets.QComboBox",
"PyQt5.QtWidgets.QPlainTextEdit",
"PyQt5.QtCore.QMetaObject.connectSlotsByName",
"PyQt5.QtWidgets.QVBoxLayout",
"PyQt5.QtWidgets.QLabel",
"PyQt5.QtWidgets.QGroupBox",
"PyQt5.QtWidgets.QSizePolicy",
"PyQt5.QtWidgets.QFormLayout",
"PyQ... | [((465, 571), 'PyQt5.QtWidgets.QSizePolicy', 'QtWidgets.QSizePolicy', (['QtWidgets.QSizePolicy.MinimumExpanding', 'QtWidgets.QSizePolicy.MinimumExpanding'], {}), '(QtWidgets.QSizePolicy.MinimumExpanding, QtWidgets.\n QSizePolicy.MinimumExpanding)\n', (486, 571), False, 'from PyQt5 import QtCore, QtGui, QtWidgets\n')... |
#!/usr/bin/env python3
# -*- coding: utf-8 -*-
# Contains a context manager for temporarily introducing an environment var.
import os
import contextlib
@contextlib.contextmanager
def use_environment_variable(key, value):
""" Used to temporarily introduce a new environment variable as if it
was set by th... | [
"os.environ.pop"
] | [((617, 636), 'os.environ.pop', 'os.environ.pop', (['key'], {}), '(key)\n', (631, 636), False, 'import os\n')] |
import matplotlib
import matplotlib.pyplot as plt
import numpy as np
from PySide2.QtWidgets import QVBoxLayout, QWidget
from traitlets import HasTraits, Instance, Bool, directional_link
from regexport.model import AppState
from regexport.views.utils import HasWidget
matplotlib.use('Qt5Agg')
from matplotlib.backends... | [
"traitlets.directional_link",
"matplotlib.backends.backend_qt5agg.NavigationToolbar2QT",
"numpy.histogram",
"regexport.views.utils.HasWidget.__init__",
"matplotlib.use",
"traitlets.Instance",
"PySide2.QtWidgets.QWidget",
"numpy.concatenate",
"matplotlib.backends.backend_qt5agg.FigureCanvasQTAgg",
... | [((269, 293), 'matplotlib.use', 'matplotlib.use', (['"""Qt5Agg"""'], {}), "('Qt5Agg')\n", (283, 293), False, 'import matplotlib\n'), ((470, 507), 'traitlets.Instance', 'Instance', (['np.ndarray'], {'allow_none': '(True)'}), '(np.ndarray, allow_none=True)\n', (478, 507), False, 'from traitlets import HasTraits, Instance... |
"""
Run all the example files and convert them to markdown files containing the output.
Uses `pweave`. It is not installed by default. To install:
pip install pweave
"""
import pweave, datetime, glob, os
def publish_to_markdown(python_file: str, output_file: str):
doc = pweave.Pweb(python_file, kernel="pyt... | [
"datetime.date.today",
"os.path.basename",
"glob.glob",
"pweave.Pweb"
] | [((284, 371), 'pweave.Pweb', 'pweave.Pweb', (['python_file'], {'kernel': '"""python3"""', 'doctype': '"""markdown"""', 'output': 'output_file'}), "(python_file, kernel='python3', doctype='markdown', output=\n output_file)\n", (295, 371), False, 'import pweave, datetime, glob, os\n'), ((841, 858), 'glob.glob', 'glob.... |
import binascii
import uuid
from collections import UserDict
from functools import cmp_to_key, wraps
from nanolib import Block as RawBlock
from nanolib import nbase32_to_bytes, get_account_id
__all__ = (
"RawBlock", "BlockProxy", "Callbacks", "CallbackSlot", "AccountIDDict"
)
class BlockProxy(object):
"""
... | [
"binascii.hexlify",
"nanolib.nbase32_to_bytes",
"uuid.uuid4"
] | [((4989, 5025), 'nanolib.nbase32_to_bytes', 'nbase32_to_bytes', (['account_id[-60:-8]'], {}), '(account_id[-60:-8])\n', (5005, 5025), False, 'from nanolib import nbase32_to_bytes, get_account_id\n'), ((3431, 3443), 'uuid.uuid4', 'uuid.uuid4', ([], {}), '()\n', (3441, 3443), False, 'import uuid\n'), ((6761, 6782), 'bina... |
import json
from copy import deepcopy
from functools import total_ordering
from typing import List, Any, Union
from canvasxpress.config.type import CXConfig, CXString, CXInt, CXFloat, CXBool, \
CXList, CXDict, CXRGBColor, CXRGBAColor
from canvasxpress.data.convert import CXDictConvertable, CXListConvertable
@tot... | [
"canvasxpress.config.type.CXRGBColor",
"copy.deepcopy",
"canvasxpress.config.type.CXRGBAColor.is_color_list",
"canvasxpress.config.type.CXString",
"canvasxpress.config.type.CXRGBAColor.is_color_dict",
"canvasxpress.config.type.CXList",
"canvasxpress.config.type.CXInt",
"canvasxpress.config.type.CXBool... | [((6342, 6361), 'canvasxpress.config.type.CXInt', 'CXInt', (['label', 'value'], {}), '(label, value)\n', (6347, 6361), False, 'from canvasxpress.config.type import CXConfig, CXString, CXInt, CXFloat, CXBool, CXList, CXDict, CXRGBColor, CXRGBAColor\n'), ((12090, 12106), 'copy.deepcopy', 'deepcopy', (['config'], {}), '(c... |
# Generated by Django 3.1.5 on 2021-01-17 15:24
import cloudinary.models
from django.conf import settings
from django.db import migrations, models
import django.db.models.deletion
class Migration(migrations.Migration):
initial = True
dependencies = [
migrations.swappable_dependency(settings.AUTH_US... | [
"django.db.models.OneToOneField",
"django.db.models.TextField",
"django.db.models.ForeignKey",
"django.db.models.IntegerField",
"django.db.models.ManyToManyField",
"django.db.models.AutoField",
"django.db.models.DateTimeField",
"django.db.migrations.swappable_dependency",
"django.db.models.CharField... | [((272, 329), 'django.db.migrations.swappable_dependency', 'migrations.swappable_dependency', (['settings.AUTH_USER_MODEL'], {}), '(settings.AUTH_USER_MODEL)\n', (303, 329), False, 'from django.db import migrations, models\n'), ((461, 554), 'django.db.models.AutoField', 'models.AutoField', ([], {'auto_created': '(True)... |
############################################################################
#
# Copyright (C) 2016 The Qt Company Ltd.
# Contact: https://www.qt.io/licensing/
#
# This file is part of Qt Creator.
#
# Commercial License Usage
# Licensees holding valid commercial Qt licenses may use this file in
# accordance with the co... | [
"__builtin__.bool"
] | [((4444, 4466), '__builtin__.bool', '__builtin__.bool', (['clip'], {}), '(clip)\n', (4460, 4466), False, 'import __builtin__\n')] |
# MIT License
# Copyright (c) 2018-2019 <NAME>, <NAME>, <NAME>, <NAME>
# Permission is hereby granted, free of charge, to any person obtaining a copy
# of this software and associated documentation files (the "Software"), to deal
# in the Software without restriction, including without limitation the rights
# to use... | [
"mathutils.Vector"
] | [((2668, 2756), 'mathutils.Vector', 'Vector', (['(rcurve.Arc.StartPoint.X, rcurve.Arc.StartPoint.Y, rcurve.Arc.StartPoint.Z)'], {}), '((rcurve.Arc.StartPoint.X, rcurve.Arc.StartPoint.Y, rcurve.Arc.\n StartPoint.Z))\n', (2674, 2756), False, 'from mathutils import Vector\n'), ((2770, 2847), 'mathutils.Vector', 'Vector... |
#!/usr/bin/env python3
import sys
import subprocess
from align_videos_by_soundtrack.align import SyncDetector, cli_common
remove = []
remove_specified = False
remove_all = False
add = []
add_specified = False
copy_subtitles = False
files = []
print_offset_only = False
for argument in sys.argv[1:]:
if argument == ... | [
"align_videos_by_soundtrack.align.cli_common.logger_config",
"align_videos_by_soundtrack.align.SyncDetector",
"subprocess.check_call"
] | [((909, 935), 'align_videos_by_soundtrack.align.cli_common.logger_config', 'cli_common.logger_config', ([], {}), '()\n', (933, 935), False, 'from align_videos_by_soundtrack.align import SyncDetector, cli_common\n'), ((942, 956), 'align_videos_by_soundtrack.align.SyncDetector', 'SyncDetector', ([], {}), '()\n', (954, 95... |
import numpy as np
import pandas as pd
import sys
import tensorflow as tf
from keras.layers import Input, Dense, Lambda, Flatten, Reshape, Activation, Dropout, Add, TimeDistributed, Multiply, Conv1D, Conv2D, MaxPooling1D, AveragePooling1D
from keras.models import Model, Sequential, load_model
from keras import backend ... | [
"keras.models.load_model",
"keras.layers.AveragePooling1D",
"pandas.read_csv",
"keras.callbacks.ModelCheckpoint",
"keras.callbacks.History",
"keras.layers.Lambda",
"pandas.DataFrame.from_dict",
"keras.layers.Add",
"keras.layers.Input",
"keras.layers.Multiply",
"keras.models.Model",
"keras.laye... | [((4182, 4234), 'keras.layers.Input', 'Input', ([], {'shape': 'self.input_shape', 'name': '"""original_input"""'}), "(shape=self.input_shape, name='original_input')\n", (4187, 4234), False, 'from keras.layers import Input, Dense, Lambda, Flatten, Reshape, Activation, Dropout, Add, TimeDistributed, Multiply, Conv1D, Con... |
# Generated by Django 3.1.6 on 2021-03-08 10:44
from django.db import migrations, models
class Migration(migrations.Migration):
initial = True
dependencies = [
]
operations = [
migrations.CreateModel(
name='Album',
fields=[
('id', models.AutoField(au... | [
"django.db.models.DateTimeField",
"django.db.models.AutoField",
"django.db.models.CharField"
] | [((301, 394), 'django.db.models.AutoField', 'models.AutoField', ([], {'auto_created': '(True)', 'primary_key': '(True)', 'serialize': '(False)', 'verbose_name': '"""ID"""'}), "(auto_created=True, primary_key=True, serialize=False,\n verbose_name='ID')\n", (317, 394), False, 'from django.db import migrations, models\... |
#!/usr/bin/env python3
import platform
import serial
import sys
from config import Settings
dev = serial.Serial(Settings.SERIAL_DEVICE, Settings.BAUD_RATE)
print("> Returned data:", file=sys.stderr)
while True:
x = dev.read()
sys.stdout.buffer.write(x)
sys.stdout.flush()
| [
"sys.stdout.flush",
"sys.stdout.buffer.write",
"serial.Serial"
] | [((99, 156), 'serial.Serial', 'serial.Serial', (['Settings.SERIAL_DEVICE', 'Settings.BAUD_RATE'], {}), '(Settings.SERIAL_DEVICE, Settings.BAUD_RATE)\n', (112, 156), False, 'import serial\n'), ((237, 263), 'sys.stdout.buffer.write', 'sys.stdout.buffer.write', (['x'], {}), '(x)\n', (260, 263), False, 'import sys\n'), ((2... |
#
# Licensed to the Apache Software Foundation (ASF) under one
# or more contributor license agreements. See the NOTICE file
# distributed with this work for additional information
# regarding copyright ownership. The ASF licenses this file
# to you under the Apache License, Version 2.0 (the
# "License"); you may not... | [
"logging.getLogger",
"json.loads",
"sqlite3.connect",
"email.utils.parseaddr",
"time.mktime",
"os.path.join",
"re.match",
"email.utils.parsedate",
"textwrap.wrap",
"email.utils.formatdate",
"quopri.decodestring",
"time.gmtime"
] | [((1087, 1117), 'logging.getLogger', '_logging.getLogger', (['"""haystack"""'], {}), "('haystack')\n", (1105, 1117), True, 'import logging as _logging\n'), ((1163, 1194), 'json.loads', '_json.loads', (["_strings['topics']"], {}), "(_strings['topics'])\n", (1174, 1194), True, 'import json as _json\n'), ((1315, 1362), 'o... |
""" Initialization file for a GDC API client.
"""
from copy import copy
import types
from .client import BaseClient
# Aliases
COMMON_ALIASES = {
"_get_cases": 'get_cases',
"_get_mappings": 'get_mappings',
"_get_genes": 'get_genes',
"_get_ssm_occurrences": 'get_ssm_occurrences',
}
# API speci... | [
"types.FunctionType",
"copy.copy"
] | [((349, 369), 'copy.copy', 'copy', (['COMMON_ALIASES'], {}), '(COMMON_ALIASES)\n', (353, 369), False, 'from copy import copy\n'), ((723, 742), 'copy.copy', 'copy', (['COMMON_KWARGS'], {}), '(COMMON_KWARGS)\n', (727, 742), False, 'from copy import copy\n'), ((1144, 1245), 'types.FunctionType', 'types.FunctionType', (['f... |
import logging
from typing import Any, Dict, List, NewType
import mlflow
import numpy as np
import pandas as pd
import torch
import transformers
from mlflow.models import ModelSignature
from mlflow.pyfunc import PythonModel
from mlflow.types import ColSpec, DataType, Schema, TensorSpec
from mlflow.utils.environment im... | [
"logging.getLogger",
"typing.NewType",
"transformers.AutoModelForSequenceClassification.from_pretrained",
"transformers.AutoTokenizer.from_pretrained",
"numpy.dtype",
"transformers.pipeline",
"mlflow.types.ColSpec"
] | [((666, 687), 'typing.NewType', 'NewType', (['"""Model"""', 'Any'], {}), "('Model', Any)\n", (673, 687), False, 'from typing import Any, Dict, List, NewType\n'), ((700, 725), 'typing.NewType', 'NewType', (['"""Tokenizer"""', 'Any'], {}), "('Tokenizer', Any)\n", (707, 725), False, 'from typing import Any, Dict, List, Ne... |
import rclpy
from rclpy.action import ActionClient
from rclpy.node import Node
from functools import partial
from my_robot_interfaces.action import Test
class TestActionClient(Node):
def __init__(self):
super().__init__('test_action_client')
self.declare_parameter("secs", 2)
self.... | [
"rclpy.spin",
"rclpy.action.ActionClient",
"functools.partial",
"my_robot_interfaces.action.Test.Goal",
"rclpy.init",
"rclpy.shutdown"
] | [((2464, 2485), 'rclpy.init', 'rclpy.init', ([], {'args': 'args'}), '(args=args)\n', (2474, 2485), False, 'import rclpy\n'), ((2534, 2564), 'rclpy.spin', 'rclpy.spin', (['test_action_client'], {}), '(test_action_client)\n', (2544, 2564), False, 'import rclpy\n'), ((2569, 2585), 'rclpy.shutdown', 'rclpy.shutdown', ([], ... |
import sys
import pygame
def draw_canvas(screen, colors):
screen.fill(colors[0])
pygame.draw.rect(screen, colors[5], (20, 20, 500, 500))
index = 1 # skip light grey
for row in range(2):
for column in range(4):
pygame.draw.rect(screen, colors[index], ((60 * column) + 20... | [
"pygame.display.update",
"pygame.draw.rect",
"pygame.font.SysFont"
] | [((97, 152), 'pygame.draw.rect', 'pygame.draw.rect', (['screen', 'colors[5]', '(20, 20, 500, 500)'], {}), '(screen, colors[5], (20, 20, 500, 500))\n', (113, 152), False, 'import pygame\n'), ((405, 461), 'pygame.draw.rect', 'pygame.draw.rect', (['screen', 'colors[5]', '(280, 530, 120, 55)'], {}), '(screen, colors[5], (2... |
from suds.client import Client
from suds import WebFault
from model.project import Project
class SoapHelper:
def __init__(self, app):
self.app = app
def can_login(self, username, password):
client = Client(self.app.base_url + "api/soap/mantisconnect.php?wsdl")
try:
client.... | [
"suds.client.Client",
"model.project.Project"
] | [((226, 287), 'suds.client.Client', 'Client', (["(self.app.base_url + 'api/soap/mantisconnect.php?wsdl')"], {}), "(self.app.base_url + 'api/soap/mantisconnect.php?wsdl')\n", (232, 287), False, 'from suds.client import Client\n'), ((549, 610), 'suds.client.Client', 'Client', (["(self.app.base_url + 'api/soap/mantisconne... |
import pprint
import requests
import argparse
def login_admin(admin_id, admin_secret, url, verbose):
headers = {'Content-Type': 'application/x-www-form-urlencoded',
'Accept': 'application/json'}
payload = {'grant_type': 'client_credentials',
'response_type': 'token',
... | [
"requests.post",
"argparse.ArgumentParser",
"requests.get",
"requests.put",
"pprint.pprint"
] | [((364, 474), 'requests.post', 'requests.post', (['f"""http://{url}/oauth/token"""'], {'headers': 'headers', 'params': 'payload', 'auth': '(admin_id, admin_secret)'}), "(f'http://{url}/oauth/token', headers=headers, params=payload,\n auth=(admin_id, admin_secret))\n", (377, 474), False, 'import requests\n'), ((1042,... |
"Code used to generate data for experiments with synthetic data"
import math
import typing as ty
import numba
import numpy as np
import torch
import torch.nn as nn
from numba.experimental import jitclass
from tqdm.auto import tqdm
class MLP(nn.Module):
def __init__(
self,
*,
d_in: int,
... | [
"numpy.ones",
"numpy.flatnonzero",
"torch.relu",
"torch.nn.init.kaiming_normal_",
"numba.experimental.jitclass",
"math.sqrt",
"torch.nn.init._calculate_fan_in_and_fan_out",
"numpy.random.randint",
"numpy.zeros",
"torch.nn.init.uniform_",
"torch.nn.Linear",
"tqdm.auto.tqdm",
"numpy.random.ran... | [((1341, 1556), 'numba.experimental.jitclass', 'jitclass', ([], {'spec': "[('left_children', numba.int64[:]), ('right_children', numba.int64[:]), (\n 'feature', numba.int64[:]), ('threshold', numba.float32[:]), ('value',\n numba.float32[:]), ('is_leaf', numba.int64[:])]"}), "(spec=[('left_children', numba.int64[:... |