code
stringlengths
22
1.05M
apis
listlengths
1
3.31k
extract_api
stringlengths
75
3.25M
text = """ ala ma kota a kot ma ale """ # ------------------------------------------------------------------------------ # TODO as class chars = list(sorted(set(text))) # stabilne indeksy len_chars = len(chars)+1 c_to_i = {c:i+1 for i,c in enumerate(chars)} i_to_c = {i+1:c for i,c in enumerate(chars)} def text_to_i...
[ "keras.models.load_model", "numpy.log", "numpy.argmax", "numpy.random.multinomial", "numpy.exp", "keras.models.Sequential", "keras.layers.LSTM", "numpy.sum", "keras.layers.Dense", "keras.optimizers.RMSprop" ]
[((1137, 1149), 'keras.models.Sequential', 'Sequential', ([], {}), '()\n', (1147, 1149), False, 'from keras.models import Sequential, load_model\n'), ((1338, 1364), 'keras.optimizers.RMSprop', 'RMSprop', ([], {'learning_rate': '(0.1)'}), '(learning_rate=0.1)\n', (1345, 1364), False, 'from keras.optimizers import RMSpro...
import tensorflow as tf from lime import lime_image from skimage.segmentation import mark_boundaries from xplainer.backend.tools.abstract_tool import AbstractTool, GeneralSettings from xplainer.backend.utils.image import prepare_for_prediction, get_base64png class Lime(AbstractTool): def name(self): retu...
[ "skimage.segmentation.mark_boundaries", "tensorflow.image.convert_image_dtype", "xplainer.backend.utils.image.prepare_for_prediction", "lime.lime_image.LimeImageExplainer", "xplainer.backend.utils.image.get_base64png", "tensorflow.cast" ]
[((2660, 2701), 'xplainer.backend.utils.image.prepare_for_prediction', 'prepare_for_prediction', (['model', 'image_path'], {}), '(model, image_path)\n', (2682, 2701), False, 'from xplainer.backend.utils.image import prepare_for_prediction, get_base64png\n'), ((2793, 2825), 'tensorflow.cast', 'tf.cast', (['image'], {'dt...
#!/usr/bin/env python # encoding: utf-8 """ @Author: yangwenhao @Contact: <EMAIL> @Software: PyCharm @File: model.py @Overview: The deep speaker model is not entirely the same as ResNet, as there are convolutional layers between blocks. """ import math import torch import torch.nn as nn from torch.autograd import Fu...
[ "torch.nn.CrossEntropyLoss", "torch.nn.Sequential", "torch.sqrt", "math.sqrt", "torch.pow", "torch.nn.init.xavier_normal", "torch.numel", "torch.nn.BatchNorm1d", "torch.sum", "torch.nn.BatchNorm2d", "torch.mean", "Define_Model.Loss.SoftmaxLoss.AngleLinear", "torch.nn.AdaptiveAvgPool2d", "t...
[((2456, 2545), 'torch.nn.Conv2d', 'nn.Conv2d', (['in_planes', 'out_planes'], {'kernel_size': '(3)', 'stride': 'stride', 'padding': '(1)', 'bias': '(False)'}), '(in_planes, out_planes, kernel_size=3, stride=stride, padding=1,\n bias=False)\n', (2465, 2545), True, 'import torch.nn as nn\n'), ((805, 823), 'torch.abs',...
import requests from datetime import datetime CLIENT_ID = '2OWPT1X5RNXEE0DGOM5VWO2FBM5R5TNTVPPLH50NCSZAX3QD' CLIENT_SECRET = '<KEY>' DATE = datetime.today().strftime('%Y%m%d') BASE_URL = 'https://api.foursquare.com/v2' AUTH_URL_PART = 'client_id={}&client_secret={}&v={}'.format(CLIENT_ID, CLIENT_SECRET, DATE) # Four...
[ "datetime.datetime.today", "requests.get" ]
[((2138, 2155), 'requests.get', 'requests.get', (['url'], {}), '(url)\n', (2150, 2155), False, 'import requests\n'), ((141, 157), 'datetime.datetime.today', 'datetime.today', ([], {}), '()\n', (155, 157), False, 'from datetime import datetime\n'), ((1832, 1849), 'requests.get', 'requests.get', (['url'], {}), '(url)\n',...
from django.conf.urls import include, url from django.contrib import admin from . import views urlpatterns = [ url(r'^$', views.ViewView.as_view(), name='profile_own_view'), url(r'^edit/', views.EditView.as_view(), name='profile_edit'), url(r'^view/$', views.ViewView.as_view(), name='profile_own_view'), ...
[ "django.conf.urls.url" ]
[((420, 481), 'django.conf.urls.url', 'url', (['"""^delete/$"""', 'views.delete_account'], {'name': '"""delete_account"""'}), "('^delete/$', views.delete_account, name='delete_account')\n", (423, 481), False, 'from django.conf.urls import include, url\n'), ((488, 577), 'django.conf.urls.url', 'url', (['"""^follow/(?P<u...
import os import sys import json import logging import requests class ApiException(Exception): """ API Exception class. """ pass logger = logging.getLogger() logger.setLevel(logging.INFO) try: api_url = os.environ['BUS_API_URL'] if api_url[-1] == '/': api_url = api_url[0:-1] except KeyError...
[ "logging.getLogger", "requests.post", "requests.get", "sys.exit" ]
[((150, 169), 'logging.getLogger', 'logging.getLogger', ([], {}), '()\n', (167, 169), False, 'import logging\n'), ((464, 496), 'requests.get', 'requests.get', (["(api_url + '/lines')"], {}), "(api_url + '/lines')\n", (476, 496), False, 'import requests\n'), ((709, 745), 'requests.get', 'requests.get', (["(api_url + '/l...
import os import os.path as osp import sys import numpy as np from sklearn.svm import LinearSVC from tqdm import tqdm sys.path.append(osp.dirname(osp.dirname(osp.abspath(__file__)))) import torch.utils.data as data from dataset.modelnet40 import LatentCapsulesModelNet40, LatentVectorsModelNet40 from utils.utils impo...
[ "sklearn.svm.LinearSVC", "os.path.join", "dataset.modelnet40.LatentCapsulesModelNet40", "numpy.zeros", "dataset.modelnet40.LatentVectorsModelNet40", "torch.utils.data.DataLoader", "os.path.abspath", "utils.utils.initialize_main", "utils.utils.create_save_folder" ]
[((392, 409), 'utils.utils.initialize_main', 'initialize_main', ([], {}), '()\n', (407, 409), False, 'from utils.utils import create_save_folder, initialize_main\n'), ((559, 599), 'os.path.join', 'os.path.join', (['logdir', "args['train_root']"], {}), "(logdir, args['train_root'])\n", (571, 599), False, 'import os\n'),...
from django.db import models, migrations class Migration(migrations.Migration): dependencies = [ ('libretto', '0009_auto_20150423_2042'), ] operations = [ migrations.AlterModelOptions( name='pupitre', options={'ordering': ('-soliste', 'partie'), 'verbose_name': 'p...
[ "django.db.migrations.AlterModelOptions", "django.db.models.ManyToManyField", "django.db.models.CharField", "django.db.models.IntegerField" ]
[((187, 348), 'django.db.migrations.AlterModelOptions', 'migrations.AlterModelOptions', ([], {'name': '"""pupitre"""', 'options': "{'ordering': ('-soliste', 'partie'), 'verbose_name': 'pupitre',\n 'verbose_name_plural': 'pupitres'}"}), "(name='pupitre', options={'ordering': (\n '-soliste', 'partie'), 'verbose_nam...
from unittest.mock import Mock from pelican.tests.support import unittest class Test_abbr_role(unittest.TestCase): def call_it(self, text): from pelican.rstdirectives import abbr_role rawtext = text lineno = 42 inliner = Mock(name='inliner') nodes, system_messages = abbr_r...
[ "pelican.rstdirectives.abbr_role", "unittest.mock.Mock" ]
[((260, 280), 'unittest.mock.Mock', 'Mock', ([], {'name': '"""inliner"""'}), "(name='inliner')\n", (264, 280), False, 'from unittest.mock import Mock\n'), ((314, 363), 'pelican.rstdirectives.abbr_role', 'abbr_role', (['"""abbr"""', 'rawtext', 'text', 'lineno', 'inliner'], {}), "('abbr', rawtext, text, lineno, inliner)\...
# Copyright 2011 <NAME> (<EMAIL>) # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writ...
[ "tmapi.exceptions.ModelConstraintException", "django.db.models.TextField", "locator.Locator", "django.db.models.ForeignKey" ]
[((1017, 1065), 'django.db.models.ForeignKey', 'models.ForeignKey', (['"""Topic"""'], {'related_name': '"""names"""'}), "('Topic', related_name='names')\n", (1034, 1065), False, 'from django.db import models\n'), ((1078, 1096), 'django.db.models.TextField', 'models.TextField', ([], {}), '()\n', (1094, 1096), False, 'fr...
#!/usr/bin/env python from __future__ import print_function from platform import node from sys import argv from os import popen from re import * def matlabRange(threads): limits = threads.split(":") if len(limits)==1: return range(int(limits[0]),int(limits[0])+1) if len(limits)==2: return...
[ "os.popen", "platform.node" ]
[((1777, 1791), 'os.popen', 'popen', (['command'], {}), '(command)\n', (1782, 1791), False, 'from os import popen\n'), ((1617, 1623), 'platform.node', 'node', ([], {}), '()\n', (1621, 1623), False, 'from platform import node\n')]
import os from typing import Dict, Optional import tomlkit def _get_project_meta(pyproj_path: str = "./pyproject.toml") -> Dict[str, str]: if os.path.exists(pyproj_path): with open(pyproj_path, "r") as pyproject: file_contents = pyproject.read() return tomlkit.parse(file_contents)["to...
[ "tomlkit.parse", "os.path.exists" ]
[((149, 176), 'os.path.exists', 'os.path.exists', (['pyproj_path'], {}), '(pyproj_path)\n', (163, 176), False, 'import os\n'), ((288, 316), 'tomlkit.parse', 'tomlkit.parse', (['file_contents'], {}), '(file_contents)\n', (301, 316), False, 'import tomlkit\n')]
import datetime as dt import logging from collections import Counter, OrderedDict, defaultdict from dataclasses import dataclass from io import StringIO from operator import itemgetter input = """8 2017-01-03,16:18:50,AAPL,142.64 2017-01-03,16:25:22,AMD,13.86 2017-01-03,16:25:25,AAPL,141.64 2017-01-03,16:25:28,AMZN,84...
[ "logging.getLogger", "datetime.time", "collections.Counter", "collections.defaultdict", "operator.itemgetter", "io.StringIO", "datetime.time.fromisoformat", "datetime.date.fromisoformat" ]
[((2052, 2104), 'collections.Counter', 'Counter', (['(feed.time.hour for feed in trading_day_feed)'], {}), '(feed.time.hour for feed in trading_day_feed)\n', (2059, 2104), False, 'from collections import Counter, OrderedDict, defaultdict\n'), ((2326, 2375), 'collections.Counter', 'Counter', (['(feed.symbol for feed in ...
from wiki_scrape_db import get_headlines headlines = [] for year in list(range(1995, 2020)): try: headlines.append(get_headlines(year)) except Exception as e: print(e) headlines.append(get_headlines(1994, start_month='July')) headlines.append(get_headlines(2020, end_month='November')) with o...
[ "wiki_scrape_db.get_headlines" ]
[((212, 251), 'wiki_scrape_db.get_headlines', 'get_headlines', (['(1994)'], {'start_month': '"""July"""'}), "(1994, start_month='July')\n", (225, 251), False, 'from wiki_scrape_db import get_headlines\n'), ((270, 311), 'wiki_scrape_db.get_headlines', 'get_headlines', (['(2020)'], {'end_month': '"""November"""'}), "(202...
#!/usr/bin/env python """Utility and General purpose functions.""" import inspect import warnings import os import re import shatter.constants as cts __author__ = '<NAME>' def read_file(absolute_path): """ :param absolute_path: string path. :return: list with lines of the file. """ return [line...
[ "os.path.exists", "inspect.stack", "shatter.constants.PARTICULAR_DEFINITION.pattern.format", "re.match", "os.remove", "re.search" ]
[((463, 487), 'os.path.exists', 'os.path.exists', (['filename'], {}), '(filename)\n', (477, 487), False, 'import os\n'), ((497, 516), 'os.remove', 'os.remove', (['filename'], {}), '(filename)\n', (506, 516), False, 'import os\n'), ((3241, 3265), 're.search', 're.search', (['pattern', 'line'], {}), '(pattern, line)\n', ...
#!/usr/bin/python # import argparse import os import pandas as pd corr = {'pH': [1, 1.02, 1.04, 1.06, 1.08, 1.1, 1.12], 'cl': [1.000000, 1.017857, 1.035714, 1.053571, 1.071429, 1.089286, 1.107143, 1.160714, 1.196429]} corr = {'pH': [1, 1.02, 1.04, 1.06, 1.08, 1.1], 'cl': [1.000000, 1....
[ "os.makedirs", "argparse.ArgumentParser", "pandas.read_csv", "os.path.join", "os.path.isdir" ]
[((490, 538), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {'description': 'description'}), '(description=description)\n', (513, 538), False, 'import argparse\n'), ((1010, 1033), 'pandas.read_csv', 'pd.read_csv', (['args.ffile'], {}), '(args.ffile)\n', (1021, 1033), True, 'import pandas as pd\n'), ((1193,...
from system import System from src.basic.sessions.cmd_session import CmdSession class CmdSystem(System): def __init__(self): super(CmdSystem, self).__init__() @classmethod def name(cls): return 'cmd' def new_session(self, agent, kb): return CmdSession(agent, kb)
[ "src.basic.sessions.cmd_session.CmdSession" ]
[((284, 305), 'src.basic.sessions.cmd_session.CmdSession', 'CmdSession', (['agent', 'kb'], {}), '(agent, kb)\n', (294, 305), False, 'from src.basic.sessions.cmd_session import CmdSession\n')]
# Import necessary modules from flask import render_template, request, redirect, url_for, flash, Blueprint from flask_app import app, db from flask_app.decorators import check_confirmed from flask_app.models import * from flask_login import login_required, current_user from sqlalchemy import or_, and_, func import rand...
[ "flask.render_template", "flask.request.args.get", "random.randint", "flask.flash", "flask_app.app.route", "flask_app.db.session.query", "requests.get", "flask.url_for", "flask.redirect", "flask_app.app.errorhandler", "flask_app.db.session.commit", "flask_app.db.session.add", "sqlalchemy.fun...
[((410, 437), 'flask.Blueprint', 'Blueprint', (['"""main"""', '__name__'], {}), "('main', __name__)\n", (419, 437), False, 'from flask import render_template, request, redirect, url_for, flash, Blueprint\n'), ((759, 773), 'flask_app.app.route', 'app.route', (['"""/"""'], {}), "('/')\n", (768, 773), False, 'from flask_a...
from datetime import timedelta import logging from typing import Union from pyschism.enums import ( IofHydroVariables, IofDvdVariables, IofWwmVariables, IofGenVariables, IofAgeVariables, IofSedVariables, IofEcoVariables, IofIcmVariables, IofCosVariables, IofFibVariables, Iof...
[ "logging.getLogger", "datetime.timedelta" ]
[((430, 457), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (447, 457), False, 'import logging\n'), ((5495, 5522), 'datetime.timedelta', 'timedelta', ([], {'hours': 'nspool_sta'}), '(hours=nspool_sta)\n', (5504, 5522), False, 'from datetime import timedelta\n'), ((6589, 6616), 'datetime....
# -*- coding: utf-8 -*- from abc import ABCMeta, abstractmethod, abstractproperty from collections import Sequence import numpy as np class MatrixBase(object): __metaclass__ = ABCMeta _base_tags = set() @abstractmethod def __init__(self, backend, ioshape, iopacking, tags): self.backend = ...
[ "numpy.unique", "numpy.any" ]
[((3740, 3757), 'numpy.unique', 'np.unique', (['matmap'], {}), '(matmap)\n', (3749, 3757), True, 'import numpy as np\n'), ((3999, 4021), 'numpy.any', 'np.any', (['(stridemap == 0)'], {}), '(stridemap == 0)\n', (4005, 4021), True, 'import numpy as np\n')]
import re from typing import List from .consts import * # =================== # # INTERNALS FUNCTIONS # # =================== # def my_re_escape(text): escape_char = r"[]" returned_text = "" for c in text: if c in escape_char: returned_text += "\\" returned_text += c retur...
[ "re.compile" ]
[((2584, 2687), 're.compile', 're.compile', (['f"""\n {ESCAPED_CSI}\n \\\\d*\n (?:\n ;\\\\d*\n )*\n m\n"""', 're.VERBOSE'], {}), '(\n f"""\n {ESCAPED_CSI}\n \\\\d*\n (?:\n ;\\\\d*\n )*\n m\n""",\n re.VERBOSE)\n', (2594, 2687), False, 'import re\n')]
from flask import Flask,render_template as render, request from models import * app = Flask(__name__,template_folder='./templates') @app.route('/GETPage',methods = ['GET']) def Gpage(): return f"You have landed on the page which allows the {request.method} method." @app.route('/POSTPage', methods = ['GET','P...
[ "flask.render_template", "json.loads", "flask.Flask" ]
[((90, 136), 'flask.Flask', 'Flask', (['__name__'], {'template_folder': '"""./templates"""'}), "(__name__, template_folder='./templates')\n", (95, 136), False, 'from flask import Flask, render_template as render, request\n'), ((478, 504), 'flask.render_template', 'render', (['"""SuccessPage.html"""'], {}), "('SuccessPa...
#!/usr/bin/env python # -*- coding: utf-8 -*- # pylint: disable=too-few-public-methods """ Web application endpoint ======================== Starts an http endpoint to serve requests """ import logging import mimetypes import os import site import sys import falcon from gunicorn.app.base import BaseApplication logg...
[ "logging.getLogger", "falcon.API", "os.path.join", "os.path.isfile", "os.path.dirname", "os.path.isdir", "mimetypes.guess_type" ]
[((325, 352), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (342, 352), False, 'import logging\n'), ((3569, 3581), 'falcon.API', 'falcon.API', ([], {}), '()\n', (3579, 3581), False, 'import falcon\n'), ((873, 925), 'os.path.join', 'os.path.join', (['sys.prefix', '"""orion-dashboard"""', ...
from __future__ import division import argparse, logging, os, math, tqdm import numpy as np import mxnet as mx from mxnet import gluon, nd, image from mxnet.gluon.data.vision import transforms import matplotlib.pyplot as plt import gluoncv as gcv from gluoncv import data from gluoncv.data import mscoco from gluoncv....
[ "gluoncv.data.transforms.presets.yolo.load_test", "mxnet.nd.sign", "gluoncv.data.transforms.pose.get_final_preds", "numpy.array", "mxnet.nd.zeros_like", "gluoncv.data.transforms.pose.get_max_pred", "argparse.ArgumentParser", "matplotlib.pyplot.scatter", "mxnet.nd.array", "matplotlib.pyplot.ylim", ...
[((415, 502), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {'description': '"""Predict ImageNet classes from a given image"""'}), "(description=\n 'Predict ImageNet classes from a given image')\n", (438, 502), False, 'import argparse, logging, os, math, tqdm\n'), ((2061, 2083), 'mxnet.nd.stack', 'nd.st...
#! /usr/bin/env python import sys import os import extargsparse import re import time ##importdebugstart sys.path.append(os.path.abspath(os.path.dirname(__file__))) from strparser import * from filehdl import * from fmthdl import * from extract_ob import * from obmaklib import * ##importdebugend REPLACE_IMPORT_LIB=1...
[ "re.split", "os.path.exists", "extargsparse.ExtArgsParse", "os.path.join", "os.path.dirname", "sys.stderr.write", "sys.exit", "extargsparse.ExtArgsOptions", "time.time", "sys.stdout.write" ]
[((1726, 1756), 'extargsparse.ExtArgsOptions', 'extargsparse.ExtArgsOptions', (['d'], {}), '(d)\n', (1753, 1756), False, 'import extargsparse\n'), ((1769, 1780), 'time.time', 'time.time', ([], {}), '()\n', (1778, 1780), False, 'import time\n'), ((1794, 1828), 'extargsparse.ExtArgsParse', 'extargsparse.ExtArgsParse', ([...
from ..cross_box import CrossBox from unittest.mock import Mock, patch from pyglet.gl import GL_LINES import unittest class TestCrossBox(unittest.TestCase): """Test rendering of cross box graphics.""" def setUp(self): """Provides the following to all tests: * ``self.rectangle``: Mock rectang...
[ "unittest.mock.patch", "unittest.mock.Mock" ]
[((1298, 1344), 'unittest.mock.patch', 'patch', (['"""engine.graphics.cross_box.vertex_list"""'], {}), "('engine.graphics.cross_box.vertex_list')\n", (1303, 1344), False, 'from unittest.mock import Mock, patch\n'), ((1689, 1735), 'unittest.mock.patch', 'patch', (['"""engine.graphics.cross_box.vertex_list"""'], {}), "('...
import pandas as pd from tqdm import tqdm from ..binarize import to_binary from cana.boolean_node import BooleanNode # set up variables n_inputs = 2**2 n_rules = 2**(2**2) df_dict = [] for rule in tqdm(range(n_rules)): canal = {} # becomes row of dataframe arr = to_binary(rule, digits=4) print(arr) ...
[ "pandas.DataFrame", "cana.boolean_node.BooleanNode.from_output_list" ]
[((857, 878), 'pandas.DataFrame', 'pd.DataFrame', (['df_dict'], {}), '(df_dict)\n', (869, 878), True, 'import pandas as pd\n'), ((367, 419), 'cana.boolean_node.BooleanNode.from_output_list', 'BooleanNode.from_output_list', ([], {'outputs': 'arr', 'name': 'rule'}), '(outputs=arr, name=rule)\n', (395, 419), False, 'from ...
from django.shortcuts import render from django.views.generic import ListView, CreateView, DetailView from django.contrib.auth.mixins import LoginRequiredMixin from posts.models import Post from posts.forms import PostForm from django.urls import reverse_lazy # Create your views here. class PostListView(LoginRequired...
[ "django.urls.reverse_lazy" ]
[((598, 624), 'django.urls.reverse_lazy', 'reverse_lazy', (['"""posts:list"""'], {}), "('posts:list')\n", (610, 624), False, 'from django.urls import reverse_lazy\n')]
import cv2 import numpy as np import random as rd import os from tensorflow.keras.models import Sequential, Model from tensorflow.keras.layers import Dense, Dropout, Activation, Flatten from tensorflow.keras.layers import Conv2D, MaxPooling2D, Input from tensorflow.keras.optimizers import Adadelta from tensorflow.keras...
[ "tensorflow.keras.layers.Conv2D", "tensorflow.keras.layers.MaxPooling2D", "os.rename", "os.path.join", "numpy.asarray", "tensorflow.keras.optimizers.Adadelta", "tensorflow.keras.layers.BatchNormalization", "tensorflow.keras.callbacks.EarlyStopping", "random.random", "tensorflow.keras.layers.Dense"...
[((1236, 1272), 'tensorflow.keras.models.Model', 'Model', (['img_input', 'x'], {'name': '"""calvonet"""'}), "(img_input, x, name='calvonet')\n", (1241, 1272), False, 'from tensorflow.keras.models import Sequential, Model\n'), ((4349, 4388), 'os.path.join', 'os.path.join', (["(output_model_path + '.h5')"], {}), "(output...
from django.views.generic import ListView, DetailView, TemplateView, CreateView, UpdateView, DeleteView from board.models import Post from django.contrib.auth.mixins import LoginRequiredMixin from django.urls import reverse_lazy from mysite.views import OwnerOnlyMixin from django.conf import settings #--- ListView c...
[ "board.models.Post.objects.filter", "django.urls.reverse_lazy" ]
[((702, 729), 'django.urls.reverse_lazy', 'reverse_lazy', (['"""board:index"""'], {}), "('board:index')\n", (714, 729), False, 'from django.urls import reverse_lazy\n'), ((1171, 1198), 'django.urls.reverse_lazy', 'reverse_lazy', (['"""board:index"""'], {}), "('board:index')\n", (1183, 1198), False, 'from django.urls im...
from django.conf import settings from django.db import models from django.dispatch import Signal from django.utils import timezone from django.utils.translation import gettext_lazy as _ from froide.foirequest.models import FoiMessage from .utils import inform_user_problem_resolved class ProblemChoices(models.TextCh...
[ "django.db.models.TextField", "django.db.models.ForeignKey", "django.utils.translation.gettext_lazy", "django.dispatch.Signal", "django.db.models.BooleanField", "django.utils.timezone.now", "django.db.models.DateTimeField", "django.db.models.CharField" ]
[((1854, 1862), 'django.dispatch.Signal', 'Signal', ([], {}), '()\n', (1860, 1862), False, 'from django.dispatch import Signal\n'), ((1873, 1881), 'django.dispatch.Signal', 'Signal', ([], {}), '()\n', (1879, 1881), False, 'from django.dispatch import Signal\n'), ((1893, 1901), 'django.dispatch.Signal', 'Signal', ([], {...
# In The Name of God # ======================================= # [] File Name : pipe.py # # [] Creation Date : 27-11-2019 # # [] Created By : <NAME> <<EMAIL>> # ======================================= import os def child(n, w): print('I am Child') f = os.fdopen(w, 'w') # the old way f.write('hello %d...
[ "os.fdopen", "os.fork", "os.pipe" ]
[((263, 280), 'os.fdopen', 'os.fdopen', (['w', '"""w"""'], {}), "(w, 'w')\n", (272, 280), False, 'import os\n'), ((405, 414), 'os.pipe', 'os.pipe', ([], {}), '()\n', (412, 414), False, 'import os\n'), ((433, 442), 'os.fork', 'os.fork', ([], {}), '()\n', (440, 442), False, 'import os\n'), ((520, 537), 'os.fdopen', 'os.f...
# -*- coding: utf-8 -*- import logging logger = logging.getLogger() logger.basicConfig = logging.basicConfig(level=logging.DEBUG) import numpy as np import matplotlib.pyplot as plt import logictensornetworks_wrapper as ltnw nr_samples=500 data=np.random.uniform([0,0],[1.,1.],(nr_samples,2)).astype(np.float32) data_A...
[ "logging.getLogger", "logging.basicConfig", "logictensornetworks_wrapper.constant", "logictensornetworks_wrapper.variable", "matplotlib.pyplot.colorbar", "logictensornetworks_wrapper.train", "numpy.square", "logictensornetworks_wrapper.predicate", "logictensornetworks_wrapper.axiom", "logictensorn...
[((48, 67), 'logging.getLogger', 'logging.getLogger', ([], {}), '()\n', (65, 67), False, 'import logging\n'), ((89, 129), 'logging.basicConfig', 'logging.basicConfig', ([], {'level': 'logging.DEBUG'}), '(level=logging.DEBUG)\n', (108, 129), False, 'import logging\n'), ((448, 476), 'logictensornetworks_wrapper.variable'...
from api.subgroups_names import ClassicalSubgroups from graph_constructor import get_graph from plotter.graph_plotter import GraphPlotter from plotter.geodesic_plotter import GeodesicPlotter from plotter.marker_plotter import MarkerPlotter from special_polygon import SpecialPolygon from fimath import Matrix, Field from...
[ "graph_constructor.get_graph", "fimath.Matrix.beautify", "plotter.geodesic_plotter.GeodesicPlotter", "fimath.Matrix.from_str", "reduction.Decomposer", "plotter.marker_plotter.MarkerPlotter", "special_polygon.SpecialPolygon", "fimath.Field", "plotter.graph_plotter.GraphPlotter" ]
[((1674, 1692), 'plotter.graph_plotter.GraphPlotter', 'GraphPlotter', (['axes'], {}), '(axes)\n', (1686, 1692), False, 'from plotter.graph_plotter import GraphPlotter\n'), ((1858, 1885), 'plotter.graph_plotter.GraphPlotter', 'GraphPlotter', ([], {'bokeh_fig': 'fig'}), '(bokeh_fig=fig)\n', (1870, 1885), False, 'from plo...
""" Module description: """ __version__ = '0.3.1' __author__ = '<NAME>, <NAME>' __email__ = '<EMAIL>, <EMAIL>' import tensorflow as tf import numpy as np import random class Sampler(): def __init__(self, indexed_ratings=None, m=None, num_users=None, num_items=None, transactions=None, batch_size=512, random_seed=...
[ "tensorflow.data.Dataset.from_generator", "numpy.random.seed", "random.seed" ]
[((333, 360), 'numpy.random.seed', 'np.random.seed', (['random_seed'], {}), '(random_seed)\n', (347, 360), True, 'import numpy as np\n'), ((369, 393), 'random.seed', 'random.seed', (['random_seed'], {}), '(random_seed)\n', (380, 393), False, 'import random\n'), ((2386, 2498), 'tensorflow.data.Dataset.from_generator', '...
#! /usr/bin/env python import socket def send_ping_data(HOST, PORT): with socket.socket(socket.AF_INET, socket.SOCK_STREAM) as s: s.connect((HOST,PORT)) s.sendall(b'Hello, world') data=s.recv(1024) return data if __name__== "__main__" : print('Received',repr(send_ping_data('127...
[ "socket.socket" ]
[((80, 129), 'socket.socket', 'socket.socket', (['socket.AF_INET', 'socket.SOCK_STREAM'], {}), '(socket.AF_INET, socket.SOCK_STREAM)\n', (93, 129), False, 'import socket\n')]
# bot.py import os import random import discord from dotenv import load_dotenv import commands as cm import qrys load_dotenv() TOKEN = os.getenv('DISCORD_TOKEN') GUILD = os.getenv('DISCORD_GUILD') client = discord.Client() prefix = '$' @client.event async def on_ready(): for guild in client.guilds: if g...
[ "qrys.connect_db", "os.getenv", "commands.m_usage", "dotenv.load_dotenv", "discord.Client" ]
[((115, 128), 'dotenv.load_dotenv', 'load_dotenv', ([], {}), '()\n', (126, 128), False, 'from dotenv import load_dotenv\n'), ((137, 163), 'os.getenv', 'os.getenv', (['"""DISCORD_TOKEN"""'], {}), "('DISCORD_TOKEN')\n", (146, 163), False, 'import os\n'), ((172, 198), 'os.getenv', 'os.getenv', (['"""DISCORD_GUILD"""'], {}...
# flake8: noqa # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distrib...
[ "sqlalchemy.true", "alembic.op.drop_table", "alembic.op.bulk_insert", "sqlalchemy.String", "sqlalchemy.CheckConstraint" ]
[((1195, 1239), 'alembic.op.bulk_insert', 'op.bulk_insert', (['table', "[{'worker_uuid': ''}]"], {}), "(table, [{'worker_uuid': ''}])\n", (1209, 1239), False, 'from alembic import op\n'), ((1277, 1306), 'alembic.op.drop_table', 'op.drop_table', (['RESOURCE_TABLE'], {}), '(RESOURCE_TABLE)\n', (1290, 1306), False, 'from ...
import torch import torch.nn.functional as F from ..models.progressive import ProGANGenerator, ProGANDiscriminator from ..modules.gan_loss import ImprovedWGANLoss from ..modules.instance_refiner import InstanceRefiner from tools.utils import to_cuda from models import load_network, save_network, print_network ...
[ "torch.mul", "torch.nn.functional.kl_div", "torch.nn.functional.mse_loss", "torch.log", "torch.nn.functional.l1_loss", "models.save_network", "torch.mean", "torch.Tensor", "torch.empty_like", "torch.tensor", "torch.sum", "models.load_network", "torch.nn.functional.interpolate", "tools.util...
[((2887, 2909), 'tools.utils.to_cuda', 'to_cuda', (['data', '"""z_seg"""'], {}), "(data, 'z_seg')\n", (2894, 2909), False, 'from tools.utils import to_cuda\n'), ((2937, 2961), 'tools.utils.to_cuda', 'to_cuda', (['data', '"""sem_seg"""'], {}), "(data, 'sem_seg')\n", (2944, 2961), False, 'from tools.utils import to_cuda\...
import os import sys TOPIC = 'studio.schoolpower.SchoolPower' PS_API = 'https://powerschool.mapleleaf.cn' CACHE_DB_LOCATION = os.environ.get("CACHE_DB_LOCATION", None) DB_LOCATION = os.environ.get('DB_LOCATION', 'users.db') PEM_FILE_PATH = os.environ.get("APNS_CERT_FILE", None) SECRET = os.environ.get("SECRET", "test"...
[ "os.environ.get" ]
[((127, 168), 'os.environ.get', 'os.environ.get', (['"""CACHE_DB_LOCATION"""', 'None'], {}), "('CACHE_DB_LOCATION', None)\n", (141, 168), False, 'import os\n'), ((183, 224), 'os.environ.get', 'os.environ.get', (['"""DB_LOCATION"""', '"""users.db"""'], {}), "('DB_LOCATION', 'users.db')\n", (197, 224), False, 'import os\...
# Generated by Django 2.1.1 on 2018-10-01 17:34 import django.contrib.postgres.fields from django.db import migrations, models import uuid class Migration(migrations.Migration): dependencies = [ ('custom', '0002_riskfield_risk_type'), ] operations = [ migrations.AlterModelOptions( ...
[ "django.db.migrations.AlterModelOptions", "django.db.models.UUIDField", "django.db.models.CharField" ]
[((285, 434), 'django.db.migrations.AlterModelOptions', 'migrations.AlterModelOptions', ([], {'name': '"""riskfield"""', 'options': "{'ordering': ('id',), 'verbose_name': 'Risk Field', 'verbose_name_plural':\n 'Risk Fields'}"}), "(name='riskfield', options={'ordering': ('id',),\n 'verbose_name': 'Risk Field', 've...
import os import sys import requests import logging import time import json import pandas as pd import numpy as np from concurrent.futures import ProcessPoolExecutor from git import Git class FPL_Gameweek: """ Get the Gameweek state """ def __init__(self, logger, season_data): """ Args: ...
[ "logging.basicConfig", "logging.getLogger", "os.path.exists", "pandas.DataFrame", "os.makedirs", "git.Git", "os.path.join", "requests.get", "time.sleep", "numpy.random.randint", "concurrent.futures.ProcessPoolExecutor", "json.load" ]
[((8774, 8849), 'logging.basicConfig', 'logging.basicConfig', ([], {'level': 'logging.INFO', 'format': '"""%(asctime)s - %(message)s"""'}), "(level=logging.INFO, format='%(asctime)s - %(message)s')\n", (8793, 8849), False, 'import logging\n'), ((8879, 8906), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}),...
from chroma_core.lib.storage_plugin.api import attributes from chroma_core.lib.storage_plugin.api.identifiers import GlobalId, ScopedId from chroma_core.lib.storage_plugin.api.plugin import Plugin from chroma_core.lib.storage_plugin.api import resources from chroma_core.lib.storage_plugin.api import relations version ...
[ "chroma_core.lib.storage_plugin.api.relations.Provide", "chroma_core.lib.storage_plugin.api.identifiers.ScopedId", "chroma_core.lib.storage_plugin.api.attributes.String", "chroma_core.lib.storage_plugin.api.relations.Subscribe", "chroma_core.lib.storage_plugin.api.attributes.Integer", "chroma_core.lib.sto...
[((445, 464), 'chroma_core.lib.storage_plugin.api.attributes.String', 'attributes.String', ([], {}), '()\n', (462, 464), False, 'from chroma_core.lib.storage_plugin.api import attributes\n'), ((572, 591), 'chroma_core.lib.storage_plugin.api.attributes.String', 'attributes.String', ([], {}), '()\n', (589, 591), False, '...
import json import time import logging import requests import websocket from . import abc from .. import events, sansio, methods, exceptions LOG = logging.getLogger(__name__) class SlackAPI(abc.SlackAPI): """ `requests` implementation of :class:`slack.io.abc.SlackAPI` Args: session: HTTP sessi...
[ "logging.getLogger", "requests.session", "json.loads", "time.sleep", "websocket.create_connection", "time.time" ]
[((150, 177), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (167, 177), False, 'import logging\n'), ((720, 752), 'websocket.create_connection', 'websocket.create_connection', (['url'], {}), '(url)\n', (747, 752), False, 'import websocket\n'), ((942, 961), 'time.sleep', 'time.sleep', (['s...
from django import forms from student.models import Major, UserProfile from django.contrib.auth.models import User class MajorForm(forms.ModelForm): code = forms.CharField(max_length=20) viName = forms.CharField(max_length=128) enName = forms.CharField(max_length=128) class Meta: model = Major fields ...
[ "django.forms.CharField", "django.forms.DateField", "django.forms.ChoiceField", "student.models.Major.objects.all", "django.forms.IntegerField" ]
[((159, 189), 'django.forms.CharField', 'forms.CharField', ([], {'max_length': '(20)'}), '(max_length=20)\n', (174, 189), False, 'from django import forms\n'), ((200, 231), 'django.forms.CharField', 'forms.CharField', ([], {'max_length': '(128)'}), '(max_length=128)\n', (215, 231), False, 'from django import forms\n'),...
# -*- coding: utf-8 -*- # Standard library imports import sys # Third party imports from Qt import QtWidgets # Local imports from .ui import Dialog if __name__ == '__main__': app = QtWidgets.QApplication(sys.argv) d = Dialog() sys.exit(d.exec_())
[ "Qt.QtWidgets.QApplication" ]
[((190, 222), 'Qt.QtWidgets.QApplication', 'QtWidgets.QApplication', (['sys.argv'], {}), '(sys.argv)\n', (212, 222), False, 'from Qt import QtWidgets\n')]
from flask import url_for from tests.conftest import normalize_spaces def test_set_inbound_sms_sets_a_number_for_service( logged_in_client, mock_add_sms_sender, multiple_available_inbound_numbers, service_one, fake_uuid, mock_no_inbound_number_for_service, mocker ): mocker.patch('app....
[ "flask.url_for" ]
[((504, 576), 'flask.url_for', 'url_for', (['"""main.service_set_inbound_number"""'], {'service_id': "service_one['id']"}), "('main.service_set_inbound_number', service_id=service_one['id'])\n", (511, 576), False, 'from flask import url_for\n')]
from ..function.node import * from ..function.tree import * import numpy as np def generate_tree(name='test'): p_branch = .2 p_infertile = .1 p_channel = 1 - p_branch - p_infertile decay = .25 branch_nodes = [Max, Sum, Mean, Min, Product, Median] infertile_nodes = [Constant, Input, Uniform, N...
[ "numpy.random.choice" ]
[((1244, 1338), 'numpy.random.choice', 'np.random.choice', (["['branch', 'infertile', 'channel']"], {'p': '[p_branch, p_infertile, p_channel]'}), "(['branch', 'infertile', 'channel'], p=[p_branch,\n p_infertile, p_channel])\n", (1260, 1338), True, 'import numpy as np\n'), ((1396, 1426), 'numpy.random.choice', 'np.ra...
from grbl import * from pyb import I2C, delay, millis from pyb_i2c_lcd import I2cLcd from lcd_v_minus import * import time """ X 01234567890123456789 X=-xxx.xx F=xxxx* W Y=-xxx.xx S=xxxx* FM Z=-xxx.xx Idle. XYZP message --------------------- * override F - feed S - spindle W - coordinate W M O FM - flood mist ...
[ "time.time_ns", "pyb_i2c_lcd.I2cLcd", "pyb.I2C", "time.sleep_ms" ]
[((1057, 1082), 'pyb.I2C', 'I2C', (['i2c_port', 'I2C.MASTER'], {}), '(i2c_port, I2C.MASTER)\n', (1060, 1082), False, 'from pyb import I2C, delay, millis\n'), ((1315, 1343), 'pyb_i2c_lcd.I2cLcd', 'I2cLcd', (['i2c', 'i2c_addr', '(4)', '(20)'], {}), '(i2c, i2c_addr, 4, 20)\n', (1321, 1343), False, 'from pyb_i2c_lcd import...
import random from django.core.management.base import BaseCommand, CommandError from django.conf import settings from django.core.cache import caches class Command(BaseCommand): help = "Loads cache with test objects" def add_arguments(self, parser): parser.add_argument("-c", "--cache", nargs="+", typ...
[ "random.choice", "django.conf.settings.CACHES.keys" ]
[((955, 977), 'django.conf.settings.CACHES.keys', 'settings.CACHES.keys', ([], {}), '()\n', (975, 977), False, 'from django.conf import settings\n'), ((859, 881), 'django.conf.settings.CACHES.keys', 'settings.CACHES.keys', ([], {}), '()\n', (879, 881), False, 'from django.conf import settings\n'), ((1465, 1489), 'rando...
import os import tensorflow as tf from .. import preprocessing def tf_parse_line(line, data_dir): line_split = tf.strings.split(line, '\t') audio_fn = line_split[1] transcription = line_split[2] audio_filepath = tf.strings.join([data_dir, 'clips', audio_fn], '/') wav_filepath = tf.strings.subst...
[ "tensorflow.strings.length", "tensorflow.strings.split", "tensorflow.data.TextLineDataset", "tensorflow.strings.join" ]
[((118, 146), 'tensorflow.strings.split', 'tf.strings.split', (['line', '"""\t"""'], {}), "(line, '\\t')\n", (134, 146), True, 'import tensorflow as tf\n'), ((233, 284), 'tensorflow.strings.join', 'tf.strings.join', (["[data_dir, 'clips', audio_fn]", '"""/"""'], {}), "([data_dir, 'clips', audio_fn], '/')\n", (248, 284)...
import numpy as np import math class virtual_factory(object): def __init__(self, blade_specs , operation, gating_ct, non_gating_ct, options): self.options = options # Blade inputs self.n_webs = blade_specs['n_webs'] ...
[ "numpy.pmt" ]
[((26268, 26325), 'numpy.pmt', 'np.pmt', (['(self.crr / 100.0 / 12.0)', '(life * 12.0)', '(-investment)'], {}), '(self.crr / 100.0 / 12.0, life * 12.0, -investment)\n', (26274, 26325), True, 'import numpy as np\n'), ((24092, 24253), 'numpy.pmt', 'np.pmt', (['(self.crr / 100.0 / 12.0)', 'self.wcp', '(-(self.wcp / 12.0 *...
from hikari import Permissions from lightbulb import Context, Check, errors from datetime import datetime, timedelta from unicodedata import normalize from aiohttp import ClientSession from typing import Union from os import environ async def api_call(link: str, headers: dict = None, post: bool = False, json: bool =...
[ "aiohttp.ClientSession", "lightbulb.errors.ConverterFailure", "datetime.datetime.utcnow", "datetime.datetime.now", "lightbulb.errors.MissingRequiredPermission", "unicodedata.normalize", "datetime.timedelta", "lightbulb.Check" ]
[((1776, 1793), 'lightbulb.Check', 'Check', (['_is_higher'], {}), '(_is_higher)\n', (1781, 1793), False, 'from lightbulb import Context, Check, errors\n'), ((1702, 1761), 'lightbulb.errors.MissingRequiredPermission', 'errors.MissingRequiredPermission', (['Permissions.ADMINISTRATOR'], {}), '(Permissions.ADMINISTRATOR)\n...
import torchvision __all__ = ["plot_compare"] def plot_compare(sr, hr, baseline, filename): """ Plot Super-Resolution and High-Resolution image comparison """ sr, hr, baseline = sr.squeeze(), hr.squeeze(), baseline.squeeze() grid = torchvision.utils.make_grid([hr, baseline, sr]) torchvision....
[ "torchvision.utils.make_grid", "torchvision.utils.save_image" ]
[((256, 303), 'torchvision.utils.make_grid', 'torchvision.utils.make_grid', (['[hr, baseline, sr]'], {}), '([hr, baseline, sr])\n', (283, 303), False, 'import torchvision\n'), ((308, 352), 'torchvision.utils.save_image', 'torchvision.utils.save_image', (['grid', 'filename'], {}), '(grid, filename)\n', (336, 352), False...
import torch from sklearn.model_selection import train_test_split from torch.utils.data import TensorDataset, DataLoader def data_loader(targets, labels): batch_size = 10 train_samples, test_samples, train_labels, test_labels = train_test_split(targets, labels, test_size=0.2) train_samples = torch.FloatTe...
[ "sklearn.model_selection.train_test_split", "torch.LongTensor", "torch.utils.data.TensorDataset", "torch.utils.data.DataLoader", "torch.FloatTensor" ]
[((238, 286), 'sklearn.model_selection.train_test_split', 'train_test_split', (['targets', 'labels'], {'test_size': '(0.2)'}), '(targets, labels, test_size=0.2)\n', (254, 286), False, 'from sklearn.model_selection import train_test_split\n'), ((307, 339), 'torch.FloatTensor', 'torch.FloatTensor', (['train_samples'], {}...
import numbers import xnmt.tensor_tools as tt import xnmt.modelparts.decoders as decoders import xnmt.transducers.recurrent as recurrent import xnmt.transducers.base as transducers_base import xnmt.expression_seqs as expr_seq import xnmt.vocabs as vocabs class SimultaneousState(decoders.AutoRegressiveDecoderState): ...
[ "xnmt.expression_seqs.ExpressionSequence", "xnmt.transducers.base.FinalTransducerState" ]
[((1684, 1727), 'xnmt.transducers.base.FinalTransducerState', 'transducers_base.FinalTransducerState', (['h', 'c'], {}), '(h, c)\n', (1721, 1727), True, 'import xnmt.transducers.base as transducers_base\n'), ((2230, 2281), 'xnmt.expression_seqs.ExpressionSequence', 'expr_seq.ExpressionSequence', ([], {'expr_list': 'src...
#!/usr/bin/env python # -*- coding: utf-8 -*- # @Time : 2019/12/27 下午6:04 # @Title : 83. 删除排序链表中的重复元素 # @Link : https://leetcode-cn.com/problems/remove-duplicates-from-sorted-list/ QUESTION = """ 给定一个排序链表,删除所有重复的元素,使得每个元素只出现一次。 示例 1: 输入: 1->1->2 输出: 1->2 示例 2: 输入: 1->1->2->3->3 输出: 1->2->3 """ THINKING = ...
[ "utils.linked_list.LinkedListGen.nodes_to_list", "utils.linked_list.LinkedListGen.list_to_nodes" ]
[((1210, 1254), 'utils.linked_list.LinkedListGen.list_to_nodes', 'LinkedListGen.list_to_nodes', (['[1, 1, 2, 3, 3]'], {}), '([1, 1, 2, 3, 3])\n', (1237, 1254), False, 'from utils.linked_list import LinkedListGen\n'), ((1312, 1356), 'utils.linked_list.LinkedListGen.nodes_to_list', 'LinkedListGen.nodes_to_list', (['dupli...
import numpy as np import pickle as pkl from envs.babyai.oracle.teacher import Teacher class XYCorrections(Teacher): def __init__(self, *args, **kwargs): super(XYCorrections, self).__init__(*args, **kwargs) self.next_state_coords = self.empty_feedback() def empty_feedback(self): """ ...
[ "pickle.dumps", "numpy.zeros", "numpy.concatenate", "numpy.random.uniform" ]
[((541, 572), 'numpy.random.uniform', 'np.random.uniform', (['(0)', '(1)'], {'size': '(8)'}), '(0, 1, size=8)\n', (558, 572), True, 'import numpy as np\n'), ((907, 947), 'numpy.concatenate', 'np.concatenate', (['[self.next_state_coords]'], {}), '([self.next_state_coords])\n', (921, 947), True, 'import numpy as np\n'), ...
TEST_TEMP_RAW = 529191 TEST_TEMP_CMP = 24.7894877676 TEST_PRES_RAW = 326816 TEST_PRES_CMP = 1006.61517564 TEST_ALT_CMP = 57.3174 def test_temperature(): from tools import SMBusFakeDevice from bmp280 import BMP280 from calibration import BMP280Calibration dev = SMBusFakeDevice(1) # Load the fake ...
[ "calibration.BMP280Calibration", "bmp280.BMP280", "tools.SMBusFakeDevice" ]
[((280, 298), 'tools.SMBusFakeDevice', 'SMBusFakeDevice', (['(1)'], {}), '(1)\n', (295, 298), False, 'from tools import SMBusFakeDevice\n'), ((530, 549), 'bmp280.BMP280', 'BMP280', ([], {'i2c_dev': 'dev'}), '(i2c_dev=dev)\n', (536, 549), False, 'from bmp280 import BMP280\n'), ((654, 673), 'calibration.BMP280Calibration...
# mnist example, Downloaded from PML github import torch import torch.nn as nn import torch.nn.functional as F import torch.optim as optim ### MNIST code originally from https://github.com/pytorch/examples/blob/master/mnist/main.py ### from torchvision import datasets, transforms from pytorch_metric_learning import d...
[ "torch.device", "torch.nn.Dropout2d", "torch.nn.Conv2d", "pytorch_metric_learning.utils.accuracy_calculator.AccuracyCalculator", "torchvision.datasets.MNIST", "torch.utils.data.DataLoader", "torch.nn.Linear", "torch.nn.functional.relu", "torch.nn.functional.max_pool2d", "torchvision.transforms.Nor...
[((2839, 2859), 'torch.device', 'torch.device', (['"""cuda"""'], {}), "('cuda')\n", (2851, 2859), False, 'import torch\n'), ((2997, 3064), 'torchvision.datasets.MNIST', 'datasets.MNIST', (['"""."""'], {'train': '(True)', 'download': '(True)', 'transform': 'transform'}), "('.', train=True, download=True, transform=trans...
import time while True: print("Test program") time.sleep(10) pass
[ "time.sleep" ]
[((54, 68), 'time.sleep', 'time.sleep', (['(10)'], {}), '(10)\n', (64, 68), False, 'import time\n')]
#!/usr/bin/env python3.6 # -*- mode: python -*- # ============================================================================= # @@-COPYRIGHT-START-@@ # # Copyright (c) 2020 of Qualcomm Innovation Center, Inc. All rights reserved. # # @@-COPYRIGHT-END-@@ # ===========================================================...
[ "logging.getLogger", "tensorflow.shape", "pycocotools.cocoeval.COCOeval", "tensorflow.transpose", "tensorflow.data.Dataset.list_files", "tensorflow.contrib.slim.tfexample_decoder.TFExampleDecoder", "tensorflow.cast", "aimet_tensorflow.quantsim.QuantizationSimModel", "tensorflow.variables_initializer...
[((1760, 1787), 'logging.getLogger', 'logging.getLogger', (['__file__'], {}), '(__file__)\n', (1777, 1787), False, 'import logging\n'), ((2079, 2111), 'tensorflow.GPUOptions', 'tf.GPUOptions', ([], {'allow_growth': '(True)'}), '(allow_growth=True)\n', (2092, 2111), True, 'import tensorflow as tf\n'), ((2125, 2191), 'te...
import torch from torchvision import datasets import shutil import argparse import os import numpy as np from tqdm import tqdm ########### Help ########### ''' #size = (h,w) python split_train_val.py \ --data_dir /Users/aman.gupta/Documents/self/datasets/blank_page_detection/letterbox_training_data/ \ --val...
[ "os.makedirs", "argparse.ArgumentParser", "tqdm.tqdm", "numpy.floor", "os.path.join", "torchvision.datasets.ImageFolder", "os.path.basename", "shutil.copy", "numpy.random.shuffle" ]
[((524, 662), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {'description': '"""this script splits classification data into train and val based on ratio provided by user"""'}), "(description=\n 'this script splits classification data into train and val based on ratio provided by user'\n )\n", (547, 6...
""" Yahoo Finance Model """ __docformat__ = "numpy" import logging import pandas as pd import yfinance as yf from gamestonk_terminal.decorators import log_start_end from gamestonk_terminal.rich_config import console logger = logging.getLogger(__name__) INDICES = { "sp500": {"name": "S&P 500", "ticker": "^GSPC"...
[ "logging.getLogger", "pandas.Series", "gamestonk_terminal.decorators.log_start_end", "gamestonk_terminal.rich_config.console.print", "yfinance.download" ]
[((229, 256), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (246, 256), False, 'import logging\n'), ((1064, 1089), 'gamestonk_terminal.decorators.log_start_end', 'log_start_end', ([], {'log': 'logger'}), '(log=logger)\n', (1077, 1089), False, 'from gamestonk_terminal.decorators import lo...
# Simple XML against XSD Validator for Python 2.7 - 3.2 # to run this script you need additionally: lxml (http://lxml.de) # author: <NAME>, 2013 import sys from lxml import etree xsd_files = [] xml_files = [] def usage(): print("Usage: ") print("python XSDValidator.py <list of xml files> <list of xsd file...
[ "lxml.etree.XMLSchema", "lxml.etree.parse", "sys.exit" ]
[((525, 553), 'lxml.etree.XMLSchema', 'etree.XMLSchema', ([], {'file': 'schema'}), '(file=schema)\n', (540, 553), False, 'from lxml import etree\n'), ((1019, 1029), 'sys.exit', 'sys.exit', ([], {}), '()\n', (1027, 1029), False, 'import sys\n'), ((1284, 1294), 'sys.exit', 'sys.exit', ([], {}), '()\n', (1292, 1294), Fals...
from __future__ import unicode_literals from django.db import models from django.utils.encoding import python_2_unicode_compatible from zshoes.stores.models import Store @python_2_unicode_compatible class Article(models.Model): """ Entity that represents the articles of the store """ #: Name of the ...
[ "django.db.models.FloatField", "django.db.models.TextField", "django.db.models.ForeignKey", "django.db.models.PositiveIntegerField", "django.db.models.CharField" ]
[((339, 370), 'django.db.models.CharField', 'models.CharField', ([], {'max_length': '(45)'}), '(max_length=45)\n', (355, 370), False, 'from django.db import models\n'), ((423, 462), 'django.db.models.TextField', 'models.TextField', ([], {'null': '(True)', 'blank': '(True)'}), '(null=True, blank=True)\n', (439, 462), Fa...
#!/usr/bin/env python3 import sys, os, json import random # Check to make sure we are running the correct version of Python assert sys.version_info >= (3,7), "This script requires at least Python 3.7" # The game and item description files (in the same folder as this script) game_file = 'game.json' # Load the content...
[ "os.path.join", "os.getcwd", "os.path.dirname", "os._exit", "json.load" ]
[((693, 713), 'json.load', 'json.load', (['json_file'], {}), '(json_file)\n', (702, 713), False, 'import sys, os, json\n'), ((829, 840), 'os._exit', 'os._exit', (['(1)'], {}), '(1)\n', (837, 840), False, 'import sys, os, json\n'), ((574, 585), 'os.getcwd', 'os.getcwd', ([], {}), '()\n', (583, 585), False, 'import sys, ...
""" This module provides a sequence class which can be used for cyclic values """ from typing import Generic, TypeVar T = TypeVar('T') class Cycle(list,Generic[T]): """This class can be used to store cyclic values""" def __getitem__(self,key:int) -> T: return super().__getitem__(key%len(self)) ...
[ "typing.TypeVar" ]
[((124, 136), 'typing.TypeVar', 'TypeVar', (['"""T"""'], {}), "('T')\n", (131, 136), False, 'from typing import Generic, TypeVar\n')]
# -*- coding: utf-8 -*- # Generated by Django 1.9.4 on 2016-03-17 12:32 from __future__ import unicode_literals from django.db import migrations, models class Migration(migrations.Migration): initial = True dependencies = [ ] operations = [ migrations.CreateModel( name='orders'...
[ "django.db.models.DateField", "django.db.models.AutoField", "django.db.models.PositiveIntegerField", "django.db.models.DecimalField", "django.db.models.CharField" ]
[((366, 459), 'django.db.models.AutoField', 'models.AutoField', ([], {'auto_created': '(True)', 'primary_key': '(True)', 'serialize': '(False)', 'verbose_name': '"""ID"""'}), "(auto_created=True, primary_key=True, serialize=False,\n verbose_name='ID')\n", (382, 459), False, 'from django.db import migrations, models\...
############################################### # <NAME> - PG Applied AI - Programming # Unit tests, for the graph algorithms ############################################### import unittest # unit testing ftw from board import Board import numpy as np import play class TestMethods(unittest.TestCas...
[ "board.Board" ]
[((379, 387), 'board.Board', 'Board', (['(4)'], {}), '(4)\n', (384, 387), False, 'from board import Board\n'), ((686, 694), 'board.Board', 'Board', (['(4)'], {}), '(4)\n', (691, 694), False, 'from board import Board\n'), ((991, 999), 'board.Board', 'Board', (['(4)'], {}), '(4)\n', (996, 999), False, 'from board import ...
import dash_bootstrap_components as dbc import dash_core_components as dcc import dash_html_components as html from dash.dependencies import Input, Output from . import dash_managers, dash_queue, dash_service from .app import app from .navbar import navbar body = dbc.Container( [ dbc.Row([ dbc...
[ "dash_bootstrap_components.Button", "dash.dependencies.Output", "dash_core_components.Location", "dash.dependencies.Input", "dash_html_components.H2", "dash_core_components.Graph", "dash_html_components.P", "dash_html_components.Div" ]
[((1018, 1052), 'dash.dependencies.Output', 'Output', (['"""page-content"""', '"""children"""'], {}), "('page-content', 'children')\n", (1024, 1052), False, 'from dash.dependencies import Input, Output\n'), ((933, 970), 'dash_core_components.Location', 'dcc.Location', ([], {'id': '"""url"""', 'refresh': '(False)'}), "(...
# -*- coding: utf-8 -*- """ Created on Mon Apr 8 11:09:33 2019 @author: 10365 """ #CreateDataSet import numpy as np import sys sys.path.append('../subway_system') sys.path.append('../ato_agent') import TrainAndRoadCharacter as trc import trainRunningModel as trm import pandas as pds import matplotlib.pyplot as pl...
[ "numpy.mat", "TrainAndRoadCharacter.getRoadGradinet", "trainRunningModel.Train_model", "TrainAndRoadCharacter.TrainAndRoadData", "pandas.DataFrame", "atoController.PidController", "TrainAndRoadCharacter.getNextSpeedLimit", "TrainAndRoadCharacter.plotSpeedLimitRoadGrad", "matplotlib.pyplot.plot", "...
[((133, 168), 'sys.path.append', 'sys.path.append', (['"""../subway_system"""'], {}), "('../subway_system')\n", (148, 168), False, 'import sys\n'), ((169, 200), 'sys.path.append', 'sys.path.append', (['"""../ato_agent"""'], {}), "('../ato_agent')\n", (184, 200), False, 'import sys\n'), ((1810, 1843), 'TrainAndRoadChara...
#!/usr/bin/python3 import boto3 import json if __name__ == '__main__': ec2_client = boto3.client('ec2') ec2_filter = [{'Name': 'tag:role', 'Values': ['ecs-cluster']}] instances=ec2_client.describe_tags(Filters=ec2_filter) #get only the instance_ids instance_ids = [] for i in instances['Tags'...
[ "json.dumps", "boto3.client" ]
[((91, 110), 'boto3.client', 'boto3.client', (['"""ec2"""'], {}), "('ec2')\n", (103, 110), False, 'import boto3\n'), ((882, 905), 'json.dumps', 'json.dumps', (['output_dict'], {}), '(output_dict)\n', (892, 905), False, 'import json\n')]
from cstream import stdwar from svgen import Point, Vector, SVG, Figure, Camera, Domain, Map, Surface, Animation from math import radians, sin, cos, pi, hypot, sqrt import sys from tqdm import tqdm from svgen.svglib.math import Transform COLOR = "#3703b3" if len(sys.argv) > 1 and sys.argv[1] == "-o": proj = C...
[ "svgen.Domain", "svgen.Surface", "math.radians", "svgen.Animation", "math.cos", "svgen.svglib.math.Transform.scale", "math.sin", "svgen.SVG" ]
[((557, 578), 'svgen.Animation', 'Animation', ([], {'delay': '(1000)'}), '(delay=1000)\n', (566, 578), False, 'from svgen import Point, Vector, SVG, Figure, Camera, Domain, Map, Surface, Animation\n'), ((584, 616), 'svgen.Domain', 'Domain', (['(-250, 250)', '(250, -250)'], {}), '((-250, 250), (250, -250))\n', (590, 616...
import sys sys.path.insert(0, '../') from mocap.settings import get_amass_validation_files, get_amass_test_files from mocap.math.amass_fk import rotmat2euclidean, exp2euclidean from mocap.visualization.sequence import SequenceVisualizer from mocap.math.mirror_smpl import mirror_p3d from mocap.datasets.dataset import Li...
[ "mocap.datasets.combined.Combined", "sys.path.insert", "mocap.settings.get_amass_validation_files", "mocap.datasets.h36m.H36M_FixedSkeleton", "mocap.datasets.amass.AMASS_SMPL3d", "numpy.array", "mocap.settings.get_amass_test_files", "mocap.visualization.sequence.SequenceVisualizer" ]
[((11, 36), 'sys.path.insert', 'sys.path.insert', (['(0)', '"""../"""'], {}), "(0, '../')\n", (26, 36), False, 'import sys\n'), ((617, 645), 'mocap.settings.get_amass_validation_files', 'get_amass_validation_files', ([], {}), '()\n', (643, 645), False, 'from mocap.settings import get_amass_validation_files, get_amass_t...
from django.test import TestCase from review.models import Review class TestReviewModel(TestCase): ''' Test suite for review modules. ''' def setUp(self): ''' Set up test data for the review model. ''' Review.objects.create( feedback='Test rev...
[ "review.models.Review.objects.all", "review.models.Review.objects.create", "review.models.Review.objects.get" ]
[((266, 351), 'review.models.Review.objects.create', 'Review.objects.create', ([], {'feedback': '"""Test review"""', 'riderReview': '"""Test review content"""'}), "(feedback='Test review', riderReview='Test review content'\n )\n", (287, 351), False, 'from review.models import Review\n'), ((651, 693), 'review.models....
"""Script to start webserving.""" from wembedder.app import create_app app = create_app() if __name__ == '__main__': app.run(debug=True)
[ "wembedder.app.create_app" ]
[((81, 93), 'wembedder.app.create_app', 'create_app', ([], {}), '()\n', (91, 93), False, 'from wembedder.app import create_app\n')]
from openprocurement.tender.core.procedure.serializers.base import ListSerializer from openprocurement.tender.core.procedure.serializers.document import ConfidentialDocumentSerializer from openprocurement.tender.core.procedure.serializers.parameter import ParameterSerializer from openprocurement.tender.esco.procedure.s...
[ "openprocurement.tender.core.procedure.serializers.base.ListSerializer" ]
[((667, 701), 'openprocurement.tender.core.procedure.serializers.base.ListSerializer', 'ListSerializer', (['LotValueSerializer'], {}), '(LotValueSerializer)\n', (681, 701), False, 'from openprocurement.tender.core.procedure.serializers.base import ListSerializer\n'), ((724, 770), 'openprocurement.tender.core.procedure....
from typing import List import argparse from detectron2.evaluation import COCOEvaluator, inference_on_dataset from detectron2.config import get_cfg from detectron2 import model_zoo from detectron2.data.datasets import register_coco_instances from detectron2.data import build_detection_test_loader from trainers import...
[ "trainers.MyTrainer", "detectron2.data.datasets.register_coco_instances", "detectron2.config.get_cfg", "argparse.ArgumentParser", "detectron2.model_zoo.get_config_file", "detectron2.evaluation.inference_on_dataset", "detectron2.data.build_detection_test_loader", "detectron2.evaluation.COCOEvaluator" ]
[((356, 435), 'detectron2.data.datasets.register_coco_instances', 'register_coco_instances', (['args.ds_test', '{}', 'args.ds_test_json', 'args.ds_test_imgs'], {}), '(args.ds_test, {}, args.ds_test_json, args.ds_test_imgs)\n', (379, 435), False, 'from detectron2.data.datasets import register_coco_instances\n'), ((446, ...
# -*- coding: utf-8 -*- # Form implementation generated from reading ui file 'npc/gui/uis/new_character.ui' # # Created by: PyQt5 UI code generator 5.7.1 # # WARNING! All changes made in this file will be lost! from PyQt5 import QtCore, QtGui, QtWidgets class Ui_NewCharacterDialog(object): def setupUi(self, NewC...
[ "PyQt5.QtWidgets.QDialogButtonBox", "PyQt5.QtWidgets.QComboBox", "PyQt5.QtWidgets.QPlainTextEdit", "PyQt5.QtCore.QMetaObject.connectSlotsByName", "PyQt5.QtWidgets.QVBoxLayout", "PyQt5.QtWidgets.QLabel", "PyQt5.QtWidgets.QGroupBox", "PyQt5.QtWidgets.QSizePolicy", "PyQt5.QtWidgets.QFormLayout", "PyQ...
[((465, 571), 'PyQt5.QtWidgets.QSizePolicy', 'QtWidgets.QSizePolicy', (['QtWidgets.QSizePolicy.MinimumExpanding', 'QtWidgets.QSizePolicy.MinimumExpanding'], {}), '(QtWidgets.QSizePolicy.MinimumExpanding, QtWidgets.\n QSizePolicy.MinimumExpanding)\n', (486, 571), False, 'from PyQt5 import QtCore, QtGui, QtWidgets\n')...
#!/usr/bin/env python3 # -*- coding: utf-8 -*- # Contains a context manager for temporarily introducing an environment var. import os import contextlib @contextlib.contextmanager def use_environment_variable(key, value): """ Used to temporarily introduce a new environment variable as if it was set by th...
[ "os.environ.pop" ]
[((617, 636), 'os.environ.pop', 'os.environ.pop', (['key'], {}), '(key)\n', (631, 636), False, 'import os\n')]
import matplotlib import matplotlib.pyplot as plt import numpy as np from PySide2.QtWidgets import QVBoxLayout, QWidget from traitlets import HasTraits, Instance, Bool, directional_link from regexport.model import AppState from regexport.views.utils import HasWidget matplotlib.use('Qt5Agg') from matplotlib.backends...
[ "traitlets.directional_link", "matplotlib.backends.backend_qt5agg.NavigationToolbar2QT", "numpy.histogram", "regexport.views.utils.HasWidget.__init__", "matplotlib.use", "traitlets.Instance", "PySide2.QtWidgets.QWidget", "numpy.concatenate", "matplotlib.backends.backend_qt5agg.FigureCanvasQTAgg", ...
[((269, 293), 'matplotlib.use', 'matplotlib.use', (['"""Qt5Agg"""'], {}), "('Qt5Agg')\n", (283, 293), False, 'import matplotlib\n'), ((470, 507), 'traitlets.Instance', 'Instance', (['np.ndarray'], {'allow_none': '(True)'}), '(np.ndarray, allow_none=True)\n', (478, 507), False, 'from traitlets import HasTraits, Instance...
""" Run all the example files and convert them to markdown files containing the output. Uses `pweave`. It is not installed by default. To install: pip install pweave """ import pweave, datetime, glob, os def publish_to_markdown(python_file: str, output_file: str): doc = pweave.Pweb(python_file, kernel="pyt...
[ "datetime.date.today", "os.path.basename", "glob.glob", "pweave.Pweb" ]
[((284, 371), 'pweave.Pweb', 'pweave.Pweb', (['python_file'], {'kernel': '"""python3"""', 'doctype': '"""markdown"""', 'output': 'output_file'}), "(python_file, kernel='python3', doctype='markdown', output=\n output_file)\n", (295, 371), False, 'import pweave, datetime, glob, os\n'), ((841, 858), 'glob.glob', 'glob....
import binascii import uuid from collections import UserDict from functools import cmp_to_key, wraps from nanolib import Block as RawBlock from nanolib import nbase32_to_bytes, get_account_id __all__ = ( "RawBlock", "BlockProxy", "Callbacks", "CallbackSlot", "AccountIDDict" ) class BlockProxy(object): """ ...
[ "binascii.hexlify", "nanolib.nbase32_to_bytes", "uuid.uuid4" ]
[((4989, 5025), 'nanolib.nbase32_to_bytes', 'nbase32_to_bytes', (['account_id[-60:-8]'], {}), '(account_id[-60:-8])\n', (5005, 5025), False, 'from nanolib import nbase32_to_bytes, get_account_id\n'), ((3431, 3443), 'uuid.uuid4', 'uuid.uuid4', ([], {}), '()\n', (3441, 3443), False, 'import uuid\n'), ((6761, 6782), 'bina...
import json from copy import deepcopy from functools import total_ordering from typing import List, Any, Union from canvasxpress.config.type import CXConfig, CXString, CXInt, CXFloat, CXBool, \ CXList, CXDict, CXRGBColor, CXRGBAColor from canvasxpress.data.convert import CXDictConvertable, CXListConvertable @tot...
[ "canvasxpress.config.type.CXRGBColor", "copy.deepcopy", "canvasxpress.config.type.CXRGBAColor.is_color_list", "canvasxpress.config.type.CXString", "canvasxpress.config.type.CXRGBAColor.is_color_dict", "canvasxpress.config.type.CXList", "canvasxpress.config.type.CXInt", "canvasxpress.config.type.CXBool...
[((6342, 6361), 'canvasxpress.config.type.CXInt', 'CXInt', (['label', 'value'], {}), '(label, value)\n', (6347, 6361), False, 'from canvasxpress.config.type import CXConfig, CXString, CXInt, CXFloat, CXBool, CXList, CXDict, CXRGBColor, CXRGBAColor\n'), ((12090, 12106), 'copy.deepcopy', 'deepcopy', (['config'], {}), '(c...
# Generated by Django 3.1.5 on 2021-01-17 15:24 import cloudinary.models from django.conf import settings from django.db import migrations, models import django.db.models.deletion class Migration(migrations.Migration): initial = True dependencies = [ migrations.swappable_dependency(settings.AUTH_US...
[ "django.db.models.OneToOneField", "django.db.models.TextField", "django.db.models.ForeignKey", "django.db.models.IntegerField", "django.db.models.ManyToManyField", "django.db.models.AutoField", "django.db.models.DateTimeField", "django.db.migrations.swappable_dependency", "django.db.models.CharField...
[((272, 329), 'django.db.migrations.swappable_dependency', 'migrations.swappable_dependency', (['settings.AUTH_USER_MODEL'], {}), '(settings.AUTH_USER_MODEL)\n', (303, 329), False, 'from django.db import migrations, models\n'), ((461, 554), 'django.db.models.AutoField', 'models.AutoField', ([], {'auto_created': '(True)...
############################################################################ # # Copyright (C) 2016 The Qt Company Ltd. # Contact: https://www.qt.io/licensing/ # # This file is part of Qt Creator. # # Commercial License Usage # Licensees holding valid commercial Qt licenses may use this file in # accordance with the co...
[ "__builtin__.bool" ]
[((4444, 4466), '__builtin__.bool', '__builtin__.bool', (['clip'], {}), '(clip)\n', (4460, 4466), False, 'import __builtin__\n')]
# MIT License # Copyright (c) 2018-2019 <NAME>, <NAME>, <NAME>, <NAME> # Permission is hereby granted, free of charge, to any person obtaining a copy # of this software and associated documentation files (the "Software"), to deal # in the Software without restriction, including without limitation the rights # to use...
[ "mathutils.Vector" ]
[((2668, 2756), 'mathutils.Vector', 'Vector', (['(rcurve.Arc.StartPoint.X, rcurve.Arc.StartPoint.Y, rcurve.Arc.StartPoint.Z)'], {}), '((rcurve.Arc.StartPoint.X, rcurve.Arc.StartPoint.Y, rcurve.Arc.\n StartPoint.Z))\n', (2674, 2756), False, 'from mathutils import Vector\n'), ((2770, 2847), 'mathutils.Vector', 'Vector...
#!/usr/bin/env python3 import sys import subprocess from align_videos_by_soundtrack.align import SyncDetector, cli_common remove = [] remove_specified = False remove_all = False add = [] add_specified = False copy_subtitles = False files = [] print_offset_only = False for argument in sys.argv[1:]: if argument == ...
[ "align_videos_by_soundtrack.align.cli_common.logger_config", "align_videos_by_soundtrack.align.SyncDetector", "subprocess.check_call" ]
[((909, 935), 'align_videos_by_soundtrack.align.cli_common.logger_config', 'cli_common.logger_config', ([], {}), '()\n', (933, 935), False, 'from align_videos_by_soundtrack.align import SyncDetector, cli_common\n'), ((942, 956), 'align_videos_by_soundtrack.align.SyncDetector', 'SyncDetector', ([], {}), '()\n', (954, 95...
import numpy as np import pandas as pd import sys import tensorflow as tf from keras.layers import Input, Dense, Lambda, Flatten, Reshape, Activation, Dropout, Add, TimeDistributed, Multiply, Conv1D, Conv2D, MaxPooling1D, AveragePooling1D from keras.models import Model, Sequential, load_model from keras import backend ...
[ "keras.models.load_model", "keras.layers.AveragePooling1D", "pandas.read_csv", "keras.callbacks.ModelCheckpoint", "keras.callbacks.History", "keras.layers.Lambda", "pandas.DataFrame.from_dict", "keras.layers.Add", "keras.layers.Input", "keras.layers.Multiply", "keras.models.Model", "keras.laye...
[((4182, 4234), 'keras.layers.Input', 'Input', ([], {'shape': 'self.input_shape', 'name': '"""original_input"""'}), "(shape=self.input_shape, name='original_input')\n", (4187, 4234), False, 'from keras.layers import Input, Dense, Lambda, Flatten, Reshape, Activation, Dropout, Add, TimeDistributed, Multiply, Conv1D, Con...
# Generated by Django 3.1.6 on 2021-03-08 10:44 from django.db import migrations, models class Migration(migrations.Migration): initial = True dependencies = [ ] operations = [ migrations.CreateModel( name='Album', fields=[ ('id', models.AutoField(au...
[ "django.db.models.DateTimeField", "django.db.models.AutoField", "django.db.models.CharField" ]
[((301, 394), 'django.db.models.AutoField', 'models.AutoField', ([], {'auto_created': '(True)', 'primary_key': '(True)', 'serialize': '(False)', 'verbose_name': '"""ID"""'}), "(auto_created=True, primary_key=True, serialize=False,\n verbose_name='ID')\n", (317, 394), False, 'from django.db import migrations, models\...
#!/usr/bin/env python3 import platform import serial import sys from config import Settings dev = serial.Serial(Settings.SERIAL_DEVICE, Settings.BAUD_RATE) print("> Returned data:", file=sys.stderr) while True: x = dev.read() sys.stdout.buffer.write(x) sys.stdout.flush()
[ "sys.stdout.flush", "sys.stdout.buffer.write", "serial.Serial" ]
[((99, 156), 'serial.Serial', 'serial.Serial', (['Settings.SERIAL_DEVICE', 'Settings.BAUD_RATE'], {}), '(Settings.SERIAL_DEVICE, Settings.BAUD_RATE)\n', (112, 156), False, 'import serial\n'), ((237, 263), 'sys.stdout.buffer.write', 'sys.stdout.buffer.write', (['x'], {}), '(x)\n', (260, 263), False, 'import sys\n'), ((2...
# # Licensed to the Apache Software Foundation (ASF) under one # or more contributor license agreements. See the NOTICE file # distributed with this work for additional information # regarding copyright ownership. The ASF licenses this file # to you under the Apache License, Version 2.0 (the # "License"); you may not...
[ "logging.getLogger", "json.loads", "sqlite3.connect", "email.utils.parseaddr", "time.mktime", "os.path.join", "re.match", "email.utils.parsedate", "textwrap.wrap", "email.utils.formatdate", "quopri.decodestring", "time.gmtime" ]
[((1087, 1117), 'logging.getLogger', '_logging.getLogger', (['"""haystack"""'], {}), "('haystack')\n", (1105, 1117), True, 'import logging as _logging\n'), ((1163, 1194), 'json.loads', '_json.loads', (["_strings['topics']"], {}), "(_strings['topics'])\n", (1174, 1194), True, 'import json as _json\n'), ((1315, 1362), 'o...
""" Initialization file for a GDC API client. """ from copy import copy import types from .client import BaseClient # Aliases COMMON_ALIASES = { "_get_cases": 'get_cases', "_get_mappings": 'get_mappings', "_get_genes": 'get_genes', "_get_ssm_occurrences": 'get_ssm_occurrences', } # API speci...
[ "types.FunctionType", "copy.copy" ]
[((349, 369), 'copy.copy', 'copy', (['COMMON_ALIASES'], {}), '(COMMON_ALIASES)\n', (353, 369), False, 'from copy import copy\n'), ((723, 742), 'copy.copy', 'copy', (['COMMON_KWARGS'], {}), '(COMMON_KWARGS)\n', (727, 742), False, 'from copy import copy\n'), ((1144, 1245), 'types.FunctionType', 'types.FunctionType', (['f...
import logging from typing import Any, Dict, List, NewType import mlflow import numpy as np import pandas as pd import torch import transformers from mlflow.models import ModelSignature from mlflow.pyfunc import PythonModel from mlflow.types import ColSpec, DataType, Schema, TensorSpec from mlflow.utils.environment im...
[ "logging.getLogger", "typing.NewType", "transformers.AutoModelForSequenceClassification.from_pretrained", "transformers.AutoTokenizer.from_pretrained", "numpy.dtype", "transformers.pipeline", "mlflow.types.ColSpec" ]
[((666, 687), 'typing.NewType', 'NewType', (['"""Model"""', 'Any'], {}), "('Model', Any)\n", (673, 687), False, 'from typing import Any, Dict, List, NewType\n'), ((700, 725), 'typing.NewType', 'NewType', (['"""Tokenizer"""', 'Any'], {}), "('Tokenizer', Any)\n", (707, 725), False, 'from typing import Any, Dict, List, Ne...
import rclpy from rclpy.action import ActionClient from rclpy.node import Node from functools import partial from my_robot_interfaces.action import Test class TestActionClient(Node): def __init__(self): super().__init__('test_action_client') self.declare_parameter("secs", 2) self....
[ "rclpy.spin", "rclpy.action.ActionClient", "functools.partial", "my_robot_interfaces.action.Test.Goal", "rclpy.init", "rclpy.shutdown" ]
[((2464, 2485), 'rclpy.init', 'rclpy.init', ([], {'args': 'args'}), '(args=args)\n', (2474, 2485), False, 'import rclpy\n'), ((2534, 2564), 'rclpy.spin', 'rclpy.spin', (['test_action_client'], {}), '(test_action_client)\n', (2544, 2564), False, 'import rclpy\n'), ((2569, 2585), 'rclpy.shutdown', 'rclpy.shutdown', ([], ...
import sys import pygame def draw_canvas(screen, colors): screen.fill(colors[0]) pygame.draw.rect(screen, colors[5], (20, 20, 500, 500)) index = 1 # skip light grey for row in range(2): for column in range(4): pygame.draw.rect(screen, colors[index], ((60 * column) + 20...
[ "pygame.display.update", "pygame.draw.rect", "pygame.font.SysFont" ]
[((97, 152), 'pygame.draw.rect', 'pygame.draw.rect', (['screen', 'colors[5]', '(20, 20, 500, 500)'], {}), '(screen, colors[5], (20, 20, 500, 500))\n', (113, 152), False, 'import pygame\n'), ((405, 461), 'pygame.draw.rect', 'pygame.draw.rect', (['screen', 'colors[5]', '(280, 530, 120, 55)'], {}), '(screen, colors[5], (2...
from suds.client import Client from suds import WebFault from model.project import Project class SoapHelper: def __init__(self, app): self.app = app def can_login(self, username, password): client = Client(self.app.base_url + "api/soap/mantisconnect.php?wsdl") try: client....
[ "suds.client.Client", "model.project.Project" ]
[((226, 287), 'suds.client.Client', 'Client', (["(self.app.base_url + 'api/soap/mantisconnect.php?wsdl')"], {}), "(self.app.base_url + 'api/soap/mantisconnect.php?wsdl')\n", (232, 287), False, 'from suds.client import Client\n'), ((549, 610), 'suds.client.Client', 'Client', (["(self.app.base_url + 'api/soap/mantisconne...
import pprint import requests import argparse def login_admin(admin_id, admin_secret, url, verbose): headers = {'Content-Type': 'application/x-www-form-urlencoded', 'Accept': 'application/json'} payload = {'grant_type': 'client_credentials', 'response_type': 'token', ...
[ "requests.post", "argparse.ArgumentParser", "requests.get", "requests.put", "pprint.pprint" ]
[((364, 474), 'requests.post', 'requests.post', (['f"""http://{url}/oauth/token"""'], {'headers': 'headers', 'params': 'payload', 'auth': '(admin_id, admin_secret)'}), "(f'http://{url}/oauth/token', headers=headers, params=payload,\n auth=(admin_id, admin_secret))\n", (377, 474), False, 'import requests\n'), ((1042,...
"Code used to generate data for experiments with synthetic data" import math import typing as ty import numba import numpy as np import torch import torch.nn as nn from numba.experimental import jitclass from tqdm.auto import tqdm class MLP(nn.Module): def __init__( self, *, d_in: int, ...
[ "numpy.ones", "numpy.flatnonzero", "torch.relu", "torch.nn.init.kaiming_normal_", "numba.experimental.jitclass", "math.sqrt", "torch.nn.init._calculate_fan_in_and_fan_out", "numpy.random.randint", "numpy.zeros", "torch.nn.init.uniform_", "torch.nn.Linear", "tqdm.auto.tqdm", "numpy.random.ran...
[((1341, 1556), 'numba.experimental.jitclass', 'jitclass', ([], {'spec': "[('left_children', numba.int64[:]), ('right_children', numba.int64[:]), (\n 'feature', numba.int64[:]), ('threshold', numba.float32[:]), ('value',\n numba.float32[:]), ('is_leaf', numba.int64[:])]"}), "(spec=[('left_children', numba.int64[:...