code
stringlengths
22
1.05M
apis
listlengths
1
3.31k
extract_api
stringlengths
75
3.25M
#!/usr/bin/env python # coding: utf-8 # /*########################################################################## # # Copyright (c) 2016-2018 European Synchrotron Radiation Facility # # Permission is hereby granted, free of charge, to any person obtaining a copy # of this software and associated documentation files ...
[ "numpy.trapz", "silx.gui.qt.QApplication", "silx.gui.plot.Plot1D", "numpy.sum", "silx.gui.plot.stats.stats.StatBase.__init__", "numpy.arange" ]
[((3016, 3035), 'silx.gui.qt.QApplication', 'qt.QApplication', (['[]'], {}), '([])\n', (3031, 3035), False, 'from silx.gui import qt\n'), ((3048, 3056), 'silx.gui.plot.Plot1D', 'Plot1D', ([], {}), '()\n', (3054, 3056), False, 'from silx.gui.plot import Plot1D\n'), ((3066, 3082), 'numpy.arange', 'numpy.arange', (['(21)'...
#loads and trains data using simple CPU data loading technique #adapted from https://colab.research.google.com/github/tensorflow/docs/blob/master/site/en/r2/tutorials/load_data/images.ipynb#scrollTo=qj_U09xpDvOg import pathlib import tensorflow as tf AUTOTUNE = tf.data.experimental.AUTOTUNE def preprocess_image(imag...
[ "matplotlib.pyplot.grid", "tensorflow.data.experimental.shuffle_and_repeat", "tensorflow.io.read_file", "tensorflow.cast", "tensorflow.keras.layers.GlobalAveragePooling2D", "matplotlib.pyplot.imshow", "tensorflow.data.Dataset.from_tensor_slices", "pathlib.Path", "matplotlib.pyplot.yticks", "tensor...
[((789, 817), 'pathlib.Path', 'pathlib.Path', (['data_root_orig'], {}), '(data_root_orig)\n', (801, 817), False, 'import pathlib\n'), ((987, 1018), 'random.shuffle', 'random.shuffle', (['all_image_paths'], {}), '(all_image_paths)\n', (1001, 1018), False, 'import random\n'), ((1680, 1695), 'matplotlib.pyplot.grid', 'plt...
from PairedNeurons import PairedNeurons from matplotlib import pyplot as plt import os import numpy as np import cv2 from xlwt import Workbook from skimage.segmentation import clear_border SMOOTH = 1e-6 def iou_numpy(outputs: np.array, labels: np.array): # outputs = outputs.squeeze(2) intersection = (outpu...
[ "numpy.uint8", "numpy.ones", "cv2.threshold", "os.path.join", "skimage.segmentation.clear_border", "matplotlib.pyplot.close", "cv2.morphologyEx", "PairedNeurons.PairedNeurons", "cv2.distanceTransform", "cv2.dilate", "cv2.subtract", "xlwt.Workbook", "matplotlib.pyplot.subplots" ]
[((748, 800), 'PairedNeurons.PairedNeurons', 'PairedNeurons', (['img_dir', 'csv_dir', '(256)'], {'is_train': '(False)'}), '(img_dir, csv_dir, 256, is_train=False)\n', (761, 800), False, 'from PairedNeurons import PairedNeurons\n'), ((870, 880), 'xlwt.Workbook', 'Workbook', ([], {}), '()\n', (878, 880), False, 'from xlw...
import unittest from .an2cn import An2Cn class An2CnTest(unittest.TestCase): def setUp(self): self.error_input_data = [ u"123.1.1", u"0.1零" ] self.integer_data = { 0: [u"零", u"零"], 1: [u"一", u"壹"], 11: [u"十一", u"拾壹"], ...
[ "unittest.main" ]
[((1549, 1564), 'unittest.main', 'unittest.main', ([], {}), '()\n', (1562, 1564), False, 'import unittest\n')]
#!/usr/bin/env python3 import argparse from game.game import Game def main(): """ Reversi game with human player vs AI player. """ parser = argparse.ArgumentParser() parser.add_argument('--timeout', help="Number of seconds the brain is allowed to think before making its move", ...
[ "game.game.Game", "argparse.ArgumentParser" ]
[((156, 181), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {}), '()\n', (179, 181), False, 'import argparse\n'), ((964, 1025), 'game.game.Game', 'Game', ([], {'timeout': 'args.timeout', 'colour': 'args.text', 'players': 'players'}), '(timeout=args.timeout, colour=args.text, players=players)\n', (968, 1025...
import os from collections import namedtuple from ConfigSpace.read_and_write import json as cs_json import nasbench301 as nb # Default dirs for models # Note: Uses 0.9 as the default models, switch to 1.0 to use 1.0 models version = '0.9' current_dir = os.path.dirname(os.path.abspath(__file__)) models_0_9_dir = os....
[ "os.path.exists", "collections.namedtuple", "os.path.join", "ConfigSpace.read_and_write.json.read", "nasbench301.download_models", "os.path.abspath", "nasbench301.load_ensemble" ]
[((317, 359), 'os.path.join', 'os.path.join', (['current_dir', '"""nb_models_0.9"""'], {}), "(current_dir, 'nb_models_0.9')\n", (329, 359), False, 'import os\n'), ((531, 573), 'os.path.join', 'os.path.join', (['current_dir', '"""nb_models_1.0"""'], {}), "(current_dir, 'nb_models_1.0')\n", (543, 573), False, 'import os\...
# emacs: -*- mode: python; py-indent-offset: 4; indent-tabs-mode: nil -*- # vi: set ft=python sts=4 ts=4 sw=4 et: ### ### ### ### ### ### ### ### ### ### ### ### ### ### ### ### ### ### ### ## # # See COPYING file distributed along with the PyMVPA package for the # copyright and license terms. # ### ### ### ### ###...
[ "mvpa2.base.dochelpers.borrowkwargs", "mvpa2.measures.adhocsearchlightbase._STATS", "numpy.unique", "numpy.ones", "mvpa2.misc.neighborhood.IndexQueryEngine", "numpy.square", "numpy.asanyarray", "numpy.sum", "numpy.zeros", "numpy.empty", "mvpa2.base.dochelpers._repr_attrs", "numpy.argmin", "m...
[((8166, 8243), 'mvpa2.base.dochelpers.borrowkwargs', 'borrowkwargs', (['M1NNSearchlight', '"""__init__"""'], {'exclude': "['roi_ids', 'queryengine']"}), "(M1NNSearchlight, '__init__', exclude=['roi_ids', 'queryengine'])\n", (8178, 8243), False, 'from mvpa2.base.dochelpers import borrowkwargs, _repr_attrs\n'), ((1046, ...
import os from django.db import migrations from django.conf import settings import privatemedia.fields import privatemedia.storage import transmittals.fileutils def move_dirs(*args): protected_root = settings.PROTECTED_ROOT if not os.path.exists(protected_root): os.makedirs(protected_root) priva...
[ "os.path.exists", "os.makedirs", "os.rename", "os.path.join", "django.db.migrations.RunPython" ]
[((446, 490), 'os.path.join', 'os.path.join', (['protected_root', '"""transmittals"""'], {}), "(protected_root, 'transmittals')\n", (458, 490), False, 'import os\n'), ((505, 547), 'os.path.join', 'os.path.join', (['private_root', '"""transmittals"""'], {}), "(private_root, 'transmittals')\n", (517, 547), False, 'import...
#!/usr/bin/env python # coding: utf8 """MMMM-Facial-Recognition-OCV3 - MagicMirror Module The MIT License (MIT) Copyright (c) 2018 <NAME> (MIT License) Based on work by <NAME> (Copyright 2016) (MIT License) """ from lib.tools.train import ToolsTrain if __name__ == '__main__': ToolsTrain().train()
[ "lib.tools.train.ToolsTrain" ]
[((285, 297), 'lib.tools.train.ToolsTrain', 'ToolsTrain', ([], {}), '()\n', (295, 297), False, 'from lib.tools.train import ToolsTrain\n')]
from django.shortcuts import render, redirect from app01 import models from app01.utils.pagination import Pagination from app01.utils.form import UserModelForm, PrettyModelForm, PrettyEditModelForm def user_list(request): """ 用户管理 """ queryset = models.UserInfo.objects.all() page_object = Pagination(re...
[ "django.shortcuts.render", "app01.utils.form.UserModelForm", "app01.utils.pagination.Pagination", "openpyxl.load_workbook", "app01.models.UserInfo.objects.all", "app01.models.Department.objects.filter", "django.shortcuts.redirect", "app01.models.UserInfo.objects.filter", "app01.models.Department.obj...
[((258, 287), 'app01.models.UserInfo.objects.all', 'models.UserInfo.objects.all', ([], {}), '()\n', (285, 287), False, 'from app01 import models\n'), ((307, 350), 'app01.utils.pagination.Pagination', 'Pagination', (['request', 'queryset'], {'page_size': '(10)'}), '(request, queryset, page_size=10)\n', (317, 350), False...
# Copyright (C) 2020 Google Inc. # Licensed under http://www.apache.org/licenses/LICENSE-2.0 <see LICENSE file> """ Add 2 new roles for scoping objects Create Date: 2018-10-23 11:02:28.166523 """ # disable Invalid constant name pylint warning for mandatory Alembic variables. # pylint: disable=invalid-name import dat...
[ "alembic.op.get_bind", "ggrc.migrations.utils.acr_propagation.propagate_roles", "alembic.op.bulk_insert", "datetime.datetime.now", "ggrc.migrations.utils.migrator.get_migration_user_id" ]
[((1230, 1243), 'alembic.op.get_bind', 'op.get_bind', ([], {}), '()\n', (1241, 1243), False, 'from alembic import op\n'), ((1256, 1298), 'ggrc.migrations.utils.migrator.get_migration_user_id', 'migrator.get_migration_user_id', (['connection'], {}), '(connection)\n', (1286, 1298), False, 'from ggrc.migrations.utils impo...
from flask import current_app from flask import request from switchmng.typing import FlaskResponse from switchmng import database from .blueprint import restbp from .errors import * @restbp.route('/switch_models', methods = ['POST']) def post_switch_model() -> FlaskResponse: session = current_app.config['SWITCHM...
[ "switchmng.database.add_vlan", "switchmng.database.add_switch", "switchmng.database.add_port_model", "switchmng.database.add_switch_model", "switchmng.database.add_connector", "switchmng.database.add_network_protocol" ]
[((885, 926), 'switchmng.database.add_switch_model', 'database.add_switch_model', (['session'], {}), '(session, **req)\n', (910, 926), False, 'from switchmng import database\n'), ((1839, 1904), 'switchmng.database.add_port_model', 'database.add_port_model', (['session', 'switch_model_resource_id'], {}), '(session, swit...
# fix redirects import os import fileinput _SRC = "source" with open("build/linkcheck/output.txt", "r") as f: link_out = f.readlines() #%% for link in link_out: if "redirected" in link: file_name = os.path.join(_SRC, link.split(":")[0]) assert os.path.exists(file_name) old, new =...
[ "os.path.exists", "fileinput.FileInput" ]
[((275, 300), 'os.path.exists', 'os.path.exists', (['file_name'], {}), '(file_name)\n', (289, 300), False, 'import os\n'), ((378, 422), 'fileinput.FileInput', 'fileinput.FileInput', (['file_name'], {'inplace': '(True)'}), '(file_name, inplace=True)\n', (397, 422), False, 'import fileinput\n')]
import pytest from app.api.services import abr_service from app.api.business.errors import AbrError import requests import mock from mock import patch class TestAbrService(): def mocked_find_business_by_abn(self): data = '<ABR><response><stateCode>NSW</stateCode><postcode>2750</postcode>'\ '<...
[ "mock.patch", "app.api.services.abr_service.call_abr_api", "requests.exceptions.SSLError", "requests.exceptions.ProxyError", "pytest.raises", "requests.exceptions.ConnectionError", "requests.exceptions.HTTPError", "requests.exceptions.RequestException" ]
[((1310, 1365), 'mock.patch', 'mock.patch', (['"""app.api.services.abr_service.call_abr_api"""'], {}), "('app.api.services.abr_service.call_abr_api')\n", (1320, 1365), False, 'import mock\n'), ((1660, 1715), 'mock.patch', 'mock.patch', (['"""app.api.services.abr_service.call_abr_api"""'], {}), "('app.api.services.abr_s...
import numpy as np from sklearn.metrics import roc_curve, auc from sklearn.metrics import confusion_matrix from sklearn import preprocessing from sklearn.preprocessing import LabelEncoder # from IPython.display import Image,display import matplotlib.pyplot as plt data = [] labels = [] alldata = [] # XORdata=np.array([...
[ "matplotlib.pyplot.ylabel", "sklearn.metrics.auc", "numpy.array", "sklearn.metrics.roc_curve", "matplotlib.pyplot.xlabel", "matplotlib.pyplot.plot", "numpy.exp", "matplotlib.pyplot.ylim", "csv.reader", "sklearn.metrics.confusion_matrix", "sklearn.model_selection.train_test_split", "numpy.argma...
[((9996, 10007), 'time.time', 'time.time', ([], {}), '()\n', (10005, 10007), False, 'import time\n'), ((10194, 10214), 'csv.reader', 'csv.reader', (['gpsTrack'], {}), '(gpsTrack)\n', (10204, 10214), False, 'import csv\n'), ((10513, 10542), 'sklearn.preprocessing.OneHotEncoder', 'preprocessing.OneHotEncoder', ([], {}), ...
#!/usr/bin/env python3 """Split PDFS by QR code and move images and PDFs to correct folder.""" import os import traceback import numpy from . import write_to_log as logger from . import submitty_ocr as scanner # try importing required modules try: from PyPDF2 import PdfFileReader, PdfFileWriter from pdf2imag...
[ "traceback.format_exc", "cv2.threshold", "os.path.join", "os.chdir", "numpy.array", "pyzbar.pyzbar.decode", "cv2.cvtColor", "os.getpid", "PyPDF2.PdfFileWriter", "traceback.print_exc", "PyPDF2.PdfFileReader" ]
[((463, 484), 'traceback.print_exc', 'traceback.print_exc', ([], {}), '()\n', (482, 484), False, 'import traceback\n'), ((854, 874), 'os.chdir', 'os.chdir', (['split_path'], {}), '(split_path)\n', (862, 874), False, 'import os\n'), ((894, 917), 'PyPDF2.PdfFileReader', 'PdfFileReader', (['filename'], {}), '(filename)\n'...
import os import cgi import re from pyramid.response import Response from pyramid.view import view_config import pyramid.httpexceptions as httpexceptions # ------------------------------------------------------------------- class ReflectorView(object): def __init__(self, request): self.request = request...
[ "os.path.dirname", "os.path.basename", "pyramid.response.Response", "cgi.escape", "pyramid.view.view_config" ]
[((327, 362), 'pyramid.view.view_config', 'view_config', ([], {'route_name': '"""reflector"""'}), "(route_name='reflector')\n", (338, 362), False, 'from pyramid.view import view_config\n'), ((925, 935), 'pyramid.response.Response', 'Response', ([], {}), '()\n', (933, 935), False, 'from pyramid.response import Response\...
# -*- coding:utf-8 -*- ''' @Author: GETF @Email: <EMAIL> @DateTime: 2017-11-09 15:12:14 @Description: Description ''' import os from PyQt5.QtGui import QIcon import sys from PyQt5.QtWidgets import (QWidget, QLCDNumber, QSlider, QVBoxLayout, QApplication) from PyQt5.QtCore...
[ "PyQt5.QtGui.QIcon", "PyQt5.QtCore.QDateTime.currentDateTime", "PyQt5.QtCore.QTimer", "os.path.dirname", "PyQt5.QtWidgets.QApplication", "PyQt5.QtWidgets.QLCDNumber" ]
[((1469, 1491), 'PyQt5.QtWidgets.QApplication', 'QApplication', (['sys.argv'], {}), '(sys.argv)\n', (1481, 1491), False, 'from PyQt5.QtWidgets import QWidget, QLCDNumber, QSlider, QVBoxLayout, QApplication\n'), ((487, 503), 'PyQt5.QtWidgets.QLCDNumber', 'QLCDNumber', (['self'], {}), '(self)\n', (497, 503), False, 'from...
"""Module to hold the Rt class definition.""" from typing import Optional from aiohttp import ClientSession from bs4 import BeautifulSoup from bs4.element import Tag from phylm.utils.web import async_soupify from phylm.utils.web import url_encode RT_BASE_MOVIE_URL = "https://www.rottentomatoes.com/search" class Rt...
[ "phylm.utils.web.url_encode", "phylm.utils.web.async_soupify" ]
[((1646, 1672), 'phylm.utils.web.url_encode', 'url_encode', (['self.raw_title'], {}), '(self.raw_title)\n', (1656, 1672), False, 'from phylm.utils.web import url_encode\n'), ((1764, 1798), 'phylm.utils.web.async_soupify', 'async_soupify', (['search_url', 'session'], {}), '(search_url, session)\n', (1777, 1798), False, ...
#!/usr/bin/env python # fake-build-for-bitcode.py - Fake build with -embed-bitcode -*- python -*- # # This source file is part of the Swift.org open source project # # Copyright (c) 2014 - 2016 Apple Inc. and the Swift project authors # Licensed under Apache License v2.0 with Runtime Library Exception # # See http://sw...
[ "sys.argv.index", "os.path.basename", "os.utime" ]
[((1052, 1078), 'os.utime', 'os.utime', (['outputFile', 'None'], {}), '(outputFile, None)\n', (1060, 1078), False, 'import os\n'), ((832, 863), 'sys.argv.index', 'sys.argv.index', (['"""-primary-file"""'], {}), "('-primary-file')\n", (846, 863), False, 'import sys\n'), ((891, 911), 'sys.argv.index', 'sys.argv.index', (...
#!/usr/bin/python # Copyright (c) 2018, Oracle and/or its affiliates. # This software is made available to you under the terms of the GPL 3.0 license or the Apache 2.0 license. # GNU General Public License v3.0+ (see COPYING or https://www.gnu.org/licenses/gpl-3.0.txt) # Apache License v2.0 # See LICENSE.TXT for detail...
[ "ansible.module_utils.basic.AnsibleModule", "ansible.module_utils.oracle.oci_lb_utils.delete_lb_resources_and_wait", "ansible.module_utils.oracle.oci_utils.create_service_client", "oci.load_balancer.models.UpdateBackendDetails", "oci.load_balancer.models.CreateBackendDetails", "oci.util.to_dict", "ansib...
[((8937, 8974), 'ansible.module_utils.oracle.oci_lb_utils.get_backend_name', 'oci_lb_utils.get_backend_name', (['module'], {}), '(module)\n', (8966, 8974), False, 'from ansible.module_utils.oracle import oci_utils, oci_lb_utils\n'), ((9004, 9026), 'oci.load_balancer.models.CreateBackendDetails', 'CreateBackendDetails',...
#coding=utf-8 # Copyright (c) 2020 PaddlePaddle Authors. All Rights Reserve. # #Licensed under the Apache License, Version 2.0 (the "License"); #you may not use this file except in compliance with the License. #You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # #Unless required...
[ "os.path.exists", "paddle.fluid.dygraph.learning_rate_scheduler.ReduceLROnPlateau", "os.listdir", "paddle.fluid.dygraph.base.to_variable", "paddle.fluid.layers.cross_entropy", "paddle.fluid.layers.mean", "opts.parse_opts", "numpy.array", "numpy.zeros", "paddle.fluid.CUDAPlace", "models.model.gen...
[((2083, 2095), 'opts.parse_opts', 'parse_opts', ([], {}), '()\n', (2093, 2095), False, 'from opts import parse_opts\n'), ((1326, 1362), 'os.path.exists', 'os.path.exists', (['opt.Flow_resume_path'], {}), '(opt.Flow_resume_path)\n', (1340, 1362), False, 'import os\n'), ((1432, 1464), 'os.listdir', 'os.listdir', (['opt....
from math import gamma from typing import Dict, List, Tuple import matplotlib.pyplot as plt import numpy as np import torch import torch.nn.functional as F import torch.optim as optim from atcenv.MASAC.buffer import ReplayBuffer from atcenv.MASAC.mactor_critic import Actor, CriticQ, CriticV from torch.nn.utils.clip_gr...
[ "numpy.clip", "torch.optim.Adam", "torch.cuda.get_device_name", "atcenv.MASAC.mactor_critic.Actor", "numpy.prod", "atcenv.MASAC.buffer.ReplayBuffer", "torch.FloatTensor", "numpy.array", "atcenv.MASAC.mactor_critic.CriticV", "torch.cuda.is_available", "atcenv.MASAC.mactor_critic.CriticQ", "torc...
[((700, 760), 'atcenv.MASAC.buffer.ReplayBuffer', 'ReplayBuffer', (['STATE_DIM', 'ACTION_DIM', 'BUFFER_SIZE', 'BATCH_SIZE'], {}), '(STATE_DIM, ACTION_DIM, BUFFER_SIZE, BATCH_SIZE)\n', (712, 760), False, 'from atcenv.MASAC.buffer import ReplayBuffer\n'), ((1207, 1261), 'torch.zeros', 'torch.zeros', (['(1)'], {'requires_...
#!/usr/bin/env python import signal import time import sys from rc522 import RFID # Sector 0 == Reserved # Sector 1 == # Sector 2 == Card Type [master,super,user][16 char], UUID [32 char] # Sector 3 == First Name [Max 48 char] # Sector 4 == Last Name [Max 48 char] # Sector 5 == employee id [Max 16 cha...
[ "signal.signal", "time.sleep", "rc522.RFID", "sys.exit" ]
[((680, 686), 'rc522.RFID', 'RFID', ([], {}), '()\n', (684, 686), False, 'from rc522 import RFID\n'), ((1446, 1484), 'signal.signal', 'signal.signal', (['signal.SIGINT', 'end_read'], {}), '(signal.SIGINT, end_read)\n', (1459, 1484), False, 'import signal\n'), ((1434, 1444), 'sys.exit', 'sys.exit', ([], {}), '()\n', (14...
# # Code borrowed and modified from https://www.esri.com/arcgis-blog/products/arcgis-pro/health/use-proximity-tracing-to-identify-possible-contact-events/ # import os import subprocess import sys import winreg from typing import Dict import arcpy import glob pro_home = arcpy.GetInstallInfo()["InstallDir"] pro_lib_dir...
[ "os.path.exists", "pyspark.sql.SparkSession.builder.getOrCreate", "sys.path.insert", "winreg.QueryValueEx", "os.environ.unsetenv", "os.getenv", "winreg.OpenKey", "winreg.ConnectRegistry", "os.path.join", "pyspark.SparkConf", "arcpy.GetInstallInfo", "pyspark.sql.SparkSession", "pyspark.SparkC...
[((323, 360), 'os.path.join', 'os.path.join', (['pro_home', '"""Java"""', '"""lib"""'], {}), "(pro_home, 'Java', 'lib')\n", (335, 360), False, 'import os\n'), ((379, 420), 'os.path.join', 'os.path.join', (['pro_home', '"""Java"""', '"""runtime"""'], {}), "(pro_home, 'Java', 'runtime')\n", (391, 420), False, 'import os\...
"Script for everything User related in the database" from models import db class User(db.Model): "Class used for configuring the User model in the database" id = db.Column(db.Integer, primary_key=True) username = db.Column(db.String(80), unique=True, nullable=False) admin = db.Column(db.Boolean) b...
[ "models.db.relationship", "models.db.String", "models.db.relation", "models.db.Column" ]
[((172, 211), 'models.db.Column', 'db.Column', (['db.Integer'], {'primary_key': '(True)'}), '(db.Integer, primary_key=True)\n', (181, 211), False, 'from models import db\n'), ((293, 314), 'models.db.Column', 'db.Column', (['db.Boolean'], {}), '(db.Boolean)\n', (302, 314), False, 'from models import db\n'), ((326, 347),...
# -*- coding: utf-8 -*- from github import Github UPSTREAM_GITHUB_REPO = 'wikimedia/pywikibot-core' GITHUB_REPO = 'magul/pywikibot-core' _pull_requests = [] def push_branch(local_branch): name = local_branch.name repo = local_branch.repo github = repo.remotes['github'] return github.push(name)[...
[ "github.Github" ]
[((515, 559), 'github.Github', 'Github', (['creds.GITHUB_USER', 'creds.GITHUB_PASS'], {}), '(creds.GITHUB_USER, creds.GITHUB_PASS)\n', (521, 559), False, 'from github import Github\n'), ((1042, 1086), 'github.Github', 'Github', (['creds.GITHUB_USER', 'creds.GITHUB_PASS'], {}), '(creds.GITHUB_USER, creds.GITHUB_PASS)\n'...
# # Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # ...
[ "migrate.changeset.constraint.ForeignKeyConstraint", "sqlalchemy.MetaData", "sqlalchemy.Column", "sqlalchemy.Table" ]
[((676, 716), 'sqlalchemy.MetaData', 'sqlalchemy.MetaData', ([], {'bind': 'migrate_engine'}), '(bind=migrate_engine)\n', (695, 716), False, 'import sqlalchemy\n'), ((733, 782), 'sqlalchemy.Table', 'sqlalchemy.Table', (['"""resource"""', 'meta'], {'autoload': '(True)'}), "('resource', meta, autoload=True)\n", (749, 782)...
import pytest import unittest from pydu.dict import AttrDict, LookupDict, CaseInsensitiveDict, OrderedDefaultDict, attrify class TestAttrDict: def test_attr_access_with_init(self): d = AttrDict(key=1) assert d['key'] == 1 assert d.key == 1 def test_attr_access_without_init(self): ...
[ "pydu.dict.LookupDict", "pydu.dict.attrify", "pytest.raises", "pydu.dict.CaseInsensitiveDict", "copy.deepcopy", "pydu.dict.OrderedDefaultDict", "pydu.dict.AttrDict" ]
[((2524, 2568), 'pydu.dict.attrify', 'attrify', (["{'a': [1, 2, {'b': 'b'}], 'c': 'c'}"], {}), "({'a': [1, 2, {'b': 'b'}], 'c': 'c'})\n", (2531, 2568), False, 'from pydu.dict import AttrDict, LookupDict, CaseInsensitiveDict, OrderedDefaultDict, attrify\n'), ((2703, 2718), 'pydu.dict.attrify', 'attrify', (['(1, 2)'], {}...
# Import needed libraries from stix2 import TAXIICollectionSource, Filter from taxii2client.v20 import Server, Collection import json import re from openpyxl import Workbook from openpyxl.styles import Font from openpyxl.styles import Alignment #_________________________________________________________________________...
[ "openpyxl.styles.Font", "taxii2client.v20.Collection", "stix2.Filter", "stix2.TAXIICollectionSource", "openpyxl.Workbook", "openpyxl.styles.Alignment" ]
[((410, 516), 'taxii2client.v20.Collection', 'Collection', (['"""https://cti-taxii.mitre.org/stix/collections/95ecc380-afe9-11e4-9b6c-751b66dd541e/"""'], {}), "(\n 'https://cti-taxii.mitre.org/stix/collections/95ecc380-afe9-11e4-9b6c-751b66dd541e/'\n )\n", (420, 516), False, 'from taxii2client.v20 import Server, ...
from typing import Any, AsyncGenerator, Optional from httpx import AsyncClient, Response from aiochclient.exceptions import ChClientError from aiochclient.http_clients.abc import HttpClientABC class HttpxHttpClient(HttpClientABC): def __init__(self, session: Optional[AsyncClient]): if session: ...
[ "httpx.AsyncClient" ]
[((389, 402), 'httpx.AsyncClient', 'AsyncClient', ([], {}), '()\n', (400, 402), False, 'from httpx import AsyncClient, Response\n')]
def options(ctx): import optparse grp = optparse.OptionGroup(ctx.parser, "architecture options") grp.add_option("--m32", action="store_true", default=False, help="compile & link in 32bits") grp.add_option("--m64", action="store_true", default=False, help="compile &...
[ "optparse.OptionGroup", "sys.platform.lower", "waflib.Errors.WafError" ]
[((49, 105), 'optparse.OptionGroup', 'optparse.OptionGroup', (['ctx.parser', '"""architecture options"""'], {}), "(ctx.parser, 'architecture options')\n", (69, 105), False, 'import optparse\n'), ((510, 564), 'waflib.Errors.WafError', 'Errors.WafError', (['"""You must choose either m32 of m64 !"""'], {}), "('You must ch...
#!/usr/bin/env python """Download the YouTube video listed in a .webloc file. The .webloc file should be specified as sys.argv[1]. The video will be downloaded to ../videos/by-youtube-id/<youtube_id>.<ext> and will be symlinked as ../videos/by-channel/<channel>/<title>.<ext>. """ import errno import os import plist...
[ "subprocess.check_output", "os.makedirs", "subprocess.check_call", "plistlib.readPlist", "os.chdir", "sys.stderr.write", "os.path.isdir", "os.unlink", "sys.exit", "sys.stdout.write" ]
[((446, 481), 'sys.stdout.write', 'sys.stdout.write', (["('Log: %s\\n' % msg)"], {}), "('Log: %s\\n' % msg)\n", (462, 481), False, 'import sys\n'), ((508, 545), 'sys.stderr.write', 'sys.stderr.write', (["('Error: %s\\n' % msg)"], {}), "('Error: %s\\n' % msg)\n", (524, 545), False, 'import sys\n'), ((2479, 2509), 'subpr...
# Copyright 2009-2017 <NAME>. # This program is distributed under the MIT license. ''' Defines the `cache` decorator. See its documentation for more details. ''' # todo: examine thread-safety import datetime as datetime_module from python_toolbox import misc_tools from python_toolbox import binary_search from pytho...
[ "python_toolbox.misc_tools.set_attributes", "python_toolbox.third_party.decorator.decorator", "python_toolbox.sleek_reffing.SleekCallArgs", "datetime.datetime.now", "python_toolbox.nifty_collections.OrderedDict", "datetime.timedelta" ]
[((755, 785), 'datetime.datetime.now', 'datetime_module.datetime.now', ([], {}), '()\n', (783, 785), True, 'import datetime as datetime_module\n'), ((6309, 6337), 'python_toolbox.third_party.decorator.decorator', 'decorator_', (['cached', 'function'], {}), '(cached, function)\n', (6319, 6337), True, 'from python_toolbo...
# coding:utf-8 from flask import Flask, render_template # 倒入蓝图 from order import app_orders # 1.循环引用-解决方案1-推迟一方的加载 # from user import get_user from goods import get_goods app = Flask(__name__) # 2.循环引用-解决方案2-使用装饰器 app.route('/get_goods')(get_goods) # 3.在app程序中注册蓝图 app.register_blueprint(app_orders, url_prefix='/orde...
[ "flask.Flask" ]
[((178, 193), 'flask.Flask', 'Flask', (['__name__'], {}), '(__name__)\n', (183, 193), False, 'from flask import Flask, render_template\n')]
"""empty message Revision ID: 4603f5ddc9cb Revises: Create Date: 2019-01-15 12:24:13.352143 """ from alembic import op import sqlalchemy as sa # revision identifiers, used by Alembic. revision = '<KEY>' down_revision = None branch_labels = None depends_on = None def upgrade(): # ### commands auto generated b...
[ "sqlalchemy.DateTime", "alembic.op.drop_table", "sqlalchemy.Text", "sqlalchemy.PrimaryKeyConstraint", "sqlalchemy.Integer" ]
[((1048, 1070), 'alembic.op.drop_table', 'op.drop_table', (['"""tasks"""'], {}), "('tasks')\n", (1061, 1070), False, 'from alembic import op\n'), ((888, 917), 'sqlalchemy.PrimaryKeyConstraint', 'sa.PrimaryKeyConstraint', (['"""id"""'], {}), "('id')\n", (911, 917), True, 'import sqlalchemy as sa\n'), ((400, 412), 'sqlal...
from mock import * import unittest from hamcrest import * from service.smjvmservice import SmJvmService from smcontext import ServiceManagerException from smprocess import SmProcess class JvmServiceMock(SmJvmService): def __init__(self, context, service_name): SmJvmService.__init__(self, context, servic...
[ "smcontext.ServiceManagerException", "service.smjvmservice.SmJvmService.__init__" ]
[((277, 331), 'service.smjvmservice.SmJvmService.__init__', 'SmJvmService.__init__', (['self', 'context', 'service_name', '""""""'], {}), "(self, context, service_name, '')\n", (298, 331), False, 'from service.smjvmservice import SmJvmService\n'), ((686, 718), 'smcontext.ServiceManagerException', 'ServiceManagerExcepti...
# Licensed to the .NET Foundation under one or more agreements. # The .NET Foundation licenses this file to you under the MIT license. # See the LICENSE file in the project root for more information. """A script to evaluate test values for special functions in high precision. This scripts looks for .csv files in /test...
[ "csv.DictWriter", "csv.DictReader", "os.scandir", "os.path.realpath", "time.time" ]
[((13503, 13518), 'os.scandir', 'os.scandir', (['dir'], {}), '(dir)\n', (13513, 13518), False, 'import os\n'), ((13400, 13426), 'os.path.realpath', 'os.path.realpath', (['__file__'], {}), '(__file__)\n', (13416, 13426), False, 'import os\n'), ((13936, 13974), 'csv.DictReader', 'csv.DictReader', (['csvfile'], {'delimite...
from django.db import models from django.contrib.auth.models import BaseUserManager, AbstractBaseUser, PermissionsMixin class UsuarioManager(BaseUserManager): def create_user(self, email, password=None): usuario = self.model(email=self.normalize_email(email)) usuario.is_active = True usua...
[ "django.db.models.EmailField", "django.db.models.BooleanField" ]
[((914, 993), 'django.db.models.EmailField', 'models.EmailField', ([], {'verbose_name': '"""Email do usuário"""', 'max_length': '(194)', 'unique': '(True)'}), "(verbose_name='Email do usuário', max_length=194, unique=True)\n", (931, 993), False, 'from django.db import models\n'), ((1010, 1078), 'django.db.models.Boolea...
from matplotlib import pyplot as plt import numpy as np def generate_and_save_images(model, epoch, test_input): # Notice `training` is set to False. # This is so all layers run in inference mode (batchnorm). predictions = model(test_input, training=False) fig = plt.figure(figsize=(10,10)) for i in ra...
[ "matplotlib.pyplot.imshow", "matplotlib.pyplot.axis", "matplotlib.pyplot.close", "numpy.max", "matplotlib.pyplot.figure", "matplotlib.pyplot.title", "matplotlib.pyplot.ylim", "matplotlib.pyplot.subplot", "matplotlib.pyplot.legend" ]
[((276, 304), 'matplotlib.pyplot.figure', 'plt.figure', ([], {'figsize': '(10, 10)'}), '(figsize=(10, 10))\n', (286, 304), True, 'from matplotlib import pyplot as plt\n'), ((547, 558), 'matplotlib.pyplot.close', 'plt.close', ([], {}), '()\n', (556, 558), True, 'from matplotlib import pyplot as plt\n'), ((594, 621), 'ma...
import numpy as np import scipy.io as scio import scipy.sparse as scsp import h5py as hp from util import read_mymat73, read_mymat, build_img_dataset, process_ad_dataset, mv_dataset, mv_tabular_collate, AverageMeter, save_roc_pr_curve_data, get_all_labels, \ load_print_results, filter_nan_grad, read_dataset, build_...
[ "models.encoder_decoder.mvae_ad", "models.encoder_decoder.mvenc", "torch.from_numpy", "models.encoder_decoder.mvae_tf", "torch.nn.MSELoss", "models.encoder_decoder.mvae_ss", "torch.cuda.is_available", "torch.sum", "util.process_ad_dataset", "util.filter_nan_grad", "models.DeepCCAModels.cca", "...
[((1981, 2030), 'torch.set_default_tensor_type', 'torch.set_default_tensor_type', (['torch.DoubleTensor'], {}), '(torch.DoubleTensor)\n', (2010, 2030), False, 'import torch\n'), ((2818, 2827), 'torch.nn.MSELoss', 'MSELoss', ([], {}), '()\n', (2825, 2827), False, 'from torch.nn import MSELoss, CrossEntropyLoss\n'), ((81...
from pathlib import Path from lib_bgp_simulator import BaseGraphSystemTester from lib_bgp_simulator import BGPSimpleAS from lib_bgp_simulator import ROVSimpleAS from lib_bgp_simulator import Graph013 from ..unstable import Unstable from ....as_classes import ROVPPV1SimpleAS from ....as_classes import ROVPPV2SimpleAS ...
[ "pathlib.Path" ]
[((616, 630), 'pathlib.Path', 'Path', (['__file__'], {}), '(__file__)\n', (620, 630), False, 'from pathlib import Path\n')]
# coding: utf-8 ######################################################################### # Name: # # Calcurate equivalent potential temperature. # # Usage: # example: # # Author: <NAME> # Date: 2021/08/13 ######################################################################### import argparse #from datetime import da...
[ "os.listdir", "argparse.ArgumentParser", "netCDF4.Dataset", "os.path.join", "math.log", "numpy.array", "os.path.abspath", "re.search" ]
[((498, 523), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {}), '()\n', (521, 523), False, 'import argparse\n'), ((1132, 1147), 'netCDF4.Dataset', 'Dataset', (['ncfile'], {}), '(ncfile)\n', (1139, 1147), False, 'from netCDF4 import Dataset\n'), ((1419, 1462), 'numpy.array', 'np.array', (['[i for i in self...
""" Servo Horns """ import cadquery as cq import cqparts from cqparts.params import * from cqparts.display import render_props, display from cqparts.constraint import Fixed, Coincident from cqparts.constraint import Mate from cqparts.utils.geometry import CoordSystem from cqparts.search import register from .multi im...
[ "cadquery.Workplane", "cqparts.search.register", "cqparts.display.display", "cqparts.utils.geometry.CoordSystem" ]
[((2960, 2984), 'cqparts.search.register', 'register', ([], {'export': '"""horns"""'}), "(export='horns')\n", (2968, 2984), False, 'from cqparts.search import register\n'), ((3403, 3427), 'cqparts.search.register', 'register', ([], {'export': '"""horns"""'}), "(export='horns')\n", (3411, 3427), False, 'from cqparts.sea...
#!/usr/bin/env python3 # -*- coding: utf-8 -*- # # This file is part of daily-wallpaper # # Copyright (c) 2017 <NAME> <a.k.a. atareao> # # Permission is hereby granted, free of charge, to any person obtaining a copy # of this software and associated documentation files (the "Software"), to deal # in the Software withou...
[ "sys.path.insert", "comun._", "lxml.html.fromstring", "requests.get", "os.getcwd", "os.path.abspath", "daily.Daily.__init__" ]
[((1361, 1409), 'sys.path.insert', 'sys.path.insert', (['(1)', '"""/usr/share/daily-wallpaper"""'], {}), "(1, '/usr/share/daily-wallpaper')\n", (1376, 1409), False, 'import sys\n'), ((1724, 1741), 'comun._', '_', (['"""Wallpapering"""'], {}), "('Wallpapering')\n", (1725, 1741), False, 'from comun import _\n'), ((1775, ...
# -*- coding: utf-8 -*- """ sphinxjp.themes.revealjs.directives ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ :author: tell-k <<EMAIL>> :copyright: tell-k. All Rights Reserved. """ from docutils import nodes from docutils.parsers.rst import directives from docutils.parsers.rst.roles import se...
[ "docutils.parsers.rst.directives.choice", "docutils.parsers.rst.roles.set_classes", "sphinxjp.themes.revealjs.compat.text" ]
[((790, 855), 'docutils.parsers.rst.directives.choice', 'directives.choice', (['argument', "('h1', 'h2', 'h3', 'h4', 'h5', 'h6')"], {}), "(argument, ('h1', 'h2', 'h3', 'h4', 'h5', 'h6'))\n", (807, 855), False, 'from docutils.parsers.rst import directives\n'), ((1939, 1964), 'docutils.parsers.rst.roles.set_classes', 'se...
import unittest import pytest import paramak class TestExtrudeHollowRectangle(unittest.TestCase): def setUp(self): self.test_shape = paramak.ExtrudeHollowRectangle( height=10, width=15, casing_thickness=1, distance=2 ) def test_default_parameters(self): """Checks that th...
[ "pytest.approx", "paramak.ExtrudeHollowRectangle" ]
[((149, 236), 'paramak.ExtrudeHollowRectangle', 'paramak.ExtrudeHollowRectangle', ([], {'height': '(10)', 'width': '(15)', 'casing_thickness': '(1)', 'distance': '(2)'}), '(height=10, width=15, casing_thickness=1,\n distance=2)\n', (179, 236), False, 'import paramak\n'), ((2146, 2186), 'pytest.approx', 'pytest.appro...
import djclick as click from azure.common import AzureMissingResourceHttpError from requests.exceptions import HTTPError from boundlexx.api.tasks import purge_static_cache from boundlexx.boundless.models import Emoji from boundlexx.utils import download_image, get_django_image_from_file, make_thumbnail @click.comman...
[ "djclick.command", "boundlexx.api.tasks.purge_static_cache", "boundlexx.utils.get_django_image_from_file", "boundlexx.utils.make_thumbnail", "djclick.echo", "boundlexx.boundless.models.Emoji.objects.filter", "boundlexx.utils.download_image" ]
[((308, 323), 'djclick.command', 'click.command', ([], {}), '()\n', (321, 323), True, 'import djclick as click\n'), ((343, 389), 'djclick.echo', 'click.echo', (['"""Adding thumbs/renmaing images..."""'], {}), "('Adding thumbs/renmaing images...')\n", (353, 389), True, 'import djclick as click\n'), ((423, 464), 'boundle...
from tortoise import Model from tortoise import fields from werkzeug.security import generate_password_hash, check_password_hash class User(Model): class Meta: table = 'users' id = fields.IntField(pk=True) username = fields.CharField(max_length=100, unique=True) email = fields.CharField(max_l...
[ "tortoise.fields.CharField", "werkzeug.security.generate_password_hash", "tortoise.fields.TextField", "tortoise.fields.IntField" ]
[((200, 224), 'tortoise.fields.IntField', 'fields.IntField', ([], {'pk': '(True)'}), '(pk=True)\n', (215, 224), False, 'from tortoise import fields\n'), ((240, 285), 'tortoise.fields.CharField', 'fields.CharField', ([], {'max_length': '(100)', 'unique': '(True)'}), '(max_length=100, unique=True)\n', (256, 285), False, ...
# coding=utf-8 # Copyright 2014 Pants project contributors (see CONTRIBUTORS.md). # Licensed under the Apache License, Version 2.0 (see LICENSE). from __future__ import (absolute_import, division, generators, nested_scopes, print_function, unicode_literals, with_statement) import itertools imp...
[ "re.compile", "os.pathsep.join", "pants.backend.python.python_requirement.PythonRequirement", "pants.base.exceptions.TestFailedTaskError", "pants.util.contextutil.temporary_dir", "textwrap.dedent", "os.path.exists", "shutil.move", "six.StringIO", "os.path.isabs", "pex.pex_info.PexInfo.default", ...
[((1404, 1425), 'logging.basicConfig', 'logging.basicConfig', ([], {}), '()\n', (1423, 1425), False, 'import logging\n'), ((9040, 9209), 'textwrap.dedent', 'dedent', (["b'\\n [run]\\n branch = True\\n timid = True\\n\\n [report]\\n exclude_lines =\\n def __repr__\\n raise NotImplementedErro...
from __future__ import print_function from __future__ import absolute_import from __future__ import division from numpy import asarray from scipy.spatial import Voronoi from scipy.spatial import Delaunay __all__ = [ 'delaunay_from_points_numpy', 'voronoi_from_points_numpy', ] def delaunay_from_points_numpy...
[ "numpy.asarray", "scipy.spatial.Voronoi", "scipy.spatial.Delaunay" ]
[((955, 970), 'numpy.asarray', 'asarray', (['points'], {}), '(points)\n', (962, 970), False, 'from numpy import asarray\n'), ((979, 1000), 'scipy.spatial.Delaunay', 'Delaunay', (['xyz[:, 0:2]'], {}), '(xyz[:, 0:2])\n', (987, 1000), False, 'from scipy.spatial import Delaunay\n'), ((1315, 1330), 'numpy.asarray', 'asarray...
# $Id: gp_unix.py,v 2.6 2003/04/21 09:44:09 mhagger Exp $ # Copyright (C) 1998-2003 <NAME> <<EMAIL>> # # This file is licensed under the GNU Lesser General Public License # (LGPL). See LICENSE.txt for details. """gp_unix -- an interface to gnuplot used for unix platforms. This file implements a low-level interface ...
[ "os.popen", "string.find" ]
[((5706, 5773), 'os.popen', 'popen', (["('echo | %s -persist 2>&1' % GnuplotOpts.gnuplot_command)", '"""r"""'], {}), "('echo | %s -persist 2>&1' % GnuplotOpts.gnuplot_command, 'r')\n", (5711, 5773), False, 'from os import popen\n'), ((7686, 7741), 'os.popen', 'popen', (["('%s -persist' % GnuplotOpts.gnuplot_command)", ...
#!/usr/bin/python """ This script will scrape the r-project.org machine learning selection and format the packages in github markdown style for this awesome-machine-learning repo. """ from pyquery import PyQuery as pq import urllib import codecs text_file = codecs.open("Packages.txt", encoding='utf-8', ...
[ "codecs.open", "urllib.urlopen" ]
[((274, 329), 'codecs.open', 'codecs.open', (['"""Packages.txt"""'], {'encoding': '"""utf-8"""', 'mode': '"""w"""'}), "('Packages.txt', encoding='utf-8', mode='w')\n", (285, 329), False, 'import codecs\n'), ((433, 452), 'urllib.urlopen', 'urllib.urlopen', (['url'], {}), '(url)\n', (447, 452), False, 'import urllib\n'),...
import torch from videoanalyst.config.config import cfg as root_cfg from videoanalyst.config.config import specify_task from videoanalyst.model import builder as model_builder from videoanalyst.pipeline import builder as pipeline_builder from imutils.video import VideoStream from imutils.video import FPS import argpa...
[ "cv2.rectangle", "cv2.__version__.split", "xml.etree.ElementTree.parse", "videoanalyst.config.config.cfg.merge_from_file", "argparse.ArgumentParser", "torch.device", "imutils.video.VideoStream", "time.sleep", "cv2.imshow", "videoanalyst.config.config.specify_task", "imutils.resize", "cv2.putTe...
[((447, 519), 'videoanalyst.config.config.cfg.merge_from_file', 'root_cfg.merge_from_file', (['"""./experiments/siamfcpp/siamfcpp_alexnet.yaml"""'], {}), "('./experiments/siamfcpp/siamfcpp_alexnet.yaml')\n", (471, 519), True, 'from videoanalyst.config.config import cfg as root_cfg\n'), ((555, 577), 'videoanalyst.config...
import sys from ..api import plot import fuc import pysam description = f""" Plot allele fraction profile from VcfFrame[Imported]. """ def create_parser(subparsers): parser = fuc.api.common._add_parser( subparsers, fuc.api.common._script_name(), description=description, help='Plo...
[ "fuc.api.common._script_name" ]
[((239, 268), 'fuc.api.common._script_name', 'fuc.api.common._script_name', ([], {}), '()\n', (266, 268), False, 'import fuc\n')]
""" File: test.py By: <NAME>, <EMAIL> Description: Test the net. """ from argmaxnet import ArgMaxNet import os #================================================================= modelPath = os.path.expanduser('~/models/argmaxnet/ArgMaxNet : StackDepth=3, Block(10,60),[111][0.02220].model') #=======================...
[ "argmaxnet.ArgMaxNet", "os.path.expanduser" ]
[((195, 305), 'os.path.expanduser', 'os.path.expanduser', (['"""~/models/argmaxnet/ArgMaxNet : StackDepth=3, Block(10,60),[111][0.02220].model"""'], {}), "(\n '~/models/argmaxnet/ArgMaxNet : StackDepth=3, Block(10,60),[111][0.02220].model'\n )\n", (213, 305), False, 'import os\n'), ((403, 423), 'argmaxnet.ArgMaxN...
from ploomber.scaffold.scaffoldloader import ScaffoldLoader from ploomber.util.util import add_to_sys_path from ploomber.util import loader from ploomber.exceptions import DAGSpecInvalidError def load_dag(): # setting lazy_import to true causes sources to be returned as paths, # instead of placeholders tr...
[ "ploomber.util.loader.create", "ploomber.util.loader._default_spec_load", "ploomber.util.util.add_to_sys_path", "ploomber.scaffold.scaffoldloader.ScaffoldLoader" ]
[((556, 586), 'ploomber.scaffold.scaffoldloader.ScaffoldLoader', 'ScaffoldLoader', (['"""ploomber_add"""'], {}), "('ploomber_add')\n", (570, 586), False, 'from ploomber.scaffold.scaffoldloader import ScaffoldLoader\n'), ((338, 383), 'ploomber.util.loader._default_spec_load', 'loader._default_spec_load', ([], {'lazy_imp...
# -*- encoding: utf-8 -*- from datetime import datetime from hashlib import md5 from logging import getLogger from redis import ConnectionPool, Redis from smsgateway.enums import DIRECTION_OUTBOUND from smsgateway.models import SMS from smsgateway.backends.base import SMSBackend from smsgateway.sms import SMSRequest ...
[ "logging.getLogger", "smsgateway.models.SMS.objects.create", "redis.ConnectionPool", "smsgateway.sms.SMSRequest", "redis.Redis", "datetime.datetime.now" ]
[((330, 349), 'logging.getLogger', 'getLogger', (['__name__'], {}), '(__name__)\n', (339, 349), False, 'from logging import getLogger\n'), ((1363, 1465), 'redis.ConnectionPool', 'ConnectionPool', ([], {'host': 'host', 'port': 'port', 'db': "account_dict['dbn']", 'password': "account_dict['<PASSWORD>']"}), "(host=host, ...
import sys import pprint pp = pprint.PrettyPrinter(); import pymongo from pymongo import MongoClient #!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Wed Oct 12 20:30:54 2016 @author: ryanlim, jpitts Requirements: - pymongo needs to be installed - mongodb needs to be running - brigade-match...
[ "pymongo.MongoClient", "pprint.PrettyPrinter", "operator.itemgetter" ]
[((30, 52), 'pprint.PrettyPrinter', 'pprint.PrettyPrinter', ([], {}), '()\n', (50, 52), False, 'import pprint\n'), ((1331, 1362), 'pymongo.MongoClient', 'MongoClient', (['"""localhost"""', '(27017)'], {}), "('localhost', 27017)\n", (1342, 1362), False, 'from pymongo import MongoClient\n'), ((10221, 10245), 'operator.it...
"""This module abstracts logging configuration""" # --------------------------------------------- # System modules # --------------------------------------------- import logging # --------------------------------------------- # External dependencies # --------------------------------------------- # -------------------...
[ "logging.basicConfig", "logging.getLogger", "coloredlogs.install" ]
[((486, 525), 'logging.basicConfig', 'logging.basicConfig', ([], {'level': 'logging.INFO'}), '(level=logging.INFO)\n', (505, 525), False, 'import logging\n'), ((593, 620), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (610, 620), False, 'import logging\n'), ((621, 669), 'coloredlogs.inst...
from swarmform import SwarmPad from fireworks import Firework, ScriptTask, FWorker from fireworks.core.rocket_launcher import launch_rocket if __name__ == "__main__": # set up the SwarmPad and reset it swarmpad = SwarmPad() swarmpad.reset('', require_password=False) # create the Firework consisting of a custom "A...
[ "fireworks.FWorker", "swarmform.SwarmPad", "fireworks.ScriptTask.from_str" ]
[((216, 226), 'swarmform.SwarmPad', 'SwarmPad', ([], {}), '()\n', (224, 226), False, 'from swarmform import SwarmPad\n'), ((355, 390), 'fireworks.ScriptTask.from_str', 'ScriptTask.from_str', (['"""echo "hello\\""""'], {}), '(\'echo "hello"\')\n', (374, 390), False, 'from fireworks import Firework, ScriptTask, FWorker\n...
"""Script for training the NormalizedEightPointNet. Example: $ python train.py to see help: $ python train.py -h """ import argparse import time import torch import torch.optim as optim from dfe.datasets import ColmapDataset from dfe.models import NormalizedEightPointNet import dfe.models.loss as L ...
[ "torch.utils.data.ConcatDataset", "argparse.ArgumentParser", "torch.utils.data.DataLoader", "dfe.datasets.ColmapDataset", "torch.optim.lr_scheduler.StepLR", "torch.cuda.is_available", "dfe.models.loss.symmetric_epipolar_distance", "dfe.models.NormalizedEightPointNet", "time.time" ]
[((844, 885), 'torch.utils.data.ConcatDataset', 'torch.utils.data.ConcatDataset', (['data_sets'], {}), '(data_sets)\n', (874, 885), False, 'import torch\n'), ((968, 1117), 'torch.utils.data.DataLoader', 'torch.utils.data.DataLoader', (['dset'], {'batch_size': 'options.batch_size', 'shuffle': '(True)', 'num_workers': 'o...
from simple_playgrounds.playground.playgrounds import * from simple_playgrounds.engine import Engine from simple_playgrounds.agent.controllers import Keyboard from simple_playgrounds.agent.agents import HeadAgent import time import cv2 my_agent = HeadAgent(controller=Keyboard(), lateral=True, interactive=True) ####...
[ "simple_playgrounds.agent.controllers.Keyboard", "cv2.waitKey", "simple_playgrounds.engine.Engine", "time.sleep" ]
[((519, 553), 'simple_playgrounds.engine.Engine', 'Engine', ([], {'playground': 'pg', 'debug': '(False)'}), '(playground=pg, debug=False)\n', (525, 553), False, 'from simple_playgrounds.engine import Engine\n'), ((271, 281), 'simple_playgrounds.agent.controllers.Keyboard', 'Keyboard', ([], {}), '()\n', (279, 281), Fals...
""" Copyright 2020 The OneFlow Authors. All rights reserved. Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agr...
[ "collections.OrderedDict", "oneflow.experimental.unittest.skip_unless_1n1d", "numpy.array", "test_util.GenArgList", "unittest.main", "oneflow.experimental.device" ]
[((2909, 2941), 'oneflow.experimental.unittest.skip_unless_1n1d', 'flow.unittest.skip_unless_1n1d', ([], {}), '()\n', (2939, 2941), True, 'import oneflow.experimental as flow\n'), ((786, 829), 'numpy.array', 'np.array', (['[[1, 2, 3], [4, 5, 6], [7, 8, 9]]'], {}), '([[1, 2, 3], [4, 5, 6], [7, 8, 9]])\n', (794, 829), Tr...
from builtins import * import argparse import numpy as np import os from bnpy.ioutil.DataReader import loadDataFromSavedTask, loadLPKwargsFromDisk from bnpy.ioutil.DataReader import loadKwargsFromDisk from bnpy.ioutil.ModelReader import loadModelForLap from bnpy.util import StateSeqUtil from bnpy.birthmove.BCreateOneP...
[ "bnpy.ioutil.ModelReader.loadModelForLap", "bnpy.ioutil.DataReader.loadDataFromSavedTask", "numpy.unique", "argparse.ArgumentParser", "bnpy.ioutil.DataReader.loadKwargsFromDisk", "bnpy.util.StateSeqUtil.alignEstimatedStateSeqToTruth", "bnpy.birthmove.BLogger.configure", "numpy.argmax", "numpy.sum", ...
[((1424, 1461), 'bnpy.ioutil.ModelReader.loadModelForLap', 'loadModelForLap', (['taskoutpath', 'lapFrac'], {}), '(taskoutpath, lapFrac)\n', (1439, 1461), False, 'from bnpy.ioutil.ModelReader import loadModelForLap\n'), ((1473, 1524), 'bnpy.ioutil.DataReader.loadDataFromSavedTask', 'loadDataFromSavedTask', (['taskoutpat...
""" To be placed into {BlenderProcRoot}/src/loader/. """ import bpy from src.loader.LoaderInterface import LoaderInterface from src.utility.Utility import Utility from src.utility.LabelIdMapping import LabelIdMapping class CustomObjectLoader(LoaderInterface): """ Custom object loader which, in addition to t...
[ "bpy.ops.object.select_all", "src.loader.LoaderInterface.LoaderInterface.__init__", "src.utility.Utility.Utility.import_objects", "src.utility.Utility.Utility.resolve_path", "src.utility.Utility.Utility.get_nodes_with_type", "src.utility.Utility.Utility.get_the_one_node_with_type", "src.loader.LoaderInt...
[((589, 627), 'src.loader.LoaderInterface.LoaderInterface.__init__', 'LoaderInterface.__init__', (['self', 'config'], {}), '(self, config)\n', (613, 627), False, 'from src.loader.LoaderInterface import LoaderInterface\n'), ((2199, 2249), 'src.loader.LoaderInterface.LoaderInterface.remove_x_axis_rotation', 'LoaderInterf...
from django.contrib import admin from .models import Quiz, Question, Response # Register your models here. class InLineResponse(admin.StackedInline): model = Response extra = 0 class InLineQuestion(admin.StackedInline): model = Question extra = 0 class QuizAdmin(admin.ModelAdmin): inlines = [...
[ "django.contrib.admin.site.register" ]
[((428, 464), 'django.contrib.admin.site.register', 'admin.site.register', (['Quiz', 'QuizAdmin'], {}), '(Quiz, QuizAdmin)\n', (447, 464), False, 'from django.contrib import admin\n'), ((465, 509), 'django.contrib.admin.site.register', 'admin.site.register', (['Question', 'QuestionAdmin'], {}), '(Question, QuestionAdmi...
import torch import torch.autograd as autograd from torch.autograd import Variable import torch.nn as nn import torch.nn.functional as F import torch.optim as optim from torch.nn.parameter import Parameter import numpy as np import datetime from transformers import EncoderDecoderModel, BertTokenizer from Hyperparameter...
[ "transformers.BertTokenizer.from_pretrained", "transformers.EncoderDecoderModel.from_encoder_decoder_pretrained" ]
[((501, 551), 'transformers.BertTokenizer.from_pretrained', 'BertTokenizer.from_pretrained', (['"""bert-base-uncased"""'], {}), "('bert-base-uncased')\n", (530, 551), False, 'from transformers import EncoderDecoderModel, BertTokenizer\n'), ((573, 670), 'transformers.EncoderDecoderModel.from_encoder_decoder_pretrained',...
import doctest def test_doctests(): doctest.testfile("../README.rst") if __name__ == "__main__": test_doctests()
[ "doctest.testfile" ]
[((42, 75), 'doctest.testfile', 'doctest.testfile', (['"""../README.rst"""'], {}), "('../README.rst')\n", (58, 75), False, 'import doctest\n')]
from __future__ import unicode_literals import re from .common import InfoExtractor from .gigya import GigyaBaseIE from ..compat import compat_HTTPError from ..utils import ( ExtractorError, clean_html, extract_attributes, float_or_none, get_element_by_class, int_or_none, merge_dicts, ...
[ "re.match" ]
[((1619, 1649), 're.match', 're.match', (['self._VALID_URL', 'url'], {}), '(self._VALID_URL, url)\n', (1627, 1649), False, 'import re\n'), ((7761, 7791), 're.match', 're.match', (['self._VALID_URL', 'url'], {}), '(self._VALID_URL, url)\n', (7769, 7791), False, 'import re\n')]
import setting set_dic = setting.setting() if set_dic["MODE"] == "WhereFrom": import logging def f1(): if False: logging.info("debug log!!") def f2(): if True: logging.info("debug log!!") if set_dic["MODE"] == "FromHere": import logging logger=logging.getLo...
[ "logging.getLogger", "logging.Formatter", "logging.handlers.RotatingFileHandler", "logging.info", "setting.setting" ]
[((26, 43), 'setting.setting', 'setting.setting', ([], {}), '()\n', (41, 43), False, 'import setting\n'), ((307, 334), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (324, 334), False, 'import logging\n'), ((660, 687), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name...
# Copyright (C) 2019-2021, TomTom (http://tomtom.com). # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law o...
[ "logging.getLogger", "pathlib.Path", "xml.etree.ElementTree.tostring", "xml.etree.ElementTree.Element", "re.sub", "xml.etree.ElementTree.SubElement" ]
[((864, 891), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (881, 891), False, 'import logging\n'), ((5084, 5137), 'xml.etree.ElementTree.Element', 'ET.Element', (['"""div"""'], {'id': '"""toc"""', 'attrib': "{'class': 'toc2'}"}), "('div', id='toc', attrib={'class': 'toc2'})\n", (5094, 5...
import requests import json import bs4 as bs import pandas as pd your_head={ 'GET':"wss: '// chat - ws.shopee.co.id / socket.io /?EIO = 3 & transport = websocket HTTP / 1.1", 'connection': 'Upgrade', 'pragma': 'no - cache', 'cache - Control': 'no - cache', 'user - Agent': 'Mozilla / 5.0(Linux;Android 6.0;Nexus 5 Build...
[ "pandas.DataFrame", "json.loads", "requests.get" ]
[((2029, 2045), 'json.loads', 'json.loads', (['data'], {}), '(data)\n', (2039, 2045), False, 'import json\n'), ((2770, 2793), 'pandas.DataFrame', 'pd.DataFrame', (['prod_data'], {}), '(prod_data)\n', (2782, 2793), True, 'import pandas as pd\n'), ((1993, 2017), 'requests.get', 'requests.get', (['shopee_url'], {}), '(sho...
import numpy as np import random from FuncionAptitud import fitness lista = [0, 1, 2, 3, 4, 5, 6, 7] # son los valores en los que puede estar la reyna poblacion = np.empty((50,8)) for i in range(50): random.shuffle(lista) for j in range(8): poblacion[i, j] = lista[j] def padres(conjunto): r1 = ra...
[ "FuncionAptitud.fitness", "random.random", "numpy.empty", "random.shuffle" ]
[((165, 182), 'numpy.empty', 'np.empty', (['(50, 8)'], {}), '((50, 8))\n', (173, 182), True, 'import numpy as np\n'), ((206, 227), 'random.shuffle', 'random.shuffle', (['lista'], {}), '(lista)\n', (220, 227), False, 'import random\n'), ((318, 333), 'random.random', 'random.random', ([], {}), '()\n', (331, 333), False, ...
# Generated by Django 3.2.11 on 2022-02-02 07:10 from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('engine', '0048_auto_20211112_1918'), ] operations = [ migrations.AlterField( model_name='labeledshape', name='type'...
[ "django.db.models.CharField" ]
[((340, 536), 'django.db.models.CharField', 'models.CharField', ([], {'choices': "[('rectangle', 'RECTANGLE'), ('polygon', 'POLYGON'), ('polyline',\n 'POLYLINE'), ('points', 'POINTS'), ('ellipse', 'ELLIPSE'), ('cuboid',\n 'CUBOID')]", 'max_length': '(16)'}), "(choices=[('rectangle', 'RECTANGLE'), ('polygon', 'POL...
import torch import torch.nn import torch.nn.functional as F import warnings from .types import DataScoreModel, ScoreMatchingLoss, ConditionalDataScoreModel from . import jacobians def _ism(tr_jac: torch.Tensor, scores: torch.Tensor) -> torch.Tensor: return (tr_jac + 0.5 * (scores ** 2).sum(-1)).mean() def ism...
[ "warnings.warn", "torch.randint", "torch.randn_like" ]
[((1568, 1587), 'torch.randn_like', 'torch.randn_like', (['x'], {}), '(x)\n', (1584, 1587), False, 'import torch\n'), ((1885, 1938), 'torch.randint', 'torch.randint', (['(0)', 'sigmae.shape[0]'], {'size': '(x.shape[0],)'}), '(0, sigmae.shape[0], size=(x.shape[0],))\n', (1898, 1938), False, 'import torch\n'), ((2027, 20...
from __future__ import absolute_import from __future__ import print_function from pysnptools.util.mapreduce1.runner import * import logging import fastlmm.pyplink.plink as plink import pysnptools.util as pstutil import pysnptools.util.pheno as pstpheno import numpy as np from fastlmm.inference import LMM import scipy.s...
[ "numpy.sqrt", "pysnptools.util.create_directory_if_necessary", "numpy.hstack", "numpy.array", "logging.info", "pysnptools.util.pheno.loadOnePhen", "numpy.arange", "pysnptools.util.pheno.loadPhen", "numpy.savez", "numpy.exp", "pysnptools.util.intersect_apply", "doctest.testmod", "numpy.ones",...
[((20159, 20170), 'time.time', 'time.time', ([], {}), '()\n', (20168, 20170), False, 'import time\n'), ((24987, 25004), 'doctest.testmod', 'doctest.testmod', ([], {}), '()\n', (25002, 25004), False, 'import doctest\n'), ((8978, 9073), 'pysnptools.util.intersect_apply', 'pstutil.intersect_apply', (['[self.test_snps, sel...
from graphene.utils.str_converters import to_snake_case def localized_field_resolver(obj, info, **kwargs): """ Custom resolver to return the user language value from localized fields """ attr = getattr(obj, to_snake_case(info.field_name)) if attr is None: return None if info.context....
[ "graphene.utils.str_converters.to_snake_case" ]
[((225, 255), 'graphene.utils.str_converters.to_snake_case', 'to_snake_case', (['info.field_name'], {}), '(info.field_name)\n', (238, 255), False, 'from graphene.utils.str_converters import to_snake_case\n')]
#!/usr/bin/env python3 from sqlalchemy.ext.declarative import declarative_base from sqlalchemy import Column, String, DateTime from sqlalchemy.sql import func from sqlalchemy.dialects.postgresql import JSONB Base = declarative_base() class JobResult(Base): __tablename__ = 'job_result' job_id = Column(Stri...
[ "sqlalchemy.String", "sqlalchemy.sql.func.now", "sqlalchemy.Column", "sqlalchemy.ext.declarative.declarative_base" ]
[((218, 236), 'sqlalchemy.ext.declarative.declarative_base', 'declarative_base', ([], {}), '()\n', (234, 236), False, 'from sqlalchemy.ext.declarative import declarative_base\n'), ((466, 480), 'sqlalchemy.Column', 'Column', (['String'], {}), '(String)\n', (472, 480), False, 'from sqlalchemy import Column, String, DateT...
""" Module containg all the necessary class for the package. FoodCategory is a Enum class representing the category of ingredient. CategorySynset is a class representing the wordnet synset of a category. """ from enum import Enum from nltk.corpus import wordnet class FoodCategory(Enum): """ Enum class used ...
[ "nltk.corpus.wordnet.synset" ]
[((829, 882), 'nltk.corpus.wordnet.synset', 'wordnet.synset', (['f"""{FoodCategory.vegetable.name}.n.01"""'], {}), "(f'{FoodCategory.vegetable.name}.n.01')\n", (843, 882), False, 'from nltk.corpus import wordnet\n'), ((892, 941), 'nltk.corpus.wordnet.synset', 'wordnet.synset', (['f"""{FoodCategory.fruit.name}.n.01"""']...
import os from nxpy.nxfile import NXFile def test_nxnode_resolve(): node = NXFile(os.path.join(os.path.dirname(__file__), 'map.nx')).get_root_node().resolve( "Tile/grassySoil.img/bsc/0") assert node.width == 90 node2 = NXFile(os.path.join(os.path.dirname(__file__), 'map.nx')).get_root_node().g...
[ "os.path.dirname" ]
[((102, 127), 'os.path.dirname', 'os.path.dirname', (['__file__'], {}), '(__file__)\n', (117, 127), False, 'import os\n'), ((264, 289), 'os.path.dirname', 'os.path.dirname', (['__file__'], {}), '(__file__)\n', (279, 289), False, 'import os\n'), ((440, 465), 'os.path.dirname', 'os.path.dirname', (['__file__'], {}), '(__...
from django.db import models from django.contrib.auth import get_user_model class Category(models.Model): title = models.CharField(max_length=255) def __str__(self): if len(str(self.title)) > 20: return self.title[:20] return self.title class Post(models.Model): title = mode...
[ "django.contrib.auth.get_user_model", "django.db.models.TextField", "django.db.models.ManyToManyField", "django.db.models.CharField" ]
[((120, 152), 'django.db.models.CharField', 'models.CharField', ([], {'max_length': '(255)'}), '(max_length=255)\n', (136, 152), False, 'from django.db import models\n'), ((316, 348), 'django.db.models.CharField', 'models.CharField', ([], {'max_length': '(255)'}), '(max_length=255)\n', (332, 348), False, 'from django.d...
from generator import Generator from discriminator import Discriminator from utils import * import torch from torch import nn from tqdm.auto import tqdm from torchvision import transforms from torchvision.datasets import MNIST from torch.utils.data import DataLoader import argparse parser = argparse.ArgumentParser( ...
[ "generator.Generator", "torch.ones_like", "argparse.ArgumentParser", "discriminator.Discriminator", "torch.zeros_like", "torchvision.transforms.Normalize", "tqdm.auto.tqdm", "torch.nn.BCEWithLogitsLoss", "torchvision.datasets.MNIST", "torchvision.transforms.ToTensor" ]
[((293, 345), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {'description': '"""DCGAN trainer"""'}), "(description='DCGAN trainer')\n", (316, 345), False, 'import argparse\n'), ((707, 718), 'generator.Generator', 'Generator', ([], {}), '()\n', (716, 718), False, 'from generator import Generator\n'), ((770,...
from dataset import CovidImageDataset from argparse import ArgumentParser import torch import torch.nn as nn from model import VGG import numpy as np import os from pytorch_lightning.utilities.seed import seed_everything import random def seed_worker(worker_id): ''' https://pytorch.org/docs/stable/notes/rando...
[ "torch.nn.CrossEntropyLoss", "argparse.ArgumentParser", "torch.utils.data.DataLoader", "torch.optim.lr_scheduler.CosineAnnealingLR", "torch.initial_seed", "os.makedirs", "torch.max", "os.path.join", "random.seed", "model.VGG", "torch.cuda.is_available", "numpy.random.seed", "torch.use_determ...
[((609, 636), 'numpy.random.seed', 'np.random.seed', (['worker_seed'], {}), '(worker_seed)\n', (623, 636), True, 'import numpy as np\n'), ((641, 665), 'random.seed', 'random.seed', (['worker_seed'], {}), '(worker_seed)\n', (652, 665), False, 'import random\n'), ((2124, 2140), 'argparse.ArgumentParser', 'ArgumentParser'...
# Copyright 2018 Google Inc. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, ...
[ "unittest.main", "unittest.mock.MagicMock", "unittest.mock.patch" ]
[((1260, 1297), 'unittest.mock.patch', 'mock.patch', (['SNIPPET_CLIENT_CLASS_PATH'], {}), '(SNIPPET_CLIENT_CLASS_PATH)\n', (1270, 1297), False, 'from unittest import mock\n'), ((1607, 1644), 'unittest.mock.patch', 'mock.patch', (['SNIPPET_CLIENT_CLASS_PATH'], {}), '(SNIPPET_CLIENT_CLASS_PATH)\n', (1617, 1644), False, '...
#!/usr/bin/env python3 """ Script to start a client connection to server. (Talker) """ import socket import json import datetime import copy import hashlib import math import logging from subprocess import check_call from Crypto.PublicKey import RSA from Crypto.Cipher import PKCS1_OAEP from Crypto.Random import get...
[ "json.loads", "socket.socket", "subprocess.check_call", "encryptlib.SimonCTR.countermode_encrypt", "Crypto.Random.get_random_bytes", "encryptlib.print_helper.PrintHelper", "json.dumps", "encryptlib.SimonCTR.countermode_decrypt", "logging.info", "datetime.datetime.now", "copy.copy", "encryptlib...
[((1003, 1016), 'encryptlib.print_helper.PrintHelper', 'PrintHelper', ([], {}), '()\n', (1014, 1016), False, 'from encryptlib.print_helper import PrintHelper\n'), ((1178, 1227), 'socket.socket', 'socket.socket', (['socket.AF_INET', 'socket.SOCK_STREAM'], {}), '(socket.AF_INET, socket.SOCK_STREAM)\n', (1191, 1227), Fals...
# Copyright (c) 2019 PaddlePaddle Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by app...
[ "logging.getLogger" ]
[((1111, 1138), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (1128, 1138), False, 'import logging\n')]
import operations def test_getOneNews_basic(): news = operations.getOneNews() print(news) assert news is not None print('test_getOneNews_basic passed!') def test_getNewsSummariesForUser_basic(): news = operations.getNewsSummariesForUser('test', 1) assert len(news) > 0 print('test_getNewsSu...
[ "operations.getOneNews", "operations.getNewsSummariesForUser" ]
[((59, 82), 'operations.getOneNews', 'operations.getOneNews', ([], {}), '()\n', (80, 82), False, 'import operations\n'), ((224, 269), 'operations.getNewsSummariesForUser', 'operations.getNewsSummariesForUser', (['"""test"""', '(1)'], {}), "('test', 1)\n", (258, 269), False, 'import operations\n'), ((417, 462), 'operati...
from setuptools import find_packages, setup LONG_DESCRIPTION = ( 'Desc.' ) setup( name='mhealth', version='0.0.3', packages=find_packages(where='src'), package_dir={'': 'src'}, url='https://github.com/callumstew/pymhealth', author='<NAME>', author_email='<E...
[ "setuptools.find_packages" ]
[((155, 181), 'setuptools.find_packages', 'find_packages', ([], {'where': '"""src"""'}), "(where='src')\n", (168, 181), False, 'from setuptools import find_packages, setup\n')]
from django.urls import path from . import views # Objektivaated from wiki.views import ArtikkelDetailView, IsikDetailView, OrganisatsioonDetailView, ObjektDetailView, KaardiobjektDetailView # Filtreerimisvaated from wiki.views import ArtikkelFilterView, IsikFilterView, OrganisatsioonFilterView, ObjektFilterView, Kaa...
[ "wiki.views.OrganisatsioonUpdate.as_view", "wiki.views.ArtikkelYearArchiveView.as_view", "wiki.views.KaardiobjektUpdate.as_view", "wiki.views.KaardiobjektDetailView.as_view", "wiki.views.IsikUpdate.as_view", "wiki.views.ArtikkelArchiveIndexView.as_view", "wiki.views.ArtikkelUpdate.as_view", "wiki.view...
[((865, 903), 'django.urls.path', 'path', (['"""info/"""', 'views.info'], {'name': '"""info"""'}), "('info/', views.info, name='info')\n", (869, 903), False, 'from django.urls import path\n'), ((909, 947), 'django.urls.path', 'path', (['"""otsi/"""', 'views.otsi'], {'name': '"""otsi"""'}), "('otsi/', views.otsi, name='...
from office365.entity import Entity from office365.runtime.client_result import ClientResult from office365.runtime.queries.delete_entity_query import DeleteEntityQuery from office365.runtime.queries.service_operation_query import ServiceOperationQuery from office365.runtime.queries.update_entity_query import UpdateEnt...
[ "office365.runtime.queries.service_operation_query.ServiceOperationQuery", "office365.runtime.client_result.ClientResult", "office365.runtime.queries.update_entity_query.UpdateEntityQuery", "office365.runtime.queries.delete_entity_query.DeleteEntityQuery" ]
[((766, 784), 'office365.runtime.client_result.ClientResult', 'ClientResult', (['None'], {}), '(None)\n', (778, 784), False, 'from office365.runtime.client_result import ClientResult\n'), ((886, 961), 'office365.runtime.queries.service_operation_query.ServiceOperationQuery', 'ServiceOperationQuery', (['self', '"""getMe...
import click from gradient.api_sdk import constants from gradient.cli import common from gradient.cli.cli import cli from gradient.cli.cli_types import ChoiceType, json_string from gradient.cli.common import validate_comma_split_option, api_key_option, ClickGroup from gradient.commands import notebooks from gradient.c...
[ "gradient.commands.notebooks.CreateNotebookCommand", "click.echo", "click.DateTime", "gradient.commands.notebooks.StreamNotebookMetricsCommand", "click.option", "gradient.commands.notebooks.StartNotebookCommand", "gradient.commands.notebooks.StopNotebookCommand", "gradient.commands.notebooks.NotebookA...
[((432, 502), 'gradient.cli.cli.cli.group', 'cli.group', (['"""notebooks"""'], {'help': '"""Manage notebooks"""', 'cls': 'common.ClickGroup'}), "('notebooks', help='Manage notebooks', cls=common.ClickGroup)\n", (441, 502), False, 'from gradient.cli.cli import cli\n'), ((843, 991), 'click.option', 'click.option', (['"""...
# coding: utf-8 """ Licensed to Cloudera, Inc. under one or more contributor license agreements. See the NOTICE file distributed with this work for additional information regarding copyright ownership. Cloudera, Inc. licenses this file to you under the Apache License, Version 2.0 (the "License"); you may not use thi...
[ "six.iteritems" ]
[((12318, 12351), 'six.iteritems', 'six.iteritems', (['self.swagger_types'], {}), '(self.swagger_types)\n', (12331, 12351), False, 'import six\n')]
import chainer import chainer as C import chainer.functions as F import chainer.links as L from chainer.backends import cuda import inspect import weakref import sys from chainer_compiler.elichika.parser import config from chainer_compiler.elichika.parser import nodes from chainer_compiler.elichika.parser import vevalu...
[ "chainer_compiler.elichika.parser.functions_builtin.VEvalContextFunction", "chainer_compiler.elichika.parser.functions_ndarray.NDArrayCumsumFunction", "chainer_compiler.elichika.parser.links_builtin.is_builtin_chainer_link", "chainer_compiler.elichika.parser.links_builtin.ChainerLinkInstance", "chainer_comp...
[((1116, 1149), 'inspect.getmembers', 'inspect.getmembers', (['parent_module'], {}), '(parent_module)\n', (1134, 1149), False, 'import inspect\n'), ((1338, 1373), 'chainer_compiler.elichika.parser.values.reset_field_and_attributes', 'values.reset_field_and_attributes', ([], {}), '()\n', (1371, 1373), False, 'from chain...
import math ####### import random import cv2 import numpy as np import matplotlib.pyplot as plt from tensorpack.dataflow.imgaug.geometry import RotationAndCropValid def crop_meta_image(image,annos,mask): _target_height=368 _target_width =368 if len(np.shape(image))==2: image = cv2.cvtColor(image, c...
[ "matplotlib.pyplot.imshow", "random.uniform", "cv2.warpAffine", "cv2.resize", "cv2.flip", "random.randrange", "cv2.copyMakeBorder", "math.cos", "numpy.random.randint", "cv2.cvtColor", "numpy.random.uniform", "tensorpack.dataflow.imgaug.geometry.RotationAndCropValid.largest_rotated_rect", "ma...
[((358, 373), 'numpy.shape', 'np.shape', (['image'], {}), '(image)\n', (366, 373), True, 'import numpy as np\n'), ((3920, 3935), 'numpy.shape', 'np.shape', (['image'], {}), '(image)\n', (3928, 3935), True, 'import numpy as np\n'), ((4659, 4744), 'cv2.resize', 'cv2.resize', (['image', '(_target_width, _target_height)'],...
from datetime import datetime import json dt = datetime.now() print(dt) json_dt = json.dumps(dt.isoformat()) with open("result.json", "wb") as f: f.write(json_dt)
[ "datetime.datetime.now" ]
[((49, 63), 'datetime.datetime.now', 'datetime.now', ([], {}), '()\n', (61, 63), False, 'from datetime import datetime\n')]
from django.db import models # Create your models here. class User_data(models.Model): """ @pony 使用者(醫生)資料 """ name = models.CharField(max_length=20) account = models.CharField(max_length=50) password = models.CharField(max_length=50) email = models.EmailField() depa...
[ "django.db.models.EmailField", "django.db.models.CharField" ]
[((151, 182), 'django.db.models.CharField', 'models.CharField', ([], {'max_length': '(20)'}), '(max_length=20)\n', (167, 182), False, 'from django.db import models\n'), ((198, 229), 'django.db.models.CharField', 'models.CharField', ([], {'max_length': '(50)'}), '(max_length=50)\n', (214, 229), False, 'from django.db im...
# -*- coding: utf-8 -*- """TcEx Framework Redis Module""" from builtins import str import redis class TcExRedis(object): """Create/Read Data in/from Redis""" def __init__(self, host, port, rhash): """Initialize the Class properties. Args: host (string): The Redis host. ...
[ "redis.StrictRedis" ]
[((456, 495), 'redis.StrictRedis', 'redis.StrictRedis', ([], {'host': 'host', 'port': 'port'}), '(host=host, port=port)\n', (473, 495), False, 'import redis\n')]
from __future__ import unicode_literals import unittest from io import StringIO from python_digest import * from python_digest.http import * from python_digest.utils import * class HttpTests(unittest.TestCase): def test_parse_quoted_string(self): test_cases = [ ('""', ''), # OK (...
[ "unittest.main", "io.StringIO" ]
[((21686, 21701), 'unittest.main', 'unittest.main', ([], {}), '()\n', (21699, 21701), False, 'import unittest\n'), ((14601, 14611), 'io.StringIO', 'StringIO', ([], {}), '()\n', (14609, 14611), False, 'from io import StringIO\n'), ((15109, 15119), 'io.StringIO', 'StringIO', ([], {}), '()\n', (15117, 15119), False, 'from...
import logging import re import shlex import threading import sshim logging.basicConfig(level='DEBUG') logger = logging.getLogger() class Device(threading.Thread): def __init__(self, script): threading.Thread.__init__(self) self.history = [] self.script = script self.start() d...
[ "logging.basicConfig", "threading.Thread.__init__", "logging.getLogger", "re.compile", "shlex.split", "sshim.Server" ]
[((69, 103), 'logging.basicConfig', 'logging.basicConfig', ([], {'level': '"""DEBUG"""'}), "(level='DEBUG')\n", (88, 103), False, 'import logging\n'), ((113, 132), 'logging.getLogger', 'logging.getLogger', ([], {}), '()\n', (130, 132), False, 'import logging\n'), ((1198, 1229), 'sshim.Server', 'sshim.Server', (['Device...