code stringlengths 22 1.05M | apis listlengths 1 3.31k | extract_api stringlengths 75 3.25M |
|---|---|---|
#!/usr/bin/env python
# coding: utf-8
# /*##########################################################################
#
# Copyright (c) 2016-2018 European Synchrotron Radiation Facility
#
# Permission is hereby granted, free of charge, to any person obtaining a copy
# of this software and associated documentation files ... | [
"numpy.trapz",
"silx.gui.qt.QApplication",
"silx.gui.plot.Plot1D",
"numpy.sum",
"silx.gui.plot.stats.stats.StatBase.__init__",
"numpy.arange"
] | [((3016, 3035), 'silx.gui.qt.QApplication', 'qt.QApplication', (['[]'], {}), '([])\n', (3031, 3035), False, 'from silx.gui import qt\n'), ((3048, 3056), 'silx.gui.plot.Plot1D', 'Plot1D', ([], {}), '()\n', (3054, 3056), False, 'from silx.gui.plot import Plot1D\n'), ((3066, 3082), 'numpy.arange', 'numpy.arange', (['(21)'... |
#loads and trains data using simple CPU data loading technique
#adapted from https://colab.research.google.com/github/tensorflow/docs/blob/master/site/en/r2/tutorials/load_data/images.ipynb#scrollTo=qj_U09xpDvOg
import pathlib
import tensorflow as tf
AUTOTUNE = tf.data.experimental.AUTOTUNE
def preprocess_image(imag... | [
"matplotlib.pyplot.grid",
"tensorflow.data.experimental.shuffle_and_repeat",
"tensorflow.io.read_file",
"tensorflow.cast",
"tensorflow.keras.layers.GlobalAveragePooling2D",
"matplotlib.pyplot.imshow",
"tensorflow.data.Dataset.from_tensor_slices",
"pathlib.Path",
"matplotlib.pyplot.yticks",
"tensor... | [((789, 817), 'pathlib.Path', 'pathlib.Path', (['data_root_orig'], {}), '(data_root_orig)\n', (801, 817), False, 'import pathlib\n'), ((987, 1018), 'random.shuffle', 'random.shuffle', (['all_image_paths'], {}), '(all_image_paths)\n', (1001, 1018), False, 'import random\n'), ((1680, 1695), 'matplotlib.pyplot.grid', 'plt... |
from PairedNeurons import PairedNeurons
from matplotlib import pyplot as plt
import os
import numpy as np
import cv2
from xlwt import Workbook
from skimage.segmentation import clear_border
SMOOTH = 1e-6
def iou_numpy(outputs: np.array, labels: np.array):
# outputs = outputs.squeeze(2)
intersection = (outpu... | [
"numpy.uint8",
"numpy.ones",
"cv2.threshold",
"os.path.join",
"skimage.segmentation.clear_border",
"matplotlib.pyplot.close",
"cv2.morphologyEx",
"PairedNeurons.PairedNeurons",
"cv2.distanceTransform",
"cv2.dilate",
"cv2.subtract",
"xlwt.Workbook",
"matplotlib.pyplot.subplots"
] | [((748, 800), 'PairedNeurons.PairedNeurons', 'PairedNeurons', (['img_dir', 'csv_dir', '(256)'], {'is_train': '(False)'}), '(img_dir, csv_dir, 256, is_train=False)\n', (761, 800), False, 'from PairedNeurons import PairedNeurons\n'), ((870, 880), 'xlwt.Workbook', 'Workbook', ([], {}), '()\n', (878, 880), False, 'from xlw... |
import unittest
from .an2cn import An2Cn
class An2CnTest(unittest.TestCase):
def setUp(self):
self.error_input_data = [
u"123.1.1",
u"0.1零"
]
self.integer_data = {
0: [u"零", u"零"],
1: [u"一", u"壹"],
11: [u"十一", u"拾壹"],
... | [
"unittest.main"
] | [((1549, 1564), 'unittest.main', 'unittest.main', ([], {}), '()\n', (1562, 1564), False, 'import unittest\n')] |
#!/usr/bin/env python3
import argparse
from game.game import Game
def main():
""" Reversi game with human player vs AI player.
"""
parser = argparse.ArgumentParser()
parser.add_argument('--timeout', help="Number of seconds the brain is allowed to think before making its move",
... | [
"game.game.Game",
"argparse.ArgumentParser"
] | [((156, 181), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {}), '()\n', (179, 181), False, 'import argparse\n'), ((964, 1025), 'game.game.Game', 'Game', ([], {'timeout': 'args.timeout', 'colour': 'args.text', 'players': 'players'}), '(timeout=args.timeout, colour=args.text, players=players)\n', (968, 1025... |
import os
from collections import namedtuple
from ConfigSpace.read_and_write import json as cs_json
import nasbench301 as nb
# Default dirs for models
# Note: Uses 0.9 as the default models, switch to 1.0 to use 1.0 models
version = '0.9'
current_dir = os.path.dirname(os.path.abspath(__file__))
models_0_9_dir = os.... | [
"os.path.exists",
"collections.namedtuple",
"os.path.join",
"ConfigSpace.read_and_write.json.read",
"nasbench301.download_models",
"os.path.abspath",
"nasbench301.load_ensemble"
] | [((317, 359), 'os.path.join', 'os.path.join', (['current_dir', '"""nb_models_0.9"""'], {}), "(current_dir, 'nb_models_0.9')\n", (329, 359), False, 'import os\n'), ((531, 573), 'os.path.join', 'os.path.join', (['current_dir', '"""nb_models_1.0"""'], {}), "(current_dir, 'nb_models_1.0')\n", (543, 573), False, 'import os\... |
# emacs: -*- mode: python; py-indent-offset: 4; indent-tabs-mode: nil -*-
# vi: set ft=python sts=4 ts=4 sw=4 et:
### ### ### ### ### ### ### ### ### ### ### ### ### ### ### ### ### ### ### ##
#
# See COPYING file distributed along with the PyMVPA package for the
# copyright and license terms.
#
### ### ### ### ###... | [
"mvpa2.base.dochelpers.borrowkwargs",
"mvpa2.measures.adhocsearchlightbase._STATS",
"numpy.unique",
"numpy.ones",
"mvpa2.misc.neighborhood.IndexQueryEngine",
"numpy.square",
"numpy.asanyarray",
"numpy.sum",
"numpy.zeros",
"numpy.empty",
"mvpa2.base.dochelpers._repr_attrs",
"numpy.argmin",
"m... | [((8166, 8243), 'mvpa2.base.dochelpers.borrowkwargs', 'borrowkwargs', (['M1NNSearchlight', '"""__init__"""'], {'exclude': "['roi_ids', 'queryengine']"}), "(M1NNSearchlight, '__init__', exclude=['roi_ids', 'queryengine'])\n", (8178, 8243), False, 'from mvpa2.base.dochelpers import borrowkwargs, _repr_attrs\n'), ((1046, ... |
import os
from django.db import migrations
from django.conf import settings
import privatemedia.fields
import privatemedia.storage
import transmittals.fileutils
def move_dirs(*args):
protected_root = settings.PROTECTED_ROOT
if not os.path.exists(protected_root):
os.makedirs(protected_root)
priva... | [
"os.path.exists",
"os.makedirs",
"os.rename",
"os.path.join",
"django.db.migrations.RunPython"
] | [((446, 490), 'os.path.join', 'os.path.join', (['protected_root', '"""transmittals"""'], {}), "(protected_root, 'transmittals')\n", (458, 490), False, 'import os\n'), ((505, 547), 'os.path.join', 'os.path.join', (['private_root', '"""transmittals"""'], {}), "(private_root, 'transmittals')\n", (517, 547), False, 'import... |
#!/usr/bin/env python
# coding: utf8
"""MMMM-Facial-Recognition-OCV3 - MagicMirror Module
The MIT License (MIT)
Copyright (c) 2018 <NAME> (MIT License)
Based on work by <NAME> (Copyright 2016) (MIT License)
"""
from lib.tools.train import ToolsTrain
if __name__ == '__main__':
ToolsTrain().train()
| [
"lib.tools.train.ToolsTrain"
] | [((285, 297), 'lib.tools.train.ToolsTrain', 'ToolsTrain', ([], {}), '()\n', (295, 297), False, 'from lib.tools.train import ToolsTrain\n')] |
from django.shortcuts import render, redirect
from app01 import models
from app01.utils.pagination import Pagination
from app01.utils.form import UserModelForm, PrettyModelForm, PrettyEditModelForm
def user_list(request):
""" 用户管理 """
queryset = models.UserInfo.objects.all()
page_object = Pagination(re... | [
"django.shortcuts.render",
"app01.utils.form.UserModelForm",
"app01.utils.pagination.Pagination",
"openpyxl.load_workbook",
"app01.models.UserInfo.objects.all",
"app01.models.Department.objects.filter",
"django.shortcuts.redirect",
"app01.models.UserInfo.objects.filter",
"app01.models.Department.obj... | [((258, 287), 'app01.models.UserInfo.objects.all', 'models.UserInfo.objects.all', ([], {}), '()\n', (285, 287), False, 'from app01 import models\n'), ((307, 350), 'app01.utils.pagination.Pagination', 'Pagination', (['request', 'queryset'], {'page_size': '(10)'}), '(request, queryset, page_size=10)\n', (317, 350), False... |
# Copyright (C) 2020 Google Inc.
# Licensed under http://www.apache.org/licenses/LICENSE-2.0 <see LICENSE file>
"""
Add 2 new roles for scoping objects
Create Date: 2018-10-23 11:02:28.166523
"""
# disable Invalid constant name pylint warning for mandatory Alembic variables.
# pylint: disable=invalid-name
import dat... | [
"alembic.op.get_bind",
"ggrc.migrations.utils.acr_propagation.propagate_roles",
"alembic.op.bulk_insert",
"datetime.datetime.now",
"ggrc.migrations.utils.migrator.get_migration_user_id"
] | [((1230, 1243), 'alembic.op.get_bind', 'op.get_bind', ([], {}), '()\n', (1241, 1243), False, 'from alembic import op\n'), ((1256, 1298), 'ggrc.migrations.utils.migrator.get_migration_user_id', 'migrator.get_migration_user_id', (['connection'], {}), '(connection)\n', (1286, 1298), False, 'from ggrc.migrations.utils impo... |
from flask import current_app
from flask import request
from switchmng.typing import FlaskResponse
from switchmng import database
from .blueprint import restbp
from .errors import *
@restbp.route('/switch_models', methods = ['POST'])
def post_switch_model() -> FlaskResponse:
session = current_app.config['SWITCHM... | [
"switchmng.database.add_vlan",
"switchmng.database.add_switch",
"switchmng.database.add_port_model",
"switchmng.database.add_switch_model",
"switchmng.database.add_connector",
"switchmng.database.add_network_protocol"
] | [((885, 926), 'switchmng.database.add_switch_model', 'database.add_switch_model', (['session'], {}), '(session, **req)\n', (910, 926), False, 'from switchmng import database\n'), ((1839, 1904), 'switchmng.database.add_port_model', 'database.add_port_model', (['session', 'switch_model_resource_id'], {}), '(session, swit... |
# fix redirects
import os
import fileinput
_SRC = "source"
with open("build/linkcheck/output.txt", "r") as f:
link_out = f.readlines()
#%%
for link in link_out:
if "redirected" in link:
file_name = os.path.join(_SRC, link.split(":")[0])
assert os.path.exists(file_name)
old, new =... | [
"os.path.exists",
"fileinput.FileInput"
] | [((275, 300), 'os.path.exists', 'os.path.exists', (['file_name'], {}), '(file_name)\n', (289, 300), False, 'import os\n'), ((378, 422), 'fileinput.FileInput', 'fileinput.FileInput', (['file_name'], {'inplace': '(True)'}), '(file_name, inplace=True)\n', (397, 422), False, 'import fileinput\n')] |
import pytest
from app.api.services import abr_service
from app.api.business.errors import AbrError
import requests
import mock
from mock import patch
class TestAbrService():
def mocked_find_business_by_abn(self):
data = '<ABR><response><stateCode>NSW</stateCode><postcode>2750</postcode>'\
'<... | [
"mock.patch",
"app.api.services.abr_service.call_abr_api",
"requests.exceptions.SSLError",
"requests.exceptions.ProxyError",
"pytest.raises",
"requests.exceptions.ConnectionError",
"requests.exceptions.HTTPError",
"requests.exceptions.RequestException"
] | [((1310, 1365), 'mock.patch', 'mock.patch', (['"""app.api.services.abr_service.call_abr_api"""'], {}), "('app.api.services.abr_service.call_abr_api')\n", (1320, 1365), False, 'import mock\n'), ((1660, 1715), 'mock.patch', 'mock.patch', (['"""app.api.services.abr_service.call_abr_api"""'], {}), "('app.api.services.abr_s... |
import numpy as np
from sklearn.metrics import roc_curve, auc
from sklearn.metrics import confusion_matrix
from sklearn import preprocessing
from sklearn.preprocessing import LabelEncoder
# from IPython.display import Image,display
import matplotlib.pyplot as plt
data = []
labels = []
alldata = []
# XORdata=np.array([... | [
"matplotlib.pyplot.ylabel",
"sklearn.metrics.auc",
"numpy.array",
"sklearn.metrics.roc_curve",
"matplotlib.pyplot.xlabel",
"matplotlib.pyplot.plot",
"numpy.exp",
"matplotlib.pyplot.ylim",
"csv.reader",
"sklearn.metrics.confusion_matrix",
"sklearn.model_selection.train_test_split",
"numpy.argma... | [((9996, 10007), 'time.time', 'time.time', ([], {}), '()\n', (10005, 10007), False, 'import time\n'), ((10194, 10214), 'csv.reader', 'csv.reader', (['gpsTrack'], {}), '(gpsTrack)\n', (10204, 10214), False, 'import csv\n'), ((10513, 10542), 'sklearn.preprocessing.OneHotEncoder', 'preprocessing.OneHotEncoder', ([], {}), ... |
#!/usr/bin/env python3
"""Split PDFS by QR code and move images and PDFs to correct folder."""
import os
import traceback
import numpy
from . import write_to_log as logger
from . import submitty_ocr as scanner
# try importing required modules
try:
from PyPDF2 import PdfFileReader, PdfFileWriter
from pdf2imag... | [
"traceback.format_exc",
"cv2.threshold",
"os.path.join",
"os.chdir",
"numpy.array",
"pyzbar.pyzbar.decode",
"cv2.cvtColor",
"os.getpid",
"PyPDF2.PdfFileWriter",
"traceback.print_exc",
"PyPDF2.PdfFileReader"
] | [((463, 484), 'traceback.print_exc', 'traceback.print_exc', ([], {}), '()\n', (482, 484), False, 'import traceback\n'), ((854, 874), 'os.chdir', 'os.chdir', (['split_path'], {}), '(split_path)\n', (862, 874), False, 'import os\n'), ((894, 917), 'PyPDF2.PdfFileReader', 'PdfFileReader', (['filename'], {}), '(filename)\n'... |
import os
import cgi
import re
from pyramid.response import Response
from pyramid.view import view_config
import pyramid.httpexceptions as httpexceptions
# -------------------------------------------------------------------
class ReflectorView(object):
def __init__(self, request):
self.request = request... | [
"os.path.dirname",
"os.path.basename",
"pyramid.response.Response",
"cgi.escape",
"pyramid.view.view_config"
] | [((327, 362), 'pyramid.view.view_config', 'view_config', ([], {'route_name': '"""reflector"""'}), "(route_name='reflector')\n", (338, 362), False, 'from pyramid.view import view_config\n'), ((925, 935), 'pyramid.response.Response', 'Response', ([], {}), '()\n', (933, 935), False, 'from pyramid.response import Response\... |
# -*- coding:utf-8 -*-
'''
@Author: GETF
@Email: <EMAIL>
@DateTime: 2017-11-09 15:12:14
@Description: Description
'''
import os
from PyQt5.QtGui import QIcon
import sys
from PyQt5.QtWidgets import (QWidget, QLCDNumber, QSlider,
QVBoxLayout, QApplication)
from PyQt5.QtCore... | [
"PyQt5.QtGui.QIcon",
"PyQt5.QtCore.QDateTime.currentDateTime",
"PyQt5.QtCore.QTimer",
"os.path.dirname",
"PyQt5.QtWidgets.QApplication",
"PyQt5.QtWidgets.QLCDNumber"
] | [((1469, 1491), 'PyQt5.QtWidgets.QApplication', 'QApplication', (['sys.argv'], {}), '(sys.argv)\n', (1481, 1491), False, 'from PyQt5.QtWidgets import QWidget, QLCDNumber, QSlider, QVBoxLayout, QApplication\n'), ((487, 503), 'PyQt5.QtWidgets.QLCDNumber', 'QLCDNumber', (['self'], {}), '(self)\n', (497, 503), False, 'from... |
"""Module to hold the Rt class definition."""
from typing import Optional
from aiohttp import ClientSession
from bs4 import BeautifulSoup
from bs4.element import Tag
from phylm.utils.web import async_soupify
from phylm.utils.web import url_encode
RT_BASE_MOVIE_URL = "https://www.rottentomatoes.com/search"
class Rt... | [
"phylm.utils.web.url_encode",
"phylm.utils.web.async_soupify"
] | [((1646, 1672), 'phylm.utils.web.url_encode', 'url_encode', (['self.raw_title'], {}), '(self.raw_title)\n', (1656, 1672), False, 'from phylm.utils.web import url_encode\n'), ((1764, 1798), 'phylm.utils.web.async_soupify', 'async_soupify', (['search_url', 'session'], {}), '(search_url, session)\n', (1777, 1798), False, ... |
#!/usr/bin/env python
# fake-build-for-bitcode.py - Fake build with -embed-bitcode -*- python -*-
#
# This source file is part of the Swift.org open source project
#
# Copyright (c) 2014 - 2016 Apple Inc. and the Swift project authors
# Licensed under Apache License v2.0 with Runtime Library Exception
#
# See http://sw... | [
"sys.argv.index",
"os.path.basename",
"os.utime"
] | [((1052, 1078), 'os.utime', 'os.utime', (['outputFile', 'None'], {}), '(outputFile, None)\n', (1060, 1078), False, 'import os\n'), ((832, 863), 'sys.argv.index', 'sys.argv.index', (['"""-primary-file"""'], {}), "('-primary-file')\n", (846, 863), False, 'import sys\n'), ((891, 911), 'sys.argv.index', 'sys.argv.index', (... |
#!/usr/bin/python
# Copyright (c) 2018, Oracle and/or its affiliates.
# This software is made available to you under the terms of the GPL 3.0 license or the Apache 2.0 license.
# GNU General Public License v3.0+ (see COPYING or https://www.gnu.org/licenses/gpl-3.0.txt)
# Apache License v2.0
# See LICENSE.TXT for detail... | [
"ansible.module_utils.basic.AnsibleModule",
"ansible.module_utils.oracle.oci_lb_utils.delete_lb_resources_and_wait",
"ansible.module_utils.oracle.oci_utils.create_service_client",
"oci.load_balancer.models.UpdateBackendDetails",
"oci.load_balancer.models.CreateBackendDetails",
"oci.util.to_dict",
"ansib... | [((8937, 8974), 'ansible.module_utils.oracle.oci_lb_utils.get_backend_name', 'oci_lb_utils.get_backend_name', (['module'], {}), '(module)\n', (8966, 8974), False, 'from ansible.module_utils.oracle import oci_utils, oci_lb_utils\n'), ((9004, 9026), 'oci.load_balancer.models.CreateBackendDetails', 'CreateBackendDetails',... |
#coding=utf-8
# Copyright (c) 2020 PaddlePaddle Authors. All Rights Reserve.
#
#Licensed under the Apache License, Version 2.0 (the "License");
#you may not use this file except in compliance with the License.
#You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
#Unless required... | [
"os.path.exists",
"paddle.fluid.dygraph.learning_rate_scheduler.ReduceLROnPlateau",
"os.listdir",
"paddle.fluid.dygraph.base.to_variable",
"paddle.fluid.layers.cross_entropy",
"paddle.fluid.layers.mean",
"opts.parse_opts",
"numpy.array",
"numpy.zeros",
"paddle.fluid.CUDAPlace",
"models.model.gen... | [((2083, 2095), 'opts.parse_opts', 'parse_opts', ([], {}), '()\n', (2093, 2095), False, 'from opts import parse_opts\n'), ((1326, 1362), 'os.path.exists', 'os.path.exists', (['opt.Flow_resume_path'], {}), '(opt.Flow_resume_path)\n', (1340, 1362), False, 'import os\n'), ((1432, 1464), 'os.listdir', 'os.listdir', (['opt.... |
from math import gamma
from typing import Dict, List, Tuple
import matplotlib.pyplot as plt
import numpy as np
import torch
import torch.nn.functional as F
import torch.optim as optim
from atcenv.MASAC.buffer import ReplayBuffer
from atcenv.MASAC.mactor_critic import Actor, CriticQ, CriticV
from torch.nn.utils.clip_gr... | [
"numpy.clip",
"torch.optim.Adam",
"torch.cuda.get_device_name",
"atcenv.MASAC.mactor_critic.Actor",
"numpy.prod",
"atcenv.MASAC.buffer.ReplayBuffer",
"torch.FloatTensor",
"numpy.array",
"atcenv.MASAC.mactor_critic.CriticV",
"torch.cuda.is_available",
"atcenv.MASAC.mactor_critic.CriticQ",
"torc... | [((700, 760), 'atcenv.MASAC.buffer.ReplayBuffer', 'ReplayBuffer', (['STATE_DIM', 'ACTION_DIM', 'BUFFER_SIZE', 'BATCH_SIZE'], {}), '(STATE_DIM, ACTION_DIM, BUFFER_SIZE, BATCH_SIZE)\n', (712, 760), False, 'from atcenv.MASAC.buffer import ReplayBuffer\n'), ((1207, 1261), 'torch.zeros', 'torch.zeros', (['(1)'], {'requires_... |
#!/usr/bin/env python
import signal
import time
import sys
from rc522 import RFID
# Sector 0 == Reserved
# Sector 1 ==
# Sector 2 == Card Type [master,super,user][16 char], UUID [32 char]
# Sector 3 == First Name [Max 48 char]
# Sector 4 == Last Name [Max 48 char]
# Sector 5 == employee id [Max 16 cha... | [
"signal.signal",
"time.sleep",
"rc522.RFID",
"sys.exit"
] | [((680, 686), 'rc522.RFID', 'RFID', ([], {}), '()\n', (684, 686), False, 'from rc522 import RFID\n'), ((1446, 1484), 'signal.signal', 'signal.signal', (['signal.SIGINT', 'end_read'], {}), '(signal.SIGINT, end_read)\n', (1459, 1484), False, 'import signal\n'), ((1434, 1444), 'sys.exit', 'sys.exit', ([], {}), '()\n', (14... |
#
# Code borrowed and modified from https://www.esri.com/arcgis-blog/products/arcgis-pro/health/use-proximity-tracing-to-identify-possible-contact-events/
#
import os
import subprocess
import sys
import winreg
from typing import Dict
import arcpy
import glob
pro_home = arcpy.GetInstallInfo()["InstallDir"]
pro_lib_dir... | [
"os.path.exists",
"pyspark.sql.SparkSession.builder.getOrCreate",
"sys.path.insert",
"winreg.QueryValueEx",
"os.environ.unsetenv",
"os.getenv",
"winreg.OpenKey",
"winreg.ConnectRegistry",
"os.path.join",
"pyspark.SparkConf",
"arcpy.GetInstallInfo",
"pyspark.sql.SparkSession",
"pyspark.SparkC... | [((323, 360), 'os.path.join', 'os.path.join', (['pro_home', '"""Java"""', '"""lib"""'], {}), "(pro_home, 'Java', 'lib')\n", (335, 360), False, 'import os\n'), ((379, 420), 'os.path.join', 'os.path.join', (['pro_home', '"""Java"""', '"""runtime"""'], {}), "(pro_home, 'Java', 'runtime')\n", (391, 420), False, 'import os\... |
"Script for everything User related in the database"
from models import db
class User(db.Model):
"Class used for configuring the User model in the database"
id = db.Column(db.Integer, primary_key=True)
username = db.Column(db.String(80), unique=True, nullable=False)
admin = db.Column(db.Boolean)
b... | [
"models.db.relationship",
"models.db.String",
"models.db.relation",
"models.db.Column"
] | [((172, 211), 'models.db.Column', 'db.Column', (['db.Integer'], {'primary_key': '(True)'}), '(db.Integer, primary_key=True)\n', (181, 211), False, 'from models import db\n'), ((293, 314), 'models.db.Column', 'db.Column', (['db.Boolean'], {}), '(db.Boolean)\n', (302, 314), False, 'from models import db\n'), ((326, 347),... |
# -*- coding: utf-8 -*-
from github import Github
UPSTREAM_GITHUB_REPO = 'wikimedia/pywikibot-core'
GITHUB_REPO = 'magul/pywikibot-core'
_pull_requests = []
def push_branch(local_branch):
name = local_branch.name
repo = local_branch.repo
github = repo.remotes['github']
return github.push(name)[... | [
"github.Github"
] | [((515, 559), 'github.Github', 'Github', (['creds.GITHUB_USER', 'creds.GITHUB_PASS'], {}), '(creds.GITHUB_USER, creds.GITHUB_PASS)\n', (521, 559), False, 'from github import Github\n'), ((1042, 1086), 'github.Github', 'Github', (['creds.GITHUB_USER', 'creds.GITHUB_PASS'], {}), '(creds.GITHUB_USER, creds.GITHUB_PASS)\n'... |
#
# Licensed under the Apache License, Version 2.0 (the "License"); you may
# not use this file except in compliance with the License. You may obtain
# a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# ... | [
"migrate.changeset.constraint.ForeignKeyConstraint",
"sqlalchemy.MetaData",
"sqlalchemy.Column",
"sqlalchemy.Table"
] | [((676, 716), 'sqlalchemy.MetaData', 'sqlalchemy.MetaData', ([], {'bind': 'migrate_engine'}), '(bind=migrate_engine)\n', (695, 716), False, 'import sqlalchemy\n'), ((733, 782), 'sqlalchemy.Table', 'sqlalchemy.Table', (['"""resource"""', 'meta'], {'autoload': '(True)'}), "('resource', meta, autoload=True)\n", (749, 782)... |
import pytest
import unittest
from pydu.dict import AttrDict, LookupDict, CaseInsensitiveDict, OrderedDefaultDict, attrify
class TestAttrDict:
def test_attr_access_with_init(self):
d = AttrDict(key=1)
assert d['key'] == 1
assert d.key == 1
def test_attr_access_without_init(self):
... | [
"pydu.dict.LookupDict",
"pydu.dict.attrify",
"pytest.raises",
"pydu.dict.CaseInsensitiveDict",
"copy.deepcopy",
"pydu.dict.OrderedDefaultDict",
"pydu.dict.AttrDict"
] | [((2524, 2568), 'pydu.dict.attrify', 'attrify', (["{'a': [1, 2, {'b': 'b'}], 'c': 'c'}"], {}), "({'a': [1, 2, {'b': 'b'}], 'c': 'c'})\n", (2531, 2568), False, 'from pydu.dict import AttrDict, LookupDict, CaseInsensitiveDict, OrderedDefaultDict, attrify\n'), ((2703, 2718), 'pydu.dict.attrify', 'attrify', (['(1, 2)'], {}... |
# Import needed libraries
from stix2 import TAXIICollectionSource, Filter
from taxii2client.v20 import Server, Collection
import json
import re
from openpyxl import Workbook
from openpyxl.styles import Font
from openpyxl.styles import Alignment
#_________________________________________________________________________... | [
"openpyxl.styles.Font",
"taxii2client.v20.Collection",
"stix2.Filter",
"stix2.TAXIICollectionSource",
"openpyxl.Workbook",
"openpyxl.styles.Alignment"
] | [((410, 516), 'taxii2client.v20.Collection', 'Collection', (['"""https://cti-taxii.mitre.org/stix/collections/95ecc380-afe9-11e4-9b6c-751b66dd541e/"""'], {}), "(\n 'https://cti-taxii.mitre.org/stix/collections/95ecc380-afe9-11e4-9b6c-751b66dd541e/'\n )\n", (420, 516), False, 'from taxii2client.v20 import Server, ... |
from typing import Any, AsyncGenerator, Optional
from httpx import AsyncClient, Response
from aiochclient.exceptions import ChClientError
from aiochclient.http_clients.abc import HttpClientABC
class HttpxHttpClient(HttpClientABC):
def __init__(self, session: Optional[AsyncClient]):
if session:
... | [
"httpx.AsyncClient"
] | [((389, 402), 'httpx.AsyncClient', 'AsyncClient', ([], {}), '()\n', (400, 402), False, 'from httpx import AsyncClient, Response\n')] |
def options(ctx):
import optparse
grp = optparse.OptionGroup(ctx.parser, "architecture options")
grp.add_option("--m32", action="store_true", default=False,
help="compile & link in 32bits")
grp.add_option("--m64", action="store_true", default=False,
help="compile &... | [
"optparse.OptionGroup",
"sys.platform.lower",
"waflib.Errors.WafError"
] | [((49, 105), 'optparse.OptionGroup', 'optparse.OptionGroup', (['ctx.parser', '"""architecture options"""'], {}), "(ctx.parser, 'architecture options')\n", (69, 105), False, 'import optparse\n'), ((510, 564), 'waflib.Errors.WafError', 'Errors.WafError', (['"""You must choose either m32 of m64 !"""'], {}), "('You must ch... |
#!/usr/bin/env python
"""Download the YouTube video listed in a .webloc file.
The .webloc file should be specified as sys.argv[1].
The video will be downloaded to ../videos/by-youtube-id/<youtube_id>.<ext>
and will be symlinked as ../videos/by-channel/<channel>/<title>.<ext>.
"""
import errno
import os
import plist... | [
"subprocess.check_output",
"os.makedirs",
"subprocess.check_call",
"plistlib.readPlist",
"os.chdir",
"sys.stderr.write",
"os.path.isdir",
"os.unlink",
"sys.exit",
"sys.stdout.write"
] | [((446, 481), 'sys.stdout.write', 'sys.stdout.write', (["('Log: %s\\n' % msg)"], {}), "('Log: %s\\n' % msg)\n", (462, 481), False, 'import sys\n'), ((508, 545), 'sys.stderr.write', 'sys.stderr.write', (["('Error: %s\\n' % msg)"], {}), "('Error: %s\\n' % msg)\n", (524, 545), False, 'import sys\n'), ((2479, 2509), 'subpr... |
# Copyright 2009-2017 <NAME>.
# This program is distributed under the MIT license.
'''
Defines the `cache` decorator.
See its documentation for more details.
'''
# todo: examine thread-safety
import datetime as datetime_module
from python_toolbox import misc_tools
from python_toolbox import binary_search
from pytho... | [
"python_toolbox.misc_tools.set_attributes",
"python_toolbox.third_party.decorator.decorator",
"python_toolbox.sleek_reffing.SleekCallArgs",
"datetime.datetime.now",
"python_toolbox.nifty_collections.OrderedDict",
"datetime.timedelta"
] | [((755, 785), 'datetime.datetime.now', 'datetime_module.datetime.now', ([], {}), '()\n', (783, 785), True, 'import datetime as datetime_module\n'), ((6309, 6337), 'python_toolbox.third_party.decorator.decorator', 'decorator_', (['cached', 'function'], {}), '(cached, function)\n', (6319, 6337), True, 'from python_toolbo... |
# coding:utf-8
from flask import Flask, render_template
# 倒入蓝图
from order import app_orders
# 1.循环引用-解决方案1-推迟一方的加载
# from user import get_user
from goods import get_goods
app = Flask(__name__)
# 2.循环引用-解决方案2-使用装饰器
app.route('/get_goods')(get_goods)
# 3.在app程序中注册蓝图
app.register_blueprint(app_orders, url_prefix='/orde... | [
"flask.Flask"
] | [((178, 193), 'flask.Flask', 'Flask', (['__name__'], {}), '(__name__)\n', (183, 193), False, 'from flask import Flask, render_template\n')] |
"""empty message
Revision ID: 4603f5ddc9cb
Revises:
Create Date: 2019-01-15 12:24:13.352143
"""
from alembic import op
import sqlalchemy as sa
# revision identifiers, used by Alembic.
revision = '<KEY>'
down_revision = None
branch_labels = None
depends_on = None
def upgrade():
# ### commands auto generated b... | [
"sqlalchemy.DateTime",
"alembic.op.drop_table",
"sqlalchemy.Text",
"sqlalchemy.PrimaryKeyConstraint",
"sqlalchemy.Integer"
] | [((1048, 1070), 'alembic.op.drop_table', 'op.drop_table', (['"""tasks"""'], {}), "('tasks')\n", (1061, 1070), False, 'from alembic import op\n'), ((888, 917), 'sqlalchemy.PrimaryKeyConstraint', 'sa.PrimaryKeyConstraint', (['"""id"""'], {}), "('id')\n", (911, 917), True, 'import sqlalchemy as sa\n'), ((400, 412), 'sqlal... |
from mock import *
import unittest
from hamcrest import *
from service.smjvmservice import SmJvmService
from smcontext import ServiceManagerException
from smprocess import SmProcess
class JvmServiceMock(SmJvmService):
def __init__(self, context, service_name):
SmJvmService.__init__(self, context, servic... | [
"smcontext.ServiceManagerException",
"service.smjvmservice.SmJvmService.__init__"
] | [((277, 331), 'service.smjvmservice.SmJvmService.__init__', 'SmJvmService.__init__', (['self', 'context', 'service_name', '""""""'], {}), "(self, context, service_name, '')\n", (298, 331), False, 'from service.smjvmservice import SmJvmService\n'), ((686, 718), 'smcontext.ServiceManagerException', 'ServiceManagerExcepti... |
# Licensed to the .NET Foundation under one or more agreements.
# The .NET Foundation licenses this file to you under the MIT license.
# See the LICENSE file in the project root for more information.
"""A script to evaluate test values for special functions in high precision.
This scripts looks for .csv files in /test... | [
"csv.DictWriter",
"csv.DictReader",
"os.scandir",
"os.path.realpath",
"time.time"
] | [((13503, 13518), 'os.scandir', 'os.scandir', (['dir'], {}), '(dir)\n', (13513, 13518), False, 'import os\n'), ((13400, 13426), 'os.path.realpath', 'os.path.realpath', (['__file__'], {}), '(__file__)\n', (13416, 13426), False, 'import os\n'), ((13936, 13974), 'csv.DictReader', 'csv.DictReader', (['csvfile'], {'delimite... |
from django.db import models
from django.contrib.auth.models import BaseUserManager, AbstractBaseUser, PermissionsMixin
class UsuarioManager(BaseUserManager):
def create_user(self, email, password=None):
usuario = self.model(email=self.normalize_email(email))
usuario.is_active = True
usua... | [
"django.db.models.EmailField",
"django.db.models.BooleanField"
] | [((914, 993), 'django.db.models.EmailField', 'models.EmailField', ([], {'verbose_name': '"""Email do usuário"""', 'max_length': '(194)', 'unique': '(True)'}), "(verbose_name='Email do usuário', max_length=194, unique=True)\n", (931, 993), False, 'from django.db import models\n'), ((1010, 1078), 'django.db.models.Boolea... |
from matplotlib import pyplot as plt
import numpy as np
def generate_and_save_images(model, epoch, test_input):
# Notice `training` is set to False.
# This is so all layers run in inference mode (batchnorm).
predictions = model(test_input, training=False)
fig = plt.figure(figsize=(10,10))
for i in ra... | [
"matplotlib.pyplot.imshow",
"matplotlib.pyplot.axis",
"matplotlib.pyplot.close",
"numpy.max",
"matplotlib.pyplot.figure",
"matplotlib.pyplot.title",
"matplotlib.pyplot.ylim",
"matplotlib.pyplot.subplot",
"matplotlib.pyplot.legend"
] | [((276, 304), 'matplotlib.pyplot.figure', 'plt.figure', ([], {'figsize': '(10, 10)'}), '(figsize=(10, 10))\n', (286, 304), True, 'from matplotlib import pyplot as plt\n'), ((547, 558), 'matplotlib.pyplot.close', 'plt.close', ([], {}), '()\n', (556, 558), True, 'from matplotlib import pyplot as plt\n'), ((594, 621), 'ma... |
import numpy as np
import scipy.io as scio
import scipy.sparse as scsp
import h5py as hp
from util import read_mymat73, read_mymat, build_img_dataset, process_ad_dataset, mv_dataset, mv_tabular_collate, AverageMeter, save_roc_pr_curve_data, get_all_labels, \
load_print_results, filter_nan_grad, read_dataset, build_... | [
"models.encoder_decoder.mvae_ad",
"models.encoder_decoder.mvenc",
"torch.from_numpy",
"models.encoder_decoder.mvae_tf",
"torch.nn.MSELoss",
"models.encoder_decoder.mvae_ss",
"torch.cuda.is_available",
"torch.sum",
"util.process_ad_dataset",
"util.filter_nan_grad",
"models.DeepCCAModels.cca",
"... | [((1981, 2030), 'torch.set_default_tensor_type', 'torch.set_default_tensor_type', (['torch.DoubleTensor'], {}), '(torch.DoubleTensor)\n', (2010, 2030), False, 'import torch\n'), ((2818, 2827), 'torch.nn.MSELoss', 'MSELoss', ([], {}), '()\n', (2825, 2827), False, 'from torch.nn import MSELoss, CrossEntropyLoss\n'), ((81... |
from pathlib import Path
from lib_bgp_simulator import BaseGraphSystemTester
from lib_bgp_simulator import BGPSimpleAS
from lib_bgp_simulator import ROVSimpleAS
from lib_bgp_simulator import Graph013
from ..unstable import Unstable
from ....as_classes import ROVPPV1SimpleAS
from ....as_classes import ROVPPV2SimpleAS
... | [
"pathlib.Path"
] | [((616, 630), 'pathlib.Path', 'Path', (['__file__'], {}), '(__file__)\n', (620, 630), False, 'from pathlib import Path\n')] |
# coding: utf-8
#########################################################################
# Name:
#
# Calcurate equivalent potential temperature.
#
# Usage:
# example:
#
# Author: <NAME>
# Date: 2021/08/13
#########################################################################
import argparse
#from datetime import da... | [
"os.listdir",
"argparse.ArgumentParser",
"netCDF4.Dataset",
"os.path.join",
"math.log",
"numpy.array",
"os.path.abspath",
"re.search"
] | [((498, 523), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {}), '()\n', (521, 523), False, 'import argparse\n'), ((1132, 1147), 'netCDF4.Dataset', 'Dataset', (['ncfile'], {}), '(ncfile)\n', (1139, 1147), False, 'from netCDF4 import Dataset\n'), ((1419, 1462), 'numpy.array', 'np.array', (['[i for i in self... |
"""
Servo Horns
"""
import cadquery as cq
import cqparts
from cqparts.params import *
from cqparts.display import render_props, display
from cqparts.constraint import Fixed, Coincident
from cqparts.constraint import Mate
from cqparts.utils.geometry import CoordSystem
from cqparts.search import register
from .multi im... | [
"cadquery.Workplane",
"cqparts.search.register",
"cqparts.display.display",
"cqparts.utils.geometry.CoordSystem"
] | [((2960, 2984), 'cqparts.search.register', 'register', ([], {'export': '"""horns"""'}), "(export='horns')\n", (2968, 2984), False, 'from cqparts.search import register\n'), ((3403, 3427), 'cqparts.search.register', 'register', ([], {'export': '"""horns"""'}), "(export='horns')\n", (3411, 3427), False, 'from cqparts.sea... |
#!/usr/bin/env python3
# -*- coding: utf-8 -*-
#
# This file is part of daily-wallpaper
#
# Copyright (c) 2017 <NAME> <a.k.a. atareao>
#
# Permission is hereby granted, free of charge, to any person obtaining a copy
# of this software and associated documentation files (the "Software"), to deal
# in the Software withou... | [
"sys.path.insert",
"comun._",
"lxml.html.fromstring",
"requests.get",
"os.getcwd",
"os.path.abspath",
"daily.Daily.__init__"
] | [((1361, 1409), 'sys.path.insert', 'sys.path.insert', (['(1)', '"""/usr/share/daily-wallpaper"""'], {}), "(1, '/usr/share/daily-wallpaper')\n", (1376, 1409), False, 'import sys\n'), ((1724, 1741), 'comun._', '_', (['"""Wallpapering"""'], {}), "('Wallpapering')\n", (1725, 1741), False, 'from comun import _\n'), ((1775, ... |
# -*- coding: utf-8 -*-
"""
sphinxjp.themes.revealjs.directives
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
:author: tell-k <<EMAIL>>
:copyright: tell-k. All Rights Reserved.
"""
from docutils import nodes
from docutils.parsers.rst import directives
from docutils.parsers.rst.roles import se... | [
"docutils.parsers.rst.directives.choice",
"docutils.parsers.rst.roles.set_classes",
"sphinxjp.themes.revealjs.compat.text"
] | [((790, 855), 'docutils.parsers.rst.directives.choice', 'directives.choice', (['argument', "('h1', 'h2', 'h3', 'h4', 'h5', 'h6')"], {}), "(argument, ('h1', 'h2', 'h3', 'h4', 'h5', 'h6'))\n", (807, 855), False, 'from docutils.parsers.rst import directives\n'), ((1939, 1964), 'docutils.parsers.rst.roles.set_classes', 'se... |
import unittest
import pytest
import paramak
class TestExtrudeHollowRectangle(unittest.TestCase):
def setUp(self):
self.test_shape = paramak.ExtrudeHollowRectangle(
height=10, width=15, casing_thickness=1, distance=2
)
def test_default_parameters(self):
"""Checks that th... | [
"pytest.approx",
"paramak.ExtrudeHollowRectangle"
] | [((149, 236), 'paramak.ExtrudeHollowRectangle', 'paramak.ExtrudeHollowRectangle', ([], {'height': '(10)', 'width': '(15)', 'casing_thickness': '(1)', 'distance': '(2)'}), '(height=10, width=15, casing_thickness=1,\n distance=2)\n', (179, 236), False, 'import paramak\n'), ((2146, 2186), 'pytest.approx', 'pytest.appro... |
import djclick as click
from azure.common import AzureMissingResourceHttpError
from requests.exceptions import HTTPError
from boundlexx.api.tasks import purge_static_cache
from boundlexx.boundless.models import Emoji
from boundlexx.utils import download_image, get_django_image_from_file, make_thumbnail
@click.comman... | [
"djclick.command",
"boundlexx.api.tasks.purge_static_cache",
"boundlexx.utils.get_django_image_from_file",
"boundlexx.utils.make_thumbnail",
"djclick.echo",
"boundlexx.boundless.models.Emoji.objects.filter",
"boundlexx.utils.download_image"
] | [((308, 323), 'djclick.command', 'click.command', ([], {}), '()\n', (321, 323), True, 'import djclick as click\n'), ((343, 389), 'djclick.echo', 'click.echo', (['"""Adding thumbs/renmaing images..."""'], {}), "('Adding thumbs/renmaing images...')\n", (353, 389), True, 'import djclick as click\n'), ((423, 464), 'boundle... |
from tortoise import Model
from tortoise import fields
from werkzeug.security import generate_password_hash, check_password_hash
class User(Model):
class Meta:
table = 'users'
id = fields.IntField(pk=True)
username = fields.CharField(max_length=100, unique=True)
email = fields.CharField(max_l... | [
"tortoise.fields.CharField",
"werkzeug.security.generate_password_hash",
"tortoise.fields.TextField",
"tortoise.fields.IntField"
] | [((200, 224), 'tortoise.fields.IntField', 'fields.IntField', ([], {'pk': '(True)'}), '(pk=True)\n', (215, 224), False, 'from tortoise import fields\n'), ((240, 285), 'tortoise.fields.CharField', 'fields.CharField', ([], {'max_length': '(100)', 'unique': '(True)'}), '(max_length=100, unique=True)\n', (256, 285), False, ... |
# coding=utf-8
# Copyright 2014 Pants project contributors (see CONTRIBUTORS.md).
# Licensed under the Apache License, Version 2.0 (see LICENSE).
from __future__ import (absolute_import, division, generators, nested_scopes, print_function,
unicode_literals, with_statement)
import itertools
imp... | [
"re.compile",
"os.pathsep.join",
"pants.backend.python.python_requirement.PythonRequirement",
"pants.base.exceptions.TestFailedTaskError",
"pants.util.contextutil.temporary_dir",
"textwrap.dedent",
"os.path.exists",
"shutil.move",
"six.StringIO",
"os.path.isabs",
"pex.pex_info.PexInfo.default",
... | [((1404, 1425), 'logging.basicConfig', 'logging.basicConfig', ([], {}), '()\n', (1423, 1425), False, 'import logging\n'), ((9040, 9209), 'textwrap.dedent', 'dedent', (["b'\\n [run]\\n branch = True\\n timid = True\\n\\n [report]\\n exclude_lines =\\n def __repr__\\n raise NotImplementedErro... |
from __future__ import print_function
from __future__ import absolute_import
from __future__ import division
from numpy import asarray
from scipy.spatial import Voronoi
from scipy.spatial import Delaunay
__all__ = [
'delaunay_from_points_numpy',
'voronoi_from_points_numpy',
]
def delaunay_from_points_numpy... | [
"numpy.asarray",
"scipy.spatial.Voronoi",
"scipy.spatial.Delaunay"
] | [((955, 970), 'numpy.asarray', 'asarray', (['points'], {}), '(points)\n', (962, 970), False, 'from numpy import asarray\n'), ((979, 1000), 'scipy.spatial.Delaunay', 'Delaunay', (['xyz[:, 0:2]'], {}), '(xyz[:, 0:2])\n', (987, 1000), False, 'from scipy.spatial import Delaunay\n'), ((1315, 1330), 'numpy.asarray', 'asarray... |
# $Id: gp_unix.py,v 2.6 2003/04/21 09:44:09 mhagger Exp $
# Copyright (C) 1998-2003 <NAME> <<EMAIL>>
#
# This file is licensed under the GNU Lesser General Public License
# (LGPL). See LICENSE.txt for details.
"""gp_unix -- an interface to gnuplot used for unix platforms.
This file implements a low-level interface ... | [
"os.popen",
"string.find"
] | [((5706, 5773), 'os.popen', 'popen', (["('echo | %s -persist 2>&1' % GnuplotOpts.gnuplot_command)", '"""r"""'], {}), "('echo | %s -persist 2>&1' % GnuplotOpts.gnuplot_command, 'r')\n", (5711, 5773), False, 'from os import popen\n'), ((7686, 7741), 'os.popen', 'popen', (["('%s -persist' % GnuplotOpts.gnuplot_command)", ... |
#!/usr/bin/python
"""
This script will scrape the r-project.org machine learning selection and
format the packages in github markdown style for this
awesome-machine-learning repo.
"""
from pyquery import PyQuery as pq
import urllib
import codecs
text_file = codecs.open("Packages.txt", encoding='utf-8', ... | [
"codecs.open",
"urllib.urlopen"
] | [((274, 329), 'codecs.open', 'codecs.open', (['"""Packages.txt"""'], {'encoding': '"""utf-8"""', 'mode': '"""w"""'}), "('Packages.txt', encoding='utf-8', mode='w')\n", (285, 329), False, 'import codecs\n'), ((433, 452), 'urllib.urlopen', 'urllib.urlopen', (['url'], {}), '(url)\n', (447, 452), False, 'import urllib\n'),... |
import torch
from videoanalyst.config.config import cfg as root_cfg
from videoanalyst.config.config import specify_task
from videoanalyst.model import builder as model_builder
from videoanalyst.pipeline import builder as pipeline_builder
from imutils.video import VideoStream
from imutils.video import FPS
import argpa... | [
"cv2.rectangle",
"cv2.__version__.split",
"xml.etree.ElementTree.parse",
"videoanalyst.config.config.cfg.merge_from_file",
"argparse.ArgumentParser",
"torch.device",
"imutils.video.VideoStream",
"time.sleep",
"cv2.imshow",
"videoanalyst.config.config.specify_task",
"imutils.resize",
"cv2.putTe... | [((447, 519), 'videoanalyst.config.config.cfg.merge_from_file', 'root_cfg.merge_from_file', (['"""./experiments/siamfcpp/siamfcpp_alexnet.yaml"""'], {}), "('./experiments/siamfcpp/siamfcpp_alexnet.yaml')\n", (471, 519), True, 'from videoanalyst.config.config import cfg as root_cfg\n'), ((555, 577), 'videoanalyst.config... |
import sys
from ..api import plot
import fuc
import pysam
description = f"""
Plot allele fraction profile from VcfFrame[Imported].
"""
def create_parser(subparsers):
parser = fuc.api.common._add_parser(
subparsers,
fuc.api.common._script_name(),
description=description,
help='Plo... | [
"fuc.api.common._script_name"
] | [((239, 268), 'fuc.api.common._script_name', 'fuc.api.common._script_name', ([], {}), '()\n', (266, 268), False, 'import fuc\n')] |
"""
File: test.py
By: <NAME>, <EMAIL>
Description:
Test the net.
"""
from argmaxnet import ArgMaxNet
import os
#=================================================================
modelPath = os.path.expanduser('~/models/argmaxnet/ArgMaxNet : StackDepth=3, Block(10,60),[111][0.02220].model')
#=======================... | [
"argmaxnet.ArgMaxNet",
"os.path.expanduser"
] | [((195, 305), 'os.path.expanduser', 'os.path.expanduser', (['"""~/models/argmaxnet/ArgMaxNet : StackDepth=3, Block(10,60),[111][0.02220].model"""'], {}), "(\n '~/models/argmaxnet/ArgMaxNet : StackDepth=3, Block(10,60),[111][0.02220].model'\n )\n", (213, 305), False, 'import os\n'), ((403, 423), 'argmaxnet.ArgMaxN... |
from ploomber.scaffold.scaffoldloader import ScaffoldLoader
from ploomber.util.util import add_to_sys_path
from ploomber.util import loader
from ploomber.exceptions import DAGSpecInvalidError
def load_dag():
# setting lazy_import to true causes sources to be returned as paths,
# instead of placeholders
tr... | [
"ploomber.util.loader.create",
"ploomber.util.loader._default_spec_load",
"ploomber.util.util.add_to_sys_path",
"ploomber.scaffold.scaffoldloader.ScaffoldLoader"
] | [((556, 586), 'ploomber.scaffold.scaffoldloader.ScaffoldLoader', 'ScaffoldLoader', (['"""ploomber_add"""'], {}), "('ploomber_add')\n", (570, 586), False, 'from ploomber.scaffold.scaffoldloader import ScaffoldLoader\n'), ((338, 383), 'ploomber.util.loader._default_spec_load', 'loader._default_spec_load', ([], {'lazy_imp... |
# -*- encoding: utf-8 -*-
from datetime import datetime
from hashlib import md5
from logging import getLogger
from redis import ConnectionPool, Redis
from smsgateway.enums import DIRECTION_OUTBOUND
from smsgateway.models import SMS
from smsgateway.backends.base import SMSBackend
from smsgateway.sms import SMSRequest
... | [
"logging.getLogger",
"smsgateway.models.SMS.objects.create",
"redis.ConnectionPool",
"smsgateway.sms.SMSRequest",
"redis.Redis",
"datetime.datetime.now"
] | [((330, 349), 'logging.getLogger', 'getLogger', (['__name__'], {}), '(__name__)\n', (339, 349), False, 'from logging import getLogger\n'), ((1363, 1465), 'redis.ConnectionPool', 'ConnectionPool', ([], {'host': 'host', 'port': 'port', 'db': "account_dict['dbn']", 'password': "account_dict['<PASSWORD>']"}), "(host=host, ... |
import sys
import pprint
pp = pprint.PrettyPrinter();
import pymongo
from pymongo import MongoClient
#!/usr/bin/env python3
# -*- coding: utf-8 -*-
"""
Created on Wed Oct 12 20:30:54 2016
@author: ryanlim, jpitts
Requirements:
- pymongo needs to be installed
- mongodb needs to be running
- brigade-match... | [
"pymongo.MongoClient",
"pprint.PrettyPrinter",
"operator.itemgetter"
] | [((30, 52), 'pprint.PrettyPrinter', 'pprint.PrettyPrinter', ([], {}), '()\n', (50, 52), False, 'import pprint\n'), ((1331, 1362), 'pymongo.MongoClient', 'MongoClient', (['"""localhost"""', '(27017)'], {}), "('localhost', 27017)\n", (1342, 1362), False, 'from pymongo import MongoClient\n'), ((10221, 10245), 'operator.it... |
"""This module abstracts logging configuration"""
# ---------------------------------------------
# System modules
# ---------------------------------------------
import logging
# ---------------------------------------------
# External dependencies
# ---------------------------------------------
# -------------------... | [
"logging.basicConfig",
"logging.getLogger",
"coloredlogs.install"
] | [((486, 525), 'logging.basicConfig', 'logging.basicConfig', ([], {'level': 'logging.INFO'}), '(level=logging.INFO)\n', (505, 525), False, 'import logging\n'), ((593, 620), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (610, 620), False, 'import logging\n'), ((621, 669), 'coloredlogs.inst... |
from swarmform import SwarmPad
from fireworks import Firework, ScriptTask, FWorker
from fireworks.core.rocket_launcher import launch_rocket
if __name__ == "__main__":
# set up the SwarmPad and reset it
swarmpad = SwarmPad()
swarmpad.reset('', require_password=False)
# create the Firework consisting of a custom "A... | [
"fireworks.FWorker",
"swarmform.SwarmPad",
"fireworks.ScriptTask.from_str"
] | [((216, 226), 'swarmform.SwarmPad', 'SwarmPad', ([], {}), '()\n', (224, 226), False, 'from swarmform import SwarmPad\n'), ((355, 390), 'fireworks.ScriptTask.from_str', 'ScriptTask.from_str', (['"""echo "hello\\""""'], {}), '(\'echo "hello"\')\n', (374, 390), False, 'from fireworks import Firework, ScriptTask, FWorker\n... |
"""Script for training the NormalizedEightPointNet.
Example:
$ python train.py
to see help:
$ python train.py -h
"""
import argparse
import time
import torch
import torch.optim as optim
from dfe.datasets import ColmapDataset
from dfe.models import NormalizedEightPointNet
import dfe.models.loss as L
... | [
"torch.utils.data.ConcatDataset",
"argparse.ArgumentParser",
"torch.utils.data.DataLoader",
"dfe.datasets.ColmapDataset",
"torch.optim.lr_scheduler.StepLR",
"torch.cuda.is_available",
"dfe.models.loss.symmetric_epipolar_distance",
"dfe.models.NormalizedEightPointNet",
"time.time"
] | [((844, 885), 'torch.utils.data.ConcatDataset', 'torch.utils.data.ConcatDataset', (['data_sets'], {}), '(data_sets)\n', (874, 885), False, 'import torch\n'), ((968, 1117), 'torch.utils.data.DataLoader', 'torch.utils.data.DataLoader', (['dset'], {'batch_size': 'options.batch_size', 'shuffle': '(True)', 'num_workers': 'o... |
from simple_playgrounds.playground.playgrounds import *
from simple_playgrounds.engine import Engine
from simple_playgrounds.agent.controllers import Keyboard
from simple_playgrounds.agent.agents import HeadAgent
import time
import cv2
my_agent = HeadAgent(controller=Keyboard(), lateral=True, interactive=True)
####... | [
"simple_playgrounds.agent.controllers.Keyboard",
"cv2.waitKey",
"simple_playgrounds.engine.Engine",
"time.sleep"
] | [((519, 553), 'simple_playgrounds.engine.Engine', 'Engine', ([], {'playground': 'pg', 'debug': '(False)'}), '(playground=pg, debug=False)\n', (525, 553), False, 'from simple_playgrounds.engine import Engine\n'), ((271, 281), 'simple_playgrounds.agent.controllers.Keyboard', 'Keyboard', ([], {}), '()\n', (279, 281), Fals... |
"""
Copyright 2020 The OneFlow Authors. All rights reserved.
Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agr... | [
"collections.OrderedDict",
"oneflow.experimental.unittest.skip_unless_1n1d",
"numpy.array",
"test_util.GenArgList",
"unittest.main",
"oneflow.experimental.device"
] | [((2909, 2941), 'oneflow.experimental.unittest.skip_unless_1n1d', 'flow.unittest.skip_unless_1n1d', ([], {}), '()\n', (2939, 2941), True, 'import oneflow.experimental as flow\n'), ((786, 829), 'numpy.array', 'np.array', (['[[1, 2, 3], [4, 5, 6], [7, 8, 9]]'], {}), '([[1, 2, 3], [4, 5, 6], [7, 8, 9]])\n', (794, 829), Tr... |
from builtins import *
import argparse
import numpy as np
import os
from bnpy.ioutil.DataReader import loadDataFromSavedTask, loadLPKwargsFromDisk
from bnpy.ioutil.DataReader import loadKwargsFromDisk
from bnpy.ioutil.ModelReader import loadModelForLap
from bnpy.util import StateSeqUtil
from bnpy.birthmove.BCreateOneP... | [
"bnpy.ioutil.ModelReader.loadModelForLap",
"bnpy.ioutil.DataReader.loadDataFromSavedTask",
"numpy.unique",
"argparse.ArgumentParser",
"bnpy.ioutil.DataReader.loadKwargsFromDisk",
"bnpy.util.StateSeqUtil.alignEstimatedStateSeqToTruth",
"bnpy.birthmove.BLogger.configure",
"numpy.argmax",
"numpy.sum",
... | [((1424, 1461), 'bnpy.ioutil.ModelReader.loadModelForLap', 'loadModelForLap', (['taskoutpath', 'lapFrac'], {}), '(taskoutpath, lapFrac)\n', (1439, 1461), False, 'from bnpy.ioutil.ModelReader import loadModelForLap\n'), ((1473, 1524), 'bnpy.ioutil.DataReader.loadDataFromSavedTask', 'loadDataFromSavedTask', (['taskoutpat... |
"""
To be placed into {BlenderProcRoot}/src/loader/.
"""
import bpy
from src.loader.LoaderInterface import LoaderInterface
from src.utility.Utility import Utility
from src.utility.LabelIdMapping import LabelIdMapping
class CustomObjectLoader(LoaderInterface):
"""
Custom object loader which, in addition to t... | [
"bpy.ops.object.select_all",
"src.loader.LoaderInterface.LoaderInterface.__init__",
"src.utility.Utility.Utility.import_objects",
"src.utility.Utility.Utility.resolve_path",
"src.utility.Utility.Utility.get_nodes_with_type",
"src.utility.Utility.Utility.get_the_one_node_with_type",
"src.loader.LoaderInt... | [((589, 627), 'src.loader.LoaderInterface.LoaderInterface.__init__', 'LoaderInterface.__init__', (['self', 'config'], {}), '(self, config)\n', (613, 627), False, 'from src.loader.LoaderInterface import LoaderInterface\n'), ((2199, 2249), 'src.loader.LoaderInterface.LoaderInterface.remove_x_axis_rotation', 'LoaderInterf... |
from django.contrib import admin
from .models import Quiz, Question, Response
# Register your models here.
class InLineResponse(admin.StackedInline):
model = Response
extra = 0
class InLineQuestion(admin.StackedInline):
model = Question
extra = 0
class QuizAdmin(admin.ModelAdmin):
inlines = [... | [
"django.contrib.admin.site.register"
] | [((428, 464), 'django.contrib.admin.site.register', 'admin.site.register', (['Quiz', 'QuizAdmin'], {}), '(Quiz, QuizAdmin)\n', (447, 464), False, 'from django.contrib import admin\n'), ((465, 509), 'django.contrib.admin.site.register', 'admin.site.register', (['Question', 'QuestionAdmin'], {}), '(Question, QuestionAdmi... |
import torch
import torch.autograd as autograd
from torch.autograd import Variable
import torch.nn as nn
import torch.nn.functional as F
import torch.optim as optim
from torch.nn.parameter import Parameter
import numpy as np
import datetime
from transformers import EncoderDecoderModel, BertTokenizer
from Hyperparameter... | [
"transformers.BertTokenizer.from_pretrained",
"transformers.EncoderDecoderModel.from_encoder_decoder_pretrained"
] | [((501, 551), 'transformers.BertTokenizer.from_pretrained', 'BertTokenizer.from_pretrained', (['"""bert-base-uncased"""'], {}), "('bert-base-uncased')\n", (530, 551), False, 'from transformers import EncoderDecoderModel, BertTokenizer\n'), ((573, 670), 'transformers.EncoderDecoderModel.from_encoder_decoder_pretrained',... |
import doctest
def test_doctests():
doctest.testfile("../README.rst")
if __name__ == "__main__":
test_doctests()
| [
"doctest.testfile"
] | [((42, 75), 'doctest.testfile', 'doctest.testfile', (['"""../README.rst"""'], {}), "('../README.rst')\n", (58, 75), False, 'import doctest\n')] |
from __future__ import unicode_literals
import re
from .common import InfoExtractor
from .gigya import GigyaBaseIE
from ..compat import compat_HTTPError
from ..utils import (
ExtractorError,
clean_html,
extract_attributes,
float_or_none,
get_element_by_class,
int_or_none,
merge_dicts,
... | [
"re.match"
] | [((1619, 1649), 're.match', 're.match', (['self._VALID_URL', 'url'], {}), '(self._VALID_URL, url)\n', (1627, 1649), False, 'import re\n'), ((7761, 7791), 're.match', 're.match', (['self._VALID_URL', 'url'], {}), '(self._VALID_URL, url)\n', (7769, 7791), False, 'import re\n')] |
import setting
set_dic = setting.setting()
if set_dic["MODE"] == "WhereFrom":
import logging
def f1():
if False:
logging.info("debug log!!")
def f2():
if True:
logging.info("debug log!!")
if set_dic["MODE"] == "FromHere":
import logging
logger=logging.getLo... | [
"logging.getLogger",
"logging.Formatter",
"logging.handlers.RotatingFileHandler",
"logging.info",
"setting.setting"
] | [((26, 43), 'setting.setting', 'setting.setting', ([], {}), '()\n', (41, 43), False, 'import setting\n'), ((307, 334), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (324, 334), False, 'import logging\n'), ((660, 687), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name... |
# Copyright (C) 2019-2021, TomTom (http://tomtom.com).
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law o... | [
"logging.getLogger",
"pathlib.Path",
"xml.etree.ElementTree.tostring",
"xml.etree.ElementTree.Element",
"re.sub",
"xml.etree.ElementTree.SubElement"
] | [((864, 891), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (881, 891), False, 'import logging\n'), ((5084, 5137), 'xml.etree.ElementTree.Element', 'ET.Element', (['"""div"""'], {'id': '"""toc"""', 'attrib': "{'class': 'toc2'}"}), "('div', id='toc', attrib={'class': 'toc2'})\n", (5094, 5... |
import requests
import json
import bs4 as bs
import pandas as pd
your_head={
'GET':"wss: '// chat - ws.shopee.co.id / socket.io /?EIO = 3 & transport = websocket HTTP / 1.1",
'connection': 'Upgrade',
'pragma': 'no - cache',
'cache - Control': 'no - cache',
'user - Agent': 'Mozilla / 5.0(Linux;Android 6.0;Nexus 5 Build... | [
"pandas.DataFrame",
"json.loads",
"requests.get"
] | [((2029, 2045), 'json.loads', 'json.loads', (['data'], {}), '(data)\n', (2039, 2045), False, 'import json\n'), ((2770, 2793), 'pandas.DataFrame', 'pd.DataFrame', (['prod_data'], {}), '(prod_data)\n', (2782, 2793), True, 'import pandas as pd\n'), ((1993, 2017), 'requests.get', 'requests.get', (['shopee_url'], {}), '(sho... |
import numpy as np
import random
from FuncionAptitud import fitness
lista = [0, 1, 2, 3, 4, 5, 6, 7] # son los valores en los que puede estar la reyna
poblacion = np.empty((50,8))
for i in range(50):
random.shuffle(lista)
for j in range(8):
poblacion[i, j] = lista[j]
def padres(conjunto):
r1 = ra... | [
"FuncionAptitud.fitness",
"random.random",
"numpy.empty",
"random.shuffle"
] | [((165, 182), 'numpy.empty', 'np.empty', (['(50, 8)'], {}), '((50, 8))\n', (173, 182), True, 'import numpy as np\n'), ((206, 227), 'random.shuffle', 'random.shuffle', (['lista'], {}), '(lista)\n', (220, 227), False, 'import random\n'), ((318, 333), 'random.random', 'random.random', ([], {}), '()\n', (331, 333), False, ... |
# Generated by Django 3.2.11 on 2022-02-02 07:10
from django.db import migrations, models
class Migration(migrations.Migration):
dependencies = [
('engine', '0048_auto_20211112_1918'),
]
operations = [
migrations.AlterField(
model_name='labeledshape',
name='type'... | [
"django.db.models.CharField"
] | [((340, 536), 'django.db.models.CharField', 'models.CharField', ([], {'choices': "[('rectangle', 'RECTANGLE'), ('polygon', 'POLYGON'), ('polyline',\n 'POLYLINE'), ('points', 'POINTS'), ('ellipse', 'ELLIPSE'), ('cuboid',\n 'CUBOID')]", 'max_length': '(16)'}), "(choices=[('rectangle', 'RECTANGLE'), ('polygon', 'POL... |
import torch
import torch.nn
import torch.nn.functional as F
import warnings
from .types import DataScoreModel, ScoreMatchingLoss, ConditionalDataScoreModel
from . import jacobians
def _ism(tr_jac: torch.Tensor, scores: torch.Tensor) -> torch.Tensor:
return (tr_jac + 0.5 * (scores ** 2).sum(-1)).mean()
def ism... | [
"warnings.warn",
"torch.randint",
"torch.randn_like"
] | [((1568, 1587), 'torch.randn_like', 'torch.randn_like', (['x'], {}), '(x)\n', (1584, 1587), False, 'import torch\n'), ((1885, 1938), 'torch.randint', 'torch.randint', (['(0)', 'sigmae.shape[0]'], {'size': '(x.shape[0],)'}), '(0, sigmae.shape[0], size=(x.shape[0],))\n', (1898, 1938), False, 'import torch\n'), ((2027, 20... |
from __future__ import absolute_import
from __future__ import print_function
from pysnptools.util.mapreduce1.runner import *
import logging
import fastlmm.pyplink.plink as plink
import pysnptools.util as pstutil
import pysnptools.util.pheno as pstpheno
import numpy as np
from fastlmm.inference import LMM
import scipy.s... | [
"numpy.sqrt",
"pysnptools.util.create_directory_if_necessary",
"numpy.hstack",
"numpy.array",
"logging.info",
"pysnptools.util.pheno.loadOnePhen",
"numpy.arange",
"pysnptools.util.pheno.loadPhen",
"numpy.savez",
"numpy.exp",
"pysnptools.util.intersect_apply",
"doctest.testmod",
"numpy.ones",... | [((20159, 20170), 'time.time', 'time.time', ([], {}), '()\n', (20168, 20170), False, 'import time\n'), ((24987, 25004), 'doctest.testmod', 'doctest.testmod', ([], {}), '()\n', (25002, 25004), False, 'import doctest\n'), ((8978, 9073), 'pysnptools.util.intersect_apply', 'pstutil.intersect_apply', (['[self.test_snps, sel... |
from graphene.utils.str_converters import to_snake_case
def localized_field_resolver(obj, info, **kwargs):
"""
Custom resolver to return the user language value from localized fields
"""
attr = getattr(obj, to_snake_case(info.field_name))
if attr is None:
return None
if info.context.... | [
"graphene.utils.str_converters.to_snake_case"
] | [((225, 255), 'graphene.utils.str_converters.to_snake_case', 'to_snake_case', (['info.field_name'], {}), '(info.field_name)\n', (238, 255), False, 'from graphene.utils.str_converters import to_snake_case\n')] |
#!/usr/bin/env python3
from sqlalchemy.ext.declarative import declarative_base
from sqlalchemy import Column, String, DateTime
from sqlalchemy.sql import func
from sqlalchemy.dialects.postgresql import JSONB
Base = declarative_base()
class JobResult(Base):
__tablename__ = 'job_result'
job_id = Column(Stri... | [
"sqlalchemy.String",
"sqlalchemy.sql.func.now",
"sqlalchemy.Column",
"sqlalchemy.ext.declarative.declarative_base"
] | [((218, 236), 'sqlalchemy.ext.declarative.declarative_base', 'declarative_base', ([], {}), '()\n', (234, 236), False, 'from sqlalchemy.ext.declarative import declarative_base\n'), ((466, 480), 'sqlalchemy.Column', 'Column', (['String'], {}), '(String)\n', (472, 480), False, 'from sqlalchemy import Column, String, DateT... |
"""
Module containg all the necessary class for the package.
FoodCategory is a Enum class representing the category of ingredient.
CategorySynset is a class representing the wordnet synset of a category.
"""
from enum import Enum
from nltk.corpus import wordnet
class FoodCategory(Enum):
"""
Enum class used ... | [
"nltk.corpus.wordnet.synset"
] | [((829, 882), 'nltk.corpus.wordnet.synset', 'wordnet.synset', (['f"""{FoodCategory.vegetable.name}.n.01"""'], {}), "(f'{FoodCategory.vegetable.name}.n.01')\n", (843, 882), False, 'from nltk.corpus import wordnet\n'), ((892, 941), 'nltk.corpus.wordnet.synset', 'wordnet.synset', (['f"""{FoodCategory.fruit.name}.n.01"""']... |
import os
from nxpy.nxfile import NXFile
def test_nxnode_resolve():
node = NXFile(os.path.join(os.path.dirname(__file__), 'map.nx')).get_root_node().resolve(
"Tile/grassySoil.img/bsc/0")
assert node.width == 90
node2 = NXFile(os.path.join(os.path.dirname(__file__), 'map.nx')).get_root_node().g... | [
"os.path.dirname"
] | [((102, 127), 'os.path.dirname', 'os.path.dirname', (['__file__'], {}), '(__file__)\n', (117, 127), False, 'import os\n'), ((264, 289), 'os.path.dirname', 'os.path.dirname', (['__file__'], {}), '(__file__)\n', (279, 289), False, 'import os\n'), ((440, 465), 'os.path.dirname', 'os.path.dirname', (['__file__'], {}), '(__... |
from django.db import models
from django.contrib.auth import get_user_model
class Category(models.Model):
title = models.CharField(max_length=255)
def __str__(self):
if len(str(self.title)) > 20:
return self.title[:20]
return self.title
class Post(models.Model):
title = mode... | [
"django.contrib.auth.get_user_model",
"django.db.models.TextField",
"django.db.models.ManyToManyField",
"django.db.models.CharField"
] | [((120, 152), 'django.db.models.CharField', 'models.CharField', ([], {'max_length': '(255)'}), '(max_length=255)\n', (136, 152), False, 'from django.db import models\n'), ((316, 348), 'django.db.models.CharField', 'models.CharField', ([], {'max_length': '(255)'}), '(max_length=255)\n', (332, 348), False, 'from django.d... |
from generator import Generator
from discriminator import Discriminator
from utils import *
import torch
from torch import nn
from tqdm.auto import tqdm
from torchvision import transforms
from torchvision.datasets import MNIST
from torch.utils.data import DataLoader
import argparse
parser = argparse.ArgumentParser(
... | [
"generator.Generator",
"torch.ones_like",
"argparse.ArgumentParser",
"discriminator.Discriminator",
"torch.zeros_like",
"torchvision.transforms.Normalize",
"tqdm.auto.tqdm",
"torch.nn.BCEWithLogitsLoss",
"torchvision.datasets.MNIST",
"torchvision.transforms.ToTensor"
] | [((293, 345), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {'description': '"""DCGAN trainer"""'}), "(description='DCGAN trainer')\n", (316, 345), False, 'import argparse\n'), ((707, 718), 'generator.Generator', 'Generator', ([], {}), '()\n', (716, 718), False, 'from generator import Generator\n'), ((770,... |
from dataset import CovidImageDataset
from argparse import ArgumentParser
import torch
import torch.nn as nn
from model import VGG
import numpy as np
import os
from pytorch_lightning.utilities.seed import seed_everything
import random
def seed_worker(worker_id):
'''
https://pytorch.org/docs/stable/notes/rando... | [
"torch.nn.CrossEntropyLoss",
"argparse.ArgumentParser",
"torch.utils.data.DataLoader",
"torch.optim.lr_scheduler.CosineAnnealingLR",
"torch.initial_seed",
"os.makedirs",
"torch.max",
"os.path.join",
"random.seed",
"model.VGG",
"torch.cuda.is_available",
"numpy.random.seed",
"torch.use_determ... | [((609, 636), 'numpy.random.seed', 'np.random.seed', (['worker_seed'], {}), '(worker_seed)\n', (623, 636), True, 'import numpy as np\n'), ((641, 665), 'random.seed', 'random.seed', (['worker_seed'], {}), '(worker_seed)\n', (652, 665), False, 'import random\n'), ((2124, 2140), 'argparse.ArgumentParser', 'ArgumentParser'... |
# Copyright 2018 Google Inc.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, ... | [
"unittest.main",
"unittest.mock.MagicMock",
"unittest.mock.patch"
] | [((1260, 1297), 'unittest.mock.patch', 'mock.patch', (['SNIPPET_CLIENT_CLASS_PATH'], {}), '(SNIPPET_CLIENT_CLASS_PATH)\n', (1270, 1297), False, 'from unittest import mock\n'), ((1607, 1644), 'unittest.mock.patch', 'mock.patch', (['SNIPPET_CLIENT_CLASS_PATH'], {}), '(SNIPPET_CLIENT_CLASS_PATH)\n', (1617, 1644), False, '... |
#!/usr/bin/env python3
"""
Script to start a client connection to server. (Talker)
"""
import socket
import json
import datetime
import copy
import hashlib
import math
import logging
from subprocess import check_call
from Crypto.PublicKey import RSA
from Crypto.Cipher import PKCS1_OAEP
from Crypto.Random import get... | [
"json.loads",
"socket.socket",
"subprocess.check_call",
"encryptlib.SimonCTR.countermode_encrypt",
"Crypto.Random.get_random_bytes",
"encryptlib.print_helper.PrintHelper",
"json.dumps",
"encryptlib.SimonCTR.countermode_decrypt",
"logging.info",
"datetime.datetime.now",
"copy.copy",
"encryptlib... | [((1003, 1016), 'encryptlib.print_helper.PrintHelper', 'PrintHelper', ([], {}), '()\n', (1014, 1016), False, 'from encryptlib.print_helper import PrintHelper\n'), ((1178, 1227), 'socket.socket', 'socket.socket', (['socket.AF_INET', 'socket.SOCK_STREAM'], {}), '(socket.AF_INET, socket.SOCK_STREAM)\n', (1191, 1227), Fals... |
# Copyright (c) 2019 PaddlePaddle Authors. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by app... | [
"logging.getLogger"
] | [((1111, 1138), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (1128, 1138), False, 'import logging\n')] |
import operations
def test_getOneNews_basic():
news = operations.getOneNews()
print(news)
assert news is not None
print('test_getOneNews_basic passed!')
def test_getNewsSummariesForUser_basic():
news = operations.getNewsSummariesForUser('test', 1)
assert len(news) > 0
print('test_getNewsSu... | [
"operations.getOneNews",
"operations.getNewsSummariesForUser"
] | [((59, 82), 'operations.getOneNews', 'operations.getOneNews', ([], {}), '()\n', (80, 82), False, 'import operations\n'), ((224, 269), 'operations.getNewsSummariesForUser', 'operations.getNewsSummariesForUser', (['"""test"""', '(1)'], {}), "('test', 1)\n", (258, 269), False, 'import operations\n'), ((417, 462), 'operati... |
from setuptools import find_packages, setup
LONG_DESCRIPTION = (
'Desc.'
)
setup(
name='mhealth',
version='0.0.3',
packages=find_packages(where='src'),
package_dir={'': 'src'},
url='https://github.com/callumstew/pymhealth',
author='<NAME>',
author_email='<E... | [
"setuptools.find_packages"
] | [((155, 181), 'setuptools.find_packages', 'find_packages', ([], {'where': '"""src"""'}), "(where='src')\n", (168, 181), False, 'from setuptools import find_packages, setup\n')] |
from django.urls import path
from . import views
# Objektivaated
from wiki.views import ArtikkelDetailView, IsikDetailView, OrganisatsioonDetailView, ObjektDetailView, KaardiobjektDetailView
# Filtreerimisvaated
from wiki.views import ArtikkelFilterView, IsikFilterView, OrganisatsioonFilterView, ObjektFilterView, Kaa... | [
"wiki.views.OrganisatsioonUpdate.as_view",
"wiki.views.ArtikkelYearArchiveView.as_view",
"wiki.views.KaardiobjektUpdate.as_view",
"wiki.views.KaardiobjektDetailView.as_view",
"wiki.views.IsikUpdate.as_view",
"wiki.views.ArtikkelArchiveIndexView.as_view",
"wiki.views.ArtikkelUpdate.as_view",
"wiki.view... | [((865, 903), 'django.urls.path', 'path', (['"""info/"""', 'views.info'], {'name': '"""info"""'}), "('info/', views.info, name='info')\n", (869, 903), False, 'from django.urls import path\n'), ((909, 947), 'django.urls.path', 'path', (['"""otsi/"""', 'views.otsi'], {'name': '"""otsi"""'}), "('otsi/', views.otsi, name='... |
from office365.entity import Entity
from office365.runtime.client_result import ClientResult
from office365.runtime.queries.delete_entity_query import DeleteEntityQuery
from office365.runtime.queries.service_operation_query import ServiceOperationQuery
from office365.runtime.queries.update_entity_query import UpdateEnt... | [
"office365.runtime.queries.service_operation_query.ServiceOperationQuery",
"office365.runtime.client_result.ClientResult",
"office365.runtime.queries.update_entity_query.UpdateEntityQuery",
"office365.runtime.queries.delete_entity_query.DeleteEntityQuery"
] | [((766, 784), 'office365.runtime.client_result.ClientResult', 'ClientResult', (['None'], {}), '(None)\n', (778, 784), False, 'from office365.runtime.client_result import ClientResult\n'), ((886, 961), 'office365.runtime.queries.service_operation_query.ServiceOperationQuery', 'ServiceOperationQuery', (['self', '"""getMe... |
import click
from gradient.api_sdk import constants
from gradient.cli import common
from gradient.cli.cli import cli
from gradient.cli.cli_types import ChoiceType, json_string
from gradient.cli.common import validate_comma_split_option, api_key_option, ClickGroup
from gradient.commands import notebooks
from gradient.c... | [
"gradient.commands.notebooks.CreateNotebookCommand",
"click.echo",
"click.DateTime",
"gradient.commands.notebooks.StreamNotebookMetricsCommand",
"click.option",
"gradient.commands.notebooks.StartNotebookCommand",
"gradient.commands.notebooks.StopNotebookCommand",
"gradient.commands.notebooks.NotebookA... | [((432, 502), 'gradient.cli.cli.cli.group', 'cli.group', (['"""notebooks"""'], {'help': '"""Manage notebooks"""', 'cls': 'common.ClickGroup'}), "('notebooks', help='Manage notebooks', cls=common.ClickGroup)\n", (441, 502), False, 'from gradient.cli.cli import cli\n'), ((843, 991), 'click.option', 'click.option', (['"""... |
# coding: utf-8
"""
Licensed to Cloudera, Inc. under one
or more contributor license agreements. See the NOTICE file
distributed with this work for additional information
regarding copyright ownership. Cloudera, Inc. licenses this file
to you under the Apache License, Version 2.0 (the
"License"); you may not use thi... | [
"six.iteritems"
] | [((12318, 12351), 'six.iteritems', 'six.iteritems', (['self.swagger_types'], {}), '(self.swagger_types)\n', (12331, 12351), False, 'import six\n')] |
import chainer
import chainer as C
import chainer.functions as F
import chainer.links as L
from chainer.backends import cuda
import inspect
import weakref
import sys
from chainer_compiler.elichika.parser import config
from chainer_compiler.elichika.parser import nodes
from chainer_compiler.elichika.parser import vevalu... | [
"chainer_compiler.elichika.parser.functions_builtin.VEvalContextFunction",
"chainer_compiler.elichika.parser.functions_ndarray.NDArrayCumsumFunction",
"chainer_compiler.elichika.parser.links_builtin.is_builtin_chainer_link",
"chainer_compiler.elichika.parser.links_builtin.ChainerLinkInstance",
"chainer_comp... | [((1116, 1149), 'inspect.getmembers', 'inspect.getmembers', (['parent_module'], {}), '(parent_module)\n', (1134, 1149), False, 'import inspect\n'), ((1338, 1373), 'chainer_compiler.elichika.parser.values.reset_field_and_attributes', 'values.reset_field_and_attributes', ([], {}), '()\n', (1371, 1373), False, 'from chain... |
import math
#######
import random
import cv2
import numpy as np
import matplotlib.pyplot as plt
from tensorpack.dataflow.imgaug.geometry import RotationAndCropValid
def crop_meta_image(image,annos,mask):
_target_height=368
_target_width =368
if len(np.shape(image))==2:
image = cv2.cvtColor(image, c... | [
"matplotlib.pyplot.imshow",
"random.uniform",
"cv2.warpAffine",
"cv2.resize",
"cv2.flip",
"random.randrange",
"cv2.copyMakeBorder",
"math.cos",
"numpy.random.randint",
"cv2.cvtColor",
"numpy.random.uniform",
"tensorpack.dataflow.imgaug.geometry.RotationAndCropValid.largest_rotated_rect",
"ma... | [((358, 373), 'numpy.shape', 'np.shape', (['image'], {}), '(image)\n', (366, 373), True, 'import numpy as np\n'), ((3920, 3935), 'numpy.shape', 'np.shape', (['image'], {}), '(image)\n', (3928, 3935), True, 'import numpy as np\n'), ((4659, 4744), 'cv2.resize', 'cv2.resize', (['image', '(_target_width, _target_height)'],... |
from datetime import datetime
import json
dt = datetime.now()
print(dt)
json_dt = json.dumps(dt.isoformat())
with open("result.json", "wb") as f:
f.write(json_dt)
| [
"datetime.datetime.now"
] | [((49, 63), 'datetime.datetime.now', 'datetime.now', ([], {}), '()\n', (61, 63), False, 'from datetime import datetime\n')] |
from django.db import models
# Create your models here.
class User_data(models.Model):
"""
@pony
使用者(醫生)資料
"""
name = models.CharField(max_length=20)
account = models.CharField(max_length=50)
password = models.CharField(max_length=50)
email = models.EmailField()
depa... | [
"django.db.models.EmailField",
"django.db.models.CharField"
] | [((151, 182), 'django.db.models.CharField', 'models.CharField', ([], {'max_length': '(20)'}), '(max_length=20)\n', (167, 182), False, 'from django.db import models\n'), ((198, 229), 'django.db.models.CharField', 'models.CharField', ([], {'max_length': '(50)'}), '(max_length=50)\n', (214, 229), False, 'from django.db im... |
# -*- coding: utf-8 -*-
"""TcEx Framework Redis Module"""
from builtins import str
import redis
class TcExRedis(object):
"""Create/Read Data in/from Redis"""
def __init__(self, host, port, rhash):
"""Initialize the Class properties.
Args:
host (string): The Redis host.
... | [
"redis.StrictRedis"
] | [((456, 495), 'redis.StrictRedis', 'redis.StrictRedis', ([], {'host': 'host', 'port': 'port'}), '(host=host, port=port)\n', (473, 495), False, 'import redis\n')] |
from __future__ import unicode_literals
import unittest
from io import StringIO
from python_digest import *
from python_digest.http import *
from python_digest.utils import *
class HttpTests(unittest.TestCase):
def test_parse_quoted_string(self):
test_cases = [
('""', ''), # OK
(... | [
"unittest.main",
"io.StringIO"
] | [((21686, 21701), 'unittest.main', 'unittest.main', ([], {}), '()\n', (21699, 21701), False, 'import unittest\n'), ((14601, 14611), 'io.StringIO', 'StringIO', ([], {}), '()\n', (14609, 14611), False, 'from io import StringIO\n'), ((15109, 15119), 'io.StringIO', 'StringIO', ([], {}), '()\n', (15117, 15119), False, 'from... |
import logging
import re
import shlex
import threading
import sshim
logging.basicConfig(level='DEBUG')
logger = logging.getLogger()
class Device(threading.Thread):
def __init__(self, script):
threading.Thread.__init__(self)
self.history = []
self.script = script
self.start()
d... | [
"logging.basicConfig",
"threading.Thread.__init__",
"logging.getLogger",
"re.compile",
"shlex.split",
"sshim.Server"
] | [((69, 103), 'logging.basicConfig', 'logging.basicConfig', ([], {'level': '"""DEBUG"""'}), "(level='DEBUG')\n", (88, 103), False, 'import logging\n'), ((113, 132), 'logging.getLogger', 'logging.getLogger', ([], {}), '()\n', (130, 132), False, 'import logging\n'), ((1198, 1229), 'sshim.Server', 'sshim.Server', (['Device... |