code
stringlengths
22
1.05M
apis
listlengths
1
3.31k
extract_api
stringlengths
75
3.25M
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ C APDCAM10G Firmware test reated on Mon Jul 2 15:46:13 2018 @author: apdcam """ def firmware_test(card=8) : import APDCAM10G_control c = APDCAM10G_control.APDCAM10G_regCom() ret = c.startReceiveAnswer() if ret != "" : print("Could...
[ "APDCAM10G_control.APDCAM10G_regCom", "APDCAM10G_control.APDCAM10G_data" ]
[((205, 241), 'APDCAM10G_control.APDCAM10G_regCom', 'APDCAM10G_control.APDCAM10G_regCom', ([], {}), '()\n', (239, 241), False, 'import APDCAM10G_control\n'), ((828, 862), 'APDCAM10G_control.APDCAM10G_data', 'APDCAM10G_control.APDCAM10G_data', ([], {}), '()\n', (860, 862), False, 'import APDCAM10G_control\n')]
import numpy as np from scipy.stats import norm def make_grid(xx, yy): """ Returns two n-by-n matrices. The first one contains all the x values and the second all the y values of a cartesian product between `xx` and `yy`. """ n = len(xx) xx, yy = np.meshgrid(xx, yy) grid = np.array([xx.r...
[ "numpy.meshgrid", "numpy.log", "scipy.stats.norm.pdf", "numpy.sum" ]
[((275, 294), 'numpy.meshgrid', 'np.meshgrid', (['xx', 'yy'], {}), '(xx, yy)\n', (286, 294), True, 'import numpy as np\n'), ((673, 747), 'scipy.stats.norm.pdf', 'norm.pdf', (['yy[:, None, None]'], {'loc': 'mus[None, :, :]', 'scale': 'sigmas[None, :, :]'}), '(yy[:, None, None], loc=mus[None, :, :], scale=sigmas[None, :,...
# -*- coding: utf-8 -*- import importlib from django.db.models.signals import post_save, post_delete from jsonmirror.conf import BACKEND if BACKEND == "rethinkdb": from jsonmirror.backends.rethinkdb.signals import model_save, model_delete def register_model(model): post_save.connect(model_save, sende...
[ "django.db.models.signals.post_save.connect", "importlib.import_module", "django.db.models.signals.post_delete.connect" ]
[((285, 328), 'django.db.models.signals.post_save.connect', 'post_save.connect', (['model_save'], {'sender': 'model'}), '(model_save, sender=model)\n', (302, 328), False, 'from django.db.models.signals import post_save, post_delete\n'), ((333, 380), 'django.db.models.signals.post_delete.connect', 'post_delete.connect',...
import sys import os import random import socket import string import time def connect( ip, port ): s = socket.socket(socket.AF_INET, socket.SOCK_STREAM) try: s.connect( (ip, port) ) except: print('Failed to connect to %s:%d' %(ip, port)) sys.exit(1) return s def readuntil( s...
[ "random.choice", "socket.socket", "time.sleep", "random.seed", "sys.exit", "socket.gethostname", "random.randint" ]
[((109, 158), 'socket.socket', 'socket.socket', (['socket.AF_INET', 'socket.SOCK_STREAM'], {}), '(socket.AF_INET, socket.SOCK_STREAM)\n', (122, 158), False, 'import socket\n'), ((28262, 28277), 'time.sleep', 'time.sleep', (['(0.3)'], {}), '(0.3)\n', (28272, 28277), False, 'import time\n'), ((30493, 30503), 'sys.exit', ...
import torch from torch import nn from torch.autograd import Variable class CovaMLoss(nn.Module): def __init__(self): super(CovaMLoss, self).__init__() # self.ranking_loss = nn.MarginRankingLoss(margin=margin) # calculate the similarity def cal_similarity(self, input, CovaMatrix_list): ...
[ "torch.norm", "torch.cuda.is_available", "torch.cat", "torch.transpose" ]
[((990, 1012), 'torch.cat', 'torch.cat', (['Cova_Sim', '(0)'], {}), '(Cova_Sim, 0)\n', (999, 1012), False, 'import torch\n'), ((513, 546), 'torch.norm', 'torch.norm', (['query_sam', '(2)', '(1)', '(True)'], {}), '(query_sam, 2, 1, True)\n', (523, 546), False, 'import torch\n'), ((616, 641), 'torch.cuda.is_available', '...
from logging import getLogger from calendar import monthrange from collections import OrderedDict from datetime import date from datetime import datetime from dateutil.parser import parse import ckan.plugins as p import ckan.logic as logic from ckan.logic.action.create import user_create as ckan_user_create import cka...
[ "logging.getLogger", "ckanext.scheming.helpers.scheming_field_choices", "ckan.lib.helpers.redirect_to", "ckan.common._", "ckanext.scheming.helpers.scheming_get_dataset_schema", "calendar.monthrange", "ckan.plugins.toolkit.add_template_directory", "datetime.datetime", "ckan.plugins.toolkit.get_action...
[((790, 809), 'logging.getLogger', 'getLogger', (['__name__'], {}), '(__name__)\n', (799, 809), False, 'from logging import getLogger\n'), ((915, 943), 'ckan.plugins.implements', 'p.implements', (['p.ITranslation'], {}), '(p.ITranslation)\n', (927, 943), True, 'import ckan.plugins as p\n'), ((1095, 1123), 'ckan.plugins...
# standard library imports import os # local application imports from atoms import Atom from exceptions import XYZfileDidNotExist from exceptions import XYZfileWrongFormat def xyz_file_to_atoms(filename): """ From an .xyz file get a list of atoms Arguments: filename (str): .xyz fil...
[ "atoms.Atom", "os.path.exists" ]
[((425, 449), 'os.path.exists', 'os.path.exists', (['filename'], {}), '(filename)\n', (439, 449), False, 'import os\n'), ((1144, 1189), 'atoms.Atom', 'Atom', ([], {'atomic_symbol': 'atom_label', 'x': 'x', 'y': 'y', 'z': 'z'}), '(atomic_symbol=atom_label, x=x, y=y, z=z)\n', (1148, 1189), False, 'from atoms import Atom\n...
#import torch.nn as nn import torch from torch.nn import functional as F #from PIL import Image import numpy as np import pandas as pd #import os import os.path as osp import shutil #import math def save_checkpoint(state,best_pred, epoch,is_best,checkpoint_path,filename='./checkpoint/checkpoint.pth.tar'): torch.sa...
[ "numpy.histogram", "torch.max", "numpy.asarray", "torch.sigmoid", "numpy.equal", "numpy.stack", "numpy.sum", "numpy.array", "torch.save", "numpy.all", "torch.argmax" ]
[((312, 339), 'torch.save', 'torch.save', (['state', 'filename'], {}), '(state, filename)\n', (322, 339), False, 'import torch\n'), ((1371, 1402), 'numpy.stack', 'np.stack', (['semantic_map'], {'axis': '(-1)'}), '(semantic_map, axis=-1)\n', (1379, 1402), True, 'import numpy as np\n'), ((4205, 4223), 'numpy.sum', 'np.su...
# -*- coding: utf-8 -*- # Copyright 2020 Google LLC # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or...
[ "proto.RepeatedField", "proto.Field", "proto.module", "proto.MapField" ]
[((941, 1280), 'proto.module', 'proto.module', ([], {'package': '"""google.cloud.datalabeling.v1beta1"""', 'manifest': "{'DataType', 'Dataset', 'InputConfig', 'TextMetadata',\n 'ClassificationMetadata', 'GcsSource', 'BigQuerySource', 'OutputConfig',\n 'GcsDestination', 'GcsFolderDestination', 'DataItem',\n 'An...
import json import math import multiprocessing import os import tempfile from pathlib import Path import gym import numpy as np import pandas as pd from ray import tune from ray.rllib.models import ModelCatalog from ray.rllib.utils import try_import_tf from examples.rllib_agent import TrainingModel from smarts.core.a...
[ "tempfile.TemporaryDirectory", "smarts.core.agent_interface.AgentInterface.from_type", "json.loads", "smarts.env.custom_observations.lane_ttc_observation_adapter.transform", "pathlib.Path", "ray.rllib.models.ModelCatalog.get_preprocessor_for_space", "os.path.join", "math.sqrt", "pandas.DataFrame.fro...
[((562, 577), 'ray.rllib.utils.try_import_tf', 'try_import_tf', ([], {}), '()\n', (575, 577), False, 'from ray.rllib.utils import try_import_tf\n'), ((1718, 1773), 'smarts.env.custom_observations.lane_ttc_observation_adapter.transform', 'lane_ttc_observation_adapter.transform', (['env_observation'], {}), '(env_observat...
""" MIT License Copyright (c) 2017 s0hvaperuna Permission is hereby granted, free of charge, to any person obtaining a copy of this software and associated documentation files (the "Software"), to deal in the Software without restriction, including without limitation the rights to use, copy, modify, merge, publish, d...
[ "logging.getLogger", "utils.utilities.get_image", "bot.bot.command", "PIL.Image.new", "bot.bot.bot_has_permissions", "io.BytesIO", "sys.exc_info", "utils.imagetools.create_shadow", "numpy.sin", "utils.imagetools.get_color", "collections.OrderedDict.fromkeys", "utils.utilities.get_picture_from_...
[((2206, 2235), 'logging.getLogger', 'logging.getLogger', (['"""terminal"""'], {}), "('terminal')\n", (2223, 2235), False, 'import logging\n'), ((8207, 8233), 'bot.bot.command', 'command', ([], {'aliases': "['stand']"}), "(aliases=['stand'])\n", (8214, 8233), False, 'from bot.bot import command, cooldown, bot_has_permi...
from django.contrib import admin from enrollment.enrollment.models import ( StudentEnrollment, CourseGroup, Schedule, ) @admin.register(CourseGroup) class CourseGroupAdmin(admin.ModelAdmin): list_display = ['course', 'name', 'teacher', 'year', 'semester'] search_fields = ['course', 'name', 'teach...
[ "django.contrib.admin.register" ]
[((136, 163), 'django.contrib.admin.register', 'admin.register', (['CourseGroup'], {}), '(CourseGroup)\n', (150, 163), False, 'from django.contrib import admin\n'), ((328, 361), 'django.contrib.admin.register', 'admin.register', (['StudentEnrollment'], {}), '(StudentEnrollment)\n', (342, 361), False, 'from django.contr...
from distutils.core import setup setup(name = "DWL5500XY", version = "0", description = "Python module to use the Digipass DWL5500XY", author = "<NAME>", author_email = "<EMAIL>", url = "https://github.com/Stormholt/DWL5500XY-Python", packages = ['DWL5500XY'], scripts = ["test.py"], ...
[ "distutils.core.setup" ]
[((36, 331), 'distutils.core.setup', 'setup', ([], {'name': '"""DWL5500XY"""', 'version': '"""0"""', 'description': '"""Python module to use the Digipass DWL5500XY"""', 'author': '"""<NAME>"""', 'author_email': '"""<EMAIL>"""', 'url': '"""https://github.com/Stormholt/DWL5500XY-Python"""', 'packages': "['DWL5500XY']", '...
# # test_siteconfig.py # # Tests for site config generation (siteconfig.py). # # Copyright (C) 2018-2021 by # <NAME>. # # This file is part of the FreeType project, and may only be used, # modified, and distributed under the terms of the FreeType project # license, LICENSE.TXT. By continuing to use, modify, o...
[ "yaml.safe_load", "docwriter.siteconfig.SiteConfig" ]
[((650, 673), 'docwriter.siteconfig.SiteConfig', 'siteconfig.SiteConfig', ([], {}), '()\n', (671, 673), False, 'from docwriter import siteconfig\n'), ((1749, 1771), 'yaml.safe_load', 'yaml.safe_load', (['result'], {}), '(result)\n', (1763, 1771), False, 'import yaml\n')]
import unittest from pywavez.zwave import Message, inboundMessageFromBytes from pywavez.zwave.Constants import LibraryType class TestGetVersion(unittest.TestCase): def test_GetVersionResponse(self): data = bytes.fromhex("01155a2d5761766520342e30350001") obj = inboundMessageFromBytes(data) ...
[ "unittest.main", "pywavez.zwave.inboundMessageFromBytes" ]
[((2482, 2497), 'unittest.main', 'unittest.main', ([], {}), '()\n', (2495, 2497), False, 'import unittest\n'), ((283, 312), 'pywavez.zwave.inboundMessageFromBytes', 'inboundMessageFromBytes', (['data'], {}), '(data)\n', (306, 312), False, 'from pywavez.zwave import Message, inboundMessageFromBytes\n'), ((793, 822), 'py...
# Generated by Django 3.2.9 on 2021-12-04 16:44 from django.db import migrations, models import django.db.models.deletion class Migration(migrations.Migration): dependencies = [ ('pokemon', '0008_auto_20211204_1644'), ('trainers', '0002_teams'), ] operations = [ migrations.Creat...
[ "django.db.models.CharField", "django.db.models.BigAutoField", "django.db.models.ForeignKey" ]
[((400, 496), 'django.db.models.BigAutoField', 'models.BigAutoField', ([], {'auto_created': '(True)', 'primary_key': '(True)', 'serialize': '(False)', 'verbose_name': '"""ID"""'}), "(auto_created=True, primary_key=True, serialize=False,\n verbose_name='ID')\n", (419, 496), False, 'from django.db import migrations, m...
from Distributions import NormalDistribution class Guest: cur_id = 0 work_time_distribution = NormalDistribution(10, 1) def __init__(self, guest_type): self.id = Guest.cur_id self.type = guest_type self.eating = False self.finished = False self.time_to_finish = 0 ...
[ "Distributions.NormalDistribution" ]
[((104, 129), 'Distributions.NormalDistribution', 'NormalDistribution', (['(10)', '(1)'], {}), '(10, 1)\n', (122, 129), False, 'from Distributions import NormalDistribution\n')]
#!/usr/bin/python import sys import os import ctypes import threading import socket import time from PySide2.QtCore import * from PySide2.QtGui import * from PySide2.QtWidgets import QWidget, QTextEdit, QTabWidget, QFormLayout class UpdateSig(QObject): sig = Signal(float) class LoggerWidget(QWidget): def __...
[ "PySide2.QtWidgets.QFormLayout", "PySide2.QtWidgets.QTabWidget", "PySide2.QtWidgets.QTextEdit" ]
[((423, 434), 'PySide2.QtWidgets.QTextEdit', 'QTextEdit', ([], {}), '()\n', (432, 434), False, 'from PySide2.QtWidgets import QWidget, QTextEdit, QTabWidget, QFormLayout\n'), ((903, 919), 'PySide2.QtWidgets.QTabWidget', 'QTabWidget', (['self'], {}), '(self)\n', (913, 919), False, 'from PySide2.QtWidgets import QWidget,...
import torch import torch.nn as nn import torch.nn.functional as F class BaseCNNModel(nn.Module): def __init__(self, num_classes): super(BaseCNNModel, self).__init__() self.conv1 = nn.Conv2d(3, 32, 5, 2) self.conv2 = nn.Conv2d(32, 32, 5, 1) self.conv3 = nn.Conv2d(32, 64, 5, 1) ...
[ "torch.nn.BatchNorm2d", "torch.nn.Dropout", "torch.nn.Conv2d", "torch.nn.MaxPool2d", "torch.nn.Linear", "torch.rand" ]
[((1732, 1760), 'torch.rand', 'torch.rand', (['(1, 3, 256, 256)'], {}), '((1, 3, 256, 256))\n', (1742, 1760), False, 'import torch\n'), ((205, 227), 'torch.nn.Conv2d', 'nn.Conv2d', (['(3)', '(32)', '(5)', '(2)'], {}), '(3, 32, 5, 2)\n', (214, 227), True, 'import torch.nn as nn\n'), ((249, 272), 'torch.nn.Conv2d', 'nn.C...
""" Support for Chuang Mi IR Remote Controller. Thank rytilahti for his great work """ import logging from datetime import timedelta import asyncio from random import randint import voluptuous as vol from socket import timeout import homeassistant.loader as loader from homeassistant.components.switch import (SwitchDe...
[ "logging.getLogger", "homeassistant.util.dt.utcnow", "voluptuous.Required", "voluptuous.Schema", "homeassistant.loader.get_component", "datetime.timedelta", "voluptuous.Length", "miio.ChuangmiIr", "asyncio.sleep", "voluptuous.Optional", "random.randint" ]
[((784, 811), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (801, 811), False, 'import logging\n'), ((2343, 2390), 'homeassistant.loader.get_component', 'loader.get_component', (['"""persistent_notification"""'], {}), "('persistent_notification')\n", (2363, 2390), True, 'import homeassis...
# Copyright (C) 2018 Google Inc. # Licensed under http://www.apache.org/licenses/LICENSE-2.0 <see LICENSE file> """This module contains common utils for integration functionality.""" import sqlalchemy as sa from ggrc import db from ggrc import settings from ggrc.models import exceptions from ggrc.models import all_m...
[ "ggrc.models.all_models.Person.email.in_", "ggrc.integrations.synchronization_jobs.assessment_sync_job.ASSESSMENT_STATUSES_MAPPING.get", "ggrc.models.exceptions.ValidationError", "ggrc.db.session.query", "sqlalchemy.and_" ]
[((3457, 3502), 'ggrc.integrations.synchronization_jobs.assessment_sync_job.ASSESSMENT_STATUSES_MAPPING.get', 'ASSESSMENT_STATUSES_MAPPING.get', (['assmt.status'], {}), '(assmt.status)\n', (3488, 3502), False, 'from ggrc.integrations.synchronization_jobs.assessment_sync_job import ASSESSMENT_STATUSES_MAPPING\n'), ((131...
# vim: set ts=4 sw=4 expandtab: from puremvc.patterns.command import SimpleCommand from model.TimeProxy import TimeProxy from view.DialogMediator import DialogMediator from view.MenuMediator import MenuMediator from view.DatePickerMediator import DatePickerMediator from view.TimeGridMediator import TimeGridMediator...
[ "view.SummaryMediator.SummaryMediator", "model.TimeProxy.TimeProxy", "view.DialogMediator.DialogMediator", "view.DatePickerMediator.DatePickerMediator", "view.MenuMediator.MenuMediator", "view.TimeGridMediator.TimeGridMediator" ]
[((470, 481), 'model.TimeProxy.TimeProxy', 'TimeProxy', ([], {}), '()\n', (479, 481), False, 'from model.TimeProxy import TimeProxy\n'), ((556, 581), 'view.DialogMediator.DialogMediator', 'DialogMediator', (['mainPanel'], {}), '(mainPanel)\n', (570, 581), False, 'from view.DialogMediator import DialogMediator\n'), ((62...
# -*- coding: utf-8 -*- from __future__ import absolute_import, division, print_function try: input = raw_input except NameError: pass import re def promptYesOrNo(message): while True: m = re.match(r'(y|yes)|(n|no)', input(message + " "), re.IGNORECASE) if m is None: print(...
[ "re.compile" ]
[((568, 614), 're.compile', 're.compile', (['"""(?P<num>\\\\d+)\\\\s*(?P<unit>\\\\w+)*"""'], {}), "('(?P<num>\\\\d+)\\\\s*(?P<unit>\\\\w+)*')\n", (578, 614), False, 'import re\n')]
from __future__ import division from __future__ import print_function from __future__ import absolute_import from builtins import zip from builtins import range from builtins import object from past.utils import old_div from nose.tools import (assert_equal, assert_not_equal, assert_almost_equal, ...
[ "logging.getLogger", "openpathsampling.IdentityPathMover", "openpathsampling.MSOuterTISInterface.from_lambdas", "past.utils.old_div", "openpathsampling.high_level.move_strategy.OrganizeByMoveGroupStrategy", "nose.tools.assert_equal", "copy.copy", "openpathsampling.high_level.move_strategy.OneWayShooti...
[((2423, 2496), 'openpathsampling.RandomChoiceMoveChange', 'paths.RandomChoiceMoveChange', ([], {'subchange': 'group_mover_change', 'mover': 'chooser'}), '(subchange=group_mover_change, mover=chooser)\n', (2451, 2496), True, 'import openpathsampling as paths\n'), ((2543, 2615), 'openpathsampling.RandomChoiceMoveChange'...
from onegov.core.utils import Bunch from onegov.core.elements import Link, Confirm, Intercooler def test_link(render_element): # text is translated result = render_element(Link(text="Settings", url='/settings')) assert result.pyquery('a').text() == "Settings" assert result.pyquery('a').attr('href') ==...
[ "onegov.core.elements.Intercooler", "onegov.core.elements.Confirm", "onegov.core.elements.Link", "onegov.core.utils.Bunch" ]
[((182, 220), 'onegov.core.elements.Link', 'Link', ([], {'text': '"""Settings"""', 'url': '"""/settings"""'}), "(text='Settings', url='/settings')\n", (186, 220), False, 'from onegov.core.elements import Link, Confirm, Intercooler\n'), ((398, 450), 'onegov.core.elements.Link', 'Link', ([], {'text': '"""foo"""', 'url': ...
import multiprocessing num_cores = multiprocessing.cpu_count() INPUT = './data/scratch/scripts.txt' BLOCK_LANG = 'data/scratch/tokens.pickle.lang' RESULTS_DIR = './results' FIGURES_DIR = './figures' START_SYMBOL = 'START' END_SYMBOL = 'END' WORD2VEC_SIZE = 8 WORD2VEC_MODEL = './results/wv-{}.model'.format(WORD2VEC_SIZ...
[ "multiprocessing.cpu_count" ]
[((36, 63), 'multiprocessing.cpu_count', 'multiprocessing.cpu_count', ([], {}), '()\n', (61, 63), False, 'import multiprocessing\n')]
import contextlib import contextvars import logging import os import sys from typing import Iterator _logging_prefix: contextvars.ContextVar[str] = contextvars.ContextVar( "logging_prefix", default="", ) @contextlib.contextmanager def logging_prefix(prefix: str) -> Iterator[str]: token = _logging_prefix....
[ "logging.basicConfig", "logging.StreamHandler", "os.getenv", "contextvars.ContextVar", "logging.root.removeHandler" ]
[((149, 201), 'contextvars.ContextVar', 'contextvars.ContextVar', (['"""logging_prefix"""'], {'default': '""""""'}), "('logging_prefix', default='')\n", (171, 201), False, 'import contextvars\n'), ((644, 672), 'os.getenv', 'os.getenv', (['"""HOSTNAME"""', '"""dev"""'], {}), "('HOSTNAME', 'dev')\n", (653, 672), False, '...
import json from web3 import Web3, HTTPProvider from web3.gas_strategies.rpc import rpc_gas_price_strategy class ContractInterface(object): def __init__(self, config): """ sets up w3 and the contract """ self.config = config self.provider = HTTPProvider(self.config.PROVIDE...
[ "json.load", "web3.HTTPProvider", "web3.Web3.toChecksumAddress", "web3.Web3" ]
[((288, 326), 'web3.HTTPProvider', 'HTTPProvider', (['self.config.PROVIDER_URI'], {}), '(self.config.PROVIDER_URI)\n', (300, 326), False, 'from web3 import Web3, HTTPProvider\n'), ((345, 364), 'web3.Web3', 'Web3', (['self.provider'], {}), '(self.provider)\n', (349, 364), False, 'from web3 import Web3, HTTPProvider\n'),...
"""This module handles all operations involving the user's settings.""" import json from os import path from PyQt5.QtWidgets import QWidget from PyQt5.QtCore import QObject, pyqtSignal from ui import SettingsTab import ManageDB from Constants import * import GeneralUtils from GeneralUtils import JsonModel class Sett...
[ "PyQt5.QtCore.pyqtSignal", "json.loads", "GeneralUtils.show_message", "ManageDB.get_all_report_files", "json.dumps", "GeneralUtils.save_json_file", "GeneralUtils.read_json_file", "ManageDB.get_all_cost_files", "os.path.abspath", "GeneralUtils.choose_directory", "ManageDB.UpdateDatabaseProgressDi...
[((3654, 3679), 'PyQt5.QtCore.pyqtSignal', 'pyqtSignal', (['SettingsModel'], {}), '(SettingsModel)\n', (3664, 3679), False, 'from PyQt5.QtCore import QObject, pyqtSignal\n'), ((3895, 3962), 'GeneralUtils.read_json_file', 'GeneralUtils.read_json_file', (['(SETTINGS_FILE_DIR + SETTINGS_FILE_NAME)'], {}), '(SETTINGS_FILE_...
"""Entries Table Seeder. You can run this seeder in order to generate journal entries. - Each time it is ran it will generate 5 random entries. - You can run the seeder by running: craft seed:run. """ from orator.seeds import Seeder from app.Entry import Entry from config.factories import factory class En...
[ "config.factories.factory" ]
[((444, 462), 'config.factories.factory', 'factory', (['Entry', '(25)'], {}), '(Entry, 25)\n', (451, 462), False, 'from config.factories import factory\n')]
# Train.py import numpy as np import pandas as pandas from Data import Data from BuildModel import BuildModel import copy import sys from keras.callbacks import EarlyStopping class Train(): @staticmethod def train(): # adam = optimizers.Adam(lr=0.001, beta_1=0.9, beta_2=0.999, epsilon=No...
[ "BuildModel.BuildModel.cnn_base_model", "keras.callbacks.EarlyStopping" ]
[((367, 394), 'BuildModel.BuildModel.cnn_base_model', 'BuildModel.cnn_base_model', ([], {}), '()\n', (392, 394), False, 'from BuildModel import BuildModel\n'), ((508, 595), 'keras.callbacks.EarlyStopping', 'EarlyStopping', ([], {'monitor': '"""val_loss"""', 'min_delta': '(0)', 'patience': '(2)', 'verbose': '(0)', 'mode...
def run(): import requests, json api = 'http://ip-api.com/json/' ip = input("\n[?] Enter IP or Domain To Lookup: ") print("[!] Sending Request... [!]") request = requests.get(api + ip).text response = json.loads(request) status = response['status'] if status == "success": ...
[ "json.loads", "requests.get" ]
[((233, 252), 'json.loads', 'json.loads', (['request'], {}), '(request)\n', (243, 252), False, 'import requests, json\n'), ((189, 211), 'requests.get', 'requests.get', (['(api + ip)'], {}), '(api + ip)\n', (201, 211), False, 'import requests, json\n')]
import os import subprocess from datetime import date from mne import __version__ as release_version # NOTE: ../codemeta.json and ../citation.cff should not be continuously # updated. Run this script only at release time. package_name = 'MNE-Python' hard_dependencies = ('numpy', 'scipy') release_date = str(dat...
[ "datetime.date.today", "subprocess.run", "os.path.join", "mne.__version__.split" ]
[((2326, 2378), 'subprocess.run', 'subprocess.run', (['args'], {'capture_output': '(True)', 'text': '(True)'}), '(args, capture_output=True, text=True)\n', (2340, 2378), False, 'import subprocess\n'), ((317, 329), 'datetime.date.today', 'date.today', ([], {}), '()\n', (327, 329), False, 'from datetime import date\n'), ...
#!/usr/bin/env python import rospy from flexbe_core import EventState, Logger from rospy.exceptions import ROSInterruptException ''' Created on 11.06.2013 @author: <NAME> ''' class MyDecisionState(EventState): ''' Evaluates a condition function in order to return one of the specified outcomes. This state can be ...
[ "rospy.sleep", "rospy.logwarn" ]
[((1396, 1412), 'rospy.sleep', 'rospy.sleep', (['(0.2)'], {}), '(0.2)\n', (1407, 1412), False, 'import rospy\n'), ((1486, 1517), 'rospy.logwarn', 'rospy.logwarn', (['"""Skipped sleep."""'], {}), "('Skipped sleep.')\n", (1499, 1517), False, 'import rospy\n')]
import glob import os import re all_files = glob.glob("*pdf") for file in all_files: if "label" in file: new_file = re.sub("label", "delivery", file) os.system(f'mv {file} {new_file}') if "Ausfahrliste.pdf" in file: new_file = re.sub("Ausfahrliste.pdf", "ausfahrliste_delivery_file.pdf",...
[ "re.sub", "os.system", "glob.glob" ]
[((44, 61), 'glob.glob', 'glob.glob', (['"""*pdf"""'], {}), "('*pdf')\n", (53, 61), False, 'import glob\n'), ((128, 161), 're.sub', 're.sub', (['"""label"""', '"""delivery"""', 'file'], {}), "('label', 'delivery', file)\n", (134, 161), False, 'import re\n'), ((170, 204), 'os.system', 'os.system', (['f"""mv {file} {new_...
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Sun Jun 25 10:44:24 2017 @author: wroscoe """ import time from threading import Thread import socket from donkeycar.parts.controller import JoystickController, PS3JoystickController, PS3Joystick class Vehicle(): def __init__(self, mem=None): ...
[ "socket.socket", "time.sleep", "donkeycar.parts.controller.PS3Joystick", "threading.Thread", "time.time" ]
[((7750, 7784), 'threading.Thread', 'Thread', ([], {'target': 'ctr.update', 'args': '()'}), '(target=ctr.update, args=())\n', (7756, 7784), False, 'from threading import Thread\n'), ((4757, 4806), 'socket.socket', 'socket.socket', (['socket.AF_INET', 'socket.SOCK_STREAM'], {}), '(socket.AF_INET, socket.SOCK_STREAM)\n',...
# _________________________________________________________________________ # # PyUtilib: A Python utility library. # Copyright (c) 2008 Sandia Corporation. # This software is distributed under the BSD License. # Under the terms of Contract DE-AC04-94AL85000 with Sandia Corporation, # the U.S. Government retains ...
[ "pyutilib.pyro.util.set_maxconnections", "socket.gethostname", "Pyro.nsc.main", "Pyro.naming.main" ]
[((641, 708), 'pyutilib.pyro.util.set_maxconnections', 'set_maxconnections', ([], {'max_allowed_connections': 'max_allowed_connections'}), '(max_allowed_connections=max_allowed_connections)\n', (659, 708), False, 'from pyutilib.pyro.util import set_maxconnections, Pyro, using_pyro4\n'), ((1061, 1083), 'Pyro.naming.main...
# -*- coding:utf-8 -*- """ 桌面截图 """ from PIL import Image import win32gui, win32ui, win32con, win32api class CaptureDesktop(object): """ 桌面截图 """ SM_XVIRTUALSCREEN = 76 SM_YVIRTUALSCREEN = 77 SM_CXVIRTUALSCREEN = 78 SM_CYVIRTUALSCREEN = 79 def __init__(self): super(CaptureDesktop,...
[ "win32gui.GetWindowDC", "win32gui.GetDesktopWindow", "win32api.GetSystemMetrics", "win32ui.CreateDCFromHandle", "win32ui.CreateBitmap" ]
[((517, 544), 'win32gui.GetDesktopWindow', 'win32gui.GetDesktopWindow', ([], {}), '()\n', (542, 544), False, 'import win32gui, win32ui, win32con, win32api\n'), ((568, 600), 'win32gui.GetWindowDC', 'win32gui.GetWindowDC', (['self.mHwnd'], {}), '(self.mHwnd)\n', (588, 600), False, 'import win32gui, win32ui, win32con, win...
# -*- coding: utf-8 -*- """ Created on Sun Jul 30 09:08:26 2017 @author: hp """ ''' SeriousDlqin2yrsY/N超过90天或更糟的逾期拖欠 RevolvingUtilizationOfUnsecuredLines 无担保放款的循环利用:除了不动产和像车贷那样除以信用额度总和的无分期付款债务的信用卡和个人信用额度总额 NumberOfTime30-59DaysPastDueNotWorse35-59天逾期但不糟糕次数 DebtRatio负债比率 NumberOfOpenCreditLinesAndLoans 开放式信贷和贷款数量,开放式贷款...
[ "pandas.read_csv", "matplotlib.pyplot.ylabel", "sklearn.metrics.auc", "numpy.log", "sklearn.metrics.roc_auc_score", "sklearn.metrics.roc_curve", "numpy.arange", "numpy.mean", "scipy.interp", "pandas.qcut", "sklearn.cross_validation.cross_val_score", "matplotlib.pyplot.xlabel", "matplotlib.py...
[((1955, 1988), 'pandas.read_csv', 'pd.read_csv', (['"""...cs-training.csv"""'], {}), "('...cs-training.csv')\n", (1966, 1988), True, 'import pandas as pd\n'), ((1999, 2028), 'pandas.read_csv', 'pd.read_csv', (['"""...cs-test.csv"""'], {}), "('...cs-test.csv')\n", (2010, 2028), True, 'import pandas as pd\n'), ((2094, 2...
from argparse import ArgumentParser import dmsbatch from dmsbatch import __version__ from dmsbatch import commands def config_generate_cmd(args): if not args.file: print('Specify config file!') return print('Generating blank config: ', args.file) commands.generate_blank_config(args.file) ...
[ "dmsbatch.commands.generate_blank_config", "argparse.ArgumentParser" ]
[((277, 318), 'dmsbatch.commands.generate_blank_config', 'commands.generate_blank_config', (['args.file'], {}), '(args.file)\n', (307, 318), False, 'from dmsbatch import commands\n'), ((348, 449), 'argparse.ArgumentParser', 'ArgumentParser', ([], {'description': '"""Azure Batch for Delta Modeling Section """', 'conflic...
# Copyright Contributors to the Pyro project. # SPDX-License-Identifier: Apache-2.0 from contextlib import ExitStack import pyro.distributions as dist from pyro import poutine from pyro.infer.autoguide.guides import AutoStructured from pyro.poutine.plate_messenger import block_plate from .reparam import Reparam cl...
[ "pyro.poutine.reparam", "pyro.poutine.get_mask", "pyro.poutine.condition", "pyro.poutine.mask", "pyro.poutine.plate_messenger.block_plate", "contextlib.ExitStack", "pyro.distributions.Delta" ]
[((2200, 2248), 'pyro.poutine.reparam', 'poutine.reparam', (['fn'], {'config': 'self._reparam_config'}), '(fn, config=self._reparam_config)\n', (2215, 2248), False, 'from pyro import poutine\n'), ((3100, 3118), 'pyro.poutine.get_mask', 'poutine.get_mask', ([], {}), '()\n', (3116, 3118), False, 'from pyro import poutine...
import cdd import numpy as np class Cone_on_Plane(object): """Attributes: Quadratic_Form (list of lists, 2x2 matrix): The quadratic form of the cone. Linear_Form (list, 2x1 vector): The linear form of the matrix. """ def __init__(self, a, b, plane_flag): self.Quadratic_Form = [[-2*n...
[ "numpy.cos", "numpy.sin", "numpy.shape", "cdd.Polyhedron", "cdd.Matrix" ]
[((1886, 1931), 'cdd.Matrix', 'cdd.Matrix', (['cdd_vertices'], {'number_type': '"""float"""'}), "(cdd_vertices, number_type='float')\n", (1896, 1931), False, 'import cdd\n'), ((1983, 2002), 'cdd.Polyhedron', 'cdd.Polyhedron', (['mat'], {}), '(mat)\n', (1997, 2002), False, 'import cdd\n'), ((2791, 2836), 'cdd.Matrix', '...
""" create_model_input.py Create Model input objects for use in tests. """ import os import seekr2.modules.common_base as base import seekr2.modules.common_prepare as common_prepare import seekr2.modules.common_cv as common_cv import seekr2.modules.filetree as filetree import seekr2.modules.check as check # Don't r...
[ "seekr2.prepare.prepare", "seekr2.modules.common_base.Amber_params", "seekr2.modules.common_cv.Spherical_cv_input", "os.path.exists", "seekr2.modules.common_cv.RMSD_cv_input", "seekr2.modules.common_prepare.Model_input", "seekr2.modules.common_prepare.MMVT_input_settings", "seekr2.modules.common_prepa...
[((433, 458), 'os.path.dirname', 'os.path.dirname', (['__file__'], {}), '(__file__)\n', (448, 458), False, 'import os\n'), ((477, 510), 'os.path.dirname', 'os.path.dirname', (['prepare.__file__'], {}), '(prepare.__file__)\n', (492, 510), False, 'import os\n'), ((623, 642), 'seekr2.modules.common_base.Amber_params', 'ba...
# Copyright (c) Facebook, Inc. and its affiliates. # # This source code is licensed under the MIT license found in the # LICENSE file in the root directory of this source tree. import torch from fairseq import utils from fairseq.data import LanguagePairDataset, TransformEosLangPairDataset from . import register_task ...
[ "logging.getLogger", "fairseq.utils.split_paths", "fairseq.data.LanguagePairDataset", "torch.no_grad", "fairseq_cli.generate.get_symbols_to_strip_from_output", "sacrebleu.corpus_bleu" ]
[((633, 660), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (650, 660), False, 'import logging\n'), ((2816, 2849), 'fairseq.utils.split_paths', 'utils.split_paths', (['self.args.data'], {}), '(self.args.data)\n', (2833, 2849), False, 'from fairseq import utils\n'), ((6135, 6268), 'fairse...
## This script will grab population and language data from Census Bureau ACS API and output to Geodatabase of user's choice. Designed as a ArcGIS script tool. ## Developed by <NAME> for Wisconsin Emergency Management Sept 2017 #import modules try: import sys, urllib2, json, string, csv, arcpy, os, socket ...
[ "string.join", "arcpy.GetSystemEnvironment", "arcpy.AddField_management", "arcpy.da.SearchCursor", "arcpy.CopyRows_management", "arcpy.ListFields", "sys.exit", "arcpy.Delete_management", "socket.setdefaulttimeout", "arcpy.da.UpdateCursor", "arcpy.CopyFeatures_management", "urllib2.urlopen", ...
[((6115, 6148), 'socket.setdefaulttimeout', 'socket.setdefaulttimeout', (['timeout'], {}), '(timeout)\n', (6139, 6148), False, 'import sys, urllib2, json, string, csv, arcpy, os, socket\n'), ((6221, 6248), 'arcpy.GetParameterAsText', 'arcpy.GetParameterAsText', (['(0)'], {}), '(0)\n', (6245, 6248), False, 'import sys, ...
import gitlab import gitlab.v4.objects as glo import zenodo_gitlab.const as const import typing as ty import zenodo_gitlab.base as base class ZenodoGitLab: """ Main class to handle the interactions between GitLab repositories and Zenodo. """ def __init__(self, *, gitlab_instance: gitlab.Gitlab, zenod...
[ "zenodo_gitlab.base.new_deposit", "zenodo_gitlab.base.get_deposition_data", "zenodo_gitlab.base.get_depositions", "zenodo_gitlab.base.publish_deposition", "zenodo_gitlab.base.extract_archive_name_from_url", "zenodo_gitlab.base.get_archive_url", "zenodo_gitlab.base.upload_archive", "zenodo_gitlab.base....
[((1774, 1881), 'zenodo_gitlab.base.new_deposit', 'base.new_deposit', ([], {'zen_token': 'self.zenodo_token', 'zen_base_url': 'self.base_url', 'deposition_metadata': 'metadata'}), '(zen_token=self.zenodo_token, zen_base_url=self.base_url,\n deposition_metadata=metadata)\n', (1790, 1881), True, 'import zenodo_gitlab....
"""Test for version file syntax.""" import re from transpose_dict.__version__ import __version__ def test_version(): pattern = re.compile(r"\d+\.\d+\.\d+") assert pattern.match(__version__)
[ "re.compile" ]
[((134, 166), 're.compile', 're.compile', (['"""\\\\d+\\\\.\\\\d+\\\\.\\\\d+"""'], {}), "('\\\\d+\\\\.\\\\d+\\\\.\\\\d+')\n", (144, 166), False, 'import re\n')]
# coding: utf-8 # import sys,os,os.path # os.environ['CUDA_VISIBLE_DEVICES']='' import json import os import sys from collections import Counter import tqdm import numpy as np import tensorflow as tf from tensorflow.contrib import layers from tensorflow.python.ops import array_ops import random import pickle import ed...
[ "tensorflow.local_variables_initializer", "tensorflow.shape", "numpy.argsort", "numpy.array", "tensorflow.contrib.layers.bias_add", "tensorflow.reduce_mean", "tensorflow.nn.embedding_lookup", "tensorflow.slice", "tensorflow.placeholder", "tensorflow.concat", "tensorflow.nn.pool", "tensorflow.C...
[((3406, 3450), 'tensorflow.placeholder', 'tf.placeholder', (['tf.int32'], {'shape': '[None, None]'}), '(tf.int32, shape=[None, None])\n', (3420, 3450), True, 'import tensorflow as tf\n'), ((3464, 3508), 'tensorflow.placeholder', 'tf.placeholder', (['tf.int32'], {'shape': '[None, None]'}), '(tf.int32, shape=[None, None...
from django.conf.urls import url from .views import index , detail urlpatterns = [ url(r'^$', view=index), #/bookr/22/ url(r'^(?P<book_id>[0-9]+)/$',view = detail) ]
[ "django.conf.urls.url" ]
[((91, 112), 'django.conf.urls.url', 'url', (['"""^$"""'], {'view': 'index'}), "('^$', view=index)\n", (94, 112), False, 'from django.conf.urls import url\n'), ((135, 177), 'django.conf.urls.url', 'url', (['"""^(?P<book_id>[0-9]+)/$"""'], {'view': 'detail'}), "('^(?P<book_id>[0-9]+)/$', view=detail)\n", (138, 177), Fal...
from setuptools import setup, find_packages setup( name="configtune", version="0.0.8", description="A package to tune parameters with config files: designed with machine learning hyperparameter tuning in mind.", long_description=None, url= "https://github.com/orionw/configtune", author="<NAME> and <NAME>", auth...
[ "setuptools.find_packages" ]
[((589, 604), 'setuptools.find_packages', 'find_packages', ([], {}), '()\n', (602, 604), False, 'from setuptools import setup, find_packages\n')]
# The MIT License (MIT) # Copyright © 2021 <NAME> # Permission is hereby granted, free of charge, to any person obtaining a copy of this software and associated # documentation files (the “Software”), to deal in the Software without restriction, including without limitation # the rights to use, copy, modify, merge, ...
[ "substrateinterface.utils.ss58.ss58_encode", "bittensor.Keypair.create_from_seed", "os.path.exists", "json.dumps", "bittensor.Keypair", "getpass.getpass", "os.chmod", "ansible_vault.Vault", "os.path.expanduser", "json.loads", "os.access", "os.path.isfile", "os.path.dirname", "cryptography....
[((4706, 4771), 'password_strength.PasswordPolicy.from_names', 'PasswordPolicy.from_names', ([], {'strength': '(0.2)', 'entropybits': '(10)', 'length': '(6)'}), '(strength=0.2, entropybits=10, length=6)\n', (4731, 4771), False, 'from password_strength import PasswordPolicy\n'), ((5118, 5159), 'getpass.getpass', 'getpas...
from setuptools import setup, find_packages def scm_version(): def local_scheme(version): if version.tag and not version.distance: return version.format_with("") else: return version.format_choice("+{node}", "+{node}.dirty") return { "relative_to": __file__, ...
[ "setuptools.find_packages", "setuptools_scm.git.parse" ]
[((545, 559), 'setuptools_scm.git.parse', 'parse_git', (['"""."""'], {}), "('.')\n", (554, 559), True, 'from setuptools_scm.git import parse as parse_git\n'), ((1555, 1598), 'setuptools.find_packages', 'find_packages', ([], {'exclude': "('tests', 'tests.*')"}), "(exclude=('tests', 'tests.*'))\n", (1568, 1598), False, '...
import os import re FILE_BEGIN = 0 FILE_END = 1 LINE = 2 LABEL = 3 CODE = 4 COMMENT = 5 PREPROC = 6 SECTION = 7 GLOBAL = 8 FINISH = 9 # Yield tuples as files in a project are scanned: # (FILE_BEGIN, full_path) # (FILE_END, full_path) # (LINE, line_num, line) # (LABEL, label) # (CODE, code) # (COM...
[ "re.findall", "os.path.join", "os.walk" ]
[((509, 526), 'os.walk', 'os.walk', (['rootpath'], {}), '(rootpath)\n', (516, 526), False, 'import os\n'), ((571, 602), 'os.path.join', 'os.path.join', (['dirname', 'filename'], {}), '(dirname, filename)\n', (583, 602), False, 'import os\n'), ((1696, 1723), 're.findall', 're.findall', (['"""[\\\\w\']+"""', 'code'], {})...
import sys import os own_dir = os.path.abspath(os.path.dirname(__name__)) repo_root = os.path.abspath(os.path.join(own_dir, os.path.pardir)) sys.path.insert(1, repo_root)
[ "os.path.dirname", "sys.path.insert", "os.path.join" ]
[((143, 172), 'sys.path.insert', 'sys.path.insert', (['(1)', 'repo_root'], {}), '(1, repo_root)\n', (158, 172), False, 'import sys\n'), ((48, 73), 'os.path.dirname', 'os.path.dirname', (['__name__'], {}), '(__name__)\n', (63, 73), False, 'import os\n'), ((103, 140), 'os.path.join', 'os.path.join', (['own_dir', 'os.path...
import math from ..lib import Image def histogram_entropy(im): """ Calculate the entropy of an images' histogram. Used for "smart cropping" in easy-thumbnails; see: https://raw.github.com/SmileyChris/easy-thumbnails/master/easy_thumbnails/utils.py """ if not isinstance(im, Image.Image): r...
[ "math.log" ]
[((521, 535), 'math.log', 'math.log', (['p', '(2)'], {}), '(p, 2)\n', (529, 535), False, 'import math\n')]
from time import time def cache(function): memo = {} time_to_cache = 60 def wrapper(*args): if args in memo and time() < memo[args]['time'] + time_to_cache: return memo[args]['record'] else: rv = function(*args) memo[args] = { 'record': r...
[ "time.time" ]
[((134, 140), 'time.time', 'time', ([], {}), '()\n', (138, 140), False, 'from time import time\n'), ((347, 353), 'time.time', 'time', ([], {}), '()\n', (351, 353), False, 'from time import time\n')]
import numpy as np import pytest from docarray import DocumentArray, Document from docarray.array.qdrant import DocumentArrayQdrant from docarray.array.sqlite import DocumentArraySqlite from docarray.array.annlite import DocumentArrayAnnlite, AnnliteConfig from docarray.array.storage.qdrant import QdrantConfig from do...
[ "docarray.array.annlite.AnnliteConfig", "docarray.array.storage.qdrant.QdrantConfig", "pytest.mark.parametrize", "docarray.Document", "pytest.fixture", "docarray.array.storage.weaviate.WeaviateConfig", "numpy.testing.assert_array_equal" ]
[((433, 465), 'pytest.fixture', 'pytest.fixture', ([], {'scope': '"""function"""'}), "(scope='function')\n", (447, 465), False, 'import pytest\n'), ((611, 657), 'pytest.mark.parametrize', 'pytest.mark.parametrize', (['"""min_freq"""', '[1, 2, 3]'], {}), "('min_freq', [1, 2, 3])\n", (634, 657), False, 'import pytest\n')...
from __future__ import division from __future__ import absolute_import from __future__ import print_function import numpy as np import os class Memory(object): """ An implementation of the replay memory. This is essential when dealing with DRL algorithms that are not multi-threaded as in A3C. """ ...
[ "os.path.exists", "os.makedirs", "os.path.join", "numpy.random.randint", "numpy.empty" ]
[((721, 763), 'numpy.empty', 'np.empty', ([], {'shape': '((memory_size,) + state_dim)'}), '(shape=(memory_size,) + state_dim)\n', (729, 763), True, 'import numpy as np\n'), ((812, 854), 'numpy.empty', 'np.empty', ([], {'shape': '((memory_size,) + state_dim)'}), '(shape=(memory_size,) + state_dim)\n', (820, 854), True, ...
from catalyst import utils from catalyst.data.cv import BlurMixin, FlareMixin, RotateMixin jpg_rgb_uri = ( "https://raw.githubusercontent.com/catalyst-team/catalyst-pics/master" "/test_images/catalyst_icon.jpg" ) image = utils.imread(jpg_rgb_uri) def test_blur_mixin(): """@TODO: Docs. Contribution is we...
[ "catalyst.utils.imread", "catalyst.data.cv.FlareMixin", "catalyst.data.cv.RotateMixin", "catalyst.data.cv.BlurMixin" ]
[((231, 256), 'catalyst.utils.imread', 'utils.imread', (['jpg_rgb_uri'], {}), '(jpg_rgb_uri)\n', (243, 256), False, 'from catalyst import utils\n'), ((390, 401), 'catalyst.data.cv.BlurMixin', 'BlurMixin', ([], {}), '()\n', (399, 401), False, 'from catalyst.data.cv import BlurMixin, FlareMixin, RotateMixin\n'), ((808, 8...
import numpy as np from proteus import Domain, Context, Comm from proteus.mprans import SpatialTools as st import proteus.TwoPhaseFlow.TwoPhaseFlowProblem as TpFlow from proteus import WaveTools as wt from proteus.Profiling import logEvent from proteus.mbd import CouplingFSI as fsi import os import pychrono rho_0 = 99...
[ "proteus.TwoPhaseFlow.TwoPhaseFlowProblem.OutputStepping", "proteus.mbd.CouplingFSI.ProtChSystem", "proteus.ctransportCoefficients.smoothedHeaviside_integral", "proteus.TwoPhaseFlow.TwoPhaseFlowProblem.TwoPhaseFlowProblem", "proteus.ctransportCoefficients.smoothedHeaviside", "proteus.mprans.SpatialTools.T...
[((728, 765), 'proteus.Domain.PiecewiseLinearComplexDomain', 'Domain.PiecewiseLinearComplexDomain', ([], {}), '()\n', (763, 765), False, 'from proteus import Domain, Context, Comm\n'), ((805, 832), 'proteus.mprans.SpatialTools.Tank3D', 'st.Tank3D', (['domain', 'tank_dim'], {}), '(domain, tank_dim)\n', (814, 832), True,...
import smtplib from email import encoders from email.mime.text import MIMEText from email.mime.base import MIMEBase from email.mime.multipart import MIMEMultipart sender_email=input("enter sender email") reciver_emil=input("add reciver email") server=smtplib.SMTP_SSL('smtp.gmail.com',465) # server.connect("smtp.gmail....
[ "email.mime.base.MIMEBase", "smtplib.SMTP_SSL", "email.encoders.encode_base64", "email.mime.multipart.MIMEMultipart", "email.mime.text.MIMEText" ]
[((252, 291), 'smtplib.SMTP_SSL', 'smtplib.SMTP_SSL', (['"""smtp.gmail.com"""', '(465)'], {}), "('smtp.gmail.com', 465)\n", (268, 291), False, 'import smtplib\n'), ((460, 475), 'email.mime.multipart.MIMEMultipart', 'MIMEMultipart', ([], {}), '()\n', (473, 475), False, 'from email.mime.multipart import MIMEMultipart\n')...
import json import os import sys from unittest import mock import gemmi from relion.cryolo_relion_it import mask_soft_edge_external_job @mock.patch("relion.cryolo_relion_it.mask_soft_edge_external_job.subprocess") def test_mask_soft_edge_main(mock_subprocess, tmpdir): # Prepare things os.chdir(tmpdir) c...
[ "unittest.mock.call", "gemmi.cif.Document", "relion.cryolo_relion_it.mask_soft_edge_external_job.main", "os.chdir", "os.path.isfile", "unittest.mock.patch", "json.dump" ]
[((141, 217), 'unittest.mock.patch', 'mock.patch', (['"""relion.cryolo_relion_it.mask_soft_edge_external_job.subprocess"""'], {}), "('relion.cryolo_relion_it.mask_soft_edge_external_job.subprocess')\n", (151, 217), False, 'from unittest import mock\n'), ((298, 314), 'os.chdir', 'os.chdir', (['tmpdir'], {}), '(tmpdir)\n...
#!/usr/bin/env python3 import os import sys import argparse import subprocess bash_output = lambda *s: subprocess.check_output(s, stderr=subprocess.STDOUT).decode() USE_ENV_BASH="""#!/bin/bash export VENV_RUNNING=1 alias pip=venv_pip_recording unvenv () { unset VENV_RUNNING unset -f unvenv unalias pip ...
[ "subprocess.check_output", "os.path.exists", "os.listdir", "argparse.ArgumentParser", "os.rmdir", "os.path.dirname", "collections.defaultdict", "os.system", "os.remove" ]
[((3467, 3574), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {'description': 'description', 'formatter_class': 'argparse.RawDescriptionHelpFormatter'}), '(description=description, formatter_class=argparse.\n RawDescriptionHelpFormatter)\n', (3490, 3574), False, 'import argparse\n'), ((1665, 1690), 'os....
import struct import datetime from simpledbf import Dbf5 class Dbf6(Dbf5): def _get_recs(self, chunk=None): '''Generator that returns individual records. Parameters ---------- chunk : int, optional Number of records to return as a single chunk. Default 'None', ...
[ "datetime.date" ]
[((3271, 3293), 'datetime.date', 'datetime.date', (['y', 'm', 'd'], {}), '(y, m, d)\n', (3284, 3293), False, 'import datetime\n')]
from playwright.sync_api import sync_playwright def crawler(username, password): with sync_playwright() as playwright: browser = playwright.chromium.launch() page = browser.new_page() page.goto("http://nginx") page.type('input[name="username"]', username) page.type('input[n...
[ "playwright.sync_api.sync_playwright" ]
[((92, 109), 'playwright.sync_api.sync_playwright', 'sync_playwright', ([], {}), '()\n', (107, 109), False, 'from playwright.sync_api import sync_playwright\n')]
# -*- coding: utf-8 -*- # # Copyright 2017 Gehirn Inc. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law...
[ "hmac.new", "hmac.compare_digest" ]
[((1690, 1725), 'hmac.compare_digest', 'hmac.compare_digest', (['signature', "b''"], {}), "(signature, b'')\n", (1709, 1725), False, 'import hmac\n'), ((2190, 2227), 'hmac.new', 'hmac.new', (['key', 'message', 'self.hash_fun'], {}), '(key, message, self.hash_fun)\n', (2198, 2227), False, 'import hmac\n')]
import kivy import pyodbc as pyodbc from kivy.app import App from kivy.uix.label import Label from kivy.uix.gridlayout import GridLayout from kivy.uix.textinput import TextInput from kivy.uix.button import Button from kivy.uix.widget import Widget from kivy.properties import StringProperty from kivy.properties import O...
[ "pyodbc.connect", "kivy.uix.label.Label", "kivy.uix.button.Button", "kivy.uix.textinput.TextInput" ]
[((687, 800), 'pyodbc.connect', 'pyodbc.connect', (['"""Trusted_Connection=yes"""'], {'driver': '"""{SQL Server}"""', 'server': '"""Server name"""', 'database': '"""FlowerShop"""'}), "('Trusted_Connection=yes', driver='{SQL Server}', server=\n 'Server name', database='FlowerShop')\n", (701, 800), True, 'import pyodb...
import matplotlib matplotlib.use('agg') import matplotlib.pyplot as plt import abc,os,pickle,sys import scipy.stats import numpy as np from datetime import datetime import tensorflow as tf from BatchIterator import PaddedDataIterator from generation import * from Plotter import get_intensity,get_integral,get_integral_...
[ "Utils.file2sequence", "tensorflow.shape", "matplotlib.pyplot.grid", "matplotlib.pyplot.ylabel", "numpy.linalg.norm", "tensorflow.reduce_mean", "tensorflow.set_random_seed", "tensorflow.GPUOptions", "tensorflow.log", "BatchIterator.PaddedDataIterator", "os.path.exists", "Utils.sequence2file", ...
[((19, 40), 'matplotlib.use', 'matplotlib.use', (['"""agg"""'], {}), "('agg')\n", (33, 40), False, 'import matplotlib\n'), ((1230, 1254), 'tensorflow.set_random_seed', 'tf.set_random_seed', (['SEED'], {}), '(SEED)\n', (1248, 1254), True, 'import tensorflow as tf\n'), ((1255, 1275), 'numpy.random.seed', 'np.random.seed'...
# Copyright (c) 2015, Frappe Technologies Pvt. Ltd. and Contributors # License: GNU General Public License v3. See license.txt from __future__ import unicode_literals import frappe from frappe.utils import cint, fmt_money, flt, nowdate, getdate no_cache = 1 no_sitemap = 1 def get_context(context): home...
[ "frappe.get_doc" ]
[((327, 374), 'frappe.get_doc', 'frappe.get_doc', (['"""Capital Care Website Settings"""'], {}), "('Capital Care Website Settings')\n", (341, 374), False, 'import frappe\n'), ((582, 628), 'frappe.get_doc', 'frappe.get_doc', (['"""Website Slideshow"""', 'slideshow'], {}), "('Website Slideshow', slideshow)\n", (596, 628)...
__author__ = "<NAME> <<EMAIL>>" import re from time import sleep from unicon.bases.routers.services import BaseService from unicon.plugins.generic.service_statements import reload_statement_list from unicon.plugins.generic.service_implementation import ReloadResult from unicon.eal.dialogs import Dialog from unicon.c...
[ "unicon.utils.AttributeDict", "unicon.eal.dialogs.Dialog", "unicon.core.errors.SubCommandFailure", "time.sleep" ]
[((1519, 1574), 'unicon.eal.dialogs.Dialog', 'Dialog', (['(reload_statement_list + [boot_from_rommon_stmt])'], {}), '(reload_statement_list + [boot_from_rommon_stmt])\n', (1525, 1574), False, 'from unicon.eal.dialogs import Dialog\n'), ((1677, 1687), 'unicon.eal.dialogs.Dialog', 'Dialog', (['[]'], {}), '([])\n', (1683,...
from django.apps import AppConfig from django.utils.translation import gettext_lazy as _ class ViscometersConfig(AppConfig): name = 'scieio.viscometers' verbose_name = _("Viscometers and Rheometers")
[ "django.utils.translation.gettext_lazy" ]
[((178, 209), 'django.utils.translation.gettext_lazy', '_', (['"""Viscometers and Rheometers"""'], {}), "('Viscometers and Rheometers')\n", (179, 209), True, 'from django.utils.translation import gettext_lazy as _\n')]
import math import cv2 import numpy as np import tensorflow as tf from scipy import ndimage import os from MNIST_data import input_data class Recognizer: def __init__(self): pass @staticmethod def shift(img, sx, sy): rows, cols = img.shape M = np.float32([[1, 0, sx], [0, 1, sy]])...
[ "math.floor", "numpy.lib.pad", "scipy.ndimage.measurements.center_of_mass", "tensorflow.cast", "tensorflow.log", "os.remove", "MNIST_data.input_data.read_data_sets", "cv2.threshold", "numpy.delete", "tensorflow.placeholder", "tensorflow.Session", "tensorflow.matmul", "numpy.round", "tensor...
[((284, 320), 'numpy.float32', 'np.float32', (['[[1, 0, sx], [0, 1, sy]]'], {}), '([[1, 0, sx], [0, 1, sy]])\n', (294, 320), True, 'import numpy as np\n'), ((339, 375), 'cv2.warpAffine', 'cv2.warpAffine', (['img', 'M', '(cols, rows)'], {}), '(img, M, (cols, rows))\n', (353, 375), False, 'import cv2\n'), ((462, 502), 's...
import matplotlib.pyplot as plt from ._compat import * class GridCurves(object): def __init__(self, curves): self._curves = curves @property def curves(self): return self._curves def __str__(self): fh = StringIO() fh.write('gridcurves objects:\n\n') fh.write(...
[ "matplotlib.pyplot.plot" ]
[((531, 576), 'matplotlib.pyplot.plot', 'plt.plot', (['curve.real', 'curve.imag'], {'color': '"""0.6"""'}), "(curve.real, curve.imag, color='0.6')\n", (539, 576), True, 'import matplotlib.pyplot as plt\n')]
#!/usr/bin/env python3 from __future__ import annotations import unittest import warnings import torch from linear_operator.test.base_test_case import BaseTestCase from linear_operator.utils.cholesky import psd_safe_cholesky from linear_operator.utils.errors import NanError from linear_operator.utils.warnings impor...
[ "linear_operator.utils.cholesky.psd_safe_cholesky", "torch.eye", "torch.cholesky", "warnings.catch_warnings", "torch.empty_like", "torch.tensor", "warnings.simplefilter", "unittest.main", "torch.allclose", "torch.arange", "torch.device" ]
[((4350, 4365), 'unittest.main', 'unittest.main', ([], {}), '()\n', (4363, 4365), False, 'import unittest\n'), ((459, 531), 'torch.tensor', 'torch.tensor', (['[[[0.25, -0.75], [-0.75, 2.25]], [[1.0, 1.0], [1.0, 1.0]]]'], {}), '([[[0.25, -0.75], [-0.75, 2.25]], [[1.0, 1.0], [1.0, 1.0]]])\n', (471, 531), False, 'import t...
from pathlib import Path from beet import BlockTag, DataPack, Function, FunctionTag, JsonFile, Structure def test_equality(): assert DataPack() == DataPack() assert DataPack("hello") == DataPack("hello") assert DataPack("hello") != DataPack("world") p1 = DataPack("foo", mcmeta=JsonFile({"pack": {"de...
[ "beet.Function", "beet.FunctionTag", "beet.BlockTag", "beet.JsonFile", "beet.DataPack" ]
[((709, 719), 'beet.DataPack', 'DataPack', ([], {}), '()\n', (717, 719), False, 'from beet import BlockTag, DataPack, Function, FunctionTag, JsonFile, Structure\n'), ((729, 739), 'beet.DataPack', 'DataPack', ([], {}), '()\n', (737, 739), False, 'from beet import BlockTag, DataPack, Function, FunctionTag, JsonFile, Stru...
# -*- coding: utf-8 -*- # Copyright 2017 <NAME> # Copyright 2018 <NAME> # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Un...
[ "ccpi.viewer.CILViewer2D.ViewerEventManager", "vtk.vtkDecimatePro", "vtk.vtkImageMapToWindowLevelColors", "vtk.vtkTextMapper", "vtk.vtkTextProperty", "vtk.vtkPNGWriter", "ccpi.viewer.utils.colormaps.CILColorMaps.get_color_transfer_function", "vtk.vtkImageActor", "vtk.vtkInteractorStyleTrackballCamer...
[((1180, 1232), 'vtk.vtkInteractorStyleTrackballCamera.__init__', 'vtk.vtkInteractorStyleTrackballCamera.__init__', (['self'], {}), '(self)\n', (1226, 1232), False, 'import vtk\n'), ((9408, 9429), 'vtk.vtkTextProperty', 'vtk.vtkTextProperty', ([], {}), '()\n', (9427, 9429), False, 'import vtk\n'), ((9809, 9828), 'vtk.v...
import os, json from django.core.exceptions import ImproperlyConfigured BASE_DIR = os.path.dirname(os.path.dirname(os.path.abspath(__file__))) ROOT_DIR = os.path.dirname(BASE_DIR) INTERNAL_IPS = ['127.0.0.1',] ROOT_URLCONF = 'djangotube.urls' TEMPLATES = [ { 'BACKEND': 'django.template.backends.django...
[ "os.path.abspath", "os.path.dirname" ]
[((156, 181), 'os.path.dirname', 'os.path.dirname', (['BASE_DIR'], {}), '(BASE_DIR)\n', (171, 181), False, 'import os, json\n'), ((117, 142), 'os.path.abspath', 'os.path.abspath', (['__file__'], {}), '(__file__)\n', (132, 142), False, 'import os, json\n')]
from random import randint import config class wall: """ Stores the state data for a wall on the map """ def __init__(self): """ Randomly generates a wall using the bounding values in config.py """ # Set lengths for the long and short sides of the wall longSide ...
[ "random.randint" ]
[((322, 401), 'random.randint', 'randint', (['config.game.wall.longSideBounds[0]', 'config.game.wall.longSideBounds[1]'], {}), '(config.game.wall.longSideBounds[0], config.game.wall.longSideBounds[1])\n', (329, 401), False, 'from random import randint\n'), ((422, 508), 'random.randint', 'randint', (['config.game.wall.s...
import json import os import time import urllib.parse import urllib.request import pandas as pd import pathlib #ルートディレクトリの設定(dataディレクトリ ) root_dir = pathlib.Path(__file__).parent.parent #都道府県一覧の取得 c = os.path.join(root_dir, 'codes/preflist.json') with open(c) as j: prefList = json.load(j) #市区町村一覧の取得 c = os.path....
[ "os.getenv", "pathlib.Path", "os.path.join", "dotenv.load_dotenv", "json.load", "json.dump" ]
[((203, 248), 'os.path.join', 'os.path.join', (['root_dir', '"""codes/preflist.json"""'], {}), "(root_dir, 'codes/preflist.json')\n", (215, 248), False, 'import os\n'), ((312, 357), 'os.path.join', 'os.path.join', (['root_dir', '"""codes/citylist.json"""'], {}), "(root_dir, 'codes/citylist.json')\n", (324, 357), False,...
from django.urls import path from .views import * from .reportes import reporte_compras urlpatterns = [ path('proveedores/', ProveedorView.as_view(), name="proveedor_list"), path('proveedores/new', ProveedorNew.as_view(), name="proveedor_new"), path('proveedores/edit/<int:pk>', ProveedorEdit.as_vi...
[ "django.urls.path" ]
[((354, 432), 'django.urls.path', 'path', (['"""proveedor/inactivar/<int:id>"""', 'proveedorInactivar'], {'name': '"""proveedor_ina"""'}), "('proveedor/inactivar/<int:id>', proveedorInactivar, name='proveedor_ina')\n", (358, 432), False, 'from django.urls import path\n'), ((521, 570), 'django.urls.path', 'path', (['"""...
from functools import reduce import urwid from ui_interface import UIInterface from feed import Feed from episode import Episode from urwid_widgets import( SelectionList, PackableLineBox, EditDialogue, InformationDialogue, ) from player import Player class UI(UIInterface): palette = ( ('...
[ "urwid.Columns", "urwid_widgets.PackableLineBox", "feed.Feed.add", "player.Player", "urwid_widgets.SelectionList", "urwid.LineBox", "urwid.SolidFill", "urwid.AsyncioEventLoop", "feed.Feed.getall", "urwid_widgets.InformationDialogue", "urwid.connect_signal", "urwid.signals.emit_signal", "feed...
[((1030, 1090), 'urwid.Columns', 'urwid.Columns', (["(('pack', self.feeds_box), self.episodes_box)"], {}), "((('pack', self.feeds_box), self.episodes_box))\n", (1043, 1090), False, 'import urwid\n'), ((1375, 1440), 'urwid.signals.emit_signal', 'urwid.signals.emit_signal', (['self.feeds_list.listwalker', '"""modified"""...
from starter_two_three_tree import TwoThreeTree, Node import unittest class NodeTest(unittest.TestCase): def test_init(self): data = 1 node = Node(data) assert node.data[0] == data assert len(node.data) == 1 assert len(node.children) == 0 assert node.parent is None...
[ "unittest.main", "starter_two_three_tree.TwoThreeTree", "starter_two_three_tree.Node" ]
[((9497, 9512), 'unittest.main', 'unittest.main', ([], {}), '()\n', (9510, 9512), False, 'import unittest\n'), ((165, 175), 'starter_two_three_tree.Node', 'Node', (['data'], {}), '(data)\n', (169, 175), False, 'from starter_two_three_tree import TwoThreeTree, Node\n'), ((400, 411), 'starter_two_three_tree.Node', 'Node'...
# from sqlite3 import connect from numpy import array import re from metfrag2 import fragmenter_rank def main(): con = connect('replicate_data_assigned_MS2.db') cur = con.cursor() pat = re.compile(r'_met[0-9]+') cmpds, ranks = [], [] qry = """SELECT cmpd, well, spectrum, adduct, monoiso, formula...
[ "metfrag2.fragmenter_rank", "sqlite3.connect", "re.compile" ]
[((127, 168), 'sqlite3.connect', 'connect', (['"""replicate_data_assigned_MS2.db"""'], {}), "('replicate_data_assigned_MS2.db')\n", (134, 168), False, 'from sqlite3 import connect\n'), ((202, 226), 're.compile', 're.compile', (['"""_met[0-9]+"""'], {}), "('_met[0-9]+')\n", (212, 226), False, 'import re\n'), ((1189, 122...
from flask import Blueprint,request,jsonify import requests import json url = 'https://flow-api.fluctuo.com/v1?access_token=eFwW53d4EiCdlmbrs8muiM5gSvOrhPwh' vehicles_info_bp = Blueprint('vehicles_info',__name__) @vehicles_info_bp.route('/vehicles_info', methods=['post']) def vehicles_info(): try: ...
[ "json.dumps", "flask.request.get_json", "flask.Blueprint", "flask.jsonify" ]
[((185, 221), 'flask.Blueprint', 'Blueprint', (['"""vehicles_info"""', '__name__'], {}), "('vehicles_info', __name__)\n", (194, 221), False, 'from flask import Blueprint, request, jsonify\n'), ((334, 352), 'flask.request.get_json', 'request.get_json', ([], {}), '()\n', (350, 352), False, 'from flask import Blueprint, r...
#!/usr/bin/env python3 import re import fire import requests import logging as logger CONFIG = { 'SERVER_ADDRESS': 'http://localhost', 'SERVER_PORT': 5050 } def interact(): """ Interactively call the model's server """ address = f"{CONFIG['SERVER_ADDRESS']}:{CONFIG['SERVER_PORT']}" wh...
[ "fire.Fire", "requests.get", "requests.head", "re.findall", "logging.info", "logging.error" ]
[((1697, 1716), 'fire.Fire', 'fire.Fire', (['interact'], {}), '(interact)\n', (1706, 1716), False, 'import fire\n'), ((712, 733), 'logging.info', 'logger.info', (['raw_text'], {}), '(raw_text)\n', (723, 733), True, 'import logging as logger\n'), ((750, 791), 're.findall', 're.findall', (['"""\\\\<\\\\|(.+?)\\\\|\\\\>""...
from __future__ import (absolute_import, division, print_function) __metaclass__ = type DOCUMENTATION = ''' name: apt_keys author: Manala (@manala) short_description: returns a curated keys list description: - Takes a keys list and returns it curated. ''' from ansible.plugins.lookup import Looku...
[ "ansible.errors.AnsibleError" ]
[((1380, 1412), 'ansible.errors.AnsibleError', 'AnsibleError', (['"""Missing "id" key"""'], {}), '(\'Missing "id" key\')\n', (1392, 1412), False, 'from ansible.errors import AnsibleError\n')]
from setuptools import setup, Extension import numpy as np from Cython.Build import cythonize from Cython.Distutils import build_ext from torch.utils.cpp_extension import BuildExtension, CUDAExtension # Obtain the numpy include directory. This logic works across numpy versions. try: numpy_include = np.get_includ...
[ "setuptools.Extension", "Cython.Build.cythonize", "numpy.get_numpy_include", "numpy.get_include" ]
[((307, 323), 'numpy.get_include', 'np.get_include', ([], {}), '()\n', (321, 323), True, 'import numpy as np\n'), ((498, 559), 'setuptools.Extension', 'Extension', (['"""nms_cpu"""'], {'sources': "['src/nms_cpu.cpp']"}), "('nms_cpu', sources=['src/nms_cpu.cpp'], **ext_args)\n", (507, 559), False, 'from setuptools impor...
from expertise.create_dataset import OpenReviewExpertise from unittest.mock import patch, MagicMock from collections import defaultdict import openreview import json import random from pathlib import Path import sys import pytest import os import time import numpy as np import shutil import expertise.service from exper...
[ "json.dumps", "time.sleep", "shutil.rmtree", "os.path.isfile", "os.path.isdir", "time.time", "pytest.fixture", "expertise.create_dataset.OpenReviewExpertise", "expertise.service.utils.mock_client", "os.remove" ]
[((509, 540), 'pytest.fixture', 'pytest.fixture', ([], {'scope': '"""session"""'}), "(scope='session')\n", (523, 540), False, 'import pytest\n'), ((965, 996), 'pytest.fixture', 'pytest.fixture', ([], {'scope': '"""session"""'}), "(scope='session')\n", (979, 996), False, 'import pytest\n'), ((1084, 1115), 'pytest.fixtur...
import numpy as np def rotX(theta): return np.array([[1, 0, 0] , [0, np.cos(theta), -np.sin(theta)] , [0, np.sin(theta), np.cos(theta)]]) def rotY(theta): return np.array([[np.cos(theta), 0, np.sin(theta)] , [0, 1, 0] , [-np.sin(theta...
[ "numpy.cross", "numpy.diag", "numpy.array", "numpy.dot", "numpy.outer", "numpy.cos", "numpy.linalg.norm", "numpy.sin" ]
[((630, 646), 'numpy.cross', 'np.cross', (['v1', 'v2'], {}), '(v1, v2)\n', (638, 646), True, 'import numpy as np\n'), ((928, 941), 'numpy.sin', 'np.sin', (['angle'], {}), '(angle)\n', (934, 941), True, 'import numpy as np\n'), ((953, 966), 'numpy.cos', 'np.cos', (['angle'], {}), '(angle)\n', (959, 966), True, 'import n...
# coding=utf-8 # Copyright 2019 The Google Research Authors. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicab...
[ "absl.app.UsageError", "tensorflow.compat.v2.io.gfile.makedirs", "uq_benchmark_2019.experiment_utils.record_config", "absl.flags.DEFINE_integer", "absl.logging.info", "uq_benchmark_2019.cifar.models_lib.build_and_train", "uq_benchmark_2019.cifar.data_lib.build_dataset", "absl.app.run", "uq_benchmark...
[((1251, 1336), 'absl.flags.DEFINE_enum', 'flags.DEFINE_enum', (['"""method"""', 'None', 'models_lib.METHODS', '"""Name of modeling method."""'], {}), "('method', None, models_lib.METHODS,\n 'Name of modeling method.')\n", (1268, 1336), False, 'from absl import flags\n'), ((1355, 1415), 'absl.flags.DEFINE_string', '...
from django.contrib.auth.models import User from django.db import models from django.template.defaultfilters import slugify class Category(models.Model): name=models.CharField(max_length=128,unique=type) views=models.IntegerField(default=0) likes=models.IntegerField(default=0) slug=models.SlugField() ...
[ "django.db.models.OneToOneField", "django.db.models.ForeignKey", "django.db.models.IntegerField", "django.db.models.ImageField", "django.template.defaultfilters.slugify", "django.db.models.SlugField", "django.db.models.URLField", "django.db.models.CharField" ]
[((163, 208), 'django.db.models.CharField', 'models.CharField', ([], {'max_length': '(128)', 'unique': 'type'}), '(max_length=128, unique=type)\n', (179, 208), False, 'from django.db import models\n'), ((218, 248), 'django.db.models.IntegerField', 'models.IntegerField', ([], {'default': '(0)'}), '(default=0)\n', (237, ...
# Resource object code (Python 3) # Created by: object code # Created by: The Resource Compiler for Qt version 5.15.2 # WARNING! All changes made in this file will be lost! from PySide2 import QtCore qt_resource_data = b"\ \x00\x00\x0f\xb6\ \x89\ PNG\x0d\x0a\x1a\x0a\x00\x00\x00\x0dIHDR\x00\ \x00\x00@\x00\x00\x00@\x08...
[ "PySide2.QtCore.qUnregisterResourceData", "PySide2.QtCore.qRegisterResourceData" ]
[((146918, 147009), 'PySide2.QtCore.qRegisterResourceData', 'QtCore.qRegisterResourceData', (['(3)', 'qt_resource_struct', 'qt_resource_name', 'qt_resource_data'], {}), '(3, qt_resource_struct, qt_resource_name,\n qt_resource_data)\n', (146946, 147009), False, 'from PySide2 import QtCore\n'), ((147039, 147132), 'PyS...
# -*- coding: utf-8 -*- # !/usr/bin/env python # Author: Mocobk # Data: 2019/03/18 # License: MIT # Modified by <NAME> # <EMAIL> # Modified At: 2020/11/15 import unittest import os from BeautifulReport import BeautifulReport report_dir = os.path.expanduser("~/") if __name__ == '__main__': test_suite = unittest.de...
[ "BeautifulReport.BeautifulReport", "os.path.expanduser", "unittest.defaultTestLoader.discover" ]
[((239, 263), 'os.path.expanduser', 'os.path.expanduser', (['"""~/"""'], {}), "('~/')\n", (257, 263), False, 'import os\n'), ((309, 380), 'unittest.defaultTestLoader.discover', 'unittest.defaultTestLoader.discover', (['"""./tests"""'], {'pattern': '"""test_demo*.py"""'}), "('./tests', pattern='test_demo*.py')\n", (344,...
import torch from torch import nn def conv_bn(in_c, out_c, kernel_size, stride=1, padding=0, groups=1): return nn.Sequential( nn.Conv2d(in_c, out_c, kernel_size, stride, padding, groups=groups), nn.BatchNorm2d(out_c) ) # Used to upscale the image (height and width) after global average pooling def scale_dim(...
[ "torch.nn.BatchNorm2d", "torch.nn.Sigmoid", "torch.nn.Dropout", "torch.randperm", "torch.nn.Sequential", "torch.nn.Flatten", "torch.nn.Conv2d", "torch.nn.SiLU", "torch.matmul", "torch.nn.AdaptiveAvgPool2d", "torch.nn.Linear", "torch.nn.AvgPool2d", "torch.nn.LogSoftmax", "torch.cat" ]
[((131, 198), 'torch.nn.Conv2d', 'nn.Conv2d', (['in_c', 'out_c', 'kernel_size', 'stride', 'padding'], {'groups': 'groups'}), '(in_c, out_c, kernel_size, stride, padding, groups=groups)\n', (140, 198), False, 'from torch import nn\n'), ((202, 223), 'torch.nn.BatchNorm2d', 'nn.BatchNorm2d', (['out_c'], {}), '(out_c)\n', ...
import matplotlib.pyplot as plt radix_len = [120, 110, 100, 90, 80, 70, 60, 50, 40, 30, 20] radix_time = [82.941, 71.651, 60.559, 47.864, 37.66, 29.485, 21.827, 14.785, 9.637, 5.537, 2.694] suffix_len = [19, 24, 37, 52, 66, 81, 90, 120] suffix_time = [0.241, 0.242, 0.244, 0.245, 0.246, 0.248, 0.250, 0.251] fig, ax =...
[ "matplotlib.pyplot.xlabel", "matplotlib.pyplot.subplots", "matplotlib.pyplot.show", "matplotlib.pyplot.ylabel" ]
[((321, 335), 'matplotlib.pyplot.subplots', 'plt.subplots', ([], {}), '()\n', (333, 335), True, 'import matplotlib.pyplot as plt\n'), ((383, 431), 'matplotlib.pyplot.xlabel', 'plt.xlabel', (['"""Document size, Kbytes"""'], {'fontsize': '(12)'}), "('Document size, Kbytes', fontsize=12)\n", (393, 431), True, 'import matp...
#<NAME> from sklearn import preprocessing import matplotlib.pyplot as plt import numpy as np from sklearn.cluster import KMeans # Data set # Datos de 13/02/2020 a 10/11/2020 # De acuerdo con: https://tablerocovid.mspas.gob.gt/ # Regiones de acuerdo con: https://aprende.guatemala.com/historia/geografia/regiones-de-gua...
[ "sklearn.cluster.KMeans", "sklearn.preprocessing.LabelEncoder", "matplotlib.pyplot.ylabel", "matplotlib.pyplot.xlabel", "numpy.array", "matplotlib.pyplot.scatter", "matplotlib.pyplot.title", "matplotlib.pyplot.show" ]
[((1468, 1496), 'sklearn.preprocessing.LabelEncoder', 'preprocessing.LabelEncoder', ([], {}), '()\n', (1494, 1496), False, 'from sklearn import preprocessing\n'), ((1591, 2612), 'numpy.array', 'np.array', (['[[region_guatemala2[0], no_fallecidos[0]], [region_guatemala2[1],\n no_fallecidos[1]], [region_guatemala2[2],...
from textwrap import dedent import attack_flow.graphviz def test_convert_attack_flow_to_dot(): flow = { "actions": [ { "id": "action1", "name": "action-one", }, { "id": "action2", "name": "action-two", ...
[ "textwrap.dedent" ]
[((853, 1187), 'textwrap.dedent', 'dedent', (['""" digraph {\n node [shape=box,style="rounded,filled,fixedsize=true,width=2,height=1"]\n\n "action1" [fillcolor=pink,label="action-one"]\n "action2" [fillcolor=pink,label="action-two"]\n\n "asset1" [fillcolor=lightblue1]\n ...
from imutils import contours from collections import deque import numpy as np import argparse import imutils import cv2 from Tkinter import Frame, Tk, BOTH, Text, Menu, END import tkFileDialog import os.path import os import sys from time import gmtime, strftime def PT(F): Time = strftime("%Y-%...
[ "os.listdir", "collections.deque", "cv2.countNonZero", "argparse.ArgumentParser", "cv2.inRange", "Tkinter.Tk", "cv2.destroyWindow", "cv2.bitwise_and", "os.path.split", "os.path.isfile", "numpy.array", "cv2.waitKey", "cv2.destroyAllWindows", "cv2.VideoCapture", "cv2.cvtColor", "time.gmt...
[((9238, 9253), 'os.listdir', 'os.listdir', (['"""."""'], {}), "('.')\n", (9248, 9253), False, 'import os\n'), ((9371, 9394), 'cv2.destroyAllWindows', 'cv2.destroyAllWindows', ([], {}), '()\n', (9392, 9394), False, 'import cv2\n'), ((420, 445), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {}), '()\n', (44...
"""LVM based backup workflows and MixIn classes.""" import copy import os import gflags import rdiff_backup_wrapper FLAGS = gflags.FLAGS gflags.DEFINE_string( 'snapshot_mount_root', '/tmp', 'root path for creating temporary directories for mounting LVM snapshots') gflags.DEFINE_string('snapshot_suffix', '-a...
[ "copy.copy", "gflags.DEFINE_string", "os.path.join", "os.path.ismount" ]
[((141, 271), 'gflags.DEFINE_string', 'gflags.DEFINE_string', (['"""snapshot_mount_root"""', '"""/tmp"""', '"""root path for creating temporary directories for mounting LVM snapshots"""'], {}), "('snapshot_mount_root', '/tmp',\n 'root path for creating temporary directories for mounting LVM snapshots')\n", (161, 271...
from pandas import DataFrame as df from pandas.io.parsers import read_csv import os import numpy as np import re from extras import toint def check_dir(dirpath): if not os.path.isdir(dirpath): raise "Directory : %s is missing" def check_file(fpath): if not os.path.isfile(fpath): r...
[ "os.path.join", "os.path.isfile", "os.path.isdir", "pandas.DataFrame", "re.findall", "pandas.io.parsers.read_csv" ]
[((459, 519), 're.findall', 're.findall', (['"""\\\\[Result "(.+)"\\\\]"""', 'everything', 're.IGNORECASE'], {}), '(\'\\\\[Result "(.+)"\\\\]\', everything, re.IGNORECASE)\n', (469, 519), False, 'import re\n'), ((539, 603), 're.findall', 're.findall', (['"""\\\\[WhiteElo "(\\\\d+)"\\\\]"""', 'everything', 're.IGNORECAS...