code
stringlengths
22
1.05M
apis
listlengths
1
3.31k
extract_api
stringlengths
75
3.25M
# Generated by Django 3.1.7 on 2021-03-24 17:18 import datetime from django.db import migrations, models import django.db.models.deletion import uuid class Migration(migrations.Migration): dependencies = [ ('xmpt', '0001_initial'), ] operations = [ migrations.CreateModel( na...
[ "django.db.models.EmailField", "django.db.models.DateField", "django.db.models.TextField", "django.db.models.IntegerField", "django.db.models.ForeignKey", "django.db.models.FileField", "django.db.models.BooleanField", "django.db.models.ImageField", "django.db.models.AutoField", "django.db.models.S...
[((14781, 14877), 'django.db.models.ForeignKey', 'models.ForeignKey', ([], {'default': '(0)', 'on_delete': 'django.db.models.deletion.CASCADE', 'to': '"""xmpt.usstate"""'}), "(default=0, on_delete=django.db.models.deletion.CASCADE,\n to='xmpt.usstate')\n", (14798, 14877), False, 'from django.db import migrations, mo...
# INFO __author__ = "<NAME>" __date__ = "26 Mar 2022" __license__ = "MIT" __version__ = "1.0" __maintainer__ = "<NAME>" __email__ = "<EMAIL>" __status__ = "Definitive version" __copyright__ = "© 2022" # SCRIPT import numpy as np from logistic_regression import * from other_functions import * print('\033[92m\033[1m\...
[ "numpy.append" ]
[((2421, 2436), 'numpy.append', 'np.append', (['w', 'b'], {}), '(w, b)\n', (2430, 2436), True, 'import numpy as np\n')]
import requests import urllib class ResolveCoreference: def __init__(self): self.resolved_sentece = [] self.input_sentence = None def process(self, corenlp_output): self.input_sentence = corenlp_output for coref in self.input_sentence['corefs']: mentions =...
[ "requests.post", "urllib.parse.quote" ]
[((2063, 2089), 'urllib.parse.quote', 'urllib.parse.quote', (['params'], {}), '(params)\n', (2081, 2089), False, 'import urllib\n'), ((2422, 2460), 'requests.post', 'requests.post', (['(url + queryString)', 'data'], {}), '(url + queryString, data)\n', (2435, 2460), False, 'import requests\n')]
#!/usr/bin/python3.8 """ Genetic Algorithm to maximize surveillance over a population for AI Assignment. Author: Sam (eremus-dev) Repo: https://github.com/eremus-dev """ import math from collections import Counter from typing import List, Dict import numpy as np import matplotlib.pyplot as plt from test_pop import te...
[ "matplotlib.pyplot.grid", "math.sqrt", "numpy.array", "numpy.mean", "matplotlib.pyplot.plot", "matplotlib.pyplot.scatter", "matplotlib.pyplot.axis", "matplotlib.pyplot.ylim", "matplotlib.pyplot.cla", "matplotlib.pyplot.savefig", "matplotlib.pyplot.Circle", "numpy.random.choice", "matplotlib....
[((1327, 1351), 'numpy.array', 'np.array', (['obs'], {'copy': '(True)'}), '(obs, copy=True)\n', (1335, 1351), True, 'import numpy as np\n'), ((4549, 4558), 'matplotlib.pyplot.cla', 'plt.cla', ([], {}), '()\n', (4556, 4558), True, 'import matplotlib.pyplot as plt\n'), ((4776, 4793), 'matplotlib.pyplot.axis', 'plt.axis',...
#################################################################### # Animus AI Developed by <NAME>ated 13th March 2018 # #################################################################### ''' THIS PROGRAM IS INTENDED FOR USE UNDER SUPERVISION OF HUMAN INTERVENTION. THE SKILL ADDS CAPABILITY OF SCANNIN...
[ "googlesearch.googlesearch.GoogleSearch" ]
[((874, 888), 'googlesearch.googlesearch.GoogleSearch', 'GoogleSearch', ([], {}), '()\n', (886, 888), False, 'from googlesearch.googlesearch import GoogleSearch\n')]
''' Copyright (c) The Dojo Foundation 2011. All Rights Reserved. Copyright (c) IBM Corporation 2008, 2011. All Rights Reserved. ''' # tornado import tornado.ioloop # std lib import logging import time # coweb from .session import BayeuxSession from .connection import BayeuxConnection from .channel import BayeuxChannel ...
[ "logging.getLogger", "time.time" ]
[((327, 361), 'logging.getLogger', 'logging.getLogger', (['"""bayeux.server"""'], {}), "('bayeux.server')\n", (344, 361), False, 'import logging\n'), ((5462, 5473), 'time.time', 'time.time', ([], {}), '()\n', (5471, 5473), False, 'import time\n'), ((5862, 5873), 'time.time', 'time.time', ([], {}), '()\n', (5871, 5873),...
"""Consumer responsible for writing to stackdriver and associated helpers.""" import logging from datetime import tzinfo, timedelta, datetime class _FixedOffsetTimeZone(tzinfo): """Hack for dealing w/ lack of %z in 2.7 strptime. See https://docs.python.org/2/library/datetime.html#datetime.tzinfo.fromutc ...
[ "datetime.datetime.strptime", "datetime.timedelta", "logging.warn", "datetime.datetime.utcnow" ]
[((382, 407), 'datetime.timedelta', 'timedelta', ([], {'seconds': 'offset'}), '(seconds=offset)\n', (391, 407), False, 'from datetime import tzinfo, timedelta, datetime\n'), ((648, 660), 'datetime.timedelta', 'timedelta', (['(0)'], {}), '(0)\n', (657, 660), False, 'from datetime import tzinfo, timedelta, datetime\n'), ...
#! /usr/bin/env python3 """ %(prog)s takes a fault injection executable and executes it Usage: %(prog)s --CLI/--GUI <fault injection executable> <the same options that you use to run the excutable before> %(prog)s --help(-h): show help information Prerequisite: 0. You need to specify --CLI or --GUI dependin...
[ "yaml.load", "sys.exit", "os.remove", "os.listdir", "subprocess.Popen", "os.path.isdir", "os.mkdir", "sys.stdout.flush", "random.randint", "os.path.isfile", "os.path.dirname", "time.time", "os.path.join", "random.seed", "os.getcwd", "os.path.realpath", "os.chdir", "os.path.basename...
[((1322, 1351), 'os.path.basename', 'os.path.basename', (['sys.argv[0]'], {}), '(sys.argv[0])\n', (1338, 1351), False, 'import sys, os, subprocess\n'), ((1579, 1595), 'sys.exit', 'sys.exit', (['retval'], {}), '(retval)\n', (1587, 1595), False, 'import sys, os, subprocess\n'), ((1715, 1740), 'os.path.realpath', 'os.path...
import os from pathlib import Path import pytest from jubox import JupyterNotebook, CodeCell from nbformat.notebooknode import NotebookNode def test_creation_from_file(notebook_file_simple): #file = f"{notebook_folder}/nb_simple.ipynb" file = notebook_file_simple nb = JupyterNotebook(file) assert nb....
[ "jubox.JupyterNotebook", "jubox.CodeCell", "pathlib.Path" ]
[((284, 305), 'jubox.JupyterNotebook', 'JupyterNotebook', (['file'], {}), '(file)\n', (299, 305), False, 'from jubox import JupyterNotebook, CodeCell\n'), ((523, 549), 'pathlib.Path', 'Path', (['notebook_file_simple'], {}), '(notebook_file_simple)\n', (527, 549), False, 'from pathlib import Path\n'), ((559, 580), 'jubo...
import spot from crome_logic.specification.temporal import LTL from crome_logic.tools.crome_io import save_to_file from crome_logic.typelement.robotic import BooleanAction, BooleanSensor from crome_logic.typeset import Typeset from crome_synthesis.controller import Controller from crome_synthesis.controller.synthesis ...
[ "crome_synthesis.controller.synthesis.generate_controller", "crome_logic.typelement.robotic.BooleanSensor", "crome_synthesis.controller.Controller", "crome_logic.tools.crome_io.save_to_file", "crome_logic.typelement.robotic.BooleanAction", "spot.automaton" ]
[((614, 653), 'crome_synthesis.controller.Controller', 'Controller', ([], {'assumptions': 'a', 'guarantees': 'g'}), '(assumptions=a, guarantees=g)\n', (624, 653), False, 'from crome_synthesis.controller import Controller\n'), ((878, 913), 'crome_synthesis.controller.synthesis.generate_controller', 'generate_controller'...
from . import db from werkzeug.security import generate_password_hash,check_password_hash from datetime import datetime import time from flask_login import UserMixin from . import login_manager @login_manager.user_loader def load_user(user_id): return User.query.get(int(user_id)) # @login_manager.user_loader # de...
[ "werkzeug.security.generate_password_hash", "werkzeug.security.check_password_hash" ]
[((1023, 1055), 'werkzeug.security.generate_password_hash', 'generate_password_hash', (['password'], {}), '(password)\n', (1045, 1055), False, 'from werkzeug.security import generate_password_hash, check_password_hash\n'), ((1120, 1167), 'werkzeug.security.check_password_hash', 'check_password_hash', (['self.pass_secur...
import logging from django.db import connection from prometheus_client import Gauge from zentral.utils.prometheus import BasePrometheusMetricsView from .models import Status logger = logging.getLogger("zentral.core.compliance_checks.metrics_views") class MetricsView(BasePrometheusMetricsView): def add_complianc...
[ "logging.getLogger", "django.db.connection.cursor", "prometheus_client.Gauge" ]
[((185, 250), 'logging.getLogger', 'logging.getLogger', (['"""zentral.core.compliance_checks.metrics_views"""'], {}), "('zentral.core.compliance_checks.metrics_views')\n", (202, 250), False, 'import logging\n'), ((354, 456), 'prometheus_client.Gauge', 'Gauge', (['"""zentral_compliance_checks"""', '"""Zentral compliance...
import click import ast import requests import pandas as pd import numpy as np from pysradb.sraweb import SRAweb import requests import pysam import re import pyfastx import pkg_resources # part of setuptools import json import warnings import sys import os from .run_workflow import make_snakes # Allows passing strin...
[ "pandas.read_csv", "pkg_resources.require", "click.File", "pysam.AlignmentFile", "click.BadParameter", "os.path.exists", "click.group", "click.option", "pysam.set_verbosity", "pandas.DataFrame", "pandas.merge", "re.match", "os.path.splitext", "os.path.dirname", "click.version_option", ...
[((1461, 1486), 'os.path.dirname', 'os.path.dirname', (['__file__'], {}), '(__file__)\n', (1476, 1486), False, 'import os\n'), ((15858, 15871), 'click.group', 'click.group', ([], {}), '()\n', (15869, 15871), False, 'import click\n'), ((15873, 15906), 'click.version_option', 'click.version_option', (['__version__'], {})...
""" Copyright (c) 2015 Red Hat, Inc All rights reserved. This software may be modified and distributed under the terms of the BSD license. See the LICENSE file for details. """ from __future__ import print_function, unicode_literals import logging import os import sys import pytest from atomic_reactor.buildimage i...
[ "logging.getLogger", "atomic_reactor.core.DockerTasker", "os.path.join", "tests.fixtures.get_uuid", "pytest.mark.parametrize", "os.path.dirname", "pytest.raises", "tests.docker_mock.mock_docker", "atomic_reactor.buildimage.BuildImageBuilder" ]
[((705, 746), 'logging.getLogger', 'logging.getLogger', (['"""atomic_reactor.tests"""'], {}), "('atomic_reactor.tests')\n", (722, 746), False, 'import logging\n'), ((780, 794), 'atomic_reactor.core.DockerTasker', 'DockerTasker', ([], {}), '()\n', (792, 794), False, 'from atomic_reactor.core import DockerTasker\n'), ((8...
# Copyright (c) 2017-2018 Wind River Systems, Inc. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law...
[ "titanium_cloud.swagger.views.APIv1SwaggerJsonView.as_view", "titanium_cloud.swagger.views.APIv1SwaggerJsonViewDepreciated.as_view", "titanium_cloud.swagger.views.SwaggerJsonViewDepreciated.as_view", "titanium_cloud.swagger.views.SwaggerJsonView.as_view", "rest_framework.urlpatterns.format_suffix_patterns" ...
[((1518, 1553), 'rest_framework.urlpatterns.format_suffix_patterns', 'format_suffix_patterns', (['URLPATTERNS'], {}), '(URLPATTERNS)\n', (1540, 1553), False, 'from rest_framework.urlpatterns import format_suffix_patterns\n'), ((1061, 1097), 'titanium_cloud.swagger.views.SwaggerJsonViewDepreciated.as_view', 'SwaggerJson...
from ofptHeader import ofptHeader def ofptBarrierReq(): header = ofptHeader(20) return header
[ "ofptHeader.ofptHeader" ]
[((67, 81), 'ofptHeader.ofptHeader', 'ofptHeader', (['(20)'], {}), '(20)\n', (77, 81), False, 'from ofptHeader import ofptHeader\n')]
import pytest from whylogs.core.metrics.nlp_metrics import NLPMetrics from whylogs.proto import NLPMetricsMessage def test_nlp_metrics(): nlp_metric = NLPMetrics() assert nlp_metric.mer is not None assert nlp_metric.mer.count == 0 nlp_metric.update("hello", "hello") assert nlp_metric.mer.histogr...
[ "whylogs.core.metrics.nlp_metrics.NLPMetrics", "whylogs.core.metrics.nlp_metrics.NLPMetrics.from_protobuf" ]
[((158, 170), 'whylogs.core.metrics.nlp_metrics.NLPMetrics', 'NLPMetrics', ([], {}), '()\n', (168, 170), False, 'from whylogs.core.metrics.nlp_metrics import NLPMetrics\n'), ((1069, 1081), 'whylogs.core.metrics.nlp_metrics.NLPMetrics', 'NLPMetrics', ([], {}), '()\n', (1079, 1081), False, 'from whylogs.core.metrics.nlp_...
# https://leetcode.com/problems/ransom-note/ from collections import Counter class Solution(object): def canConstruct(self, ransomNote, magazine): """ :type ransomNote: str :type magazine: str :rtype: bool """ available = Counter(magazine) for c in ransom...
[ "collections.Counter" ]
[((278, 295), 'collections.Counter', 'Counter', (['magazine'], {}), '(magazine)\n', (285, 295), False, 'from collections import Counter\n')]
# InteractionSubmission: import tensorflow as tf import numpy as np from waymo_open_dataset.protos import motion_submission_pb2 def eval_and_generate_submission(model, eval_dataset, num_modes = 6): submission = motion_submission_pb2.MotionChallengeSubmission() submission.submission_type = 2 submission.affiliatio...
[ "waymo_open_dataset.protos.motion_submission_pb2.MotionChallengeSubmission" ]
[((214, 263), 'waymo_open_dataset.protos.motion_submission_pb2.MotionChallengeSubmission', 'motion_submission_pb2.MotionChallengeSubmission', ([], {}), '()\n', (261, 263), False, 'from waymo_open_dataset.protos import motion_submission_pb2\n')]
from django.conf import settings from rest_framework.permissions import IsAdminUser from rest_framework import status, viewsets, decorators from quser.permissions import CURDPermissionsOrReadOnly from rest_framework.response import Response from . import models, serializers from .filters import FileFilter class TagV...
[ "rest_framework.response.Response", "rest_framework.decorators.action" ]
[((836, 979), 'rest_framework.decorators.action', 'decorators.action', ([], {'methods': "['delete']", 'detail': '(False)', 'serializer_class': 'serializers.BulkDestroySerializer', 'permission_classes': '(IsAdminUser,)'}), "(methods=['delete'], detail=False, serializer_class=\n serializers.BulkDestroySerializer, perm...
""" GitFeederController tests """ from django.test import TestCase from django.contrib.auth.models import User from app.logic.gitfeeder.models.FeedModel import FeedEntry from app.logic.gitrepo.models.GitProjectModel import GitProjectEntry from app.logic.gitrepo.models.GitUserModel import GitUserEntry from app.logic.bl...
[ "app.logic.gitfeeder.models.FeedModel.FeedEntry.objects.all", "app.logic.gitrepo.models.GitUserModel.GitUserEntry.objects.create", "app.logic.commandrepo.models.CommandGroupModel.CommandGroupEntry.objects.create", "app.logic.commandrepo.models.CommandGroupModel.CommandGroupEntry.objects.all", "app.logic.git...
[((528, 588), 'django.contrib.auth.models.User.objects.create_user', 'User.objects.create_user', (['"""<EMAIL>"""', '"""<EMAIL>"""', '"""<PASSWORD>"""'], {}), "('<EMAIL>', '<EMAIL>', '<PASSWORD>')\n", (552, 588), False, 'from django.contrib.auth.models import User\n'), ((613, 785), 'app.logic.bluesteelworker.models.Wor...
from django.conf.urls import patterns, url from irods_browser_app import views urlpatterns = patterns('', url(r'^login/$',views.login, name='irods_login'), url(r'^store/$',views.store, name='irods_store'), url(r'^upload/$',views.upload, name='irods_upload'), url(r'^upload_add/$',views.upload_add, name=...
[ "django.conf.urls.url" ]
[((111, 159), 'django.conf.urls.url', 'url', (['"""^login/$"""', 'views.login'], {'name': '"""irods_login"""'}), "('^login/$', views.login, name='irods_login')\n", (114, 159), False, 'from django.conf.urls import patterns, url\n'), ((165, 213), 'django.conf.urls.url', 'url', (['"""^store/$"""', 'views.store'], {'name':...
#!/usr/bin/python """ Script to initialise CROP database. """ import sys from crop.constants import SQL_CONNECTION_STRING, SQL_DBNAME from crop.db import create_database def confirm(question): """ Ask user to enter Y or N (case-insensitive). :return: True if the answer is Y. """ answer = "" ...
[ "crop.db.create_database" ]
[((518, 568), 'crop.db.create_database', 'create_database', (['SQL_CONNECTION_STRING', 'SQL_DBNAME'], {}), '(SQL_CONNECTION_STRING, SQL_DBNAME)\n', (533, 568), False, 'from crop.db import create_database\n')]
import rules # Projectroles dependency from projectroles import rules as pr_rules # To access common predicates # Predicates ------------------------------------------------------------- # TODO: If we need to assign new predicates, we do it here # Rules ----------------------------------------------------------...
[ "rules.add_perm" ]
[((526, 695), 'rules.add_perm', 'rules.add_perm', (['"""samplesheets.view_sheet"""', '(pr_rules.is_project_owner | pr_rules.is_project_delegate | pr_rules.\n is_project_contributor | pr_rules.is_project_guest)'], {}), "('samplesheets.view_sheet', pr_rules.is_project_owner |\n pr_rules.is_project_delegate | pr_rul...
from django.conf.urls.defaults import patterns, url, include from pycash.controllers import TaxController as controller urlpatterns = patterns('', (r'^upcomingList$', controller.upcomingList), (r'^upcoming$', controller.upcoming), url(r'^pay$', controller.pay, name="tax_pay"), (r'^list$', controller.li...
[ "django.conf.urls.defaults.url" ]
[((244, 288), 'django.conf.urls.defaults.url', 'url', (['"""^pay$"""', 'controller.pay'], {'name': '"""tax_pay"""'}), "('^pay$', controller.pay, name='tax_pay')\n", (247, 288), False, 'from django.conf.urls.defaults import patterns, url, include\n'), ((329, 386), 'django.conf.urls.defaults.url', 'url', (['"""^save$"""'...
from abc import abstractmethod from functools import wraps from dd.api.workflow.utils import normalize_columns class CallableBuilder(object): # TODO : remove reference to dataset # replace with lighter reference to dataset.output_table def __init__(self, context, dataset=None): self.context = con...
[ "dd.api.workflow.utils.normalize_columns", "functools.wraps" ]
[((1409, 1435), 'functools.wraps', 'wraps', (['self.transformation'], {}), '(self.transformation)\n', (1414, 1435), False, 'from functools import wraps\n'), ((14081, 14102), 'functools.wraps', 'wraps', (['self.operation'], {}), '(self.operation)\n', (14086, 14102), False, 'from functools import wraps\n'), ((5911, 5940)...
# Generated by Django 2.1.15 on 2021-05-21 08:24 from django.db import migrations, models import django.db.models.deletion class Migration(migrations.Migration): initial = True dependencies = [ ] operations = [ migrations.CreateModel( name='Company', fields=[ ...
[ "django.db.models.ForeignKey", "django.db.models.AutoField", "django.db.models.CharField", "django.db.models.IntegerField" ]
[((337, 430), 'django.db.models.AutoField', 'models.AutoField', ([], {'auto_created': '(True)', 'primary_key': '(True)', 'serialize': '(False)', 'verbose_name': '"""ID"""'}), "(auto_created=True, primary_key=True, serialize=False,\n verbose_name='ID')\n", (353, 430), False, 'from django.db import migrations, models\...
r""" Echelle Spectrum ---------------- An abstract base class for a high resolution spectrum, for some echelle order :math:`m \in ` out of :math:`M` total orders, each with vectors for wavelength, flux, and uncertainty, e.g. :math:`F_m(\lambda)`. This class is a subclass of specutils' Spectrum1D and is intended to ha...
[ "logging.getLogger", "numpy.sqrt", "celerite2.GaussianProcess", "numpy.log", "scipy.signal.savgol_filter", "scipy.interpolate.interp1d", "numpy.isfinite", "specutils.spectra.spectral_region.SpectralRegion", "copy.deepcopy", "numpy.arange", "numpy.mean", "numpy.exp", "astropy.nddata.StdDevUnc...
[((1464, 1491), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (1481, 1491), False, 'import logging\n'), ((1543, 1598), 'warnings.simplefilter', 'warnings.simplefilter', (['"""ignore"""'], {'category': 'VerifyWarning'}), "('ignore', category=VerifyWarning)\n", (1564, 1598), False, 'import...
import nltk import string import requests import json from io import StringIO from html.parser import HTMLParser import os import time from sys import platform path = "" if platform == "linux" or platform == "linux2": path = os.path.dirname(os.path.realpath(__file__)).replace("test","data/nltk") elif ...
[ "os.path.realpath", "nltk.data.find", "nltk.data.path.append", "nltk.download" ]
[((541, 568), 'nltk.data.path.append', 'nltk.data.path.append', (['path'], {}), '(path)\n', (562, 568), False, 'import nltk\n'), ((627, 647), 'nltk.data.find', 'nltk.data.find', (['find'], {}), '(find)\n', (641, 647), False, 'import nltk\n'), ((683, 721), 'nltk.download', 'nltk.download', (['name'], {'download_dir': 'p...
#!/usr/bin/env python # coding=utf-8 from db import Model __doc__ = "记录数据库层封装" recordModel = Model('record') #类添加函数 # join查找
[ "db.Model" ]
[((96, 111), 'db.Model', 'Model', (['"""record"""'], {}), "('record')\n", (101, 111), False, 'from db import Model\n')]
# Licensed under a 3-clause BSD style license - see LICENSE.rst """Hillas shower parametrization. TODO: ----- - Should have a separate function or option to compute 3rd order moments + asymmetry (which are not always needed) - remove alpha calculation (which is only about (0,0), and make a get alpha function tha...
[ "numpy.abs", "collections.namedtuple", "numpy.sqrt", "numpy.asanyarray", "numpy.sum", "numpy.arctan2", "numpy.row_stack", "numpy.hypot" ]
[((608, 686), 'collections.namedtuple', 'namedtuple', (['"""MomentParameters"""', '"""size,cen_x,cen_y,length,width,r,phi,psi,miss"""'], {}), "('MomentParameters', 'size,cen_x,cen_y,length,width,r,phi,psi,miss')\n", (618, 686), False, 'from collections import namedtuple\n'), ((863, 933), 'collections.namedtuple', 'name...
# Generated by Django 4.0.1 on 2022-01-12 23:44 from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('points', '0003_spend_alter_transaction_timestamp'), ] operations = [ migrations.AddField( model_name='spend', name='...
[ "django.db.models.TextField" ]
[((348, 387), 'django.db.models.TextField', 'models.TextField', ([], {'blank': '(True)', 'null': '(True)'}), '(blank=True, null=True)\n', (364, 387), False, 'from django.db import migrations, models\n')]
import json from multiprocessing import Pool, cpu_count from catti.IO.portfolio import loadPortfolioInfo from catti.IO.general import * from catti.indicators.calculations import indicatorsCalculator from catti.indicators.signals import signalsCalculator from catti.indicators.signalsValidation import signalsValida...
[ "multiprocessing.cpu_count", "catti.IO.portfolio.loadPortfolioInfo", "tqdm.tqdm", "json.dump" ]
[((976, 995), 'catti.IO.portfolio.loadPortfolioInfo', 'loadPortfolioInfo', ([], {}), '()\n', (993, 995), False, 'from catti.IO.portfolio import loadPortfolioInfo\n'), ((2184, 2240), 'tqdm.tqdm', 'tqdm.tqdm', (['chunks'], {'desc': '"""Chunk completion"""', 'total': 'splits'}), "(chunks, desc='Chunk completion', total=sp...
import indicoio from celery import Celery from flask import Flask from flask_bootstrap import Bootstrap from flask_googlemaps import GoogleMaps from flask_mail import Mail from flask_mongoengine import MongoEngine from app.celery.factory import init_celery from config import config, DEVELOPMENT_CONFIG_NAME celery = C...
[ "flask_mail.Mail", "flask.Flask", "celery.Celery", "flask_googlemaps.GoogleMaps", "app.celery.factory.init_celery", "flask_bootstrap.Bootstrap", "flask_mongoengine.MongoEngine" ]
[((319, 332), 'celery.Celery', 'Celery', (['"""app"""'], {}), "('app')\n", (325, 332), False, 'from celery import Celery\n'), ((340, 346), 'flask_mail.Mail', 'Mail', ([], {}), '()\n', (344, 346), False, 'from flask_mail import Mail\n'), ((352, 365), 'flask_mongoengine.MongoEngine', 'MongoEngine', ([], {}), '()\n', (363...
# -*- coding:utf-8 -*- import logging import stores release = "0.1.1" logging.getLogger("dbsync") class DBSync(): def __init__(self): pass def syncer(self, syncer): return self def validator(self, validator): return self def serializer(self, serializer): return se...
[ "logging.getLogger" ]
[((72, 99), 'logging.getLogger', 'logging.getLogger', (['"""dbsync"""'], {}), "('dbsync')\n", (89, 99), False, 'import logging\n')]
# -*- coding: utf-8 -*- import asyncio from config import (CHECK_SERVER_INTERVAL, CHECK_SERVER_INTERVAL_MAX, CRON_LOOP_INTERVAL) from discord import Activity, ActivityType from discord.errors import Forbidden, NotFound from discord.ext import commands, tasks from modules.db import Servers from modu...
[ "modules.utils.embed_generator", "asyncio.get_event_loop", "modules.utils.get_server_info", "modules.utils.stop_server", "discord.Activity", "asyncio.sleep", "modules.logging.logger.info", "discord.ext.tasks.loop", "modules.db.Servers.filter" ]
[((635, 689), 'discord.ext.tasks.loop', 'tasks.loop', ([], {'seconds': 'CRON_LOOP_INTERVAL', 'reconnect': '(True)'}), '(seconds=CRON_LOOP_INTERVAL, reconnect=True)\n', (645, 689), False, 'from discord.ext import commands, tasks\n'), ((539, 563), 'asyncio.get_event_loop', 'asyncio.get_event_loop', ([], {}), '()\n', (561...
from django.urls import path from dashboards.views import DashboardView urlpatterns = [ path( "<slug:organisation>/dashboards/<slug:dashboard>/", DashboardView.as_view(), name="dashboard_detail", ), ]
[ "dashboards.views.DashboardView.as_view" ]
[((168, 191), 'dashboards.views.DashboardView.as_view', 'DashboardView.as_view', ([], {}), '()\n', (189, 191), False, 'from dashboards.views import DashboardView\n')]
from datetime import date from django.http import JsonResponse from rest_framework import viewsets from rest_framework.permissions import IsAdminUser from .models import CampPlan from .serializers import CampPlanSerializer # =========================== # Camp Plan ViewSet # =========================== class CampPla...
[ "datetime.date.today", "django.http.JsonResponse" ]
[((1544, 1603), 'django.http.JsonResponse', 'JsonResponse', (["{'success': 'Successfully created camp plan'}"], {}), "({'success': 'Successfully created camp plan'})\n", (1556, 1603), False, 'from django.http import JsonResponse\n'), ((1637, 1679), 'django.http.JsonResponse', 'JsonResponse', (["{'error': serializer.err...
# coding=utf-8 # 导入自己的函数包d2lzh_pytorch,注意要先将目标包的父路径添加到系统路径中 import sys sys.path.append(r".") from d2lzh_pytorch import data_process, plot, train, rnn import torch import time import torchvision from PIL import Image from torch import nn,optim from torch.utils.data import Dataset,DataLoader """ 这一节介绍了如何对图像数据集进行增广以获得更...
[ "PIL.Image.open", "torch.nn.CrossEntropyLoss", "torch.utils.data.DataLoader", "torchvision.transforms.RandomHorizontalFlip", "d2lzh_pytorch.plot.set_figsize", "torch.cuda.is_available", "d2lzh_pytorch.rnn.resnet18", "torchvision.datasets.CIFAR10", "d2lzh_pytorch.plot.show_images", "torchvision.tra...
[((72, 92), 'sys.path.append', 'sys.path.append', (['"""."""'], {}), "('.')\n", (87, 92), False, 'import sys\n'), ((496, 514), 'd2lzh_pytorch.plot.set_figsize', 'plot.set_figsize', ([], {}), '()\n', (512, 514), False, 'from d2lzh_pytorch import data_process, plot, train, rnn\n'), ((525, 567), 'PIL.Image.open', 'Image.o...
import sphinx_rtd_theme project = 'Bitcoin DCA' copyright = '2021, <NAME>' author = '<NAME>' extensions = [] templates_path = ['_templates'] exclude_patterns = ['_build', 'Thumbs.db', '.DS_Store'] html_theme = 'sphinx_rtd_theme' html_static_path = ['_static'] pygments_style = 'sphinx' html_theme_path = [sphinx_rtd_the...
[ "sphinx_rtd_theme.get_html_theme_path" ]
[((306, 344), 'sphinx_rtd_theme.get_html_theme_path', 'sphinx_rtd_theme.get_html_theme_path', ([], {}), '()\n', (342, 344), False, 'import sphinx_rtd_theme\n')]
from django.contrib import admin from app_restaurantes.models import Restaurante # Register your models here. admin.site.register(Restaurante)
[ "django.contrib.admin.site.register" ]
[((112, 144), 'django.contrib.admin.site.register', 'admin.site.register', (['Restaurante'], {}), '(Restaurante)\n', (131, 144), False, 'from django.contrib import admin\n')]
# -*- coding: utf-8 -*- # Generated by Django 1.9.9 on 2016-09-09 15:44 from __future__ import unicode_literals from django.db import migrations, models import django.db.models.deletion class Migration(migrations.Migration): dependencies = [ ('wagtailcore', '0029_unicode_slugfield_dj19'), ('peop...
[ "django.db.models.OneToOneField", "django.db.models.ForeignKey", "django.db.models.ManyToManyField", "django.db.migrations.AlterModelOptions", "django.db.models.AutoField", "django.db.models.CharField" ]
[((1558, 1644), 'django.db.migrations.AlterModelOptions', 'migrations.AlterModelOptions', ([], {'name': '"""person"""', 'options': "{'ordering': ('last_name',)}"}), "(name='person', options={'ordering': (\n 'last_name',)})\n", (1586, 1644), False, 'from django.db import migrations, models\n'), ((1782, 1837), 'django...
import pandas as pd from matplotlib import pyplot as plt from scipy.interpolate import spline from scipy.ndimage.filters import gaussian_filter1d import numpy as np #%% # TZ numbers via https://www.fangraphs.com/leaders.aspx?pos=3b&stats=fld&lg=all&qual=y&type=0&season=2017&month=0&season1=1961&ind=1&team=0&rost=0&ag...
[ "numpy.ones", "pandas.read_csv", "matplotlib.pyplot.ylabel", "matplotlib.pyplot.xlabel", "matplotlib.pyplot.plot", "matplotlib.pyplot.figure", "scipy.ndimage.filters.gaussian_filter1d", "matplotlib.pyplot.tight_layout", "matplotlib.pyplot.title", "matplotlib.pyplot.xlim", "pandas.concat" ]
[((347, 368), 'pandas.read_csv', 'pd.read_csv', (['"""TZ.csv"""'], {}), "('TZ.csv')\n", (358, 368), True, 'import pandas as pd\n'), ((666, 688), 'pandas.read_csv', 'pd.read_csv', (['"""UZR.csv"""'], {}), "('UZR.csv')\n", (677, 688), True, 'import pandas as pd\n'), ((791, 811), 'pandas.concat', 'pd.concat', (['[df, df2]...
import pytest import numpy as np import pandas as pd from endaq.calc import rotation @pytest.mark.parametrize( 'quat, euler', [ ((0., 0., 0., 1.), (0., 0., 0.)), ((0., 0., 0., -1.), (0., 0., 0.)), ((1., 0., 0., 0.), (np.pi, 0., 0.)), (...
[ "pandas.DataFrame", "pytest.mark.parametrize", "endaq.calc.rotation.quaternion_to_euler", "pytest.raises" ]
[((90, 357), 'pytest.mark.parametrize', 'pytest.mark.parametrize', (['"""quat, euler"""', '[((0.0, 0.0, 0.0, 1.0), (0.0, 0.0, 0.0)), ((0.0, 0.0, 0.0, -1.0), (0.0, 0.0,\n 0.0)), ((1.0, 0.0, 0.0, 0.0), (np.pi, 0.0, 0.0)), ((0.0, 1.0, 0.0, 0.0),\n (np.pi, 0.0, np.pi)), ((0.0, 0.0, 1.0, 0.0), (0.0, 0.0, np.pi))]'], {...
# Copyright 2021 <NAME> # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # http://www.apache.org/licenses/LICENSE-2.0 # Unless required by applicable law or agreed to in writing, software...
[ "toshiba_ac.amqp_api.ToshibaAcAmqpApi", "toshiba_ac.device.ToshibaAcDevice", "toshiba_ac.http_api.ToshibaAcHttpApi" ]
[((823, 855), 'toshiba_ac.http_api.ToshibaAcHttpApi', 'ToshibaAcHttpApi', (['user', 'password'], {}), '(user, password)\n', (839, 855), False, 'from toshiba_ac.http_api import ToshibaAcHttpApi\n'), ((1096, 1137), 'toshiba_ac.amqp_api.ToshibaAcAmqpApi', 'ToshibaAcAmqpApi', (['self.reg_info.sas_token'], {}), '(self.reg_i...
"""JSON-LD utilities. """ __all__ = ('encode_jsonld', 'JsonLdEncoder', 'decode_jsonld') import datetime import json def encode_jsonld(jsonld_dataset, **kwargs): """Encode a JSON-LD dataset into a string. Parameters ---------- jsonld_dataset : `dict` A JSON-LD dataset. kwargs Key...
[ "datetime.datetime.strptime", "json.JSONDecoder", "json.JSONEncoder.default" ]
[((2180, 2236), 'json.JSONDecoder', 'json.JSONDecoder', ([], {'object_pairs_hook': '_decode_object_pairs'}), '(object_pairs_hook=_decode_object_pairs)\n', (2196, 2236), False, 'import json\n'), ((1044, 1079), 'json.JSONEncoder.default', 'json.JSONEncoder.default', (['self', 'obj'], {}), '(self, obj)\n', (1068, 1079), F...
from __future__ import absolute_import from __future__ import division from __future__ import print_function import tensorflow as tf def loss2(logits, labels, num_classes, scope, head=None): with tf.name_scope(scope): logits = tf.reshape(logits, (-1, num_classes)) softmax = tf.nn.softmax(logits) + ...
[ "tensorflow.name_scope", "tensorflow.nn.softmax", "tensorflow.reshape", "tensorflow.reduce_mean", "tensorflow.log" ]
[((838, 867), 'tensorflow.reduce_mean', 'tf.reduce_mean', (['cross_entropy'], {}), '(cross_entropy)\n', (852, 867), True, 'import tensorflow as tf\n'), ((201, 221), 'tensorflow.name_scope', 'tf.name_scope', (['scope'], {}), '(scope)\n', (214, 221), True, 'import tensorflow as tf\n'), ((240, 277), 'tensorflow.reshape', ...
import pytest from pretalx.event.models import Event @pytest.mark.django_db def test_shred_used_event(resource, answered_choice_question, personal_answer, rejected_submission, deleted_submission, mail, sent_mail, room_availability, slot, unreleased_slot, past_slot, feedback, canceled_talk, review, information, other...
[ "pretalx.event.models.Event.objects.count" ]
[((340, 361), 'pretalx.event.models.Event.objects.count', 'Event.objects.count', ([], {}), '()\n', (359, 361), False, 'from pretalx.event.models import Event\n'), ((426, 447), 'pretalx.event.models.Event.objects.count', 'Event.objects.count', ([], {}), '()\n', (445, 447), False, 'from pretalx.event.models import Event\...
# -*- coding: utf-8 -*- """ Setup for pytrafik """ from distutils.core import setup setup( name='pytrafik', version='0.2.1', description='PyTrafik', long_description='Wrapper for Västtrafik public API.', url='https://github.com/axelniklasson/PyTrafik', download_url = 'https://github.com/axelni...
[ "distutils.core.setup" ]
[((86, 901), 'distutils.core.setup', 'setup', ([], {'name': '"""pytrafik"""', 'version': '"""0.2.1"""', 'description': '"""PyTrafik"""', 'long_description': '"""Wrapper for Västtrafik public API."""', 'url': '"""https://github.com/axelniklasson/PyTrafik"""', 'download_url': '"""https://github.com/axelniklasson/PyTrafik...
#coding=utf-8 ''' Created on 2015-10-23 @author: Devuser ''' from doraemon.ci.models import CITask,CITaskHistory from gatesidelib.common.simplelogger import SimpleLogger from django.contrib.admin.models import DELETION,CHANGE,ADDITION from doraemon.project.models import Product,Project,Tag from business.project.projec...
[ "doraemon.ci.models.CITaskHistory", "doraemon.project.models.Project.objects.all", "doraemon.ci.models.CITask", "doraemon.home.models.TaskQueue.objects.all", "business.ci.ci_task_config_service.CITaskConfigService.copy_config", "business.project.project_service.ProjectService.get_projects_include_me", "...
[((1414, 1461), 'business.project.project_service.ProjectService.get_projects_include_me', 'ProjectService.get_projects_include_me', (['request'], {}), '(request)\n', (1452, 1461), False, 'from business.project.project_service import ProjectService\n'), ((5529, 5537), 'doraemon.ci.models.CITask', 'CITask', ([], {}), '(...
import struct from pox.lib.addresses import EthAddr, IPAddr """ Utility to convert MAC and IP to/from integers. """ def eth_to_int(addr): if not isinstance(addr, EthAddr): return None value = 0 raw = addr.toRaw() for i in range(len(raw)): byte_shift = 5 - i byte = raw[i] ...
[ "struct.unpack", "pox.lib.addresses.IPAddr", "struct.pack" ]
[((821, 833), 'pox.lib.addresses.IPAddr', 'IPAddr', (['addr'], {}), '(addr)\n', (827, 833), False, 'from pox.lib.addresses import EthAddr, IPAddr\n'), ((337, 361), 'struct.unpack', 'struct.unpack', (['"""B"""', 'byte'], {}), "('B', byte)\n", (350, 361), False, 'import struct\n'), ((536, 564), 'struct.pack', 'struct.pac...
import os import json import glob from rcnn.utils import get_path_with_annotation,get_path_with_annotation_ratio from rcnn.utils import get_weight_path __disease__ = ['Covid-Seg','Lung_Tumor'] __net__ = ['rcnn_unet'] __mode__ = ['cls','seg','mtl'] json_path = { 'Cervical':'/staff/shijun/torch_projects/Med_Seg/c...
[ "json.load", "rcnn.utils.get_weight_path", "os.path.join" ]
[((2639, 2665), 'rcnn.utils.get_weight_path', 'get_weight_path', (['CKPT_PATH'], {}), '(CKPT_PATH)\n', (2654, 2665), False, 'from rcnn.utils import get_weight_path\n'), ((1402, 1415), 'json.load', 'json.load', (['fp'], {}), '(fp)\n', (1411, 1415), False, 'import json\n'), ((2092, 2132), 'os.path.join', 'os.path.join', ...
# -*- python -*- import math import numpy import Shadow from Shadow.ShadowPreprocessorsXraylib import prerefl, pre_mlayer, bragg from srxraylib.sources import srfunc from sirepo.template import transfer_mat_bl from pykern.pkcollections import PKDict from pykern import pkjson sigmax = 0.0045000000000000005 sigdix = 2....
[ "pykern.pkjson.dump_pretty", "Shadow.OE", "Shadow.ShadowTools.plotxy", "sirepo.template.transfer_mat_bl.create_mat_rays", "math.sqrt", "numpy.matrix", "numpy.transpose" ]
[((418, 458), 'sirepo.template.transfer_mat_bl.create_mat_rays', 'transfer_mat_bl.create_mat_rays', (['epsilon'], {}), '(epsilon)\n', (449, 458), False, 'from sirepo.template import transfer_mat_bl\n'), ((472, 586), 'numpy.matrix', 'numpy.matrix', (['[[sigmax ** 2, 0, 0, 0], [0, sigdix ** 2, 0, 0], [0, 0, sigmaz ** 2, ...
import json import logging import os import re from collections import namedtuple from copy import deepcopy from typing import Any, Dict, List, Tuple import numpy as np import pandas as pd import spacy from scirex_utilities.analyse_pwc_entity_results import * from scirex_utilities.entity_utils import * from spacy.toke...
[ "logging.basicConfig", "json.loads", "collections.namedtuple", "os.listdir", "argparse.ArgumentParser", "spacy.load", "tqdm.tqdm", "json.dumps", "spacy.tokens.Doc", "os.path.join", "os.path.isdir", "pandas.concat", "re.finditer", "copy.deepcopy", "pandas.DataFrame", "numpy.cumsum", "...
[((357, 370), 'tqdm.tqdm.pandas', 'tqdm.pandas', ([], {}), '()\n', (368, 370), False, 'from tqdm import tqdm\n'), ((383, 482), 'collections.namedtuple', 'namedtuple', (['"""Span"""', "['start', 'end', 'token_start', 'token_end', 'entity', 'links', 'modified']"], {}), "('Span', ['start', 'end', 'token_start', 'token_end...
# Copyright (C) 2018 Innoviz Technologies # All rights reserved. # # This software may be modified and distributed under the terms # of the BSD 3-Clause license. See the LICENSE file for details. import pandas as pd import os import numpy as np from utilities.math_utils import RotationTranslationData from visualizatio...
[ "utilities.data_utils.enumerate_frames", "os.path.join", "os.getcwd", "utilities.math_utils.RotationTranslationData", "numpy.concatenate", "numpy.linalg.norm", "utilities.data_utils.read_all_data", "utilities.data_utils.frame_to_filename", "visualizations.vis.pcshow" ]
[((463, 516), 'os.path.join', 'os.path.join', (['base_dir', '"""data_examples"""', '"""test_video"""'], {}), "(base_dir, 'data_examples', 'test_video')\n", (475, 516), False, 'import os\n'), ((642, 680), 'utilities.data_utils.enumerate_frames', 'data_utils.enumerate_frames', (['video_dir'], {}), '(video_dir)\n', (669, ...
from typing import Sequence, Optional import pandas as pd import numpy as np def formatted_corr_df(df: pd.DataFrame, cols: Optional[Sequence[str]] = None) -> pd.DataFrame: """ Calculates correlations on a DataFrame and displays only the lower triangular of the resulting correlation DataFrame. :param...
[ "numpy.tril" ]
[((738, 749), 'numpy.tril', 'np.tril', (['df'], {}), '(df)\n', (745, 749), True, 'import numpy as np\n')]
# -*- coding: utf-8 -*- __doc__="返回选择物体的类型" import rpw from rpw import revit, DB, UI,db,doc #from System.Collections.Generic import List #import json #from scriptutils import this_script #from scriptutils.userinput import CommandSwitchWindow #import subprocess as sp #Change Selected Grid From 3D to 2D #selection = rpw...
[ "rpw.ui.Selection", "rpw.db.Transaction.ensure" ]
[((465, 483), 'rpw.ui.Selection', 'rpw.ui.Selection', ([], {}), '()\n', (481, 483), False, 'import rpw\n'), ((635, 680), 'rpw.db.Transaction.ensure', 'rpw.db.Transaction.ensure', (['"""Hide_Grid_Bubble"""'], {}), "('Hide_Grid_Bubble')\n", (660, 680), False, 'import rpw\n')]
import json import urllib.request import urllib.parse import csv import login class Action: def __init__(self, tenantId, appId, appSecret, body, url, filename, column): self.tenantId = tenantId self.appId = appId self.appSecret = appSecret self.body = body self.url = url ...
[ "login.Login", "csv.DictReader", "json.dumps" ]
[((426, 480), 'login.Login', 'login.Login', (['self.tenantId', 'self.appId', 'self.appSecret'], {}), '(self.tenantId, self.appId, self.appSecret)\n', (437, 480), False, 'import login\n'), ((598, 621), 'csv.DictReader', 'csv.DictReader', (['csvfile'], {}), '(csvfile)\n', (612, 621), False, 'import csv\n'), ((1061, 1082)...
from flask import Flask from flask_login import LoginManager from flask_wtf import CsrfProtect login_manager = LoginManager() def create_app(): app = Flask(__name__) app.config.from_pyfile("config/setting.py") csrf = CsrfProtect() csrf.init_app(app) register_blueprint(app) create_db(app) ...
[ "flask_login.LoginManager", "flask_wtf.CsrfProtect", "flask.Flask" ]
[((112, 126), 'flask_login.LoginManager', 'LoginManager', ([], {}), '()\n', (124, 126), False, 'from flask_login import LoginManager\n'), ((157, 172), 'flask.Flask', 'Flask', (['__name__'], {}), '(__name__)\n', (162, 172), False, 'from flask import Flask\n'), ((232, 245), 'flask_wtf.CsrfProtect', 'CsrfProtect', ([], {}...
#! /usr/bin/env python3 # -*- coding: utf-8 -*- """ parse_bib_file.py: Convert raw bibtex file to yml format and md pages. """ import os import yaml import bibtexparser as bb def get_link(entry): if 'url' in entry.keys(): return entry['url'] elif 'doi' in entry.keys(): return f"https://doi....
[ "os.path.exists", "yaml.dump", "os.path.join", "sys.exit", "bibtexparser.load" ]
[((1113, 1134), 'os.path.exists', 'os.path.exists', (['fname'], {}), '(fname)\n', (1127, 1134), False, 'import os\n'), ((1545, 1555), 'bibtexparser.load', 'bb.load', (['f'], {}), '(f)\n', (1552, 1555), True, 'import bibtexparser as bb\n'), ((2510, 2561), 'os.path.join', 'os.path.join', (['out_folder_pages', "(entry['ID...
# Generated by Django 2.2.13 on 2020-07-22 12:47 from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('apis_vocabularies', '0001_initial'), ] operations = [ migrations.AddField( model_name='texttype', name='lang', ...
[ "django.db.models.CharField" ]
[((334, 501), 'django.db.models.CharField', 'models.CharField', ([], {'blank': '(True)', 'default': '"""deu"""', 'help_text': '"""The ISO 639-3 (or 2) code for the label\'s language."""', 'max_length': '(3)', 'null': '(True)', 'verbose_name': '"""ISO Code"""'}), '(blank=True, default=\'deu\', help_text=\n "The ISO 6...
import torch import get_data import numpy as np import torchaudio def number_of_correct(pred, target): return pred.squeeze().eq(target).sum().item() def get_likely_index(tensor): return tensor.argmax(dim=-1) def compute_accuracy(model, data_loader, device): model.eval() correct = 0 for data, t...
[ "torch.nn.Softmax", "get_data.prepare_wav", "torchaudio.load", "torch.unsqueeze", "get_data.idx_to_label", "numpy.argsort" ]
[((707, 750), 'get_data.prepare_wav', 'get_data.prepare_wav', (['waveform', 'sample_rate'], {}), '(waveform, sample_rate)\n', (727, 750), False, 'import get_data\n'), ((938, 956), 'numpy.argsort', 'np.argsort', (['(-probs)'], {}), '(-probs)\n', (948, 956), True, 'import numpy as np\n'), ((1156, 1181), 'torchaudio.load'...
import os import sys import numpy as np from sklearn import linear_model from sklearn.metrics import classification_report from sklearn.metrics import confusion_matrix from scipy.sparse import csr_matrix import csv def ReadCsv(fileName, mode): in_file = open(fileName) reader = csv.reader(in_file, delimiter='\t...
[ "csv.reader" ]
[((287, 337), 'csv.reader', 'csv.reader', (['in_file'], {'delimiter': '"""\t"""', 'quotechar': '"""\\""""'}), '(in_file, delimiter=\'\\t\', quotechar=\'"\')\n', (297, 337), False, 'import csv\n')]
import string from src import config def password_generator(): alphabet = string.digits alphabet += string.ascii_lowercase base = len(alphabet) shift = 0 length = int(config.get('minimal_password_length')) while True: password = '' temp = shift while temp > 0: ...
[ "src.config.get" ]
[((191, 228), 'src.config.get', 'config.get', (['"""minimal_password_length"""'], {}), "('minimal_password_length')\n", (201, 228), False, 'from src import config\n')]
from django.db import models from django.contrib.auth import get_user_model class TextAnswer(models.Model): """Model definition for TextAnswer.""" user = models.ForeignKey(get_user_model(), on_delete=models.CASCADE) question = models.ForeignKey('Question', on_delete=models.CASCADE) answer = models.Ch...
[ "django.contrib.auth.get_user_model", "django.db.models.CharField", "django.db.models.ForeignKey" ]
[((242, 297), 'django.db.models.ForeignKey', 'models.ForeignKey', (['"""Question"""'], {'on_delete': 'models.CASCADE'}), "('Question', on_delete=models.CASCADE)\n", (259, 297), False, 'from django.db import models\n'), ((311, 353), 'django.db.models.CharField', 'models.CharField', (['"""Answer"""'], {'max_length': '(10...
from django.conf.urls import patterns, url from qs import views urlpatterns = patterns('', url(r'^$', views.index, name='index'), url(r'^sendqs/$', views.sendqs, name='sendqs'), url(r'^show/$', views.showall, name='showall'), )
[ "django.conf.urls.url" ]
[((97, 133), 'django.conf.urls.url', 'url', (['"""^$"""', 'views.index'], {'name': '"""index"""'}), "('^$', views.index, name='index')\n", (100, 133), False, 'from django.conf.urls import patterns, url\n'), ((140, 185), 'django.conf.urls.url', 'url', (['"""^sendqs/$"""', 'views.sendqs'], {'name': '"""sendqs"""'}), "('^...
import cv2 import numpy as np from plantcv.plantcv.transform import nonuniform_illumination def test_nonuniform_illumination_rgb(transform_test_data): """Test for PlantCV.""" # Load rgb image rgb_img = cv2.imread(transform_test_data.small_rgb_img) corrected = nonuniform_illumination(img=rgb_img, ksize...
[ "numpy.mean", "plantcv.plantcv.transform.nonuniform_illumination", "cv2.imread" ]
[((216, 261), 'cv2.imread', 'cv2.imread', (['transform_test_data.small_rgb_img'], {}), '(transform_test_data.small_rgb_img)\n', (226, 261), False, 'import cv2\n'), ((278, 324), 'plantcv.plantcv.transform.nonuniform_illumination', 'nonuniform_illumination', ([], {'img': 'rgb_img', 'ksize': '(11)'}), '(img=rgb_img, ksize...
def main(): import numpy as np import matplotlib.pyplot as plt import torchvision from torch.autograd import Variable import torch.nn as nn import pickle from random import randint, randrange import sys from tqdm import tqdm import cv2 print("CUDA available: {}".format(torch...
[ "models.DSCLRCN_OldContext.DSCLRCN", "cv2.resize", "tqdm.tqdm.write", "torch.load", "util.data_utils.get_SALICON_datasets", "tqdm.tqdm", "models.CoSADUV_NoTemporal.CoSADUV_NoTemporal", "util.data_utils.get_video_datasets", "torch.from_numpy", "models.CoSADUV.CoSADUV", "torch.cuda.is_available", ...
[((12055, 12107), 'torch.multiprocessing.set_start_method', 'torch.multiprocessing.set_start_method', (['"""forkserver"""'], {}), "('forkserver')\n", (12093, 12107), False, 'import torch\n'), ((1562, 1627), 'util.data_utils.get_SALICON_datasets', 'get_SALICON_datasets', (['dataset_root_dir', 'mean_image_name', 'img_siz...
#!/usr/bin/env python3 from pyglet.gl import * from pyglet.window import NoSuchConfigException from rubiks import CubeController, CubeView if __name__ == '__main__': controller = CubeController() platform = pyglet.window.get_platform() display = platform.get_default_display() screen = display.get_de...
[ "rubiks.CubeController", "rubiks.CubeView" ]
[((186, 202), 'rubiks.CubeController', 'CubeController', ([], {}), '()\n', (200, 202), False, 'from rubiks import CubeController, CubeView\n'), ((729, 757), 'rubiks.CubeView', 'CubeView', (['controller', 'window'], {}), '(controller, window)\n', (737, 757), False, 'from rubiks import CubeController, CubeView\n')]
import sys import os from PyQt5.QtWidgets import (QTabWidget, QMessageBox) from codeeditor import CodeEditor from widgets import MessageBox class TabWidget(QTabWidget): def __init__(self, parent=None): super().__init__() self.mainWindow = parent self.setStyleSheet( ''...
[ "os.path.basename", "codeeditor.CodeEditor", "widgets.MessageBox" ]
[((997, 1031), 'codeeditor.CodeEditor', 'CodeEditor', ([], {'parent': 'self.mainWindow'}), '(parent=self.mainWindow)\n', (1007, 1031), False, 'from codeeditor import CodeEditor\n'), ((1811, 1925), 'widgets.MessageBox', 'MessageBox', (['QMessageBox.Warning', '"""Warning"""', '"""File not saved\n\nSave now ?"""', '(QMess...
#!/usr/bin/python3 #Title: headers.py #Author: ApexPredator #License: MIT #Github: https://github.com/ApexPredator-InfoSec/header_check #Description: This script take a URL or list or URLs as arguments and tests for the headers: 'Strict-Transport-Security', 'Content-Security-Policy', 'X-Frame-Options', and 'Server' imp...
[ "socket.gethostbyname", "requests.session", "requests.packages.urllib3.disable_warnings", "argparse.ArgumentParser" ]
[((427, 632), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {'prog': '"""headers.py"""', 'usage': '"""python3 -t <target> -f <file contianing target list> -d\npython3 headers.py -t https://securityheaders.com -d\npython3 headers.py -f urls.txt"""'}), '(prog=\'headers.py\', usage=\n """python3 -t <target...
# Generated by Django 2.1 on 2018-08-06 02:11 from django.db import migrations, models import django.db.models.deletion class Migration(migrations.Migration): initial = True dependencies = [ ('sites', '0002_alter_domain_unique'), ] operations = [ migrations.CreateModel( ...
[ "django.db.models.OneToOneField", "django.db.models.TextField", "django.db.models.ForeignKey", "django.db.models.ManyToManyField", "django.db.models.BooleanField", "django.db.models.SlugField", "django.db.models.AutoField", "django.db.models.DateTimeField", "django.db.models.DecimalField", "django...
[((3818, 3935), 'django.db.models.ForeignKey', 'models.ForeignKey', ([], {'on_delete': 'django.db.models.deletion.CASCADE', 'related_name': '"""usage"""', 'to': '"""features.FeatureValue"""'}), "(on_delete=django.db.models.deletion.CASCADE, related_name\n ='usage', to='features.FeatureValue')\n", (3835, 3935), False...
import argparse import os from os.path import join import sys import joblib import pandas as pd import matplotlib.pyplot as plt import matplotlib matplotlib.use('TkAgg') sys.path.append('.') from project.models.common import get_errors, get_model_details_for_algorithm, get_color, init_scale_from_train_set from proj...
[ "project.utils.logger.logger.info", "project.models.common.init_scale_from_train_set", "project.models.details.ModelDetails", "project.models.details.get_model_filepath", "sys.path.append", "matplotlib.pyplot.margins", "project.models.common.get_errors", "argparse.ArgumentParser", "project.models.co...
[((148, 171), 'matplotlib.use', 'matplotlib.use', (['"""TkAgg"""'], {}), "('TkAgg')\n", (162, 171), False, 'import matplotlib\n'), ((173, 193), 'sys.path.append', 'sys.path.append', (['"""."""'], {}), "('.')\n", (188, 193), False, 'import sys\n'), ((663, 732), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], ...
import copy as _copy import math as _math import os as _os import cv2 as _cv2 import numpy as _np from PIL import Image as _IMG from easytorch.utils.logger import * """ ################################################################################################## Very useful image related utilities ##############...
[ "PIL.Image.open", "copy.deepcopy", "numpy.ones", "numpy.where", "scipy.ndimage.measurements.label", "os.path.join", "math.sqrt", "numpy.max", "cv2.createCLAHE", "numpy.array", "numpy.zeros", "numpy.pad", "numpy.bincount", "numpy.min", "copy.copy", "numpy.zeros_like" ]
[((3470, 3535), 'numpy.zeros', '_np.zeros', (['[arr_2d.shape[0], arr_2d.shape[1], 3]'], {'dtype': '_np.uint8'}), '([arr_2d.shape[0], arr_2d.shape[1], 3], dtype=_np.uint8)\n', (3479, 3535), True, 'import numpy as _np\n'), ((4755, 4767), 'numpy.max', '_np.max', (['arr'], {}), '(arr)\n', (4762, 4767), True, 'import numpy ...
from django.test import TestCase from django.test import TransactionTestCase from main import tasks from celery.contrib.testing.worker import start_worker from main.models import User from src.celery import app class UserModelTests(TestCase): @classmethod def setUpTestData(cls): cls.user = User.object...
[ "main.models.User.objects.create_user" ]
[((309, 358), 'main.models.User.objects.create_user', 'User.objects.create_user', (['"""<EMAIL>"""', '"""<PASSWORD>"""'], {}), "('<EMAIL>', '<PASSWORD>')\n", (333, 358), False, 'from main.models import User\n')]
from pexpect import pxssh s = pxssh.pxssh() s.force_password=True if not s.login ('192.168.127.12', 'root', '<PASSWORD>!'): print ("SSH session failed on login.") else: print ("SSH session login successful") s.sendline ('cp /root/some.txt /root/some2.txt') s.prompt() # match the prompt prin...
[ "pexpect.pxssh.pxssh" ]
[((30, 43), 'pexpect.pxssh.pxssh', 'pxssh.pxssh', ([], {}), '()\n', (41, 43), False, 'from pexpect import pxssh\n')]
from zeus.config import db from zeus.db.mixins import ApiTokenMixin, RepositoryMixin, StandardAttributes from zeus.db.utils import model_repr class RepositoryApiToken(StandardAttributes, RepositoryMixin, ApiTokenMixin, db.Model): """ An API token associated to a repository. """ __tablename__ = "repos...
[ "zeus.db.utils.model_repr" ]
[((352, 386), 'zeus.db.utils.model_repr', 'model_repr', (['"""repository_id"""', '"""key"""'], {}), "('repository_id', 'key')\n", (362, 386), False, 'from zeus.db.utils import model_repr\n')]
import numpy as np from PySide import QtGui, QtCore import sharppy.sharptab as tab from sharppy.sharptab.constants import * ## Written by <NAME> - OU School of Meteorology ## and <NAME> - CIMMS __all__ = ['backgroundWatch', 'plotWatch'] class backgroundWatch(QtGui.QFrame): ''' Draw the background frame and ...
[ "PySide.QtGui.QFont", "PySide.QtGui.QFontMetrics", "PySide.QtGui.QColor", "PySide.QtGui.QPen", "PySide.QtGui.QPainter", "PySide.QtCore.QRect" ]
[((996, 1027), 'PySide.QtGui.QFont', 'QtGui.QFont', (['"""Helvetica"""', 'fsize'], {}), "('Helvetica', fsize)\n", (1007, 1027), False, 'from PySide import QtGui, QtCore\n'), ((1053, 1084), 'PySide.QtGui.QFont', 'QtGui.QFont', (['"""Helvetica"""', 'fsize'], {}), "('Helvetica', fsize)\n", (1064, 1084), False, 'from PySid...
#!/usr/bin/python # Copyright (c) 2011 GeometryFactory Sarl (France) # # $URL$ # $Id$ # SPDX-License-Identifier: LGPL-3.0-or-later OR LicenseRef-Commercial # # Author(s) : <NAME> import sys import os import gdb sys.path.insert(0, os.getcwd() + '/python') import CGAL.printers
[ "os.getcwd" ]
[((239, 250), 'os.getcwd', 'os.getcwd', ([], {}), '()\n', (248, 250), False, 'import os\n')]
import os import dill from copy import deepcopy from itertools import tee from abc import ABCMeta, abstractmethod from malemba import ModelBase from malemba.ds_tools import ArrayHandler, group_array class ModelDualStage(ModelBase): class Concatenator(object, metaclass=ABCMeta): @abstractmethod ...
[ "os.path.exists", "os.makedirs", "os.path.join", "malemba.ds_tools.ArrayHandler", "itertools.tee", "dill.dump", "dill.load" ]
[((5840, 5846), 'itertools.tee', 'tee', (['X'], {}), '(X)\n', (5843, 5846), False, 'from itertools import tee\n'), ((8255, 8287), 'dill.dump', 'dill.dump', (['self.__dict__', 'meta_f'], {}), '(self.__dict__, meta_f)\n', (8264, 8287), False, 'import dill\n'), ((5542, 5556), 'malemba.ds_tools.ArrayHandler', 'ArrayHandler...
import xml.etree.ElementTree as ET from typing import NewType MethodCall = NewType("MethodCall", ET.Element) ConfigurableMethodTerminal = NewType("ConfigurableMethodTerminal", ET.Element) MethodTerminal = NewType("MethodTerminal", ET.Element) WhileLoop = NewType("WhileLoop", ET.Element) Terminal = NewType("Terminal", ...
[ "typing.NewType" ]
[((76, 109), 'typing.NewType', 'NewType', (['"""MethodCall"""', 'ET.Element'], {}), "('MethodCall', ET.Element)\n", (83, 109), False, 'from typing import NewType\n'), ((139, 188), 'typing.NewType', 'NewType', (['"""ConfigurableMethodTerminal"""', 'ET.Element'], {}), "('ConfigurableMethodTerminal', ET.Element)\n", (146,...
#!/usr/bin/env python3 # -*- coding: utf-8 -*- from pylookyloo import Lookyloo import json # lookyloo_url = "https://lookyloo.circl.lu/" lookyloo_url = "http://0.0.0.0:5100" lookyloo = Lookyloo(lookyloo_url) lookyloo.init_apikey(username='admin', password='<PASSWORD>') event = lookyloo.misp_push('6ae2afdc-4d90-41ce-...
[ "json.dumps", "pylookyloo.Lookyloo" ]
[((188, 210), 'pylookyloo.Lookyloo', 'Lookyloo', (['lookyloo_url'], {}), '(lookyloo_url)\n', (196, 210), False, 'from pylookyloo import Lookyloo\n'), ((347, 374), 'json.dumps', 'json.dumps', (['event'], {'indent': '(2)'}), '(event, indent=2)\n', (357, 374), False, 'import json\n')]
# Copyright (C) 2015-2019 <NAME> # SPDX-License-Identifier: Apache-2.0 # Check for presence of FEniCS etc from .verify_environment import verify_env verify_env() __version__ = '2019.0.2' # This should potentially be made local to the mesh creation routines import dolfin dolfin.parameters['ghost_mode'] = 'shared_...
[ "os.path.abspath", "subprocess.check_output", "os.path.join" ]
[((656, 681), 'os.path.abspath', 'os.path.abspath', (['__file__'], {}), '(__file__)\n', (671, 681), False, 'import os\n'), ((714, 742), 'os.path.join', 'os.path.join', (['this_dir', '""".."""'], {}), "(this_dir, '..')\n", (726, 742), False, 'import os\n'), ((765, 795), 'os.path.join', 'os.path.join', (['proj_dir', '"""...
''' Created on July 7, 2019 @author: Terry @email:<EMAIL> ''' print(__doc__) import numpy as np import matplotlib.pyplot as plt from sklearn.datasets import load_iris from sklearn.tree import DecisionTreeClassifier, plot_tree # Parameter n_classes = 3 plot_colors = "ryb" plot_step = 0.02 # Load data iris = load_...
[ "sklearn.datasets.load_iris", "matplotlib.pyplot.contourf", "matplotlib.pyplot.ylabel", "numpy.arange", "numpy.where", "matplotlib.pyplot.xlabel", "sklearn.tree.DecisionTreeClassifier", "matplotlib.pyplot.figure", "sklearn.tree.plot_tree", "matplotlib.pyplot.tight_layout", "matplotlib.pyplot.sca...
[((315, 326), 'sklearn.datasets.load_iris', 'load_iris', ([], {}), '()\n', (324, 326), False, 'from sklearn.datasets import load_iris\n'), ((1404, 1477), 'matplotlib.pyplot.suptitle', 'plt.suptitle', (['"""Decision surface of a decision tree using paired features"""'], {}), "('Decision surface of a decision tree using ...
# encoding: utf-8 from application import static_manager from static_bundle import (JsBundle, CssBundle) css1 = CssBundle("css") css1.add_file("example1.less") css1.add_file("example2.css") js1 = JsBundle("js") js1.add_file("vendors/example1.js") js1.add_file("vendors/example2.js") js2 = ...
[ "application.static_manager.create_builder", "static_bundle.CssBundle", "static_bundle.JsBundle" ]
[((141, 157), 'static_bundle.CssBundle', 'CssBundle', (['"""css"""'], {}), "('css')\n", (150, 157), False, 'from static_bundle import JsBundle, CssBundle\n'), ((226, 240), 'static_bundle.JsBundle', 'JsBundle', (['"""js"""'], {}), "('js')\n", (234, 240), False, 'from static_bundle import JsBundle, CssBundle\n'), ((320, ...
""" Intermediate Factors @author: <NAME> This module computes the interpolated features between the principal vectors -- the one linking source to target following the geodesics on the Grassmannian. We use the equivalent formulation derived in [1] and represent this geodesics for each pair of principal components. E...
[ "numpy.block", "numpy.diag", "joblib.Parallel", "numpy.array", "numpy.linspace", "numpy.isnan", "numpy.cos", "precise.principal_vectors.PVComputation", "numpy.min", "numpy.sin", "joblib.delayed", "numpy.isinf" ]
[((2327, 2403), 'numpy.min', 'np.min', (['[self.source_components_.shape[0], self.target_components_.shape[0]]'], {}), '([self.source_components_.shape[0], self.target_components_.shape[0]])\n', (2333, 2403), True, 'import numpy as np\n'), ((2602, 2632), 'precise.principal_vectors.PVComputation', 'PVComputation', (['n_...
import logging import os import hydra import pytorch_lightning as pl from hydra.utils import instantiate from omegaconf import DictConfig, OmegaConf from deep_learning_template import BaseDataModule, BaseTask from deep_learning_template.core import initialization as init from deep_learning_template.utils.config impor...
[ "logging.getLogger", "deep_learning_template.core.initialization.initialize_loggers", "hydra.main", "hydra.utils.instantiate", "omegaconf.OmegaConf.to_yaml", "deep_learning_template.core.initialization.initialize_task", "deep_learning_template.core.initialization.validate_cfg", "deep_learning_template...
[((381, 408), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (398, 408), False, 'import logging\n'), ((412, 464), 'hydra.main', 'hydra.main', ([], {'config_path': '"""conf"""', 'config_name': '"""config"""'}), "(config_path='conf', config_name='config')\n", (422, 464), False, 'import hydr...
import numpy as np from sklearn.preprocessing import MinMaxScaler, StandardScaler from sklearn.cross_validation import train_test_split import theanets import climate climate.enable_default_logging() X_orig = np.load('/Users/bzamecnik/Documents/music-processing/music-processing-experiments/c-scale-piano_spectrogram_2...
[ "theanets.Experiment", "climate.enable_default_logging", "sklearn.preprocessing.StandardScaler", "matplotlib.animation.ArtistAnimation", "numpy.zeros", "sklearn.cross_validation.train_test_split", "numpy.load", "sklearn.preprocessing.MinMaxScaler" ]
[((168, 200), 'climate.enable_default_logging', 'climate.enable_default_logging', ([], {}), '()\n', (198, 200), False, 'import climate\n'), ((211, 347), 'numpy.load', 'np.load', (['"""/Users/bzamecnik/Documents/music-processing/music-processing-experiments/c-scale-piano_spectrogram_2048_hamming.npy"""'], {}), "(\n '...
import os import glob import argparse import numpy as np def parse_args(): parser = argparse.ArgumentParser(description='Display datas') parser.add_argument('--data-dir', default='C:/Users/junya/Documents/plant_segmentation_data', help='dataset directory') parser.add_argument('--va...
[ "numpy.random.choice", "numpy.savetxt", "os.path.join", "argparse.ArgumentParser" ]
[((90, 142), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {'description': '"""Display datas"""'}), "(description='Display datas')\n", (113, 142), False, 'import argparse\n'), ((548, 582), 'os.path.join', 'os.path.join', (['args.data_dir', 'phase'], {}), '(args.data_dir, phase)\n', (560, 582), False, 'impo...
''' This file gets the cities from the Deutscher Wetterdienst website file (a file of the name 'TU_Stundenwerte_Beschreibung_Stationen.txt' must already be in the current folder) and saves it into the file 'current_cityfile.dump'.''' import pickle import re # regular expressions package import csv #### F...
[ "re.sub" ]
[((1146, 1177), 're.sub', 're.sub', (['"""[^\\\\w]"""', '""" """', 'textline'], {}), "('[^\\\\w]', ' ', textline)\n", (1152, 1177), False, 'import re\n')]
# python3 # Copyright 2021 InstaDeep Ltd. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applic...
[ "copy.deepcopy", "acme.specs.BoundedArray", "dm_env.TimeStep", "numpy.asarray", "mava.utils.wrapper_utils.apply_env_wrapper_preprocessers", "acme.specs.Array", "mava.utils.wrapper_utils.convert_np_type", "acme.wrappers.gym_wrapper._convert_to_spec", "mava.utils.wrapper_utils.parameterized_restart", ...
[((2552, 2610), 'mava.utils.wrapper_utils.parameterized_restart', 'parameterized_restart', (['reward', 'self._discount', 'observation'], {}), '(reward, self._discount, observation)\n', (2573, 2610), False, 'from mava.utils.wrapper_utils import apply_env_wrapper_preprocessers, convert_np_type, parameterized_restart\n'),...
import os import numpy as np from matplotlib import pyplot as plt from mpl_toolkits import mplot3d from mpl_toolkits.mplot3d import Axes3D from sklearn.decomposition import IncrementalPCA ## set paths inputpath = './input/' ## Toggle PCA for better visualization of clusters pca_flag = int(input("\nPerform ...
[ "os.path.exists", "numpy.shape", "numpy.unique", "numpy.where", "os.path.realpath", "matplotlib.pyplot.figure", "matplotlib.pyplot.axes", "matplotlib.pyplot.title", "sklearn.decomposition.IncrementalPCA", "numpy.genfromtxt", "matplotlib.pyplot.legend", "matplotlib.pyplot.show" ]
[((935, 971), 'os.path.exists', 'os.path.exists', (["(inputpath + 'kx.txt')"], {}), "(inputpath + 'kx.txt')\n", (949, 971), False, 'import os\n'), ((1027, 1062), 'numpy.genfromtxt', 'np.genfromtxt', (["(inputpath + 'kx.txt')"], {}), "(inputpath + 'kx.txt')\n", (1040, 1062), True, 'import numpy as np\n'), ((1071, 1107),...
import unittest from database import Database from search_database import SearchDatabase class SearchDatabaseTestCase(unittest.TestCase): def test_update(self): database = Database('purplePolitics', 'events') events = database.get_events(False) search_database = SearchDatabase() s...
[ "unittest.main", "search_database.SearchDatabase", "database.Database" ]
[((415, 430), 'unittest.main', 'unittest.main', ([], {}), '()\n', (428, 430), False, 'import unittest\n'), ((187, 223), 'database.Database', 'Database', (['"""purplePolitics"""', '"""events"""'], {}), "('purplePolitics', 'events')\n", (195, 223), False, 'from database import Database\n'), ((294, 310), 'search_database....
from sys import stdin def main_s(): print("---------------------------------------------------------------------") print("Bienvenido al simulador de Pac-Man") print("OPCIONES:") print("1. Aleatorio") print("2. Manual") val = str(stdin.readline().strip()) print("----------------------------...
[ "sys.stdin.readline", "Español.No_aleatorio.no_aleatorio", "Español.aleatorio.aleatorio" ]
[((445, 459), 'Español.No_aleatorio.no_aleatorio', 'no_aleatorio', ([], {}), '()\n', (457, 459), False, 'from Español.No_aleatorio import no_aleatorio\n'), ((537, 548), 'Español.aleatorio.aleatorio', 'aleatorio', ([], {}), '()\n', (546, 548), False, 'from Español.aleatorio import aleatorio\n'), ((255, 271), 'sys.stdin....
import discord from bs4 import BeautifulSoup import asyncio import datetime from io import BytesIO class Archiver: def __init__(self, bot): self.bot = bot async def on_message(self, message): await self.archive_message(message) async def on_message_edit(self, before, aft...
[ "discord.utils.get", "discord.Embed", "io.BytesIO", "discord.File" ]
[((1493, 1568), 'discord.utils.get', 'discord.utils.get', (['self.bot.main_server.channels'], {'name': 'archive_channel_name'}), '(self.bot.main_server.channels, name=archive_channel_name)\n', (1510, 1568), False, 'import discord\n'), ((1608, 1685), 'discord.utils.get', 'discord.utils.get', (['self.bot.backup_server.ch...
# Copyright (c) 2012-2018 SoftBank Robotics. All rights reserved. # Use of this source code is governed by a BSD-style license that can be # found in the COPYING file. from qisrc.test.conftest import git_server, qisrc_action # pylint: disable=unused-import # pylint: disable=redefined-outer-name def test_info(qibui...
[ "qisrc.test.conftest.qisrc_action", "qisrc.test.conftest.git_server.create_repo" ]
[((379, 412), 'qisrc.test.conftest.git_server.create_repo', 'git_server.create_repo', (['"""foo.git"""'], {}), "('foo.git')\n", (401, 412), False, 'from qisrc.test.conftest import git_server, qisrc_action\n'), ((417, 462), 'qisrc.test.conftest.qisrc_action', 'qisrc_action', (['"""init"""', 'git_server.manifest_url'], {...
import sqlite3 mem_db = sqlite3.connect(':memory:') db = sqlite3.connect('db/mydb') cr = db.cursor() """ Syntax or Purpose ------ --------- ---------------------- ------------ SELECT DISTINCT Exclude duplicate records for fields selected. SELECT DISTINCT CASE WHEN expression Conditional expression SELECT DIST...
[ "sqlite3.connect" ]
[((25, 52), 'sqlite3.connect', 'sqlite3.connect', (['""":memory:"""'], {}), "(':memory:')\n", (40, 52), False, 'import sqlite3\n'), ((58, 84), 'sqlite3.connect', 'sqlite3.connect', (['"""db/mydb"""'], {}), "('db/mydb')\n", (73, 84), False, 'import sqlite3\n')]
#!/usr/bin/python # ref: https://gist.github.com/bruienne/f81ea88253629abaf5f9 import objc import plistlib class attrdict(dict): __getattr__ = dict.__getitem__ __setattr__ = dict.__setitem__ ServerInformation = attrdict() ServerInformation_bundle = objc.loadBundle('ServerInformation', ServerInformation, \ ...
[ "plistlib.readPlist", "objc.loadBundle" ]
[((260, 397), 'objc.loadBundle', 'objc.loadBundle', (['"""ServerInformation"""', 'ServerInformation'], {'bundle_path': '"""/System/Library/PrivateFrameworks/ServerInformation.framework"""'}), "('ServerInformation', ServerInformation, bundle_path=\n '/System/Library/PrivateFrameworks/ServerInformation.framework')\n",...
import numpy as np import matplotlib.pyplot as plt import pickle import pandas as pd import cv2 as cv import os def read_pkl(file_path): obj = pd.read_pickle(file_path) return obj def show_results(obj , folder): for im_num in obj.keys(): data = obj[im_num] file = data['fileName'][12:] ...
[ "pandas.read_pickle", "cv2.imwrite", "cv2.rectangle", "matplotlib.pyplot.imshow", "os.path.join", "numpy.asarray", "matplotlib.pyplot.pause", "cv2.imread" ]
[((149, 174), 'pandas.read_pickle', 'pd.read_pickle', (['file_path'], {}), '(file_path)\n', (163, 174), True, 'import pandas as pd\n'), ((1781, 1828), 'os.path.join', 'os.path.join', (['folder', '"""YoloV3_res\\\\res_pkl.pkl"""'], {}), "(folder, 'YoloV3_res\\\\res_pkl.pkl')\n", (1793, 1828), False, 'import os\n'), ((33...
from os import path import sqlite3 as sql ROOT = path.dirname(path.relpath(__file__)) dbname = 'database.sqlite' def create_post(user, post, parent, location, item, tags, ats, created, status): con = sql.connect(path.join(ROOT, dbname)) cur = con.cursor() cur.execute( """insert into post ...
[ "os.path.join", "os.path.relpath" ]
[((63, 85), 'os.path.relpath', 'path.relpath', (['__file__'], {}), '(__file__)\n', (75, 85), False, 'from os import path\n'), ((218, 241), 'os.path.join', 'path.join', (['ROOT', 'dbname'], {}), '(ROOT, dbname)\n', (227, 241), False, 'from os import path\n'), ((605, 628), 'os.path.join', 'path.join', (['ROOT', 'dbname']...