code stringlengths 22 1.05M | apis listlengths 1 3.31k | extract_api stringlengths 75 3.25M |
|---|---|---|
# Generated by Django 3.1.7 on 2021-03-24 17:18
import datetime
from django.db import migrations, models
import django.db.models.deletion
import uuid
class Migration(migrations.Migration):
dependencies = [
('xmpt', '0001_initial'),
]
operations = [
migrations.CreateModel(
na... | [
"django.db.models.EmailField",
"django.db.models.DateField",
"django.db.models.TextField",
"django.db.models.IntegerField",
"django.db.models.ForeignKey",
"django.db.models.FileField",
"django.db.models.BooleanField",
"django.db.models.ImageField",
"django.db.models.AutoField",
"django.db.models.S... | [((14781, 14877), 'django.db.models.ForeignKey', 'models.ForeignKey', ([], {'default': '(0)', 'on_delete': 'django.db.models.deletion.CASCADE', 'to': '"""xmpt.usstate"""'}), "(default=0, on_delete=django.db.models.deletion.CASCADE,\n to='xmpt.usstate')\n", (14798, 14877), False, 'from django.db import migrations, mo... |
# INFO
__author__ = "<NAME>"
__date__ = "26 Mar 2022"
__license__ = "MIT"
__version__ = "1.0"
__maintainer__ = "<NAME>"
__email__ = "<EMAIL>"
__status__ = "Definitive version"
__copyright__ = "© 2022"
# SCRIPT
import numpy as np
from logistic_regression import *
from other_functions import *
print('\033[92m\033[1m\... | [
"numpy.append"
] | [((2421, 2436), 'numpy.append', 'np.append', (['w', 'b'], {}), '(w, b)\n', (2430, 2436), True, 'import numpy as np\n')] |
import requests
import urllib
class ResolveCoreference:
def __init__(self):
self.resolved_sentece = []
self.input_sentence = None
def process(self, corenlp_output):
self.input_sentence = corenlp_output
for coref in self.input_sentence['corefs']:
mentions =... | [
"requests.post",
"urllib.parse.quote"
] | [((2063, 2089), 'urllib.parse.quote', 'urllib.parse.quote', (['params'], {}), '(params)\n', (2081, 2089), False, 'import urllib\n'), ((2422, 2460), 'requests.post', 'requests.post', (['(url + queryString)', 'data'], {}), '(url + queryString, data)\n', (2435, 2460), False, 'import requests\n')] |
#!/usr/bin/python3.8
"""
Genetic Algorithm to maximize surveillance over a population for AI Assignment.
Author: Sam (eremus-dev)
Repo: https://github.com/eremus-dev
"""
import math
from collections import Counter
from typing import List, Dict
import numpy as np
import matplotlib.pyplot as plt
from test_pop import te... | [
"matplotlib.pyplot.grid",
"math.sqrt",
"numpy.array",
"numpy.mean",
"matplotlib.pyplot.plot",
"matplotlib.pyplot.scatter",
"matplotlib.pyplot.axis",
"matplotlib.pyplot.ylim",
"matplotlib.pyplot.cla",
"matplotlib.pyplot.savefig",
"matplotlib.pyplot.Circle",
"numpy.random.choice",
"matplotlib.... | [((1327, 1351), 'numpy.array', 'np.array', (['obs'], {'copy': '(True)'}), '(obs, copy=True)\n', (1335, 1351), True, 'import numpy as np\n'), ((4549, 4558), 'matplotlib.pyplot.cla', 'plt.cla', ([], {}), '()\n', (4556, 4558), True, 'import matplotlib.pyplot as plt\n'), ((4776, 4793), 'matplotlib.pyplot.axis', 'plt.axis',... |
####################################################################
# Animus AI Developed by <NAME>ated 13th March 2018 #
####################################################################
'''
THIS PROGRAM IS INTENDED FOR USE UNDER SUPERVISION OF HUMAN
INTERVENTION. THE SKILL ADDS CAPABILITY OF SCANNIN... | [
"googlesearch.googlesearch.GoogleSearch"
] | [((874, 888), 'googlesearch.googlesearch.GoogleSearch', 'GoogleSearch', ([], {}), '()\n', (886, 888), False, 'from googlesearch.googlesearch import GoogleSearch\n')] |
'''
Copyright (c) The Dojo Foundation 2011. All Rights Reserved.
Copyright (c) IBM Corporation 2008, 2011. All Rights Reserved.
'''
# tornado
import tornado.ioloop
# std lib
import logging
import time
# coweb
from .session import BayeuxSession
from .connection import BayeuxConnection
from .channel import BayeuxChannel
... | [
"logging.getLogger",
"time.time"
] | [((327, 361), 'logging.getLogger', 'logging.getLogger', (['"""bayeux.server"""'], {}), "('bayeux.server')\n", (344, 361), False, 'import logging\n'), ((5462, 5473), 'time.time', 'time.time', ([], {}), '()\n', (5471, 5473), False, 'import time\n'), ((5862, 5873), 'time.time', 'time.time', ([], {}), '()\n', (5871, 5873),... |
"""Consumer responsible for writing to stackdriver and associated helpers."""
import logging
from datetime import tzinfo, timedelta, datetime
class _FixedOffsetTimeZone(tzinfo):
"""Hack for dealing w/ lack of %z in 2.7 strptime.
See https://docs.python.org/2/library/datetime.html#datetime.tzinfo.fromutc
... | [
"datetime.datetime.strptime",
"datetime.timedelta",
"logging.warn",
"datetime.datetime.utcnow"
] | [((382, 407), 'datetime.timedelta', 'timedelta', ([], {'seconds': 'offset'}), '(seconds=offset)\n', (391, 407), False, 'from datetime import tzinfo, timedelta, datetime\n'), ((648, 660), 'datetime.timedelta', 'timedelta', (['(0)'], {}), '(0)\n', (657, 660), False, 'from datetime import tzinfo, timedelta, datetime\n'), ... |
#! /usr/bin/env python3
"""
%(prog)s takes a fault injection executable and executes it
Usage: %(prog)s --CLI/--GUI <fault injection executable> <the same options that you use to run the excutable before>
%(prog)s --help(-h): show help information
Prerequisite:
0. You need to specify --CLI or --GUI dependin... | [
"yaml.load",
"sys.exit",
"os.remove",
"os.listdir",
"subprocess.Popen",
"os.path.isdir",
"os.mkdir",
"sys.stdout.flush",
"random.randint",
"os.path.isfile",
"os.path.dirname",
"time.time",
"os.path.join",
"random.seed",
"os.getcwd",
"os.path.realpath",
"os.chdir",
"os.path.basename... | [((1322, 1351), 'os.path.basename', 'os.path.basename', (['sys.argv[0]'], {}), '(sys.argv[0])\n', (1338, 1351), False, 'import sys, os, subprocess\n'), ((1579, 1595), 'sys.exit', 'sys.exit', (['retval'], {}), '(retval)\n', (1587, 1595), False, 'import sys, os, subprocess\n'), ((1715, 1740), 'os.path.realpath', 'os.path... |
import os
from pathlib import Path
import pytest
from jubox import JupyterNotebook, CodeCell
from nbformat.notebooknode import NotebookNode
def test_creation_from_file(notebook_file_simple):
#file = f"{notebook_folder}/nb_simple.ipynb"
file = notebook_file_simple
nb = JupyterNotebook(file)
assert nb.... | [
"jubox.JupyterNotebook",
"jubox.CodeCell",
"pathlib.Path"
] | [((284, 305), 'jubox.JupyterNotebook', 'JupyterNotebook', (['file'], {}), '(file)\n', (299, 305), False, 'from jubox import JupyterNotebook, CodeCell\n'), ((523, 549), 'pathlib.Path', 'Path', (['notebook_file_simple'], {}), '(notebook_file_simple)\n', (527, 549), False, 'from pathlib import Path\n'), ((559, 580), 'jubo... |
import spot
from crome_logic.specification.temporal import LTL
from crome_logic.tools.crome_io import save_to_file
from crome_logic.typelement.robotic import BooleanAction, BooleanSensor
from crome_logic.typeset import Typeset
from crome_synthesis.controller import Controller
from crome_synthesis.controller.synthesis ... | [
"crome_synthesis.controller.synthesis.generate_controller",
"crome_logic.typelement.robotic.BooleanSensor",
"crome_synthesis.controller.Controller",
"crome_logic.tools.crome_io.save_to_file",
"crome_logic.typelement.robotic.BooleanAction",
"spot.automaton"
] | [((614, 653), 'crome_synthesis.controller.Controller', 'Controller', ([], {'assumptions': 'a', 'guarantees': 'g'}), '(assumptions=a, guarantees=g)\n', (624, 653), False, 'from crome_synthesis.controller import Controller\n'), ((878, 913), 'crome_synthesis.controller.synthesis.generate_controller', 'generate_controller'... |
from . import db
from werkzeug.security import generate_password_hash,check_password_hash
from datetime import datetime
import time
from flask_login import UserMixin
from . import login_manager
@login_manager.user_loader
def load_user(user_id):
return User.query.get(int(user_id))
# @login_manager.user_loader
# de... | [
"werkzeug.security.generate_password_hash",
"werkzeug.security.check_password_hash"
] | [((1023, 1055), 'werkzeug.security.generate_password_hash', 'generate_password_hash', (['password'], {}), '(password)\n', (1045, 1055), False, 'from werkzeug.security import generate_password_hash, check_password_hash\n'), ((1120, 1167), 'werkzeug.security.check_password_hash', 'check_password_hash', (['self.pass_secur... |
import logging
from django.db import connection
from prometheus_client import Gauge
from zentral.utils.prometheus import BasePrometheusMetricsView
from .models import Status
logger = logging.getLogger("zentral.core.compliance_checks.metrics_views")
class MetricsView(BasePrometheusMetricsView):
def add_complianc... | [
"logging.getLogger",
"django.db.connection.cursor",
"prometheus_client.Gauge"
] | [((185, 250), 'logging.getLogger', 'logging.getLogger', (['"""zentral.core.compliance_checks.metrics_views"""'], {}), "('zentral.core.compliance_checks.metrics_views')\n", (202, 250), False, 'import logging\n'), ((354, 456), 'prometheus_client.Gauge', 'Gauge', (['"""zentral_compliance_checks"""', '"""Zentral compliance... |
import click
import ast
import requests
import pandas as pd
import numpy as np
from pysradb.sraweb import SRAweb
import requests
import pysam
import re
import pyfastx
import pkg_resources # part of setuptools
import json
import warnings
import sys
import os
from .run_workflow import make_snakes
# Allows passing strin... | [
"pandas.read_csv",
"pkg_resources.require",
"click.File",
"pysam.AlignmentFile",
"click.BadParameter",
"os.path.exists",
"click.group",
"click.option",
"pysam.set_verbosity",
"pandas.DataFrame",
"pandas.merge",
"re.match",
"os.path.splitext",
"os.path.dirname",
"click.version_option",
... | [((1461, 1486), 'os.path.dirname', 'os.path.dirname', (['__file__'], {}), '(__file__)\n', (1476, 1486), False, 'import os\n'), ((15858, 15871), 'click.group', 'click.group', ([], {}), '()\n', (15869, 15871), False, 'import click\n'), ((15873, 15906), 'click.version_option', 'click.version_option', (['__version__'], {})... |
"""
Copyright (c) 2015 Red Hat, Inc
All rights reserved.
This software may be modified and distributed under the terms
of the BSD license. See the LICENSE file for details.
"""
from __future__ import print_function, unicode_literals
import logging
import os
import sys
import pytest
from atomic_reactor.buildimage i... | [
"logging.getLogger",
"atomic_reactor.core.DockerTasker",
"os.path.join",
"tests.fixtures.get_uuid",
"pytest.mark.parametrize",
"os.path.dirname",
"pytest.raises",
"tests.docker_mock.mock_docker",
"atomic_reactor.buildimage.BuildImageBuilder"
] | [((705, 746), 'logging.getLogger', 'logging.getLogger', (['"""atomic_reactor.tests"""'], {}), "('atomic_reactor.tests')\n", (722, 746), False, 'import logging\n'), ((780, 794), 'atomic_reactor.core.DockerTasker', 'DockerTasker', ([], {}), '()\n', (792, 794), False, 'from atomic_reactor.core import DockerTasker\n'), ((8... |
# Copyright (c) 2017-2018 Wind River Systems, Inc.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law... | [
"titanium_cloud.swagger.views.APIv1SwaggerJsonView.as_view",
"titanium_cloud.swagger.views.APIv1SwaggerJsonViewDepreciated.as_view",
"titanium_cloud.swagger.views.SwaggerJsonViewDepreciated.as_view",
"titanium_cloud.swagger.views.SwaggerJsonView.as_view",
"rest_framework.urlpatterns.format_suffix_patterns"
... | [((1518, 1553), 'rest_framework.urlpatterns.format_suffix_patterns', 'format_suffix_patterns', (['URLPATTERNS'], {}), '(URLPATTERNS)\n', (1540, 1553), False, 'from rest_framework.urlpatterns import format_suffix_patterns\n'), ((1061, 1097), 'titanium_cloud.swagger.views.SwaggerJsonViewDepreciated.as_view', 'SwaggerJson... |
from ofptHeader import ofptHeader
def ofptBarrierReq():
header = ofptHeader(20)
return header | [
"ofptHeader.ofptHeader"
] | [((67, 81), 'ofptHeader.ofptHeader', 'ofptHeader', (['(20)'], {}), '(20)\n', (77, 81), False, 'from ofptHeader import ofptHeader\n')] |
import pytest
from whylogs.core.metrics.nlp_metrics import NLPMetrics
from whylogs.proto import NLPMetricsMessage
def test_nlp_metrics():
nlp_metric = NLPMetrics()
assert nlp_metric.mer is not None
assert nlp_metric.mer.count == 0
nlp_metric.update("hello", "hello")
assert nlp_metric.mer.histogr... | [
"whylogs.core.metrics.nlp_metrics.NLPMetrics",
"whylogs.core.metrics.nlp_metrics.NLPMetrics.from_protobuf"
] | [((158, 170), 'whylogs.core.metrics.nlp_metrics.NLPMetrics', 'NLPMetrics', ([], {}), '()\n', (168, 170), False, 'from whylogs.core.metrics.nlp_metrics import NLPMetrics\n'), ((1069, 1081), 'whylogs.core.metrics.nlp_metrics.NLPMetrics', 'NLPMetrics', ([], {}), '()\n', (1079, 1081), False, 'from whylogs.core.metrics.nlp_... |
# https://leetcode.com/problems/ransom-note/
from collections import Counter
class Solution(object):
def canConstruct(self, ransomNote, magazine):
"""
:type ransomNote: str
:type magazine: str
:rtype: bool
"""
available = Counter(magazine)
for c in ransom... | [
"collections.Counter"
] | [((278, 295), 'collections.Counter', 'Counter', (['magazine'], {}), '(magazine)\n', (285, 295), False, 'from collections import Counter\n')] |
# InteractionSubmission:
import tensorflow as tf
import numpy as np
from waymo_open_dataset.protos import motion_submission_pb2
def eval_and_generate_submission(model, eval_dataset, num_modes = 6):
submission = motion_submission_pb2.MotionChallengeSubmission()
submission.submission_type = 2
submission.affiliatio... | [
"waymo_open_dataset.protos.motion_submission_pb2.MotionChallengeSubmission"
] | [((214, 263), 'waymo_open_dataset.protos.motion_submission_pb2.MotionChallengeSubmission', 'motion_submission_pb2.MotionChallengeSubmission', ([], {}), '()\n', (261, 263), False, 'from waymo_open_dataset.protos import motion_submission_pb2\n')] |
from django.conf import settings
from rest_framework.permissions import IsAdminUser
from rest_framework import status, viewsets, decorators
from quser.permissions import CURDPermissionsOrReadOnly
from rest_framework.response import Response
from . import models, serializers
from .filters import FileFilter
class TagV... | [
"rest_framework.response.Response",
"rest_framework.decorators.action"
] | [((836, 979), 'rest_framework.decorators.action', 'decorators.action', ([], {'methods': "['delete']", 'detail': '(False)', 'serializer_class': 'serializers.BulkDestroySerializer', 'permission_classes': '(IsAdminUser,)'}), "(methods=['delete'], detail=False, serializer_class=\n serializers.BulkDestroySerializer, perm... |
""" GitFeederController tests """
from django.test import TestCase
from django.contrib.auth.models import User
from app.logic.gitfeeder.models.FeedModel import FeedEntry
from app.logic.gitrepo.models.GitProjectModel import GitProjectEntry
from app.logic.gitrepo.models.GitUserModel import GitUserEntry
from app.logic.bl... | [
"app.logic.gitfeeder.models.FeedModel.FeedEntry.objects.all",
"app.logic.gitrepo.models.GitUserModel.GitUserEntry.objects.create",
"app.logic.commandrepo.models.CommandGroupModel.CommandGroupEntry.objects.create",
"app.logic.commandrepo.models.CommandGroupModel.CommandGroupEntry.objects.all",
"app.logic.git... | [((528, 588), 'django.contrib.auth.models.User.objects.create_user', 'User.objects.create_user', (['"""<EMAIL>"""', '"""<EMAIL>"""', '"""<PASSWORD>"""'], {}), "('<EMAIL>', '<EMAIL>', '<PASSWORD>')\n", (552, 588), False, 'from django.contrib.auth.models import User\n'), ((613, 785), 'app.logic.bluesteelworker.models.Wor... |
from django.conf.urls import patterns, url
from irods_browser_app import views
urlpatterns = patterns('',
url(r'^login/$',views.login, name='irods_login'),
url(r'^store/$',views.store, name='irods_store'),
url(r'^upload/$',views.upload, name='irods_upload'),
url(r'^upload_add/$',views.upload_add, name=... | [
"django.conf.urls.url"
] | [((111, 159), 'django.conf.urls.url', 'url', (['"""^login/$"""', 'views.login'], {'name': '"""irods_login"""'}), "('^login/$', views.login, name='irods_login')\n", (114, 159), False, 'from django.conf.urls import patterns, url\n'), ((165, 213), 'django.conf.urls.url', 'url', (['"""^store/$"""', 'views.store'], {'name':... |
#!/usr/bin/python
"""
Script to initialise CROP database.
"""
import sys
from crop.constants import SQL_CONNECTION_STRING, SQL_DBNAME
from crop.db import create_database
def confirm(question):
"""
Ask user to enter Y or N (case-insensitive).
:return: True if the answer is Y.
"""
answer = ""
... | [
"crop.db.create_database"
] | [((518, 568), 'crop.db.create_database', 'create_database', (['SQL_CONNECTION_STRING', 'SQL_DBNAME'], {}), '(SQL_CONNECTION_STRING, SQL_DBNAME)\n', (533, 568), False, 'from crop.db import create_database\n')] |
import rules
# Projectroles dependency
from projectroles import rules as pr_rules # To access common predicates
# Predicates -------------------------------------------------------------
# TODO: If we need to assign new predicates, we do it here
# Rules ----------------------------------------------------------... | [
"rules.add_perm"
] | [((526, 695), 'rules.add_perm', 'rules.add_perm', (['"""samplesheets.view_sheet"""', '(pr_rules.is_project_owner | pr_rules.is_project_delegate | pr_rules.\n is_project_contributor | pr_rules.is_project_guest)'], {}), "('samplesheets.view_sheet', pr_rules.is_project_owner |\n pr_rules.is_project_delegate | pr_rul... |
from django.conf.urls.defaults import patterns, url, include
from pycash.controllers import TaxController as controller
urlpatterns = patterns('',
(r'^upcomingList$', controller.upcomingList),
(r'^upcoming$', controller.upcoming),
url(r'^pay$', controller.pay, name="tax_pay"),
(r'^list$', controller.li... | [
"django.conf.urls.defaults.url"
] | [((244, 288), 'django.conf.urls.defaults.url', 'url', (['"""^pay$"""', 'controller.pay'], {'name': '"""tax_pay"""'}), "('^pay$', controller.pay, name='tax_pay')\n", (247, 288), False, 'from django.conf.urls.defaults import patterns, url, include\n'), ((329, 386), 'django.conf.urls.defaults.url', 'url', (['"""^save$"""'... |
from abc import abstractmethod
from functools import wraps
from dd.api.workflow.utils import normalize_columns
class CallableBuilder(object):
# TODO : remove reference to dataset
# replace with lighter reference to dataset.output_table
def __init__(self, context, dataset=None):
self.context = con... | [
"dd.api.workflow.utils.normalize_columns",
"functools.wraps"
] | [((1409, 1435), 'functools.wraps', 'wraps', (['self.transformation'], {}), '(self.transformation)\n', (1414, 1435), False, 'from functools import wraps\n'), ((14081, 14102), 'functools.wraps', 'wraps', (['self.operation'], {}), '(self.operation)\n', (14086, 14102), False, 'from functools import wraps\n'), ((5911, 5940)... |
# Generated by Django 2.1.15 on 2021-05-21 08:24
from django.db import migrations, models
import django.db.models.deletion
class Migration(migrations.Migration):
initial = True
dependencies = [
]
operations = [
migrations.CreateModel(
name='Company',
fields=[
... | [
"django.db.models.ForeignKey",
"django.db.models.AutoField",
"django.db.models.CharField",
"django.db.models.IntegerField"
] | [((337, 430), 'django.db.models.AutoField', 'models.AutoField', ([], {'auto_created': '(True)', 'primary_key': '(True)', 'serialize': '(False)', 'verbose_name': '"""ID"""'}), "(auto_created=True, primary_key=True, serialize=False,\n verbose_name='ID')\n", (353, 430), False, 'from django.db import migrations, models\... |
r"""
Echelle Spectrum
----------------
An abstract base class for a high resolution spectrum, for some echelle order :math:`m \in ` out of :math:`M` total orders, each with vectors for wavelength, flux, and uncertainty, e.g. :math:`F_m(\lambda)`. This class is a subclass of specutils' Spectrum1D and is intended to ha... | [
"logging.getLogger",
"numpy.sqrt",
"celerite2.GaussianProcess",
"numpy.log",
"scipy.signal.savgol_filter",
"scipy.interpolate.interp1d",
"numpy.isfinite",
"specutils.spectra.spectral_region.SpectralRegion",
"copy.deepcopy",
"numpy.arange",
"numpy.mean",
"numpy.exp",
"astropy.nddata.StdDevUnc... | [((1464, 1491), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (1481, 1491), False, 'import logging\n'), ((1543, 1598), 'warnings.simplefilter', 'warnings.simplefilter', (['"""ignore"""'], {'category': 'VerifyWarning'}), "('ignore', category=VerifyWarning)\n", (1564, 1598), False, 'import... |
import nltk
import string
import requests
import json
from io import StringIO
from html.parser import HTMLParser
import os
import time
from sys import platform
path = ""
if platform == "linux" or platform == "linux2":
path = os.path.dirname(os.path.realpath(__file__)).replace("test","data/nltk")
elif ... | [
"os.path.realpath",
"nltk.data.find",
"nltk.data.path.append",
"nltk.download"
] | [((541, 568), 'nltk.data.path.append', 'nltk.data.path.append', (['path'], {}), '(path)\n', (562, 568), False, 'import nltk\n'), ((627, 647), 'nltk.data.find', 'nltk.data.find', (['find'], {}), '(find)\n', (641, 647), False, 'import nltk\n'), ((683, 721), 'nltk.download', 'nltk.download', (['name'], {'download_dir': 'p... |
#!/usr/bin/env python
# coding=utf-8
from db import Model
__doc__ = "记录数据库层封装"
recordModel = Model('record')
#类添加函数
# join查找 | [
"db.Model"
] | [((96, 111), 'db.Model', 'Model', (['"""record"""'], {}), "('record')\n", (101, 111), False, 'from db import Model\n')] |
# Licensed under a 3-clause BSD style license - see LICENSE.rst
"""Hillas shower parametrization.
TODO:
-----
- Should have a separate function or option to compute 3rd order
moments + asymmetry (which are not always needed)
- remove alpha calculation (which is only about (0,0), and make a get
alpha function tha... | [
"numpy.abs",
"collections.namedtuple",
"numpy.sqrt",
"numpy.asanyarray",
"numpy.sum",
"numpy.arctan2",
"numpy.row_stack",
"numpy.hypot"
] | [((608, 686), 'collections.namedtuple', 'namedtuple', (['"""MomentParameters"""', '"""size,cen_x,cen_y,length,width,r,phi,psi,miss"""'], {}), "('MomentParameters', 'size,cen_x,cen_y,length,width,r,phi,psi,miss')\n", (618, 686), False, 'from collections import namedtuple\n'), ((863, 933), 'collections.namedtuple', 'name... |
# Generated by Django 4.0.1 on 2022-01-12 23:44
from django.db import migrations, models
class Migration(migrations.Migration):
dependencies = [
('points', '0003_spend_alter_transaction_timestamp'),
]
operations = [
migrations.AddField(
model_name='spend',
name='... | [
"django.db.models.TextField"
] | [((348, 387), 'django.db.models.TextField', 'models.TextField', ([], {'blank': '(True)', 'null': '(True)'}), '(blank=True, null=True)\n', (364, 387), False, 'from django.db import migrations, models\n')] |
import json
from multiprocessing import Pool, cpu_count
from catti.IO.portfolio import loadPortfolioInfo
from catti.IO.general import *
from catti.indicators.calculations import indicatorsCalculator
from catti.indicators.signals import signalsCalculator
from catti.indicators.signalsValidation import signalsValida... | [
"multiprocessing.cpu_count",
"catti.IO.portfolio.loadPortfolioInfo",
"tqdm.tqdm",
"json.dump"
] | [((976, 995), 'catti.IO.portfolio.loadPortfolioInfo', 'loadPortfolioInfo', ([], {}), '()\n', (993, 995), False, 'from catti.IO.portfolio import loadPortfolioInfo\n'), ((2184, 2240), 'tqdm.tqdm', 'tqdm.tqdm', (['chunks'], {'desc': '"""Chunk completion"""', 'total': 'splits'}), "(chunks, desc='Chunk completion', total=sp... |
import indicoio
from celery import Celery
from flask import Flask
from flask_bootstrap import Bootstrap
from flask_googlemaps import GoogleMaps
from flask_mail import Mail
from flask_mongoengine import MongoEngine
from app.celery.factory import init_celery
from config import config, DEVELOPMENT_CONFIG_NAME
celery = C... | [
"flask_mail.Mail",
"flask.Flask",
"celery.Celery",
"flask_googlemaps.GoogleMaps",
"app.celery.factory.init_celery",
"flask_bootstrap.Bootstrap",
"flask_mongoengine.MongoEngine"
] | [((319, 332), 'celery.Celery', 'Celery', (['"""app"""'], {}), "('app')\n", (325, 332), False, 'from celery import Celery\n'), ((340, 346), 'flask_mail.Mail', 'Mail', ([], {}), '()\n', (344, 346), False, 'from flask_mail import Mail\n'), ((352, 365), 'flask_mongoengine.MongoEngine', 'MongoEngine', ([], {}), '()\n', (363... |
# -*- coding:utf-8 -*-
import logging
import stores
release = "0.1.1"
logging.getLogger("dbsync")
class DBSync():
def __init__(self):
pass
def syncer(self, syncer):
return self
def validator(self, validator):
return self
def serializer(self, serializer):
return se... | [
"logging.getLogger"
] | [((72, 99), 'logging.getLogger', 'logging.getLogger', (['"""dbsync"""'], {}), "('dbsync')\n", (89, 99), False, 'import logging\n')] |
# -*- coding: utf-8 -*-
import asyncio
from config import (CHECK_SERVER_INTERVAL, CHECK_SERVER_INTERVAL_MAX,
CRON_LOOP_INTERVAL)
from discord import Activity, ActivityType
from discord.errors import Forbidden, NotFound
from discord.ext import commands, tasks
from modules.db import Servers
from modu... | [
"modules.utils.embed_generator",
"asyncio.get_event_loop",
"modules.utils.get_server_info",
"modules.utils.stop_server",
"discord.Activity",
"asyncio.sleep",
"modules.logging.logger.info",
"discord.ext.tasks.loop",
"modules.db.Servers.filter"
] | [((635, 689), 'discord.ext.tasks.loop', 'tasks.loop', ([], {'seconds': 'CRON_LOOP_INTERVAL', 'reconnect': '(True)'}), '(seconds=CRON_LOOP_INTERVAL, reconnect=True)\n', (645, 689), False, 'from discord.ext import commands, tasks\n'), ((539, 563), 'asyncio.get_event_loop', 'asyncio.get_event_loop', ([], {}), '()\n', (561... |
from django.urls import path
from dashboards.views import DashboardView
urlpatterns = [
path(
"<slug:organisation>/dashboards/<slug:dashboard>/",
DashboardView.as_view(),
name="dashboard_detail",
),
]
| [
"dashboards.views.DashboardView.as_view"
] | [((168, 191), 'dashboards.views.DashboardView.as_view', 'DashboardView.as_view', ([], {}), '()\n', (189, 191), False, 'from dashboards.views import DashboardView\n')] |
from datetime import date
from django.http import JsonResponse
from rest_framework import viewsets
from rest_framework.permissions import IsAdminUser
from .models import CampPlan
from .serializers import CampPlanSerializer
# ===========================
# Camp Plan ViewSet
# ===========================
class CampPla... | [
"datetime.date.today",
"django.http.JsonResponse"
] | [((1544, 1603), 'django.http.JsonResponse', 'JsonResponse', (["{'success': 'Successfully created camp plan'}"], {}), "({'success': 'Successfully created camp plan'})\n", (1556, 1603), False, 'from django.http import JsonResponse\n'), ((1637, 1679), 'django.http.JsonResponse', 'JsonResponse', (["{'error': serializer.err... |
# coding=utf-8
# 导入自己的函数包d2lzh_pytorch,注意要先将目标包的父路径添加到系统路径中
import sys
sys.path.append(r".")
from d2lzh_pytorch import data_process, plot, train, rnn
import torch
import time
import torchvision
from PIL import Image
from torch import nn,optim
from torch.utils.data import Dataset,DataLoader
"""
这一节介绍了如何对图像数据集进行增广以获得更... | [
"PIL.Image.open",
"torch.nn.CrossEntropyLoss",
"torch.utils.data.DataLoader",
"torchvision.transforms.RandomHorizontalFlip",
"d2lzh_pytorch.plot.set_figsize",
"torch.cuda.is_available",
"d2lzh_pytorch.rnn.resnet18",
"torchvision.datasets.CIFAR10",
"d2lzh_pytorch.plot.show_images",
"torchvision.tra... | [((72, 92), 'sys.path.append', 'sys.path.append', (['"""."""'], {}), "('.')\n", (87, 92), False, 'import sys\n'), ((496, 514), 'd2lzh_pytorch.plot.set_figsize', 'plot.set_figsize', ([], {}), '()\n', (512, 514), False, 'from d2lzh_pytorch import data_process, plot, train, rnn\n'), ((525, 567), 'PIL.Image.open', 'Image.o... |
import sphinx_rtd_theme
project = 'Bitcoin DCA'
copyright = '2021, <NAME>'
author = '<NAME>'
extensions = []
templates_path = ['_templates']
exclude_patterns = ['_build', 'Thumbs.db', '.DS_Store']
html_theme = 'sphinx_rtd_theme'
html_static_path = ['_static']
pygments_style = 'sphinx'
html_theme_path = [sphinx_rtd_the... | [
"sphinx_rtd_theme.get_html_theme_path"
] | [((306, 344), 'sphinx_rtd_theme.get_html_theme_path', 'sphinx_rtd_theme.get_html_theme_path', ([], {}), '()\n', (342, 344), False, 'import sphinx_rtd_theme\n')] |
from django.contrib import admin
from app_restaurantes.models import Restaurante
# Register your models here.
admin.site.register(Restaurante)
| [
"django.contrib.admin.site.register"
] | [((112, 144), 'django.contrib.admin.site.register', 'admin.site.register', (['Restaurante'], {}), '(Restaurante)\n', (131, 144), False, 'from django.contrib import admin\n')] |
# -*- coding: utf-8 -*-
# Generated by Django 1.9.9 on 2016-09-09 15:44
from __future__ import unicode_literals
from django.db import migrations, models
import django.db.models.deletion
class Migration(migrations.Migration):
dependencies = [
('wagtailcore', '0029_unicode_slugfield_dj19'),
('peop... | [
"django.db.models.OneToOneField",
"django.db.models.ForeignKey",
"django.db.models.ManyToManyField",
"django.db.migrations.AlterModelOptions",
"django.db.models.AutoField",
"django.db.models.CharField"
] | [((1558, 1644), 'django.db.migrations.AlterModelOptions', 'migrations.AlterModelOptions', ([], {'name': '"""person"""', 'options': "{'ordering': ('last_name',)}"}), "(name='person', options={'ordering': (\n 'last_name',)})\n", (1586, 1644), False, 'from django.db import migrations, models\n'), ((1782, 1837), 'django... |
import pandas as pd
from matplotlib import pyplot as plt
from scipy.interpolate import spline
from scipy.ndimage.filters import gaussian_filter1d
import numpy as np
#%%
# TZ numbers via https://www.fangraphs.com/leaders.aspx?pos=3b&stats=fld&lg=all&qual=y&type=0&season=2017&month=0&season1=1961&ind=1&team=0&rost=0&ag... | [
"numpy.ones",
"pandas.read_csv",
"matplotlib.pyplot.ylabel",
"matplotlib.pyplot.xlabel",
"matplotlib.pyplot.plot",
"matplotlib.pyplot.figure",
"scipy.ndimage.filters.gaussian_filter1d",
"matplotlib.pyplot.tight_layout",
"matplotlib.pyplot.title",
"matplotlib.pyplot.xlim",
"pandas.concat"
] | [((347, 368), 'pandas.read_csv', 'pd.read_csv', (['"""TZ.csv"""'], {}), "('TZ.csv')\n", (358, 368), True, 'import pandas as pd\n'), ((666, 688), 'pandas.read_csv', 'pd.read_csv', (['"""UZR.csv"""'], {}), "('UZR.csv')\n", (677, 688), True, 'import pandas as pd\n'), ((791, 811), 'pandas.concat', 'pd.concat', (['[df, df2]... |
import pytest
import numpy as np
import pandas as pd
from endaq.calc import rotation
@pytest.mark.parametrize(
'quat, euler',
[
((0., 0., 0., 1.), (0., 0., 0.)),
((0., 0., 0., -1.), (0., 0., 0.)),
((1., 0., 0., 0.), (np.pi, 0., 0.)),
(... | [
"pandas.DataFrame",
"pytest.mark.parametrize",
"endaq.calc.rotation.quaternion_to_euler",
"pytest.raises"
] | [((90, 357), 'pytest.mark.parametrize', 'pytest.mark.parametrize', (['"""quat, euler"""', '[((0.0, 0.0, 0.0, 1.0), (0.0, 0.0, 0.0)), ((0.0, 0.0, 0.0, -1.0), (0.0, 0.0,\n 0.0)), ((1.0, 0.0, 0.0, 0.0), (np.pi, 0.0, 0.0)), ((0.0, 1.0, 0.0, 0.0),\n (np.pi, 0.0, np.pi)), ((0.0, 0.0, 1.0, 0.0), (0.0, 0.0, np.pi))]'], {... |
# Copyright 2021 <NAME>
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
# http://www.apache.org/licenses/LICENSE-2.0
# Unless required by applicable law or agreed to in writing, software... | [
"toshiba_ac.amqp_api.ToshibaAcAmqpApi",
"toshiba_ac.device.ToshibaAcDevice",
"toshiba_ac.http_api.ToshibaAcHttpApi"
] | [((823, 855), 'toshiba_ac.http_api.ToshibaAcHttpApi', 'ToshibaAcHttpApi', (['user', 'password'], {}), '(user, password)\n', (839, 855), False, 'from toshiba_ac.http_api import ToshibaAcHttpApi\n'), ((1096, 1137), 'toshiba_ac.amqp_api.ToshibaAcAmqpApi', 'ToshibaAcAmqpApi', (['self.reg_info.sas_token'], {}), '(self.reg_i... |
"""JSON-LD utilities.
"""
__all__ = ('encode_jsonld', 'JsonLdEncoder', 'decode_jsonld')
import datetime
import json
def encode_jsonld(jsonld_dataset, **kwargs):
"""Encode a JSON-LD dataset into a string.
Parameters
----------
jsonld_dataset : `dict`
A JSON-LD dataset.
kwargs
Key... | [
"datetime.datetime.strptime",
"json.JSONDecoder",
"json.JSONEncoder.default"
] | [((2180, 2236), 'json.JSONDecoder', 'json.JSONDecoder', ([], {'object_pairs_hook': '_decode_object_pairs'}), '(object_pairs_hook=_decode_object_pairs)\n', (2196, 2236), False, 'import json\n'), ((1044, 1079), 'json.JSONEncoder.default', 'json.JSONEncoder.default', (['self', 'obj'], {}), '(self, obj)\n', (1068, 1079), F... |
from __future__ import absolute_import
from __future__ import division
from __future__ import print_function
import tensorflow as tf
def loss2(logits, labels, num_classes, scope, head=None):
with tf.name_scope(scope):
logits = tf.reshape(logits, (-1, num_classes))
softmax = tf.nn.softmax(logits) + ... | [
"tensorflow.name_scope",
"tensorflow.nn.softmax",
"tensorflow.reshape",
"tensorflow.reduce_mean",
"tensorflow.log"
] | [((838, 867), 'tensorflow.reduce_mean', 'tf.reduce_mean', (['cross_entropy'], {}), '(cross_entropy)\n', (852, 867), True, 'import tensorflow as tf\n'), ((201, 221), 'tensorflow.name_scope', 'tf.name_scope', (['scope'], {}), '(scope)\n', (214, 221), True, 'import tensorflow as tf\n'), ((240, 277), 'tensorflow.reshape', ... |
import pytest
from pretalx.event.models import Event
@pytest.mark.django_db
def test_shred_used_event(resource, answered_choice_question, personal_answer, rejected_submission, deleted_submission, mail, sent_mail, room_availability, slot, unreleased_slot, past_slot, feedback, canceled_talk, review, information, other... | [
"pretalx.event.models.Event.objects.count"
] | [((340, 361), 'pretalx.event.models.Event.objects.count', 'Event.objects.count', ([], {}), '()\n', (359, 361), False, 'from pretalx.event.models import Event\n'), ((426, 447), 'pretalx.event.models.Event.objects.count', 'Event.objects.count', ([], {}), '()\n', (445, 447), False, 'from pretalx.event.models import Event\... |
# -*- coding: utf-8 -*-
""" Setup for pytrafik """
from distutils.core import setup
setup(
name='pytrafik',
version='0.2.1',
description='PyTrafik',
long_description='Wrapper for Västtrafik public API.',
url='https://github.com/axelniklasson/PyTrafik',
download_url = 'https://github.com/axelni... | [
"distutils.core.setup"
] | [((86, 901), 'distutils.core.setup', 'setup', ([], {'name': '"""pytrafik"""', 'version': '"""0.2.1"""', 'description': '"""PyTrafik"""', 'long_description': '"""Wrapper for Västtrafik public API."""', 'url': '"""https://github.com/axelniklasson/PyTrafik"""', 'download_url': '"""https://github.com/axelniklasson/PyTrafik... |
#coding=utf-8
'''
Created on 2015-10-23
@author: Devuser
'''
from doraemon.ci.models import CITask,CITaskHistory
from gatesidelib.common.simplelogger import SimpleLogger
from django.contrib.admin.models import DELETION,CHANGE,ADDITION
from doraemon.project.models import Product,Project,Tag
from business.project.projec... | [
"doraemon.ci.models.CITaskHistory",
"doraemon.project.models.Project.objects.all",
"doraemon.ci.models.CITask",
"doraemon.home.models.TaskQueue.objects.all",
"business.ci.ci_task_config_service.CITaskConfigService.copy_config",
"business.project.project_service.ProjectService.get_projects_include_me",
"... | [((1414, 1461), 'business.project.project_service.ProjectService.get_projects_include_me', 'ProjectService.get_projects_include_me', (['request'], {}), '(request)\n', (1452, 1461), False, 'from business.project.project_service import ProjectService\n'), ((5529, 5537), 'doraemon.ci.models.CITask', 'CITask', ([], {}), '(... |
import struct
from pox.lib.addresses import EthAddr, IPAddr
"""
Utility to convert MAC and IP to/from integers.
"""
def eth_to_int(addr):
if not isinstance(addr, EthAddr):
return None
value = 0
raw = addr.toRaw()
for i in range(len(raw)):
byte_shift = 5 - i
byte = raw[i]
... | [
"struct.unpack",
"pox.lib.addresses.IPAddr",
"struct.pack"
] | [((821, 833), 'pox.lib.addresses.IPAddr', 'IPAddr', (['addr'], {}), '(addr)\n', (827, 833), False, 'from pox.lib.addresses import EthAddr, IPAddr\n'), ((337, 361), 'struct.unpack', 'struct.unpack', (['"""B"""', 'byte'], {}), "('B', byte)\n", (350, 361), False, 'import struct\n'), ((536, 564), 'struct.pack', 'struct.pac... |
import os
import json
import glob
from rcnn.utils import get_path_with_annotation,get_path_with_annotation_ratio
from rcnn.utils import get_weight_path
__disease__ = ['Covid-Seg','Lung_Tumor']
__net__ = ['rcnn_unet']
__mode__ = ['cls','seg','mtl']
json_path = {
'Cervical':'/staff/shijun/torch_projects/Med_Seg/c... | [
"json.load",
"rcnn.utils.get_weight_path",
"os.path.join"
] | [((2639, 2665), 'rcnn.utils.get_weight_path', 'get_weight_path', (['CKPT_PATH'], {}), '(CKPT_PATH)\n', (2654, 2665), False, 'from rcnn.utils import get_weight_path\n'), ((1402, 1415), 'json.load', 'json.load', (['fp'], {}), '(fp)\n', (1411, 1415), False, 'import json\n'), ((2092, 2132), 'os.path.join', 'os.path.join', ... |
# -*- python -*-
import math
import numpy
import Shadow
from Shadow.ShadowPreprocessorsXraylib import prerefl, pre_mlayer, bragg
from srxraylib.sources import srfunc
from sirepo.template import transfer_mat_bl
from pykern.pkcollections import PKDict
from pykern import pkjson
sigmax = 0.0045000000000000005
sigdix = 2.... | [
"pykern.pkjson.dump_pretty",
"Shadow.OE",
"Shadow.ShadowTools.plotxy",
"sirepo.template.transfer_mat_bl.create_mat_rays",
"math.sqrt",
"numpy.matrix",
"numpy.transpose"
] | [((418, 458), 'sirepo.template.transfer_mat_bl.create_mat_rays', 'transfer_mat_bl.create_mat_rays', (['epsilon'], {}), '(epsilon)\n', (449, 458), False, 'from sirepo.template import transfer_mat_bl\n'), ((472, 586), 'numpy.matrix', 'numpy.matrix', (['[[sigmax ** 2, 0, 0, 0], [0, sigdix ** 2, 0, 0], [0, 0, sigmaz ** 2, ... |
import json
import logging
import os
import re
from collections import namedtuple
from copy import deepcopy
from typing import Any, Dict, List, Tuple
import numpy as np
import pandas as pd
import spacy
from scirex_utilities.analyse_pwc_entity_results import *
from scirex_utilities.entity_utils import *
from spacy.toke... | [
"logging.basicConfig",
"json.loads",
"collections.namedtuple",
"os.listdir",
"argparse.ArgumentParser",
"spacy.load",
"tqdm.tqdm",
"json.dumps",
"spacy.tokens.Doc",
"os.path.join",
"os.path.isdir",
"pandas.concat",
"re.finditer",
"copy.deepcopy",
"pandas.DataFrame",
"numpy.cumsum",
"... | [((357, 370), 'tqdm.tqdm.pandas', 'tqdm.pandas', ([], {}), '()\n', (368, 370), False, 'from tqdm import tqdm\n'), ((383, 482), 'collections.namedtuple', 'namedtuple', (['"""Span"""', "['start', 'end', 'token_start', 'token_end', 'entity', 'links', 'modified']"], {}), "('Span', ['start', 'end', 'token_start', 'token_end... |
# Copyright (C) 2018 Innoviz Technologies
# All rights reserved.
#
# This software may be modified and distributed under the terms
# of the BSD 3-Clause license. See the LICENSE file for details.
import pandas as pd
import os
import numpy as np
from utilities.math_utils import RotationTranslationData
from visualizatio... | [
"utilities.data_utils.enumerate_frames",
"os.path.join",
"os.getcwd",
"utilities.math_utils.RotationTranslationData",
"numpy.concatenate",
"numpy.linalg.norm",
"utilities.data_utils.read_all_data",
"utilities.data_utils.frame_to_filename",
"visualizations.vis.pcshow"
] | [((463, 516), 'os.path.join', 'os.path.join', (['base_dir', '"""data_examples"""', '"""test_video"""'], {}), "(base_dir, 'data_examples', 'test_video')\n", (475, 516), False, 'import os\n'), ((642, 680), 'utilities.data_utils.enumerate_frames', 'data_utils.enumerate_frames', (['video_dir'], {}), '(video_dir)\n', (669, ... |
from typing import Sequence, Optional
import pandas as pd
import numpy as np
def formatted_corr_df(df: pd.DataFrame, cols: Optional[Sequence[str]] = None) -> pd.DataFrame:
"""
Calculates correlations on a DataFrame and displays only the lower triangular of the
resulting correlation DataFrame.
:param... | [
"numpy.tril"
] | [((738, 749), 'numpy.tril', 'np.tril', (['df'], {}), '(df)\n', (745, 749), True, 'import numpy as np\n')] |
# -*- coding: utf-8 -*-
__doc__="返回选择物体的类型"
import rpw
from rpw import revit, DB, UI,db,doc
#from System.Collections.Generic import List
#import json
#from scriptutils import this_script
#from scriptutils.userinput import CommandSwitchWindow
#import subprocess as sp
#Change Selected Grid From 3D to 2D
#selection = rpw... | [
"rpw.ui.Selection",
"rpw.db.Transaction.ensure"
] | [((465, 483), 'rpw.ui.Selection', 'rpw.ui.Selection', ([], {}), '()\n', (481, 483), False, 'import rpw\n'), ((635, 680), 'rpw.db.Transaction.ensure', 'rpw.db.Transaction.ensure', (['"""Hide_Grid_Bubble"""'], {}), "('Hide_Grid_Bubble')\n", (660, 680), False, 'import rpw\n')] |
import json
import urllib.request
import urllib.parse
import csv
import login
class Action:
def __init__(self, tenantId, appId, appSecret, body, url, filename, column):
self.tenantId = tenantId
self.appId = appId
self.appSecret = appSecret
self.body = body
self.url = url
... | [
"login.Login",
"csv.DictReader",
"json.dumps"
] | [((426, 480), 'login.Login', 'login.Login', (['self.tenantId', 'self.appId', 'self.appSecret'], {}), '(self.tenantId, self.appId, self.appSecret)\n', (437, 480), False, 'import login\n'), ((598, 621), 'csv.DictReader', 'csv.DictReader', (['csvfile'], {}), '(csvfile)\n', (612, 621), False, 'import csv\n'), ((1061, 1082)... |
from flask import Flask
from flask_login import LoginManager
from flask_wtf import CsrfProtect
login_manager = LoginManager()
def create_app():
app = Flask(__name__)
app.config.from_pyfile("config/setting.py")
csrf = CsrfProtect()
csrf.init_app(app)
register_blueprint(app)
create_db(app)
... | [
"flask_login.LoginManager",
"flask_wtf.CsrfProtect",
"flask.Flask"
] | [((112, 126), 'flask_login.LoginManager', 'LoginManager', ([], {}), '()\n', (124, 126), False, 'from flask_login import LoginManager\n'), ((157, 172), 'flask.Flask', 'Flask', (['__name__'], {}), '(__name__)\n', (162, 172), False, 'from flask import Flask\n'), ((232, 245), 'flask_wtf.CsrfProtect', 'CsrfProtect', ([], {}... |
#! /usr/bin/env python3
# -*- coding: utf-8 -*-
"""
parse_bib_file.py: Convert raw bibtex file to yml format and md pages.
"""
import os
import yaml
import bibtexparser as bb
def get_link(entry):
if 'url' in entry.keys():
return entry['url']
elif 'doi' in entry.keys():
return f"https://doi.... | [
"os.path.exists",
"yaml.dump",
"os.path.join",
"sys.exit",
"bibtexparser.load"
] | [((1113, 1134), 'os.path.exists', 'os.path.exists', (['fname'], {}), '(fname)\n', (1127, 1134), False, 'import os\n'), ((1545, 1555), 'bibtexparser.load', 'bb.load', (['f'], {}), '(f)\n', (1552, 1555), True, 'import bibtexparser as bb\n'), ((2510, 2561), 'os.path.join', 'os.path.join', (['out_folder_pages', "(entry['ID... |
# Generated by Django 2.2.13 on 2020-07-22 12:47
from django.db import migrations, models
class Migration(migrations.Migration):
dependencies = [
('apis_vocabularies', '0001_initial'),
]
operations = [
migrations.AddField(
model_name='texttype',
name='lang',
... | [
"django.db.models.CharField"
] | [((334, 501), 'django.db.models.CharField', 'models.CharField', ([], {'blank': '(True)', 'default': '"""deu"""', 'help_text': '"""The ISO 639-3 (or 2) code for the label\'s language."""', 'max_length': '(3)', 'null': '(True)', 'verbose_name': '"""ISO Code"""'}), '(blank=True, default=\'deu\', help_text=\n "The ISO 6... |
import torch
import get_data
import numpy as np
import torchaudio
def number_of_correct(pred, target):
return pred.squeeze().eq(target).sum().item()
def get_likely_index(tensor):
return tensor.argmax(dim=-1)
def compute_accuracy(model, data_loader, device):
model.eval()
correct = 0
for data, t... | [
"torch.nn.Softmax",
"get_data.prepare_wav",
"torchaudio.load",
"torch.unsqueeze",
"get_data.idx_to_label",
"numpy.argsort"
] | [((707, 750), 'get_data.prepare_wav', 'get_data.prepare_wav', (['waveform', 'sample_rate'], {}), '(waveform, sample_rate)\n', (727, 750), False, 'import get_data\n'), ((938, 956), 'numpy.argsort', 'np.argsort', (['(-probs)'], {}), '(-probs)\n', (948, 956), True, 'import numpy as np\n'), ((1156, 1181), 'torchaudio.load'... |
import os
import sys
import numpy as np
from sklearn import linear_model
from sklearn.metrics import classification_report
from sklearn.metrics import confusion_matrix
from scipy.sparse import csr_matrix
import csv
def ReadCsv(fileName, mode):
in_file = open(fileName)
reader = csv.reader(in_file, delimiter='\t... | [
"csv.reader"
] | [((287, 337), 'csv.reader', 'csv.reader', (['in_file'], {'delimiter': '"""\t"""', 'quotechar': '"""\\""""'}), '(in_file, delimiter=\'\\t\', quotechar=\'"\')\n', (297, 337), False, 'import csv\n')] |
import string
from src import config
def password_generator():
alphabet = string.digits
alphabet += string.ascii_lowercase
base = len(alphabet)
shift = 0
length = int(config.get('minimal_password_length'))
while True:
password = ''
temp = shift
while temp > 0:
... | [
"src.config.get"
] | [((191, 228), 'src.config.get', 'config.get', (['"""minimal_password_length"""'], {}), "('minimal_password_length')\n", (201, 228), False, 'from src import config\n')] |
from django.db import models
from django.contrib.auth import get_user_model
class TextAnswer(models.Model):
"""Model definition for TextAnswer."""
user = models.ForeignKey(get_user_model(), on_delete=models.CASCADE)
question = models.ForeignKey('Question', on_delete=models.CASCADE)
answer = models.Ch... | [
"django.contrib.auth.get_user_model",
"django.db.models.CharField",
"django.db.models.ForeignKey"
] | [((242, 297), 'django.db.models.ForeignKey', 'models.ForeignKey', (['"""Question"""'], {'on_delete': 'models.CASCADE'}), "('Question', on_delete=models.CASCADE)\n", (259, 297), False, 'from django.db import models\n'), ((311, 353), 'django.db.models.CharField', 'models.CharField', (['"""Answer"""'], {'max_length': '(10... |
from django.conf.urls import patterns, url
from qs import views
urlpatterns = patterns('',
url(r'^$', views.index, name='index'),
url(r'^sendqs/$', views.sendqs, name='sendqs'),
url(r'^show/$', views.showall, name='showall'),
)
| [
"django.conf.urls.url"
] | [((97, 133), 'django.conf.urls.url', 'url', (['"""^$"""', 'views.index'], {'name': '"""index"""'}), "('^$', views.index, name='index')\n", (100, 133), False, 'from django.conf.urls import patterns, url\n'), ((140, 185), 'django.conf.urls.url', 'url', (['"""^sendqs/$"""', 'views.sendqs'], {'name': '"""sendqs"""'}), "('^... |
import cv2
import numpy as np
from plantcv.plantcv.transform import nonuniform_illumination
def test_nonuniform_illumination_rgb(transform_test_data):
"""Test for PlantCV."""
# Load rgb image
rgb_img = cv2.imread(transform_test_data.small_rgb_img)
corrected = nonuniform_illumination(img=rgb_img, ksize... | [
"numpy.mean",
"plantcv.plantcv.transform.nonuniform_illumination",
"cv2.imread"
] | [((216, 261), 'cv2.imread', 'cv2.imread', (['transform_test_data.small_rgb_img'], {}), '(transform_test_data.small_rgb_img)\n', (226, 261), False, 'import cv2\n'), ((278, 324), 'plantcv.plantcv.transform.nonuniform_illumination', 'nonuniform_illumination', ([], {'img': 'rgb_img', 'ksize': '(11)'}), '(img=rgb_img, ksize... |
def main():
import numpy as np
import matplotlib.pyplot as plt
import torchvision
from torch.autograd import Variable
import torch.nn as nn
import pickle
from random import randint, randrange
import sys
from tqdm import tqdm
import cv2
print("CUDA available: {}".format(torch... | [
"models.DSCLRCN_OldContext.DSCLRCN",
"cv2.resize",
"tqdm.tqdm.write",
"torch.load",
"util.data_utils.get_SALICON_datasets",
"tqdm.tqdm",
"models.CoSADUV_NoTemporal.CoSADUV_NoTemporal",
"util.data_utils.get_video_datasets",
"torch.from_numpy",
"models.CoSADUV.CoSADUV",
"torch.cuda.is_available",
... | [((12055, 12107), 'torch.multiprocessing.set_start_method', 'torch.multiprocessing.set_start_method', (['"""forkserver"""'], {}), "('forkserver')\n", (12093, 12107), False, 'import torch\n'), ((1562, 1627), 'util.data_utils.get_SALICON_datasets', 'get_SALICON_datasets', (['dataset_root_dir', 'mean_image_name', 'img_siz... |
#!/usr/bin/env python3
from pyglet.gl import *
from pyglet.window import NoSuchConfigException
from rubiks import CubeController, CubeView
if __name__ == '__main__':
controller = CubeController()
platform = pyglet.window.get_platform()
display = platform.get_default_display()
screen = display.get_de... | [
"rubiks.CubeController",
"rubiks.CubeView"
] | [((186, 202), 'rubiks.CubeController', 'CubeController', ([], {}), '()\n', (200, 202), False, 'from rubiks import CubeController, CubeView\n'), ((729, 757), 'rubiks.CubeView', 'CubeView', (['controller', 'window'], {}), '(controller, window)\n', (737, 757), False, 'from rubiks import CubeController, CubeView\n')] |
import sys
import os
from PyQt5.QtWidgets import (QTabWidget, QMessageBox)
from codeeditor import CodeEditor
from widgets import MessageBox
class TabWidget(QTabWidget):
def __init__(self, parent=None):
super().__init__()
self.mainWindow = parent
self.setStyleSheet(
''... | [
"os.path.basename",
"codeeditor.CodeEditor",
"widgets.MessageBox"
] | [((997, 1031), 'codeeditor.CodeEditor', 'CodeEditor', ([], {'parent': 'self.mainWindow'}), '(parent=self.mainWindow)\n', (1007, 1031), False, 'from codeeditor import CodeEditor\n'), ((1811, 1925), 'widgets.MessageBox', 'MessageBox', (['QMessageBox.Warning', '"""Warning"""', '"""File not saved\n\nSave now ?"""', '(QMess... |
#!/usr/bin/python3
#Title: headers.py
#Author: ApexPredator
#License: MIT
#Github: https://github.com/ApexPredator-InfoSec/header_check
#Description: This script take a URL or list or URLs as arguments and tests for the headers: 'Strict-Transport-Security', 'Content-Security-Policy', 'X-Frame-Options', and 'Server'
imp... | [
"socket.gethostbyname",
"requests.session",
"requests.packages.urllib3.disable_warnings",
"argparse.ArgumentParser"
] | [((427, 632), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {'prog': '"""headers.py"""', 'usage': '"""python3 -t <target> -f <file contianing target list> -d\npython3 headers.py -t https://securityheaders.com -d\npython3 headers.py -f urls.txt"""'}), '(prog=\'headers.py\', usage=\n """python3 -t <target... |
# Generated by Django 2.1 on 2018-08-06 02:11
from django.db import migrations, models
import django.db.models.deletion
class Migration(migrations.Migration):
initial = True
dependencies = [
('sites', '0002_alter_domain_unique'),
]
operations = [
migrations.CreateModel(
... | [
"django.db.models.OneToOneField",
"django.db.models.TextField",
"django.db.models.ForeignKey",
"django.db.models.ManyToManyField",
"django.db.models.BooleanField",
"django.db.models.SlugField",
"django.db.models.AutoField",
"django.db.models.DateTimeField",
"django.db.models.DecimalField",
"django... | [((3818, 3935), 'django.db.models.ForeignKey', 'models.ForeignKey', ([], {'on_delete': 'django.db.models.deletion.CASCADE', 'related_name': '"""usage"""', 'to': '"""features.FeatureValue"""'}), "(on_delete=django.db.models.deletion.CASCADE, related_name\n ='usage', to='features.FeatureValue')\n", (3835, 3935), False... |
import argparse
import os
from os.path import join
import sys
import joblib
import pandas as pd
import matplotlib.pyplot as plt
import matplotlib
matplotlib.use('TkAgg')
sys.path.append('.')
from project.models.common import get_errors, get_model_details_for_algorithm, get_color, init_scale_from_train_set
from proj... | [
"project.utils.logger.logger.info",
"project.models.common.init_scale_from_train_set",
"project.models.details.ModelDetails",
"project.models.details.get_model_filepath",
"sys.path.append",
"matplotlib.pyplot.margins",
"project.models.common.get_errors",
"argparse.ArgumentParser",
"project.models.co... | [((148, 171), 'matplotlib.use', 'matplotlib.use', (['"""TkAgg"""'], {}), "('TkAgg')\n", (162, 171), False, 'import matplotlib\n'), ((173, 193), 'sys.path.append', 'sys.path.append', (['"""."""'], {}), "('.')\n", (188, 193), False, 'import sys\n'), ((663, 732), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], ... |
import copy as _copy
import math as _math
import os as _os
import cv2 as _cv2
import numpy as _np
from PIL import Image as _IMG
from easytorch.utils.logger import *
"""
##################################################################################################
Very useful image related utilities
##############... | [
"PIL.Image.open",
"copy.deepcopy",
"numpy.ones",
"numpy.where",
"scipy.ndimage.measurements.label",
"os.path.join",
"math.sqrt",
"numpy.max",
"cv2.createCLAHE",
"numpy.array",
"numpy.zeros",
"numpy.pad",
"numpy.bincount",
"numpy.min",
"copy.copy",
"numpy.zeros_like"
] | [((3470, 3535), 'numpy.zeros', '_np.zeros', (['[arr_2d.shape[0], arr_2d.shape[1], 3]'], {'dtype': '_np.uint8'}), '([arr_2d.shape[0], arr_2d.shape[1], 3], dtype=_np.uint8)\n', (3479, 3535), True, 'import numpy as _np\n'), ((4755, 4767), 'numpy.max', '_np.max', (['arr'], {}), '(arr)\n', (4762, 4767), True, 'import numpy ... |
from django.test import TestCase
from django.test import TransactionTestCase
from main import tasks
from celery.contrib.testing.worker import start_worker
from main.models import User
from src.celery import app
class UserModelTests(TestCase):
@classmethod
def setUpTestData(cls):
cls.user = User.object... | [
"main.models.User.objects.create_user"
] | [((309, 358), 'main.models.User.objects.create_user', 'User.objects.create_user', (['"""<EMAIL>"""', '"""<PASSWORD>"""'], {}), "('<EMAIL>', '<PASSWORD>')\n", (333, 358), False, 'from main.models import User\n')] |
from pexpect import pxssh
s = pxssh.pxssh()
s.force_password=True
if not s.login ('192.168.127.12', 'root', '<PASSWORD>!'):
print ("SSH session failed on login.")
else:
print ("SSH session login successful")
s.sendline ('cp /root/some.txt /root/some2.txt')
s.prompt() # match the prompt
prin... | [
"pexpect.pxssh.pxssh"
] | [((30, 43), 'pexpect.pxssh.pxssh', 'pxssh.pxssh', ([], {}), '()\n', (41, 43), False, 'from pexpect import pxssh\n')] |
from zeus.config import db
from zeus.db.mixins import ApiTokenMixin, RepositoryMixin, StandardAttributes
from zeus.db.utils import model_repr
class RepositoryApiToken(StandardAttributes, RepositoryMixin, ApiTokenMixin, db.Model):
"""
An API token associated to a repository.
"""
__tablename__ = "repos... | [
"zeus.db.utils.model_repr"
] | [((352, 386), 'zeus.db.utils.model_repr', 'model_repr', (['"""repository_id"""', '"""key"""'], {}), "('repository_id', 'key')\n", (362, 386), False, 'from zeus.db.utils import model_repr\n')] |
import numpy as np
from PySide import QtGui, QtCore
import sharppy.sharptab as tab
from sharppy.sharptab.constants import *
## Written by <NAME> - OU School of Meteorology
## and <NAME> - CIMMS
__all__ = ['backgroundWatch', 'plotWatch']
class backgroundWatch(QtGui.QFrame):
'''
Draw the background frame and ... | [
"PySide.QtGui.QFont",
"PySide.QtGui.QFontMetrics",
"PySide.QtGui.QColor",
"PySide.QtGui.QPen",
"PySide.QtGui.QPainter",
"PySide.QtCore.QRect"
] | [((996, 1027), 'PySide.QtGui.QFont', 'QtGui.QFont', (['"""Helvetica"""', 'fsize'], {}), "('Helvetica', fsize)\n", (1007, 1027), False, 'from PySide import QtGui, QtCore\n'), ((1053, 1084), 'PySide.QtGui.QFont', 'QtGui.QFont', (['"""Helvetica"""', 'fsize'], {}), "('Helvetica', fsize)\n", (1064, 1084), False, 'from PySid... |
#!/usr/bin/python
# Copyright (c) 2011 GeometryFactory Sarl (France)
#
# $URL$
# $Id$
# SPDX-License-Identifier: LGPL-3.0-or-later OR LicenseRef-Commercial
#
# Author(s) : <NAME>
import sys
import os
import gdb
sys.path.insert(0, os.getcwd() + '/python')
import CGAL.printers
| [
"os.getcwd"
] | [((239, 250), 'os.getcwd', 'os.getcwd', ([], {}), '()\n', (248, 250), False, 'import os\n')] |
import os
import dill
from copy import deepcopy
from itertools import tee
from abc import ABCMeta, abstractmethod
from malemba import ModelBase
from malemba.ds_tools import ArrayHandler, group_array
class ModelDualStage(ModelBase):
class Concatenator(object, metaclass=ABCMeta):
@abstractmethod
... | [
"os.path.exists",
"os.makedirs",
"os.path.join",
"malemba.ds_tools.ArrayHandler",
"itertools.tee",
"dill.dump",
"dill.load"
] | [((5840, 5846), 'itertools.tee', 'tee', (['X'], {}), '(X)\n', (5843, 5846), False, 'from itertools import tee\n'), ((8255, 8287), 'dill.dump', 'dill.dump', (['self.__dict__', 'meta_f'], {}), '(self.__dict__, meta_f)\n', (8264, 8287), False, 'import dill\n'), ((5542, 5556), 'malemba.ds_tools.ArrayHandler', 'ArrayHandler... |
import xml.etree.ElementTree as ET
from typing import NewType
MethodCall = NewType("MethodCall", ET.Element)
ConfigurableMethodTerminal = NewType("ConfigurableMethodTerminal", ET.Element)
MethodTerminal = NewType("MethodTerminal", ET.Element)
WhileLoop = NewType("WhileLoop", ET.Element)
Terminal = NewType("Terminal", ... | [
"typing.NewType"
] | [((76, 109), 'typing.NewType', 'NewType', (['"""MethodCall"""', 'ET.Element'], {}), "('MethodCall', ET.Element)\n", (83, 109), False, 'from typing import NewType\n'), ((139, 188), 'typing.NewType', 'NewType', (['"""ConfigurableMethodTerminal"""', 'ET.Element'], {}), "('ConfigurableMethodTerminal', ET.Element)\n", (146,... |
#!/usr/bin/env python3
# -*- coding: utf-8 -*-
from pylookyloo import Lookyloo
import json
# lookyloo_url = "https://lookyloo.circl.lu/"
lookyloo_url = "http://0.0.0.0:5100"
lookyloo = Lookyloo(lookyloo_url)
lookyloo.init_apikey(username='admin', password='<PASSWORD>')
event = lookyloo.misp_push('6ae2afdc-4d90-41ce-... | [
"json.dumps",
"pylookyloo.Lookyloo"
] | [((188, 210), 'pylookyloo.Lookyloo', 'Lookyloo', (['lookyloo_url'], {}), '(lookyloo_url)\n', (196, 210), False, 'from pylookyloo import Lookyloo\n'), ((347, 374), 'json.dumps', 'json.dumps', (['event'], {'indent': '(2)'}), '(event, indent=2)\n', (357, 374), False, 'import json\n')] |
# Copyright (C) 2015-2019 <NAME>
# SPDX-License-Identifier: Apache-2.0
# Check for presence of FEniCS etc
from .verify_environment import verify_env
verify_env()
__version__ = '2019.0.2'
# This should potentially be made local to the mesh creation routines
import dolfin
dolfin.parameters['ghost_mode'] = 'shared_... | [
"os.path.abspath",
"subprocess.check_output",
"os.path.join"
] | [((656, 681), 'os.path.abspath', 'os.path.abspath', (['__file__'], {}), '(__file__)\n', (671, 681), False, 'import os\n'), ((714, 742), 'os.path.join', 'os.path.join', (['this_dir', '""".."""'], {}), "(this_dir, '..')\n", (726, 742), False, 'import os\n'), ((765, 795), 'os.path.join', 'os.path.join', (['proj_dir', '"""... |
'''
Created on July 7, 2019
@author: Terry
@email:<EMAIL>
'''
print(__doc__)
import numpy as np
import matplotlib.pyplot as plt
from sklearn.datasets import load_iris
from sklearn.tree import DecisionTreeClassifier, plot_tree
# Parameter
n_classes = 3
plot_colors = "ryb"
plot_step = 0.02
# Load data
iris = load_... | [
"sklearn.datasets.load_iris",
"matplotlib.pyplot.contourf",
"matplotlib.pyplot.ylabel",
"numpy.arange",
"numpy.where",
"matplotlib.pyplot.xlabel",
"sklearn.tree.DecisionTreeClassifier",
"matplotlib.pyplot.figure",
"sklearn.tree.plot_tree",
"matplotlib.pyplot.tight_layout",
"matplotlib.pyplot.sca... | [((315, 326), 'sklearn.datasets.load_iris', 'load_iris', ([], {}), '()\n', (324, 326), False, 'from sklearn.datasets import load_iris\n'), ((1404, 1477), 'matplotlib.pyplot.suptitle', 'plt.suptitle', (['"""Decision surface of a decision tree using paired features"""'], {}), "('Decision surface of a decision tree using ... |
# encoding: utf-8
from application import static_manager
from static_bundle import (JsBundle,
CssBundle)
css1 = CssBundle("css")
css1.add_file("example1.less")
css1.add_file("example2.css")
js1 = JsBundle("js")
js1.add_file("vendors/example1.js")
js1.add_file("vendors/example2.js")
js2 = ... | [
"application.static_manager.create_builder",
"static_bundle.CssBundle",
"static_bundle.JsBundle"
] | [((141, 157), 'static_bundle.CssBundle', 'CssBundle', (['"""css"""'], {}), "('css')\n", (150, 157), False, 'from static_bundle import JsBundle, CssBundle\n'), ((226, 240), 'static_bundle.JsBundle', 'JsBundle', (['"""js"""'], {}), "('js')\n", (234, 240), False, 'from static_bundle import JsBundle, CssBundle\n'), ((320, ... |
""" Intermediate Factors
@author: <NAME>
This module computes the interpolated features between the principal vectors -- the one
linking source to target following the geodesics on the Grassmannian. We use the
equivalent formulation derived in [1] and represent this geodesics for each pair
of principal components.
E... | [
"numpy.block",
"numpy.diag",
"joblib.Parallel",
"numpy.array",
"numpy.linspace",
"numpy.isnan",
"numpy.cos",
"precise.principal_vectors.PVComputation",
"numpy.min",
"numpy.sin",
"joblib.delayed",
"numpy.isinf"
] | [((2327, 2403), 'numpy.min', 'np.min', (['[self.source_components_.shape[0], self.target_components_.shape[0]]'], {}), '([self.source_components_.shape[0], self.target_components_.shape[0]])\n', (2333, 2403), True, 'import numpy as np\n'), ((2602, 2632), 'precise.principal_vectors.PVComputation', 'PVComputation', (['n_... |
import logging
import os
import hydra
import pytorch_lightning as pl
from hydra.utils import instantiate
from omegaconf import DictConfig, OmegaConf
from deep_learning_template import BaseDataModule, BaseTask
from deep_learning_template.core import initialization as init
from deep_learning_template.utils.config impor... | [
"logging.getLogger",
"deep_learning_template.core.initialization.initialize_loggers",
"hydra.main",
"hydra.utils.instantiate",
"omegaconf.OmegaConf.to_yaml",
"deep_learning_template.core.initialization.initialize_task",
"deep_learning_template.core.initialization.validate_cfg",
"deep_learning_template... | [((381, 408), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (398, 408), False, 'import logging\n'), ((412, 464), 'hydra.main', 'hydra.main', ([], {'config_path': '"""conf"""', 'config_name': '"""config"""'}), "(config_path='conf', config_name='config')\n", (422, 464), False, 'import hydr... |
import numpy as np
from sklearn.preprocessing import MinMaxScaler, StandardScaler
from sklearn.cross_validation import train_test_split
import theanets
import climate
climate.enable_default_logging()
X_orig = np.load('/Users/bzamecnik/Documents/music-processing/music-processing-experiments/c-scale-piano_spectrogram_2... | [
"theanets.Experiment",
"climate.enable_default_logging",
"sklearn.preprocessing.StandardScaler",
"matplotlib.animation.ArtistAnimation",
"numpy.zeros",
"sklearn.cross_validation.train_test_split",
"numpy.load",
"sklearn.preprocessing.MinMaxScaler"
] | [((168, 200), 'climate.enable_default_logging', 'climate.enable_default_logging', ([], {}), '()\n', (198, 200), False, 'import climate\n'), ((211, 347), 'numpy.load', 'np.load', (['"""/Users/bzamecnik/Documents/music-processing/music-processing-experiments/c-scale-piano_spectrogram_2048_hamming.npy"""'], {}), "(\n '... |
import os
import glob
import argparse
import numpy as np
def parse_args():
parser = argparse.ArgumentParser(description='Display datas')
parser.add_argument('--data-dir', default='C:/Users/junya/Documents/plant_segmentation_data',
help='dataset directory')
parser.add_argument('--va... | [
"numpy.random.choice",
"numpy.savetxt",
"os.path.join",
"argparse.ArgumentParser"
] | [((90, 142), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {'description': '"""Display datas"""'}), "(description='Display datas')\n", (113, 142), False, 'import argparse\n'), ((548, 582), 'os.path.join', 'os.path.join', (['args.data_dir', 'phase'], {}), '(args.data_dir, phase)\n', (560, 582), False, 'impo... |
''' This file gets the cities from the Deutscher Wetterdienst
website file (a file of the name
'TU_Stundenwerte_Beschreibung_Stationen.txt' must already be in the
current folder) and saves it into the file 'current_cityfile.dump'.'''
import pickle
import re # regular expressions package
import csv
#### F... | [
"re.sub"
] | [((1146, 1177), 're.sub', 're.sub', (['"""[^\\\\w]"""', '""" """', 'textline'], {}), "('[^\\\\w]', ' ', textline)\n", (1152, 1177), False, 'import re\n')] |
# python3
# Copyright 2021 InstaDeep Ltd. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applic... | [
"copy.deepcopy",
"acme.specs.BoundedArray",
"dm_env.TimeStep",
"numpy.asarray",
"mava.utils.wrapper_utils.apply_env_wrapper_preprocessers",
"acme.specs.Array",
"mava.utils.wrapper_utils.convert_np_type",
"acme.wrappers.gym_wrapper._convert_to_spec",
"mava.utils.wrapper_utils.parameterized_restart",
... | [((2552, 2610), 'mava.utils.wrapper_utils.parameterized_restart', 'parameterized_restart', (['reward', 'self._discount', 'observation'], {}), '(reward, self._discount, observation)\n', (2573, 2610), False, 'from mava.utils.wrapper_utils import apply_env_wrapper_preprocessers, convert_np_type, parameterized_restart\n'),... |
import os
import numpy as np
from matplotlib import pyplot as plt
from mpl_toolkits import mplot3d
from mpl_toolkits.mplot3d import Axes3D
from sklearn.decomposition import IncrementalPCA
## set paths
inputpath = './input/'
## Toggle PCA for better visualization of clusters
pca_flag = int(input("\nPerform ... | [
"os.path.exists",
"numpy.shape",
"numpy.unique",
"numpy.where",
"os.path.realpath",
"matplotlib.pyplot.figure",
"matplotlib.pyplot.axes",
"matplotlib.pyplot.title",
"sklearn.decomposition.IncrementalPCA",
"numpy.genfromtxt",
"matplotlib.pyplot.legend",
"matplotlib.pyplot.show"
] | [((935, 971), 'os.path.exists', 'os.path.exists', (["(inputpath + 'kx.txt')"], {}), "(inputpath + 'kx.txt')\n", (949, 971), False, 'import os\n'), ((1027, 1062), 'numpy.genfromtxt', 'np.genfromtxt', (["(inputpath + 'kx.txt')"], {}), "(inputpath + 'kx.txt')\n", (1040, 1062), True, 'import numpy as np\n'), ((1071, 1107),... |
import unittest
from database import Database
from search_database import SearchDatabase
class SearchDatabaseTestCase(unittest.TestCase):
def test_update(self):
database = Database('purplePolitics', 'events')
events = database.get_events(False)
search_database = SearchDatabase()
s... | [
"unittest.main",
"search_database.SearchDatabase",
"database.Database"
] | [((415, 430), 'unittest.main', 'unittest.main', ([], {}), '()\n', (428, 430), False, 'import unittest\n'), ((187, 223), 'database.Database', 'Database', (['"""purplePolitics"""', '"""events"""'], {}), "('purplePolitics', 'events')\n", (195, 223), False, 'from database import Database\n'), ((294, 310), 'search_database.... |
from sys import stdin
def main_s():
print("---------------------------------------------------------------------")
print("Bienvenido al simulador de Pac-Man")
print("OPCIONES:")
print("1. Aleatorio")
print("2. Manual")
val = str(stdin.readline().strip())
print("----------------------------... | [
"sys.stdin.readline",
"Español.No_aleatorio.no_aleatorio",
"Español.aleatorio.aleatorio"
] | [((445, 459), 'Español.No_aleatorio.no_aleatorio', 'no_aleatorio', ([], {}), '()\n', (457, 459), False, 'from Español.No_aleatorio import no_aleatorio\n'), ((537, 548), 'Español.aleatorio.aleatorio', 'aleatorio', ([], {}), '()\n', (546, 548), False, 'from Español.aleatorio import aleatorio\n'), ((255, 271), 'sys.stdin.... |
import discord
from bs4 import BeautifulSoup
import asyncio
import datetime
from io import BytesIO
class Archiver:
def __init__(self, bot):
self.bot = bot
async def on_message(self, message):
await self.archive_message(message)
async def on_message_edit(self, before, aft... | [
"discord.utils.get",
"discord.Embed",
"io.BytesIO",
"discord.File"
] | [((1493, 1568), 'discord.utils.get', 'discord.utils.get', (['self.bot.main_server.channels'], {'name': 'archive_channel_name'}), '(self.bot.main_server.channels, name=archive_channel_name)\n', (1510, 1568), False, 'import discord\n'), ((1608, 1685), 'discord.utils.get', 'discord.utils.get', (['self.bot.backup_server.ch... |
# Copyright (c) 2012-2018 SoftBank Robotics. All rights reserved.
# Use of this source code is governed by a BSD-style license that can be
# found in the COPYING file.
from qisrc.test.conftest import git_server, qisrc_action # pylint: disable=unused-import
# pylint: disable=redefined-outer-name
def test_info(qibui... | [
"qisrc.test.conftest.qisrc_action",
"qisrc.test.conftest.git_server.create_repo"
] | [((379, 412), 'qisrc.test.conftest.git_server.create_repo', 'git_server.create_repo', (['"""foo.git"""'], {}), "('foo.git')\n", (401, 412), False, 'from qisrc.test.conftest import git_server, qisrc_action\n'), ((417, 462), 'qisrc.test.conftest.qisrc_action', 'qisrc_action', (['"""init"""', 'git_server.manifest_url'], {... |
import sqlite3
mem_db = sqlite3.connect(':memory:')
db = sqlite3.connect('db/mydb')
cr = db.cursor()
"""
Syntax or Purpose
------ --------- ---------------------- ------------
SELECT DISTINCT Exclude duplicate records for fields selected.
SELECT DISTINCT CASE WHEN expression Conditional expression
SELECT DIST... | [
"sqlite3.connect"
] | [((25, 52), 'sqlite3.connect', 'sqlite3.connect', (['""":memory:"""'], {}), "(':memory:')\n", (40, 52), False, 'import sqlite3\n'), ((58, 84), 'sqlite3.connect', 'sqlite3.connect', (['"""db/mydb"""'], {}), "('db/mydb')\n", (73, 84), False, 'import sqlite3\n')] |
#!/usr/bin/python
# ref: https://gist.github.com/bruienne/f81ea88253629abaf5f9
import objc
import plistlib
class attrdict(dict):
__getattr__ = dict.__getitem__
__setattr__ = dict.__setitem__
ServerInformation = attrdict()
ServerInformation_bundle = objc.loadBundle('ServerInformation', ServerInformation, \
... | [
"plistlib.readPlist",
"objc.loadBundle"
] | [((260, 397), 'objc.loadBundle', 'objc.loadBundle', (['"""ServerInformation"""', 'ServerInformation'], {'bundle_path': '"""/System/Library/PrivateFrameworks/ServerInformation.framework"""'}), "('ServerInformation', ServerInformation, bundle_path=\n '/System/Library/PrivateFrameworks/ServerInformation.framework')\n",... |
import numpy as np
import matplotlib.pyplot as plt
import pickle
import pandas as pd
import cv2 as cv
import os
def read_pkl(file_path):
obj = pd.read_pickle(file_path)
return obj
def show_results(obj , folder):
for im_num in obj.keys():
data = obj[im_num]
file = data['fileName'][12:]
... | [
"pandas.read_pickle",
"cv2.imwrite",
"cv2.rectangle",
"matplotlib.pyplot.imshow",
"os.path.join",
"numpy.asarray",
"matplotlib.pyplot.pause",
"cv2.imread"
] | [((149, 174), 'pandas.read_pickle', 'pd.read_pickle', (['file_path'], {}), '(file_path)\n', (163, 174), True, 'import pandas as pd\n'), ((1781, 1828), 'os.path.join', 'os.path.join', (['folder', '"""YoloV3_res\\\\res_pkl.pkl"""'], {}), "(folder, 'YoloV3_res\\\\res_pkl.pkl')\n", (1793, 1828), False, 'import os\n'), ((33... |
from os import path
import sqlite3 as sql
ROOT = path.dirname(path.relpath(__file__))
dbname = 'database.sqlite'
def create_post(user, post, parent, location, item, tags, ats, created, status):
con = sql.connect(path.join(ROOT, dbname))
cur = con.cursor()
cur.execute(
"""insert into post
... | [
"os.path.join",
"os.path.relpath"
] | [((63, 85), 'os.path.relpath', 'path.relpath', (['__file__'], {}), '(__file__)\n', (75, 85), False, 'from os import path\n'), ((218, 241), 'os.path.join', 'path.join', (['ROOT', 'dbname'], {}), '(ROOT, dbname)\n', (227, 241), False, 'from os import path\n'), ((605, 628), 'os.path.join', 'path.join', (['ROOT', 'dbname']... |