code stringlengths 22 1.05M | apis listlengths 1 3.31k | extract_api stringlengths 75 3.25M |
|---|---|---|
from backend.serializers.user_model_serializer import UserModelDetailSerializer
def jwt_response_payload_handler(token, user=None, request=None):
return {
'token': token,
'user': UserModelDetailSerializer(user, context={'request':request}).data
} | [
"backend.serializers.user_model_serializer.UserModelDetailSerializer"
] | [((207, 268), 'backend.serializers.user_model_serializer.UserModelDetailSerializer', 'UserModelDetailSerializer', (['user'], {'context': "{'request': request}"}), "(user, context={'request': request})\n", (232, 268), False, 'from backend.serializers.user_model_serializer import UserModelDetailSerializer\n')] |
from lyrebird import application
from .. import checker
class CustomDecoder:
def __call__(self, rules=None, *args, **kw):
def func(origin_func):
func_type = checker.TYPE_DECODER
if not checker.scripts_tmp_storage.get(func_type):
checker.scripts_tmp_storage[func_typ... | [
"lyrebird.application.decoder.append",
"lyrebird.application.decoder.remove"
] | [((624, 661), 'lyrebird.application.decoder.append', 'application.decoder.append', (['func_info'], {}), '(func_info)\n', (650, 661), False, 'from lyrebird import application\n'), ((769, 806), 'lyrebird.application.decoder.remove', 'application.decoder.remove', (['func_info'], {}), '(func_info)\n', (795, 806), False, 'f... |
from selene import have
from tests.src import BasePage
class SelectList(BasePage):
def __init__(self, name=None):
super().__init__()
loc = f'//select[@name="{name}"]'
self._element = self.element(loc)
def open(self):
self._element.click()
return self
def _options... | [
"selene.have.value",
"selene.have.exact_text",
"selene.have.text"
] | [((445, 462), 'selene.have.value', 'have.value', (['value'], {}), '(value)\n', (455, 462), False, 'from selene import have\n'), ((564, 579), 'selene.have.text', 'have.text', (['text'], {}), '(text)\n', (573, 579), False, 'from selene import have\n'), ((687, 708), 'selene.have.exact_text', 'have.exact_text', (['text'], ... |
from mongoengine import Document, IntField, DoesNotExist, MultipleObjectsReturned, StringField
class SwapTrackerObject(Document):
nonce = IntField(required=True)
src = StringField(required=True, unique=True)
@classmethod
def last_processed(cls, src: str):
"""
Returns last processed co... | [
"mongoengine.StringField",
"mongoengine.IntField"
] | [((144, 167), 'mongoengine.IntField', 'IntField', ([], {'required': '(True)'}), '(required=True)\n', (152, 167), False, 'from mongoengine import Document, IntField, DoesNotExist, MultipleObjectsReturned, StringField\n'), ((178, 217), 'mongoengine.StringField', 'StringField', ([], {'required': '(True)', 'unique': '(True... |
"""Account Model."""
from masoniteorm.relationships import has_many
from masoniteorm.models import Model
from masoniteorm.scopes import scope
class Account(Model):
"""Account Model."""
__fillable__ = ["name"]
@has_many('id', 'account_id')
def users(self):
from app.User import User
r... | [
"masoniteorm.relationships.has_many"
] | [((226, 254), 'masoniteorm.relationships.has_many', 'has_many', (['"""id"""', '"""account_id"""'], {}), "('id', 'account_id')\n", (234, 254), False, 'from masoniteorm.relationships import has_many\n'), ((337, 365), 'masoniteorm.relationships.has_many', 'has_many', (['"""id"""', '"""account_id"""'], {}), "('id', 'accoun... |
import flask
import hashlib
import requests
import os
from datetime import datetime
from models.models import db
from models.models import User
from models.models import Posts
from models.models import Connection
from decorator import decorator
from sqlalchemy.sql import select
from sqlalchemy.orm import session
'''
I... | [
"models.models.Posts.user_id.in_",
"sqlalchemy.sql.select",
"datetime.datetime.utcnow",
"flask.jsonify",
"models.models.db.session.commit",
"models.models.db.session.query",
"models.models.db.session.add"
] | [((966, 1037), 'flask.jsonify', 'flask.jsonify', (["{'code': 401, 'message': 'You need to be authenticated'}"], {}), "({'code': 401, 'message': 'You need to be authenticated'})\n", (979, 1037), False, 'import flask\n'), ((3113, 3167), 'flask.jsonify', 'flask.jsonify', (["{'code': 400, 'message': 'Bad request'}"], {}), ... |
"""
Replace markdown tags by html tags.
TO DO:
* replace ### \|+ headers not by <hx> but by <div><hx>title</hx>content</div>?
"""
import re
from bs4 import BeautifulSoup
if __name__ == '__main__':
from os import sys, path
root_folder = path.dirname(path.dirname(path.abspath(__file__)))
root_folder = path.... | [
"re.split",
"os.sys.path.append",
"bs4.BeautifulSoup",
"os.path.dirname",
"doctest.testmod",
"os.path.abspath",
"re.sub",
"re.findall",
"re.search"
] | [((359, 387), 'os.sys.path.append', 'sys.path.append', (['root_folder'], {}), '(root_folder)\n', (374, 387), False, 'from os import sys, path\n'), ((720, 768), 're.sub', 're.sub', (['""".+?#META#HEADER#END"""', 'HTML_HEADER', 'text'], {}), "('.+?#META#HEADER#END', HTML_HEADER, text)\n", (726, 768), False, 'import re\n'... |
#%%
import numpy as np
import torch
import torch.nn.functional as F
import dgl
#%%%
# collect single small graphs as a batch
def collate(graphs):
graph = dgl.batch(graphs)
return graph
def evalEdge(model, nfeat, efeat, subgraph, labels, n_classes):
"This function can be fed with nod... | [
"dgl.batch",
"torch.sum",
"torch.nn.functional.cross_entropy",
"torch.no_grad",
"torch.argmax"
] | [((172, 189), 'dgl.batch', 'dgl.batch', (['graphs'], {}), '(graphs)\n', (181, 189), False, 'import dgl\n'), ((396, 411), 'torch.no_grad', 'torch.no_grad', ([], {}), '()\n', (409, 411), False, 'import torch\n'), ((619, 646), 'torch.argmax', 'torch.argmax', (['labels'], {'dim': '(1)'}), '(labels, dim=1)\n', (631, 646), F... |
"""Converter from TypeDoc output to IR format"""
from codecs import getreader
from errno import ENOENT
from json import load, dump
from os.path import basename, join, normpath, relpath, sep, splitext
import re
import subprocess
from tempfile import NamedTemporaryFile
from typing import List, Optional, Tuple, Union
fr... | [
"codecs.getreader",
"os.path.join",
"re.match",
"os.path.splitext",
"sphinx.errors.SphinxError",
"os.path.basename",
"tempfile.NamedTemporaryFile",
"json.dump",
"os.path.relpath"
] | [((15158, 15188), 'tempfile.NamedTemporaryFile', 'NamedTemporaryFile', ([], {'mode': '"""w+b"""'}), "(mode='w+b')\n", (15176, 15188), False, 'from tempfile import NamedTemporaryFile\n'), ((1361, 1402), 'json.dump', 'dump', (['self._name_to_filename', 'f'], {'indent': '(2)'}), '(self._name_to_filename, f, indent=2)\n', ... |
from collections import defaultdict, namedtuple
from math import ceil, floor
from time import time
import sublime
import sublime_plugin
from ..core import RequestCommandMixin
from ..core.parsers import parse_requests
from ..core.responses import prepare_request
ResponseMetrics = namedtuple("ResponseMetrics", "elapse... | [
"collections.namedtuple",
"sublime.set_timeout_async",
"math.floor",
"collections.defaultdict",
"time.time",
"sublime.error_message"
] | [((283, 354), 'collections.namedtuple', 'namedtuple', (['"""ResponseMetrics"""', '"""elapsed, sent, received, code, success"""'], {}), "('ResponseMetrics', 'elapsed, sent, received, code, success')\n", (293, 354), False, 'from collections import defaultdict, namedtuple\n'), ((374, 517), 'collections.namedtuple', 'named... |
'''OpenGL extension EXT.draw_range_elements
This module customises the behaviour of the
OpenGL.raw.GL.EXT.draw_range_elements to provide a more
Python-friendly API
'''
from OpenGL import platform, constants, constant, arrays
from OpenGL import extensions, wrapper
from OpenGL.GL import glget
import ctypes
from OpenGL... | [
"OpenGL.wrapper.returnPyArgument",
"OpenGL.arrays.AsArrayOfType",
"OpenGL.wrapper.wrapper"
] | [((551, 586), 'OpenGL.wrapper.returnPyArgument', 'wrapper.returnPyArgument', (['"""indices"""'], {}), "('indices')\n", (575, 586), False, 'from OpenGL import extensions, wrapper\n'), ((487, 526), 'OpenGL.arrays.AsArrayOfType', 'arrays.AsArrayOfType', (['"""indices"""', '"""type"""'], {}), "('indices', 'type')\n", (507,... |
# Copyright (c) Facebook, Inc. and its affiliates.
#
# This source code is licensed under the MIT license found in the
# LICENSE file in the root directory of this source tree.
"""Evaluate the logs of a random run."""
import json
from pathlib import Path
import humanize
import numpy as np
from absl import app, flags
... | [
"absl.app.UsageError",
"compiler_gym.util.logs.ProgressLogEntry.from_csv",
"pathlib.Path",
"compiler_gym.util.statistics.geometric_mean",
"absl.app.run",
"numpy.array",
"humanize.naturaldelta",
"json.load",
"humanize.intcomma",
"compiler_gym.util.tabulate.tabulate"
] | [((3583, 3596), 'absl.app.run', 'app.run', (['main'], {}), '(main)\n', (3590, 3596), False, 'from absl import app, flags\n'), ((1303, 1345), 'compiler_gym.util.logs.ProgressLogEntry.from_csv', 'logs.ProgressLogEntry.from_csv', (['final_line'], {}), '(final_line)\n', (1333, 1345), False, 'from compiler_gym.util import l... |
"""
Unit and regression test for the kissim.api.compare module.
"""
from pathlib import Path
import pytest
from kissim.api import encode, compare
from kissim.encoding import FingerprintGenerator
@pytest.mark.parametrize(
"output_path, feature_weights, n_cores",
[
(None, None, 1),
(None, Non... | [
"pytest.mark.parametrize",
"kissim.api.compare",
"pathlib.Path"
] | [((201, 489), 'pytest.mark.parametrize', 'pytest.mark.parametrize', (['"""output_path, feature_weights, n_cores"""', "[(None, None, 1), (None, None, 2), (None, [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, \n 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0], 1), ('.', [1.0, 0.0, 0.0,\n 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0... |
# Altere o Programa 7.2, o jogo da forca
# Utilize um arquivo em que uma palavra seja gravada a cada linha
# Use um editor de textos para gerar o arquivo
# Ao iniciar o programa, utilize esse arquivo para carregar (ler) a lista de palavras
# Experimente também perguntar o nome do jogador e gerar um arquivo com o número... | [
"sys.exit"
] | [((795, 806), 'sys.exit', 'sys.exit', (['(1)'], {}), '(1)\n', (803, 806), False, 'import sys\n')] |
from openslides_backend.action.generics.create import CreateAction
from openslides_backend.action.generics.update import UpdateAction
from openslides_backend.action.relations.single_relation_handler import (
SingleRelationHandler,
)
from openslides_backend.action.relations.typing import RelationFieldUpdates
from op... | [
"openslides_backend.models.fields.IntegerField",
"openslides_backend.shared.patterns.Collection",
"openslides_backend.action.util.register.register_action"
] | [((3835, 3914), 'openslides_backend.action.util.register.register_action', 'register_action', (['"""fake_model_a.create"""'], {'action_type': 'ActionType.BACKEND_INTERNAL'}), "('fake_model_a.create', action_type=ActionType.BACKEND_INTERNAL)\n", (3850, 3914), False, 'from openslides_backend.action.util.register import r... |
#!/usr/bin/python
import sys, traceback
DATA = b'Lorem ipsum dolor sit amet, consectetur adipiscing elit, sed do eiusmod tempor incididunt ut labore et dolore magna aliqua. Ut enim ad minim veniam, quis nostrud exercitation ullamco laboris nisi ut aliquip ex ea commodo consequat. Duis aute irure dolor in reprehenderit... | [
"traceback.print_exc",
"sys.exit"
] | [((968, 979), 'sys.exit', 'sys.exit', (['(0)'], {}), '(0)\n', (976, 979), False, 'import sys, traceback\n'), ((1015, 1036), 'traceback.print_exc', 'traceback.print_exc', ([], {}), '()\n', (1034, 1036), False, 'import sys, traceback\n'), ((1045, 1056), 'sys.exit', 'sys.exit', (['(1)'], {}), '(1)\n', (1053, 1056), False,... |
"""Register koalas accessor for pandera schema metadata."""
from databricks.koalas.extensions import (
register_dataframe_accessor,
register_series_accessor,
)
from pandera.pandas_accessor import (
PanderaDataFrameAccessor,
PanderaSeriesAccessor,
)
register_dataframe_accessor("pandera")(PanderaDataFr... | [
"databricks.koalas.extensions.register_dataframe_accessor",
"databricks.koalas.extensions.register_series_accessor"
] | [((268, 306), 'databricks.koalas.extensions.register_dataframe_accessor', 'register_dataframe_accessor', (['"""pandera"""'], {}), "('pandera')\n", (295, 306), False, 'from databricks.koalas.extensions import register_dataframe_accessor, register_series_accessor\n'), ((333, 368), 'databricks.koalas.extensions.register_s... |
from astroid import MANAGER
from astroid import scoped_nodes
NEED_FIX = ('Headers', 'NonRepeatKeyValueContainer', 'KeyValueContainer',
'FormParameters')
FIX_MEMBERS = ('update', 'items', 'iteritems', 'keys', '__setitem__',
'setdefault', 'get')
def register(linter):
pass
def transform... | [
"astroid.scoped_nodes.Class",
"astroid.MANAGER.register_transform"
] | [((676, 733), 'astroid.MANAGER.register_transform', 'MANAGER.register_transform', (['scoped_nodes.Class', 'transform'], {}), '(scoped_nodes.Class, transform)\n', (702, 733), False, 'from astroid import MANAGER\n'), ((645, 672), 'astroid.scoped_nodes.Class', 'scoped_nodes.Class', (['f', 'None'], {}), '(f, None)\n', (663... |
import glob
import subprocess
nbImages = '500'
with open("trainList.txt", "w") as ftrain:
for image in glob.glob('../pretrained/res/*.t7'):
subprocess.call(['th', 'createListImage.lua', '-data', image, '-nImages', nbImages], stdout=ftrain)
with open("valList.txt", "w") as f:
for image in glob.glob('../... | [
"subprocess.call",
"glob.glob"
] | [((112, 147), 'glob.glob', 'glob.glob', (['"""../pretrained/res/*.t7"""'], {}), "('../pretrained/res/*.t7')\n", (121, 147), False, 'import glob\n'), ((306, 341), 'glob.glob', 'glob.glob', (['"""../pretrained/val/*.t7"""'], {}), "('../pretrained/val/*.t7')\n", (315, 341), False, 'import glob\n'), ((152, 255), 'subproces... |
from django import forms
from selectable.forms import AutoCompleteSelectMultipleField
from website.apps.taxonomy.lookups import TaxonomyLookup
from .models import Package
class PackageCreateEditForm(forms.ModelForm):
# Use a radio select field rather than the default.
pkg_type = forms.ChoiceField(label='Pa... | [
"selectable.forms.AutoCompleteSelectMultipleField",
"django.forms.CharField",
"django.forms.ValidationError"
] | [((471, 569), 'selectable.forms.AutoCompleteSelectMultipleField', 'AutoCompleteSelectMultipleField', ([], {'lookup_class': 'TaxonomyLookup', 'required': '(False)', 'label': '"""Taxonomy"""'}), "(lookup_class=TaxonomyLookup, required=False,\n label='Taxonomy')\n", (502, 569), False, 'from selectable.forms import Auto... |
##############################################################################
#
# Copyright (c) 2002 Zope Foundation and Contributors.
#
# This software is subject to the provisions of the Zope Public License,
# Version 2.1 (ZPL). A copy of the ZPL should accompany this distribution.
# THIS SOFTWARE IS PROVIDED "AS I... | [
"warnings.filterwarnings"
] | [((744, 806), 'warnings.filterwarnings', 'warnings.filterwarnings', (['"""ignore"""'], {'category': 'DeprecationWarning'}), "('ignore', category=DeprecationWarning)\n", (767, 806), False, 'import warnings\n')] |
from django.urls import path
from . import views
app_name = 'tickets'
urlpatterns = [
path('', views.ticket_list, name='list'),
path('detail/<int:id>', views.ticket_detail, name='detail'),
path('create/', views.ticket_create, name='create'),
path('edit/<int:id>', views.ticket_edit, name='edit'),
p... | [
"django.urls.path"
] | [((92, 132), 'django.urls.path', 'path', (['""""""', 'views.ticket_list'], {'name': '"""list"""'}), "('', views.ticket_list, name='list')\n", (96, 132), False, 'from django.urls import path\n'), ((138, 197), 'django.urls.path', 'path', (['"""detail/<int:id>"""', 'views.ticket_detail'], {'name': '"""detail"""'}), "('det... |
# -*- mode:python; coding:utf-8 -*-
# Copyright (c) 2020 IBM Corp. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
#... | [
"trestle.oscal.target.Party",
"datetime.datetime.now",
"trestle.core.models.actions.UpdateAction",
"trestle.core.models.elements.ElementPath",
"trestle.core.models.elements.Element"
] | [((973, 999), 'trestle.core.models.elements.Element', 'Element', (['sample_target_def'], {}), '(sample_target_def)\n', (980, 999), False, 'from trestle.core.models.elements import Element, ElementPath\n'), ((1265, 1306), 'trestle.core.models.elements.ElementPath', 'ElementPath', (['"""target-definition.metadata"""'], {... |
"""
BetterFindAllUnder is an implementation of the same command from
Sublime Selection Tools, https://github.com/simonrad/sublime-selection-tools,
but I didn't use any code from there so licensing issues are void.
This is hereby released completely and irrevocably into the Public Domain.
- <NAME> <<EMAIL>>
"""
impor... | [
"sublime.Region"
] | [((704, 728), 'sublime.Region', 'sublime.Region', (['new', 'new'], {}), '(new, new)\n', (718, 728), False, 'import sublime, sublime_plugin\n'), ((1327, 1351), 'sublime.Region', 'sublime.Region', (['new', 'new'], {}), '(new, new)\n', (1341, 1351), False, 'import sublime, sublime_plugin\n'), ((2150, 2174), 'sublime.Regio... |
#!venv/bin/python
from my_site.app import create_app
from my_site import config
my_app = create_app()
my_app.run(debug=True, host='0.0.0.0', port=config.port)
| [
"my_site.app.create_app"
] | [((91, 103), 'my_site.app.create_app', 'create_app', ([], {}), '()\n', (101, 103), False, 'from my_site.app import create_app\n')] |
# -*- coding: utf-8 -*-
# Generated by Django 1.10.5 on 2017-03-07 18:40
from __future__ import unicode_literals
from django.db import migrations
class Migration(migrations.Migration):
dependencies = [
('catalogue', '0014_auto_20170307_2340'),
]
operations = [
migrations.RemoveField(
... | [
"django.db.migrations.RemoveField"
] | [((294, 355), 'django.db.migrations.RemoveField', 'migrations.RemoveField', ([], {'model_name': '"""post"""', 'name': '"""description"""'}), "(model_name='post', name='description')\n", (316, 355), False, 'from django.db import migrations\n'), ((400, 459), 'django.db.migrations.RemoveField', 'migrations.RemoveField', (... |
import argparse
from src.generator import Generator
def main(args):
mode = args.mode
amount = args.number_amount
bits_number = args.bits
if mode not in range(3):
raise Exception('Mode ' + str(mode) + ' not allowed.')
if amount < 0:
raise Exception('Amount ' + str(amount) + ' not... | [
"src.generator.Generator",
"argparse.ArgumentParser"
] | [((452, 488), 'src.generator.Generator', 'Generator', (['mode', 'amount', 'bits_number'], {}), '(mode, amount, bits_number)\n', (461, 488), False, 'from src.generator import Generator\n'), ((960, 985), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {}), '()\n', (983, 985), False, 'import argparse\n')] |
"""
Symbolic transformation usage example
"""
from utils.robo_math import SymbolicTransformation as st
def main():
T_ik = st("TxTyTzRzRyRx", ['x', 'y', 'z', 'alpha', 'theta', 'gamma'])
T_0 = st("Tz", ['l_0']).inv() * T_ik * st("Tx", ["l_6"]).inv()
T_012 = st("RzTzTxRyTxRyRyiTx",
["q_0", "l... | [
"utils.robo_math.SymbolicTransformation"
] | [((128, 190), 'utils.robo_math.SymbolicTransformation', 'st', (['"""TxTyTzRzRyRx"""', "['x', 'y', 'z', 'alpha', 'theta', 'gamma']"], {}), "('TxTyTzRzRyRx', ['x', 'y', 'z', 'alpha', 'theta', 'gamma'])\n", (130, 190), True, 'from utils.robo_math import SymbolicTransformation as st\n'), ((271, 349), 'utils.robo_math.Symbo... |
import re
import pathlib
import tempfile
from collections import OrderedDict
from enum import Enum
from functools import lru_cache
from typing import List, Union, Dict
from hikari.resources import cif_core_dict
from hikari.utility import make_abspath
class CifBlock(OrderedDict):
"""
CifBlock object handles ... | [
"tempfile.TemporaryDirectory",
"collections.OrderedDict",
"re.compile",
"pathlib.Path",
"hikari.utility.make_abspath",
"functools.lru_cache"
] | [((4993, 5042), 're.compile', 're.compile', (['"""(?<=\\\\s)(#.*)(?=$)|(?<=^)(#.*)(?=$)"""'], {}), "('(?<=\\\\s)(#.*)(?=$)|(?<=^)(#.*)(?=$)')\n", (5003, 5042), False, 'import re\n'), ((5085, 5153), 're.compile', 're.compile', (['"""(?<!\\\\b)(["\'])((?:\\\\\\\\\\\\1|(?!\\\\1\\\\s).)*.)(\\\\1)(?!\\\\b)"""'], {}), '(\'(?... |
"""
Objetivo: Resolver questão 1 do segundo laboratorio.
"""
from math import exp
import matplotlib.pyplot as plt
import numpy as np
def f(n): #Calcula o valor de um I passado
e = exp(1)
I = (1 / e) * (e - 1) #I0
soma = 0
for c in range(0, n + 1): #Calcula cada I ate In de maneira sucessiva
... | [
"matplotlib.pyplot.ylabel",
"matplotlib.pyplot.xlabel",
"matplotlib.pyplot.plot",
"matplotlib.pyplot.style.use",
"matplotlib.pyplot.figure",
"matplotlib.pyplot.tight_layout",
"math.exp",
"numpy.arange",
"matplotlib.pyplot.show"
] | [((485, 505), 'numpy.arange', 'np.arange', (['(0)', '(301)', '(1)'], {}), '(0, 301, 1)\n', (494, 505), True, 'import numpy as np\n'), ((609, 632), 'matplotlib.pyplot.style.use', 'plt.style.use', (['"""ggplot"""'], {}), "('ggplot')\n", (622, 632), True, 'import matplotlib.pyplot as plt\n'), ((633, 659), 'matplotlib.pypl... |
#!/usr/bin/python3
""" Redis Note Connector module
"""
import re
import time
import redis
import uuid
import ashaw_notes.utils.search
import ashaw_notes.utils.configuration
CONFIG_SECTION = 'redis_notes'
logger = ashaw_notes.utils.configuration.get_logger()
def is_enabled():
"""Checks if connector is enabled""... | [
"re.findall",
"time.gmtime",
"uuid.uuid4"
] | [((5161, 5183), 'time.gmtime', 'time.gmtime', (['timestamp'], {}), '(timestamp)\n', (5172, 5183), False, 'import time\n'), ((5525, 5551), 're.findall', 're.findall', (['"""(\\\\w+)"""', 'line'], {}), "('(\\\\w+)', line)\n", (5535, 5551), False, 'import re\n'), ((5717, 5751), 're.findall', 're.findall', (['"""(#[A-z0-9-... |
import numpy as np
import pandas as pd
from tensorflow import keras
from ENUtransform import WGS84toENU, ENUtoWGS84
from PythonCode.Trajectory_Prediction.process import scale_data, reshape_data, get_inverse_transform
tREF = {"lon": 12.114733,
"lat": 54.145409,
"ECEF": np.array([[3660725], [7857... | [
"ENUtransform.WGS84toENU",
"PythonCode.Trajectory_Prediction.process.get_inverse_transform",
"pandas.read_csv",
"numpy.array",
"tensorflow.keras.models.load_model",
"PythonCode.Trajectory_Prediction.process.scale_data",
"numpy.full",
"numpy.transpose"
] | [((600, 723), 'tensorflow.keras.models.load_model', 'keras.models.load_model', (["('/home/sing_sd/Desktop/anomaly_detection/PythonCode/Trajectory_Prediction/' +\n model_name)"], {}), "(\n '/home/sing_sd/Desktop/anomaly_detection/PythonCode/Trajectory_Prediction/'\n + model_name)\n", (623, 723), False, 'from t... |
import unittest
import numpy as np
import matplotlib
matplotlib.use('Agg')
import openmdao.api as om
from openmdao.utils.assert_utils import assert_near_equal
from openmdao.utils.testing_utils import use_tempdirs
import dymos as dm
class _BrachistochroneTestODE(om.ExplicitComponent):
def initialize(self):
... | [
"numpy.reshape",
"numpy.ones",
"matplotlib.use",
"dymos.Phase",
"unittest.main",
"dymos.GaussLobatto",
"openmdao.api.Group",
"openmdao.api.DirectSolver",
"numpy.zeros",
"numpy.cos",
"openmdao.utils.assert_utils.assert_near_equal",
"openmdao.api.ScipyOptimizeDriver",
"numpy.sin",
"dymos.Rad... | [((54, 75), 'matplotlib.use', 'matplotlib.use', (['"""Agg"""'], {}), "('Agg')\n", (68, 75), False, 'import matplotlib\n'), ((9739, 9754), 'unittest.main', 'unittest.main', ([], {}), '()\n', (9752, 9754), False, 'import unittest\n'), ((1538, 1574), 'numpy.arange', 'np.arange', (["self.options['num_nodes']"], {}), "(self... |
import time
from csv import DictReader
from pathlib import Path
from types import FunctionType
from typing import List, Dict
from util.jtl_convertor.validation_exception import ValidationException
from util.jtl_convertor.validation_funcs import is_not_none, is_number, is_not_blank
CONNECT = 'Connect'
HOSTNAME = 'Host... | [
"csv.DictReader",
"util.jtl_convertor.validation_exception.ValidationException",
"time.time"
] | [((2229, 2259), 'util.jtl_convertor.validation_exception.ValidationException', 'ValidationException', (['error_msg'], {}), '(error_msg)\n', (2248, 2259), False, 'from util.jtl_convertor.validation_exception import ValidationException\n'), ((2640, 2651), 'time.time', 'time.time', ([], {}), '()\n', (2649, 2651), False, '... |
from __future__ import absolute_import, division, print_function
import time
from mmtbx.validation import validate_ligands
from mmtbx.validation.validate_ligands import master_params_str
import mmtbx.model
import iotbx.pdb
from libtbx.utils import null_out
import libtbx.load_env
from libtbx.test_utils import approx_equ... | [
"libtbx.test_utils.approx_equal",
"mmtbx.validation.validate_ligands.manager",
"time.time",
"libtbx.utils.null_out"
] | [((6229, 6330), 'mmtbx.validation.validate_ligands.manager', 'validate_ligands.manager', ([], {'model': 'model', 'fmodel': 'None', 'params': 'params.validate_ligands', 'log': 'null_out'}), '(model=model, fmodel=None, params=params.\n validate_ligands, log=null_out)\n', (6253, 6330), False, 'from mmtbx.validation imp... |
# Generated by Django 2.2.13 on 2020-10-22 19:36
import autoslug.fields
from django.db import migrations
class Migration(migrations.Migration):
dependencies = [
('versus', '0007_auto_20200516_1849'),
]
operations = [
migrations.AlterModelOptions(
name='joueur',
o... | [
"django.db.migrations.AlterModelOptions"
] | [((250, 325), 'django.db.migrations.AlterModelOptions', 'migrations.AlterModelOptions', ([], {'name': '"""joueur"""', 'options': "{'ordering': ('nom',)}"}), "(name='joueur', options={'ordering': ('nom',)})\n", (278, 325), False, 'from django.db import migrations\n')] |
import pytest
from exchange_calendars.exchange_calendar_xbse import XBSEExchangeCalendar
from .test_exchange_calendar import ExchangeCalendarTestBaseNew
class TestXBSECalendar(ExchangeCalendarTestBaseNew):
@pytest.fixture(scope="class")
def calendar_cls(self):
yield XBSEExchangeCalendar
@pytest.... | [
"pytest.fixture"
] | [((214, 243), 'pytest.fixture', 'pytest.fixture', ([], {'scope': '"""class"""'}), "(scope='class')\n", (228, 243), False, 'import pytest\n')] |
import os
import sys
import json
import tempfile
import pdb
from nose.tools import assert_equal
from nose.tools import assert_not_equal
from nose.tools import assert_raises
from nose.tools import raises
from nose import with_setup
from collections import namedtuple
from ansible.parsing.dataloader import DataLoader
from... | [
"json.dumps",
"os.path.realpath",
"nose.with_setup",
"os.removedirs",
"linchpin_utils.module_utils.run_module",
"tempfile.mkdtemp",
"os.unlink",
"tempfile.NamedTemporaryFile",
"nose.tools.assert_equal",
"os.remove"
] | [((1279, 1296), 'nose.with_setup', 'with_setup', (['setup'], {}), '(setup)\n', (1289, 1296), False, 'from nose import with_setup\n'), ((1682, 1699), 'nose.with_setup', 'with_setup', (['setup'], {}), '(setup)\n', (1692, 1699), False, 'from nose import with_setup\n'), ((2122, 2139), 'nose.with_setup', 'with_setup', (['se... |
"""Cases for eeadm/file_state.py."""
from unittest.mock import MagicMock, patch
import pytest # noqa
import core
from core.eeadm.recall import EEADM_Recall
@patch("sh.eeadm", create=True)
def test_EEADM_Recall(monkeypatch):
"""Call EEADM_Recall it should return right away and find command in background."""
... | [
"unittest.mock.MagicMock",
"core.eeadm.recall.EEADM_Recall",
"unittest.mock.patch"
] | [((162, 192), 'unittest.mock.patch', 'patch', (['"""sh.eeadm"""'], {'create': '(True)'}), "('sh.eeadm', create=True)\n", (167, 192), False, 'from unittest.mock import MagicMock, patch\n'), ((332, 357), 'unittest.mock.MagicMock', 'MagicMock', ([], {'name': '"""sh-echo"""'}), "(name='sh-echo')\n", (341, 357), False, 'fro... |
from canonical_args.frontend.html import generate
if __name__ == "__main__":
stuff = {
"args": [
{
"name": "arg1",
"type": "one([int, float, str, dict])",
"values": {
"int": ">=0",
"float": "<=0",
"str": ["A", "B", "C"],
"dict": {
"subkey1": {
"type": "one([int, float... | [
"canonical_args.frontend.html.generate.generate_html"
] | [((892, 921), 'canonical_args.frontend.html.generate.generate_html', 'generate.generate_html', (['stuff'], {}), '(stuff)\n', (914, 921), False, 'from canonical_args.frontend.html import generate\n')] |
from django.db import models
# Create your models here.
class Tank(models.Model):
name = models.CharField(verbose_name=u'装备名称',max_length = 100)
country = models.CharField(verbose_name =u'研制国家',max_length = 20)
start = models.IntegerField(verbose_name =u'起始制造年份')
end = models.IntegerField(verbose_name ... | [
"django.db.models.ImageField",
"django.db.models.DecimalField",
"django.db.models.CharField",
"django.db.models.IntegerField"
] | [((94, 148), 'django.db.models.CharField', 'models.CharField', ([], {'verbose_name': 'u"""装备名称"""', 'max_length': '(100)'}), "(verbose_name=u'装备名称', max_length=100)\n", (110, 148), False, 'from django.db import models\n'), ((164, 217), 'django.db.models.CharField', 'models.CharField', ([], {'verbose_name': 'u"""研制国家"""... |
from bge import logic
class HeartContainer:
def __init__(self, startHeart, maxHeart):
self.isLow = False
# if heartContainer not exist init then
if not 'heartContainer' in logic.globalDict['Player']:
logic.globalDict['Player']['heartContainer'] = {'heart' : startHeart, 'maxHeart' : maxHeart}
def calculLow... | [
"bge.logic.playerHUD.updateRupee",
"bge.logic.playerHUD.updateHeart",
"bge.logic.playerHUD.low_healt"
] | [((537, 574), 'bge.logic.playerHUD.low_healt', 'logic.playerHUD.low_healt', (['self.isLow'], {}), '(self.isLow)\n', (562, 574), False, 'from bge import logic\n'), ((1170, 1199), 'bge.logic.playerHUD.updateHeart', 'logic.playerHUD.updateHeart', ([], {}), '()\n', (1197, 1199), False, 'from bge import logic\n'), ((1573, 1... |
import time
import majoranaJJ.lattice.shapes as shps #lattice shapes
import majoranaJJ.junk.lattice.neighbors as nb2
import majoranaJJ.lattice.nbrs as nb #neighbor arrays
import majoranaJJ.modules.plots as plots #plotting functions
print("")
N = 45
#Making square lattice, nothing has changed with this method
coor = ... | [
"majoranaJJ.modules.plots.lattice",
"majoranaJJ.junk.lattice.neighbors.NN_Arr",
"majoranaJJ.lattice.nbrs.NN_Arr",
"majoranaJJ.lattice.shapes.square",
"time.time"
] | [((320, 337), 'majoranaJJ.lattice.shapes.square', 'shps.square', (['N', 'N'], {}), '(N, N)\n', (331, 337), True, 'import majoranaJJ.lattice.shapes as shps\n'), ((494, 505), 'time.time', 'time.time', ([], {}), '()\n', (503, 505), False, 'import time\n'), ((515, 531), 'majoranaJJ.junk.lattice.neighbors.NN_Arr', 'nb2.NN_A... |
"""
# Controller for Line-Following Robot
# This runs on an Adafruit Feather M4, with a MiniTFT board.
# It drives a TB6612 to control 2 DC Motors (in blue servo case)
# and talks over I2C to an ItsyBitsy that interfaces a Pololu
# line following sensor
#
# Author(s): <NAME>
# Module: mode_config.py generates and m... | [
"adafruit_display_text.label.Label",
"time.sleep",
"displayio.Group"
] | [((3512, 3540), 'displayio.Group', 'displayio.Group', ([], {'max_size': '(10)'}), '(max_size=10)\n', (3527, 3540), False, 'import displayio\n'), ((3567, 3656), 'adafruit_display_text.label.Label', 'label.Label', (['terminalio.FONT'], {'text': '""""""', 'max_glyphs': '(36)', 'color': 'mycolors.YELLOW', 'x': '(2)', 'y': ... |
import rclpy
from rclpy.node import Node
from spherov2 import scanner
from spherov2.sphero_edu import SpheroEduAPI
from spherov2.types import Color
from datetime import datetime
from datetime import timedelta
import argparse
from sensor_msgs.msg import Imu
from nav_msgs.msg import Odometry
from geometry_msgs.msg imp... | [
"spherov2.types.Color",
"rclpy.init",
"rclpy.spin",
"argparse.ArgumentParser",
"diagnostic_msgs.msg.DiagnosticStatus",
"sensor_msgs.msg.Imu",
"datetime.datetime.now",
"diagnostic_msgs.msg.DiagnosticArray",
"spherov2.scanner.find_toys",
"spherov2.sphero_edu.SpheroEduAPI",
"datetime.timedelta",
... | [((8720, 8775), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {'description': '"""Pass MAC Address"""'}), "(description='Pass MAC Address')\n", (8743, 8775), False, 'import argparse\n'), ((9034, 9053), 'spherov2.scanner.find_toys', 'scanner.find_toys', ([], {}), '()\n', (9051, 9053), False, 'from spherov2 ... |
#!/usr/bin/env python
# -*- coding: utf-8 -*-
"""Download a single year of 6-Hour Reanalysis V2 data from
https://www.esrl.noaa.gov/psd/thredds/dodsC/Datasets/ncep.reanalysis2
"""
import argparse
import os
import xarray as xr
import numpy as np
parser = argparse.ArgumentParser() # pylint: disable=C0103
parser.add_ar... | [
"argparse.ArgumentParser",
"os.path.isfile",
"os.path.abspath",
"xarray.open_dataset",
"xarray.decode_cf",
"os.remove"
] | [((256, 281), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {}), '()\n', (279, 281), False, 'import argparse\n'), ((1399, 1422), 'os.path.isfile', 'os.path.isfile', (['outfile'], {}), '(outfile)\n', (1413, 1422), False, 'import os\n'), ((1639, 1668), 'os.path.abspath', 'os.path.abspath', (['args.outfile'],... |
#!/usr/bin/env python
from iris_sdk.models.base_resource import BaseData, BaseResourceSimpleList
from iris_sdk.models.maps.import_tn_checker_list import TnListMap
class TelephoneNumbers(TnListMap, BaseData):
@property
def items(self):
return self.telephone_number.items
def __init__(self):
... | [
"iris_sdk.models.base_resource.BaseResourceSimpleList"
] | [((346, 370), 'iris_sdk.models.base_resource.BaseResourceSimpleList', 'BaseResourceSimpleList', ([], {}), '()\n', (368, 370), False, 'from iris_sdk.models.base_resource import BaseData, BaseResourceSimpleList\n')] |
#!/usr/bin/env python
# coding=UTF-8
import os
import sys
import getpass
import traceback
from operators.op import op
from globalConfig import globalConfig
from db.controllers.ServidorController import ServidorController
serverManager = None
class ServerManager():
'''
# Base Server Manager
Description ... | [
"operators.op.op.startLog",
"operators.op.op.printLog",
"operators.op.op.createClass",
"db.controllers.ServidorController.ServidorController",
"sys.exit",
"getpass.getuser",
"traceback.print_exc"
] | [((17693, 17704), 'sys.exit', 'sys.exit', (['(0)'], {}), '(0)\n', (17701, 17704), False, 'import sys\n'), ((701, 721), 'db.controllers.ServidorController.ServidorController', 'ServidorController', ([], {}), '()\n', (719, 721), False, 'from db.controllers.ServidorController import ServidorController\n'), ((4931, 5051), ... |
"""
mcpython - a minecraft clone written in python licenced under the MIT-licence
(https://github.com/mcpython4-coding/core)
Contributors: uuk, xkcdjerry (inactive)
Based on the game of fogleman (https://github.com/fogleman/Minecraft), licenced under the MIT-licence
Original game "minecraft" by Mojang Studios (www.m... | [
"mcpython.shared.world.dimensions.values",
"mcpython.common.network.packages.WorldDataExchangePackage.ChunkDataPackage",
"mcpython.shared.event_handler.call_async",
"mcpython.shared.world.get_active_dimension"
] | [((2798, 2830), 'mcpython.shared.world.dimensions.values', 'shared.world.dimensions.values', ([], {}), '()\n', (2828, 2830), False, 'from mcpython import shared\n'), ((10433, 10498), 'mcpython.shared.event_handler.call_async', 'shared.event_handler.call_async', (['"""worldgen:chunk:finished"""', 'chunk'], {}), "('world... |
from django import template
from django.template.loader import get_template
register = template.Library()
@register.simple_tag(takes_context=True)
def activity_item(context, item):
template_name = f"barriers/activity/partials/{item.model}/{item.field}.html"
try:
item_template = get_template(template_... | [
"django.template.loader.get_template",
"django.template.Library"
] | [((88, 106), 'django.template.Library', 'template.Library', ([], {}), '()\n', (104, 106), False, 'from django import template\n'), ((298, 325), 'django.template.loader.get_template', 'get_template', (['template_name'], {}), '(template_name)\n', (310, 325), False, 'from django.template.loader import get_template\n')] |
# -*- coding: utf-8 -*-
from __future__ import unicode_literals
import os, sys
from datetime import datetime, timezone
sys.path.append(os.path.join(os.path.abspath(os.path.dirname(__file__)), "parsers"))
import parser_interface
from utility import carpe_db
class WindowsInstallation(parser_interface.ParserInterface):... | [
"os.path.dirname",
"utility.carpe_db.Mariadb"
] | [((751, 769), 'utility.carpe_db.Mariadb', 'carpe_db.Mariadb', ([], {}), '()\n', (767, 769), False, 'from utility import carpe_db\n'), ((165, 190), 'os.path.dirname', 'os.path.dirname', (['__file__'], {}), '(__file__)\n', (180, 190), False, 'import os, sys\n')] |
#!/usr/bin/env python
# -*- coding: utf-8 -*-
import requests, json
class AOSSAuth():
"""
This class requests and stores an authentication cookie for the Aruba AOS
Switch Software.
"""
def __init__(self, switchip, username, password, version="v3"):
url_login = "http://" + switchip + "/rest... | [
"json.dumps",
"requests.Session",
"requests.delete"
] | [((421, 439), 'requests.Session', 'requests.Session', ([], {}), '()\n', (437, 439), False, 'import requests, json\n'), ((750, 797), 'requests.delete', 'requests.delete', (['url_login'], {'headers': 'self.cookie'}), '(url_login, headers=self.cookie)\n', (765, 797), False, 'import requests, json\n'), ((482, 507), 'json.d... |
# -*- coding: utf-8 -*-
# vim:set ts=4 sw=4 et:
#
# Copyright 2014 <NAME>.
#
# This file is part of the mavros package and subject to the license terms
# in the top-level LICENSE file of the mavros repository.
# https://github.com/mavlink/mavros/tree/master/LICENSE.md
import csv
import time
import rospy
import mavros
... | [
"mavros_msgs.msg.ParamValue",
"mavros.get_topic",
"csv.writer",
"time.strftime",
"csv.reader"
] | [((1441, 1475), 'csv.reader', 'csv.reader', (['file_', 'self.CSVDialect'], {}), '(file_, self.CSVDialect)\n', (1451, 1475), False, 'import csv\n'), ((1770, 1804), 'csv.writer', 'csv.writer', (['file_', 'self.CSVDialect'], {}), '(file_, self.CSVDialect)\n', (1780, 1804), False, 'import csv\n'), ((2330, 2364), 'csv.reade... |
from django.template.loader import render_to_string
from django.core import mail
from django.conf import settings
from emails.text_alt import render_as_text
from emails.models import MailoutUser, MailoutCategory
import logging
UNSUBSCRIBE_EMAIL = 'unsubscribe@' + settings.EMAIL_DOMAIN
def email(user, subj, template,... | [
"logging.basicConfig",
"django.core.mail.EmailMultiAlternatives",
"emails.text_alt.render_as_text",
"logging.warning",
"emails.models.MailoutUser.objects.filter",
"django.template.loader.render_to_string",
"emails.models.MailoutCategory.objects.get",
"emails.models.Email"
] | [((988, 1028), 'emails.models.Email', 'Email', ([], {'to': 'user', 'subject': 'subject', 'body': '""""""'}), "(to=user, subject=subject, body='')\n", (993, 1028), False, 'from emails.models import Email\n'), ((1358, 1410), 'django.template.loader.render_to_string', 'render_to_string', (['f"""emails/{template}.html"""',... |
from flask import (
render_template,
request,
redirect,
session,
url_for,
Blueprint,
make_response,
)
from models.user import User
from utils import log
main = Blueprint('index', __name__)
def current_user():
uid = session.get('user_id', -1)
u = User.find_by(id=uid)
return u
... | [
"models.user.User.find_by",
"flask.render_template",
"models.user.User.register",
"flask.session.get",
"models.user.User.validate_login",
"flask.url_for",
"flask.session.pop",
"flask.make_response",
"flask.Blueprint"
] | [((189, 217), 'flask.Blueprint', 'Blueprint', (['"""index"""', '__name__'], {}), "('index', __name__)\n", (198, 217), False, 'from flask import render_template, request, redirect, session, url_for, Blueprint, make_response\n'), ((250, 276), 'flask.session.get', 'session.get', (['"""user_id"""', '(-1)'], {}), "('user_id... |
# This file is part of the REMOTE API
#
# Copyright 2006-2017 Coppelia Robotics GmbH. All rights reserved.
# <EMAIL>
# www.coppeliarobotics.com
#
# The REMOTE API is licensed under the terms of GNU GPL:
#
# -------------------------------------------------------------------
# The REMOTE API is free software: you ca... | [
"ctypes.CFUNCTYPE",
"ctypes.byref",
"ctypes.POINTER",
"ctypes.c_ubyte",
"struct.pack",
"os.path.dirname",
"platform.system",
"struct.unpack",
"ctypes.c_int",
"ctypes.CDLL",
"ctypes.c_float"
] | [((44719, 44739), 'ctypes.CDLL', 'ct.CDLL', (['libfullpath'], {}), '(libfullpath)\n', (44726, 44739), True, 'import ctypes as ct\n'), ((45290, 45362), 'ctypes.CFUNCTYPE', 'ct.CFUNCTYPE', (['ct.c_int32', 'ct.c_int32', 'ct.c_int32', 'ct.c_float', 'ct.c_int32'], {}), '(ct.c_int32, ct.c_int32, ct.c_int32, ct.c_float, ct.c_... |
from RockStar import RockStar
arnoldc_code = """IT'S SHOWTIME
TALK TO THE HAND "Hello World"
YOU HAVE BEEN TERMINATED"""
rock_it_bro = RockStar(days=400, file_name='hello.arnoldc', code=arnoldc_code)
rock_it_bro.make_me_a_rockstar()
| [
"RockStar.RockStar"
] | [((137, 201), 'RockStar.RockStar', 'RockStar', ([], {'days': '(400)', 'file_name': '"""hello.arnoldc"""', 'code': 'arnoldc_code'}), "(days=400, file_name='hello.arnoldc', code=arnoldc_code)\n", (145, 201), False, 'from RockStar import RockStar\n')] |
from django.urls import path
from . import views
urlpatterns = [
path('dashboard/servicos/', views.servicos_list, name='servicos_list'),
path('dashboard/servicos/upload/fotolito/', views.fotolito_upload,
name='upload_fotolito'),
]
| [
"django.urls.path"
] | [((71, 141), 'django.urls.path', 'path', (['"""dashboard/servicos/"""', 'views.servicos_list'], {'name': '"""servicos_list"""'}), "('dashboard/servicos/', views.servicos_list, name='servicos_list')\n", (75, 141), False, 'from django.urls import path\n'), ((147, 242), 'django.urls.path', 'path', (['"""dashboard/servicos... |
from logging import info as logging_info
from random import randint
from firebase_admin import firestore
from discord.ext import commands
dab = firestore.client()
class work(commands.Cog):
def __init__(self, bot):
self.bot = bot
@commands.command(invoke_without_command=True, )
@commands.coo... | [
"discord.ext.commands.group",
"firebase_admin.firestore.client",
"discord.ext.commands.cooldown",
"discord.ext.commands.command",
"logging.info",
"random.randint"
] | [((147, 165), 'firebase_admin.firestore.client', 'firestore.client', ([], {}), '()\n', (163, 165), False, 'from firebase_admin import firestore\n'), ((255, 300), 'discord.ext.commands.command', 'commands.command', ([], {'invoke_without_command': '(True)'}), '(invoke_without_command=True)\n', (271, 300), False, 'from di... |
#!/usr/bin/env python3
# Author: <NAME> (<EMAIL>)
# License: BSD-3-Clause
import logging, os, time
import numpy as np
import pandas as pd
import matplotlib.pyplot as plt
from astropy.time import Time
from astropy import constants as const
from astropy import units as u
from astropy.cosmology import FlatLambdaCDM
imp... | [
"logging.getLogger",
"numpy.log10",
"matplotlib.pyplot.savefig",
"matplotlib.rcParams.update",
"pandas.read_csv",
"os.path.join",
"astropy.cosmology.FlatLambdaCDM",
"os.path.dirname",
"matplotlib.pyplot.figure",
"matplotlib.pyplot.tight_layout",
"pandas.DataFrame",
"matplotlib.pyplot.subplot",... | [((443, 481), 'matplotlib.rcParams.update', 'matplotlib.rcParams.update', (['nice_fonts'], {}), '(nice_fonts)\n', (469, 481), False, 'import matplotlib\n'), ((503, 532), 'astropy.cosmology.FlatLambdaCDM', 'FlatLambdaCDM', ([], {'H0': '(70)', 'Om0': '(0.3)'}), '(H0=70, Om0=0.3)\n', (516, 532), False, 'from astropy.cosmo... |
###############################################################################
# Copyright (c) 2012 <NAME>. Licensed under the MIT License. #
# See LICENSE.txt for full details. #
###############################################################################
# I M ... | [
"logging.getLogger",
"re.split",
"logging.StreamHandler",
"os.listdir",
"lib.markdown2.markdown",
"re.compile",
"lib.bottle.static_file",
"os.environ.get",
"lib.bottle.abort",
"ConfigParser.ConfigParser",
"lib.bottle.route",
"lib.bottle.run",
"datetime.date",
"lib.bottle.view",
"lib.bott... | [((728, 751), 'logging.StreamHandler', 'logging.StreamHandler', ([], {}), '()\n', (749, 751), False, 'import logging\n'), ((788, 807), 'logging.getLogger', 'logging.getLogger', ([], {}), '()\n', (805, 807), False, 'import logging\n'), ((880, 917), 're.compile', 're.compile', (['"""(?:.md|.markdown|.txt)$"""'], {}), "('... |
"""Top-level class for all object run objects and collections thereof."""
from abc import ABC
from typing import TypeVar
from citrine.resources.data_objects import DataObject, DataObjectCollection
class ObjectRun(DataObject, ABC):
"""
An abstract object run object.
ObjectRun must be extended along with ... | [
"typing.TypeVar"
] | [((365, 416), 'typing.TypeVar', 'TypeVar', (['"""ObjectRunResourceType"""'], {'bound': '"""ObjectRun"""'}), "('ObjectRunResourceType', bound='ObjectRun')\n", (372, 416), False, 'from typing import TypeVar\n')] |
from django.contrib.auth.models import AbstractUser
from django.db import models
from django.urls import reverse
class User(AbstractUser):
# Fields beginning with `contact_` are only intende for use by
# administrators, not for displaying on the web.
contact_name = models.CharField(max_length=100)
cont... | [
"django.db.models.DateTimeField",
"django.urls.reverse",
"django.db.models.CharField",
"django.db.models.BooleanField"
] | [((279, 311), 'django.db.models.CharField', 'models.CharField', ([], {'max_length': '(100)'}), '(max_length=100)\n', (295, 311), False, 'from django.db import models\n'), ((334, 389), 'django.db.models.CharField', 'models.CharField', ([], {'max_length': '(100)', 'null': '(True)', 'blank': '(True)'}), '(max_length=100, ... |
import visa
import numpy
def IsConnected(func):
def func_wrapper(self, *args, **kwargs):
#print("SELF: {}\nFUNCTION: {}\nARGS: {}\nKWARGS: {}".format(self, func, args, kwargs))
if self._connected:
return func(self, *args, **kwargs)
else:
pass # Not connect... | [
"visa.ResourceManager"
] | [((454, 476), 'visa.ResourceManager', 'visa.ResourceManager', ([], {}), '()\n', (474, 476), False, 'import visa\n')] |
"""
Tests the Critical Line Algorithm (CLA).
"""
import unittest
import os
import numpy as np
import pandas as pd
from mlfinlab.portfolio_optimization.cla import CLA
from mlfinlab.portfolio_optimization.returns_estimators import ReturnsEstimation
class TestCLA(unittest.TestCase):
# pylint: disable=too-many-publi... | [
"pandas.read_csv",
"mlfinlab.portfolio_optimization.cla.CLA",
"os.path.dirname",
"numpy.sum",
"mlfinlab.portfolio_optimization.returns_estimators.ReturnsEstimation"
] | [((513, 538), 'os.path.dirname', 'os.path.dirname', (['__file__'], {}), '(__file__)\n', (528, 538), False, 'import os\n'), ((624, 682), 'pandas.read_csv', 'pd.read_csv', (['data_path'], {'parse_dates': '(True)', 'index_col': '"""Date"""'}), "(data_path, parse_dates=True, index_col='Date')\n", (635, 682), True, 'import ... |
import csv
# Objeto usuarios tanto ADM quanto OPR
class Usuario:
numero_id = 0 # Ao reiniciar o APP o contador volta a posição 1
def __init__(self, nome, senha, cargo):
self.__id = Usuario.numero_id + 1
self.__nome = nome
self.__senha = senha
self.__cargo = cargo
Usua... | [
"csv.DictWriter",
"csv.DictReader"
] | [((898, 942), 'csv.DictWriter', 'csv.DictWriter', (['salvar'], {'fieldnames': 'cabecalho'}), '(salvar, fieldnames=cabecalho)\n', (912, 942), False, 'import csv\n'), ((1167, 1188), 'csv.DictReader', 'csv.DictReader', (['abrir'], {}), '(abrir)\n', (1181, 1188), False, 'import csv\n'), ((1430, 1472), 'csv.DictWriter', 'cs... |
"""
当用户登录的时候,将cookie数据合并到redis中
"""
import base64
import pickle
from django_redis import get_redis_connection
"""
1. 获取cookie数据
cookie:
{
1:{count:100,selected:True},
3:{count:100,selected:False},
}
2.获取redis数据
redis:
hash: {2:50,3:50}
set : [2]
3.初始化工作
初始化redis的hash数据(把hash数据都获取出来)
初始化选中的列表
4.... | [
"django_redis.get_redis_connection",
"base64.b64decode"
] | [((971, 999), 'django_redis.get_redis_connection', 'get_redis_connection', (['"""cart"""'], {}), "('cart')\n", (991, 999), False, 'from django_redis import get_redis_connection\n'), ((897, 925), 'base64.b64decode', 'base64.b64decode', (['cookie_str'], {}), '(cookie_str)\n', (913, 925), False, 'import base64\n')] |
"""MolecularAI Implementation of sample generation, randomizing scaffolds as well as fetching unique sample sequences
The source of this file is
https://raw.githubusercontent.com/MolecularAI/Reinvent/982b26dd6cfeb8aa84b6d7e4a8c2a7edde2bad36/running_modes/lib_invent/rl_actions/sample_model.py
and it was only minimally ... | [
"logging.getLogger",
"logging.NullHandler",
"reinvent_chemistry.library_design.AttachmentPoints",
"numpy.array",
"reinvent_models.lib_invent.models.dataset.Dataset",
"reinvent_chemistry.library_design.BondMaker",
"reinvent_chemistry.utils.get_indices_of_unique_smiles",
"reinvent_chemistry.Conversions"... | [((800, 827), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (817, 827), False, 'import logging\n'), ((846, 867), 'logging.NullHandler', 'logging.NullHandler', ([], {}), '()\n', (865, 867), False, 'import logging\n'), ((1400, 1411), 'reinvent_chemistry.library_design.BondMaker', 'BondMake... |
#!/usr/bin/env python
# -*- coding: utf-8 -*-
"""AWS Lambda to copy an AWS image."""
from __future__ import annotations
import json
import os
from typing import Any, Dict
import boto3
from migrationstate import MigrationState, MigrationStateHandler
print("Loading function copy_image")
# {
# "ami_id" : "ami-1... | [
"json.dumps",
"boto3.client",
"os.environ.get",
"migrationstate.MigrationStateHandler"
] | [((872, 891), 'boto3.client', 'boto3.client', (['"""sts"""'], {}), "('sts')\n", (884, 891), False, 'import boto3\n'), ((1105, 1300), 'boto3.client', 'boto3.client', (['"""ec2"""'], {'region_name': 'region', 'aws_access_key_id': "credentials['AccessKeyId']", 'aws_secret_access_key': "credentials['SecretAccessKey']", 'aw... |
import sqlite3
conn = sqlite3.connect("contacts.sqlite")
name = input("Please enter a name to search for: ")
# The sqlite placeholder is ?, this will possible be different depending on database platform
# Also, the variable has to be passed to the SQL statment as a tuple
for row in conn.execute("SELECT * FROM contac... | [
"sqlite3.connect"
] | [((23, 57), 'sqlite3.connect', 'sqlite3.connect', (['"""contacts.sqlite"""'], {}), "('contacts.sqlite')\n", (38, 57), False, 'import sqlite3\n')] |
import bsdiff4
import yaml
import os
import lzma
import hashlib
import threading
import concurrent.futures
import zipfile
import sys
from typing import Tuple, Optional
import Utils
from Rom import JAP10HASH, read_rom
def get_base_rom_path(file_name: str = "") -> str:
options = Utils.get_options()
if not file... | [
"os.path.exists",
"json.loads",
"hashlib.md5",
"zipfile.ZipFile",
"yaml.dump",
"Utils.get_public_ipv4",
"threading.Lock",
"Utils.local_path",
"os.path.splitext",
"lzma.LZMAFile",
"os.replace",
"os.path.split",
"Utils.get_options",
"lzma.decompress",
"Utils.get_adjuster_settings",
"trac... | [((285, 304), 'Utils.get_options', 'Utils.get_options', ([], {}), '()\n', (302, 304), False, 'import Utils\n'), ((1227, 1321), 'yaml.dump', 'yaml.dump', (["{'meta': metadata, 'patch': patch, 'game': 'alttp', 'base_checksum': JAP10HASH}"], {}), "({'meta': metadata, 'patch': patch, 'game': 'alttp',\n 'base_checksum': ... |
import secrets
from typing import Any, Union
from pydantic import AnyHttpUrl, BaseSettings, PostgresDsn, validator
class Settings(BaseSettings):
class Config:
env_file = ".env"
env_file_encoding = "utf-8"
case_sensitive = True
SECRET_KEY: str = secrets.token_urlsafe(32)
ACCESS_TO... | [
"secrets.token_urlsafe",
"pydantic.validator"
] | [((281, 306), 'secrets.token_urlsafe', 'secrets.token_urlsafe', (['(32)'], {}), '(32)\n', (302, 306), False, 'import secrets\n'), ((458, 501), 'pydantic.validator', 'validator', (['"""BACKEND_CORS_ORIGINS"""'], {'pre': '(True)'}), "('BACKEND_CORS_ORIGINS', pre=True)\n", (467, 501), False, 'from pydantic import AnyHttpU... |
# !/usr/bin/env python3
#######################################################################################
# #
# Program purpose: Get the size of an object in bytes. #
# Program Author : <NAME... | [
"sys.getsizeof"
] | [((772, 791), 'sys.getsizeof', 'sys.getsizeof', (['str1'], {}), '(str1)\n', (785, 791), False, 'import sys\n'), ((842, 861), 'sys.getsizeof', 'sys.getsizeof', (['str2'], {}), '(str2)\n', (855, 861), False, 'import sys\n'), ((912, 931), 'sys.getsizeof', 'sys.getsizeof', (['str3'], {}), '(str3)\n', (925, 931), False, 'im... |
import itertools
import logging
import numpy as np
import pandas as pd
import scipy.stats
def create_regression_dataset(metafeatures, experiments):
X = []
X_indices = []
Y = []
for dataset_name in experiments:
experiment = experiments[dataset_name]
mf = metafeatures.loc[dataset_name]
... | [
"pandas.Series",
"numpy.hstack",
"itertools.combinations_with_replacement",
"numpy.isfinite",
"pandas.DataFrame",
"itertools.permutations",
"logging.info",
"numpy.random.RandomState"
] | [((655, 687), 'pandas.DataFrame', 'pd.DataFrame', (['X'], {'index': 'X_indices'}), '(X, index=X_indices)\n', (667, 687), True, 'import pandas as pd\n'), ((696, 728), 'pandas.DataFrame', 'pd.DataFrame', (['Y'], {'index': 'X_indices'}), '(Y, index=X_indices)\n', (708, 728), True, 'import pandas as pd\n'), ((733, 768), 'l... |
# Copyright 2020 Google LLC
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, ... | [
"unittest.main",
"import_data.StateGDPDataLoader",
"re.match"
] | [((3560, 3575), 'unittest.main', 'unittest.main', ([], {}), '()\n', (3573, 3575), False, 'import unittest\n'), ((1231, 1251), 'import_data.StateGDPDataLoader', 'StateGDPDataLoader', ([], {}), '()\n', (1249, 1251), False, 'from import_data import StateGDPDataLoader\n'), ((2196, 2216), 'import_data.StateGDPDataLoader', '... |
# The MIT License (MIT)
#
# Copyright (c) 2020-2022 <NAME>
#
# Use of this source code is governed by The MIT License (MIT)
# that can be found in the LICENSE.txt file.
"""
APIs for working with Debian packages
"""
import os
import os.path
import re
import magic
from shlibvischeck.analysis.header import *
from shl... | [
"shlibvischeck.common.error.error",
"magic.from_file",
"shlibvischeck.common.process.run",
"os.path.splitext",
"os.path.isfile",
"re.search"
] | [((653, 691), 'shlibvischeck.common.process.run', 'run', (['f"""{cc} -E -v -x {lang} /dev/null"""'], {}), "(f'{cc} -E -v -x {lang} /dev/null')\n", (656, 691), False, 'from shlibvischeck.common.process import run\n'), ((1203, 1220), 'os.path.isfile', 'os.path.isfile', (['f'], {}), '(f)\n', (1217, 1220), False, 'import o... |
import logging
from dbnd._core.utils.timezone import utcnow
from dbnd_airflow_export.dag_processing import get_dags, load_dags_models
from dbnd_airflow_export.datetime_utils import pendulum_max_dt
from dbnd_airflow_export.plugin_old.db_queries import (
get_completed_task_instances_and_dag_runs,
get_dag_runs_wi... | [
"dbnd_airflow_export.plugin_old.db_queries.get_completed_task_instances_and_dag_runs",
"dbnd_airflow_export.plugin_old.model.ExportData",
"dbnd_airflow_export.plugin_old.db_queries.get_dag_runs_within_time_window",
"dbnd_airflow_export.plugin_old.db_queries.get_incomplete_task_instances_from_completed_dag_run... | [((863, 888), 'dbnd_airflow_export.dag_processing.load_dags_models', 'load_dags_models', (['session'], {}), '(session)\n', (879, 888), False, 'from dbnd_airflow_export.dag_processing import get_dags, load_dags_models\n'), ((905, 994), 'dbnd_airflow_export.dag_processing.get_dags', 'get_dags', ([], {'dagbag': 'dagbag', ... |
"""
Copyright (c) 2018, National Institute of Informatics
All rights reserved.
Author: <NAME>
-----------------------------------------------------
Script for evaluating the network on full-size dataset using the LDA classifier
"""
import argparse
import os
import random
import torch
import torch.nn as... | [
"torch.nn.ReLU",
"torchvision.models.vgg19",
"numpy.array",
"torch.cuda.is_available",
"torch.nn.BatchNorm2d",
"argparse.ArgumentParser",
"torch.mean",
"numpy.vstack",
"numpy.concatenate",
"torchvision.transforms.ToTensor",
"torch.autograd.Variable",
"torchvision.transforms.Normalize",
"torc... | [((863, 888), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {}), '()\n', (886, 888), False, 'import argparse\n'), ((1901, 1933), 'os.path.join', 'os.path.join', (['opt.outf', 'opt.name'], {}), '(opt.outf, opt.name)\n', (1913, 1933), False, 'import os\n'), ((3101, 3130), 'torchvision.models.vgg19', 'models.... |
from django import template
from ..models import Portfolio
from ..helpers import grouped
register = template.Library()
@register.simple_tag
def get_projects(count=None):
portfolio = None
if count:
portfolio = Portfolio.objects.all()[:count]
portfolio = Portfolio.objects.all()
return grouped... | [
"django.template.Library"
] | [((103, 121), 'django.template.Library', 'template.Library', ([], {}), '()\n', (119, 121), False, 'from django import template\n')] |
"""
Util files for TorchServe
"""
import inspect
import os
import json
import itertools
import logging
logger = logging.getLogger(__name__)
def list_classes_from_module(module, parent_class=None):
"""
Parse user defined module to get all model service classes in it.
:param module:
:param parent_class... | [
"logging.getLogger",
"json.load",
"inspect.isclass",
"os.path.isfile"
] | [((113, 140), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (130, 140), False, 'import logging\n'), ((961, 994), 'os.path.isfile', 'os.path.isfile', (['mapping_file_path'], {}), '(mapping_file_path)\n', (975, 994), False, 'import os\n'), ((1183, 1195), 'json.load', 'json.load', (['f'], {... |
"""
riperpki.py : Imports RIS WHOIS Dumps for both IPv4 and IPv6 into an SQL Lite database
:author: <EMAIL>
:date: 20160526
:date: 20170328
:date: 20170719
Changelog:
- 20170719 fixed prefixlen in roas not getting updated
"""
#########################################################################################
... | [
"cm2c.commons.gen.utils.get_tmp_fn",
"cm2c.commons.gen.getfile.getfile",
"cm2c.csvimport.sql3load.sql3load",
"cm2c.commons.debug.dprint.dprint",
"lacniclabs.utils.addr.pfxExplode"
] | [((1823, 1879), 'cm2c.csvimport.sql3load.sql3load', 'sql3load', (['csv_template', 'self.db_filename', '""","""', '"""roadata"""'], {}), "(csv_template, self.db_filename, ',', 'roadata')\n", (1831, 1879), False, 'from cm2c.csvimport.sql3load import sql3load\n'), ((2681, 2689), 'cm2c.commons.debug.dprint.dprint', 'dprint... |
"""CSC110 Fall 2020 Prep 3: Programming Exercises
Instructions (READ THIS FIRST!)
===============================
This Python module contains several function headers and descriptions.
Your task is to complete this module by doing the following for EACH function below:
1. Write precondition expressions in each funct... | [
"python_ta.check_all",
"python_ta.contracts.check_all_contracts",
"doctest.testmod"
] | [((4380, 4421), 'python_ta.contracts.check_all_contracts', 'python_ta.contracts.check_all_contracts', ([], {}), '()\n', (4419, 4421), False, 'import python_ta\n'), ((4446, 4475), 'doctest.testmod', 'doctest.testmod', ([], {'verbose': '(True)'}), '(verbose=True)\n', (4461, 4475), False, 'import doctest\n'), ((4598, 4718... |
#!/usr/bin/env python
#
# Copyright 2007 Google Inc.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law o... | [
"os.path.exists",
"google.appengine.tools.devappserver2.vm_runtime_proxy.VMRuntimeProxy",
"os.readlink",
"subprocess.check_call",
"os.path.join",
"shutil.copytree",
"tempfile.mkdtemp",
"shutil.rmtree",
"os.path.abspath",
"os.path.islink",
"logging.info"
] | [((2929, 3194), 'google.appengine.tools.devappserver2.vm_runtime_proxy.VMRuntimeProxy', 'vm_runtime_proxy.VMRuntimeProxy', ([], {'docker_client': 'docker_client', 'runtime_config_getter': 'runtime_config_getter', 'module_configuration': 'module_configuration', 'default_port': 'default_port', 'port_bindings': 'port_bind... |
import sys
import os
sys.path.append(os.getcwd())
from training_structures.architecture_search import train,test
from fusions.common_fusions import Concat
from datasets.mimic.get_data import get_dataloader
from unimodals.common_models import LeNet,MLP,Constant
from torch import nn
import torch
import utils.surrogate as... | [
"datasets.mimic.get_data.get_dataloader",
"torch.load",
"training_structures.architecture_search.test",
"os.getcwd"
] | [((360, 414), 'datasets.mimic.get_data.get_dataloader', 'get_dataloader', (['(1)'], {'imputed_path': '"""datasets/mimic/im.pk"""'}), "(1, imputed_path='datasets/mimic/im.pk')\n", (374, 414), False, 'from datasets.mimic.get_data import get_dataloader\n'), ((510, 562), 'training_structures.architecture_search.test', 'tes... |
#!/usr/bin/python3
# Create a single header file from many sources
# (C) 2015 <NAME> http://www.nedprod.com/
# Created: June 2015
import sys, os, re
if len(sys.argv)<2:
print("Usage: "+sys.argv[0]+" [-Iincludepath...] [-Dmacro...] [-Aalwaysinclude...] [-Ealwaysexclude] header1 [header2...]", file=sys.stderr)
... | [
"os.path.exists",
"os.path.join",
"re.match",
"os.getcwd",
"os.chdir",
"os.path.dirname",
"sys.exit",
"os.path.abspath"
] | [((320, 331), 'sys.exit', 'sys.exit', (['(1)'], {}), '(1)\n', (328, 331), False, 'import sys, os, re\n'), ((782, 804), 'os.path.exists', 'os.path.exists', (['header'], {}), '(header)\n', (796, 804), False, 'import sys, os, re\n'), ((1118, 1129), 'os.getcwd', 'os.getcwd', ([], {}), '()\n', (1127, 1129), False, 'import s... |
from libcloud.dns.types import Provider, RecordType
from libcloud.dns.providers import get_driver
cls = get_driver(Provider.GODADDY)
driver = cls("customer_id", "api_key", "api_secret")
record = driver.get_record("waffle-machines.com", "www:A")
record = driver.update_record(
record=record, name="www", type=Record... | [
"libcloud.dns.providers.get_driver"
] | [((105, 133), 'libcloud.dns.providers.get_driver', 'get_driver', (['Provider.GODADDY'], {}), '(Provider.GODADDY)\n', (115, 133), False, 'from libcloud.dns.providers import get_driver\n')] |
#!/usr/bin/python3
from agent_with_DQN import *
import gym
def main():
env = gym.make("MountainCar-v0")
cartpoleagent = Agent(env = env,max_episodes = 10000,learning_rate = 0.0005,hidden_dim = 20,replay_size = 1000,batch_size = 64,max_step_each_episode = 1000,epsilon_decay = False)
cartpoleagent.learning... | [
"gym.make"
] | [((83, 109), 'gym.make', 'gym.make', (['"""MountainCar-v0"""'], {}), "('MountainCar-v0')\n", (91, 109), False, 'import gym\n')] |
#!/usr/bin/env python3
# encoding: utf-8
# Copyright 2020 Nagoya University (<NAME>)
# Apache 2.0 (http://www.apache.org/licenses/LICENSE-2.0)
# Calculate MCD using converted waveform.
import argparse
import fnmatch
import multiprocessing as mp
import os
import numpy as np
import pysptk
import pyworld as pw
impor... | [
"numpy.log10",
"numpy.sqrt",
"multiprocessing.Process",
"numpy.log",
"numpy.array_split",
"numpy.array",
"os.walk",
"numpy.mean",
"argparse.ArgumentParser",
"numpy.where",
"os.path.split",
"pyworld.d4c",
"scipy.signal.firwin",
"pyworld.harvest",
"scipy.io.wavfile.read",
"multiprocessin... | [((805, 840), 'os.walk', 'os.walk', (['root_dir'], {'followlinks': '(True)'}), '(root_dir, followlinks=True)\n', (812, 840), False, 'import os\n'), ((1464, 1505), 'scipy.signal.firwin', 'firwin', (['(255)', 'norm_cutoff'], {'pass_zero': '(False)'}), '(255, norm_cutoff, pass_zero=False)\n', (1470, 1505), False, 'from sc... |
#!/usr/bin/env python3
# -*- coding: utf-8 -*-
# This file is part of CbM (https://github.com/ec-jrc/cbm).
# Author : <NAME>
# Credits : GTCAP Team
# Copyright : 2021 European Commission, Joint Research Centre
# License : 3-Clause BSD
"""
Project: Copernicus DIAS for CAP 'checks by monitoring'.
Functions for... | [
"rasterio.features.shapes",
"osgeo.osr.SpatialReference",
"rasterio.open",
"rasterio.warp.transform_geom",
"osgeo.ogr.CreateGeometryFromJson",
"osgeo.osr.CoordinateTransformation"
] | [((3861, 3893), 'osgeo.ogr.CreateGeometryFromJson', 'ogr.CreateGeometryFromJson', (['geom'], {}), '(geom)\n', (3887, 3893), False, 'from osgeo import ogr, osr\n'), ((3911, 3933), 'osgeo.osr.SpatialReference', 'osr.SpatialReference', ([], {}), '()\n', (3931, 3933), False, 'from osgeo import ogr, osr\n'), ((4002, 4024), ... |
"""
l_system_plugin
"""
from __future__ import absolute_import, division, print_function
import logging
from mcedit2.editortools.generate import GeneratePlugin
from mcedit2.synth.l_system import renderBlocks, renderSceneNodes, applyReplacementsIterated
from mcedit2.util.showprogress import showProgress
from mcedit... | [
"logging.getLogger",
"mcedit2.widgets.spinslider.SpinSlider",
"mcedit2.synth.l_system.renderSceneNodes",
"mcedit2.synth.l_system.renderBlocks",
"mcedit2.synth.l_system.applyReplacementsIterated"
] | [((366, 393), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (383, 393), False, 'import logging\n'), ((1813, 1825), 'mcedit2.widgets.spinslider.SpinSlider', 'SpinSlider', ([], {}), '()\n', (1823, 1825), False, 'from mcedit2.widgets.spinslider import SpinSlider\n'), ((3515, 3544), 'mcedit2... |
#!/usr/bin/python
# Copyright 2017 Google Inc.
#
# Use of this source code is governed by a BSD-style
# license that can be found in the LICENSE file or at
# https://developers.google.com/open-source/licenses/bsd
# cromwell_driver.py
#
# This script provides a library interface to Cromwell, namely:
# * Start the Cro... | [
"requests.post",
"subprocess.Popen",
"os.path.join",
"time.sleep",
"requests.get",
"sys_util.exit_with_error",
"logging.info"
] | [((880, 996), 'subprocess.Popen', 'subprocess.Popen', (["['java', '-Dconfig.file=' + self.cromwell_conf, '-Xmx4g', '-jar', self.\n cromwell_jar, 'server']"], {}), "(['java', '-Dconfig.file=' + self.cromwell_conf, '-Xmx4g',\n '-jar', self.cromwell_jar, 'server'])\n", (896, 996), False, 'import subprocess\n'), ((10... |
import sys
import click
from globus_cli.parsing.excepthook import custom_except_hook
class GlobusCommand(click.Command):
"""
A custom command class which stores the special attributes
of the form "adoc_*" with defaults of None. This lets us pass additional info to the
adoc generator.
It also au... | [
"sys.exc_info"
] | [((3055, 3069), 'sys.exc_info', 'sys.exc_info', ([], {}), '()\n', (3067, 3069), False, 'import sys\n')] |
#!/usr/bin/python
#
# Copyright 2002-2021 Barcelona Supercomputing Center (www.bsc.es)
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
... | [
"logging.getLogger",
"pycompss.runtime.task.parameter.get_compss_type",
"typing.get_type_hints",
"pycompss.util.context.add_to_register_later",
"pycompss.runtime.commons.get_object_conversion",
"pycompss.runtime.binding.register_ce",
"inspect.getfullargspec",
"pycompss.util.objects.properties.is_dict"... | [((5151, 5178), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (5168, 5178), False, 'import logging\n'), ((9787, 9803), 'threading.Lock', 'threading.Lock', ([], {}), '()\n', (9801, 9803), False, 'import threading\n'), ((84979, 84996), 'pycompss.util.storages.persistent.get_id', 'get_id', ... |
# -*- coding: utf-8 -*-
# Copyright 2020 Green Valley Belgium NV
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by appl... | [
"mcfw.rpc.arguments",
"mcfw.rpc.returns",
"StringIO.StringIO"
] | [((1223, 1235), 'mcfw.rpc.returns', 'returns', (['str'], {}), '(str)\n', (1230, 1235), False, 'from mcfw.rpc import arguments, returns\n'), ((1237, 1255), 'mcfw.rpc.arguments', 'arguments', ([], {'val': 'int'}), '(val=int)\n', (1246, 1255), False, 'from mcfw.rpc import arguments, returns\n'), ((1408, 1420), 'mcfw.rpc.r... |
#!/usr/bin/env python
# coding: utf-8
import logging
import os
import pickle
import shutil
import numpy as np
import pandas as pd
from sklearn.metrics import mean_squared_error
from sklearn.model_selection import train_test_split
from lightautoml.automl.presets.text_presets import TabularNLPAutoML
from lightautoml.ta... | [
"logging.basicConfig",
"lightautoml.automl.presets.text_presets.TabularNLPAutoML",
"lightautoml.tasks.Task",
"pickle.dump",
"logging.debug",
"pandas.read_csv",
"sklearn.model_selection.train_test_split",
"pickle.load",
"sklearn.metrics.mean_squared_error",
"numpy.isnan",
"numpy.random.seed",
"... | [((365, 383), 'numpy.random.seed', 'np.random.seed', (['(42)'], {}), '(42)\n', (379, 383), True, 'import numpy as np\n'), ((388, 485), 'logging.basicConfig', 'logging.basicConfig', ([], {'format': '"""[%(asctime)s] (%(levelname)s): %(message)s"""', 'level': 'logging.DEBUG'}), "(format='[%(asctime)s] (%(levelname)s): %(... |
"""
Script to keep the ES eventlog index upto date with the aggregated objects like -
DataONE portal objects, DataONE User profiles, etc...
Note: This script needs Python 3.6 or higher.
"""
import os
import re
import sys
import json
import time
import asyncio
import hashlib
import logging
import argparse
import ... | [
"logging.getLogger",
"d1_metrics.metricselasticsearch.MetricsElasticSearch",
"d1_metrics.metricsdatabase.MetricsDatabase",
"asyncio.Semaphore",
"datetime.timedelta",
"argparse.ArgumentParser",
"asyncio.new_event_loop",
"json.dumps",
"logging.FileHandler",
"asyncio.gather",
"asyncio.get_event_loo... | [((1017, 1038), 'datetime.timedelta', 'datetime.timedelta', (['(0)'], {}), '(0)\n', (1035, 1038), False, 'import datetime\n'), ((1062, 1092), 'datetime.timedelta', 'datetime.timedelta', ([], {'minutes': '(10)'}), '(minutes=10)\n', (1080, 1092), False, 'import datetime\n'), ((4443, 4466), 'logging.getLogger', 'logging.g... |
import boto3
import json
import time
from crhelper import CfnResource
import logging
logger = logging.getLogger(__name__)
helper = CfnResource(json_logging=True, log_level='DEBUG', boto_level='CRITICAL')
def get_dns_configuration(domain_name):
acm = boto3.client('acm')
while True:
response = acm.list... | [
"logging.getLogger",
"boto3.client",
"json.dumps",
"time.sleep",
"crhelper.CfnResource"
] | [((95, 122), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (112, 122), False, 'import logging\n'), ((132, 204), 'crhelper.CfnResource', 'CfnResource', ([], {'json_logging': '(True)', 'log_level': '"""DEBUG"""', 'boto_level': '"""CRITICAL"""'}), "(json_logging=True, log_level='DEBUG', bot... |
# -----------------------------------------------------------
# MIT Licence (Expat License Wording)
# -----------------------------------------------------------
# Copyright © 2020, <NAME>
#
# Permission is hereby granted, free of charge, to any person obtaining a copy
# of this software and associated documentation f... | [
"weyland.regex.Regex",
"html.escape"
] | [((1932, 1959), 'weyland.regex.Regex', 'Regex', (['pattern'], {'debug': '(False)'}), '(pattern, debug=False)\n', (1937, 1959), False, 'from weyland.regex import Regex\n'), ((5693, 5713), 'html.escape', 'html.escape', (['tok.val'], {}), '(tok.val)\n', (5704, 5713), False, 'import html\n')] |
import tensorflow as tf
import cv2
import time
import argparse
import torch
from omegaconf import OmegaConf
from models.networks.LSTM import LSTM
import numpy as np
import posenet
#csvへの書き込み
import csv
import pprint
parser = argparse.ArgumentParser()
parser.add_argument('--model', type=int, default=101)
parser.add_a... | [
"argparse.ArgumentParser",
"posenet.draw_skel_and_kp",
"tensorflow.Session",
"csv.writer",
"posenet.read_cap",
"omegaconf.OmegaConf.load",
"cv2.imshow",
"posenet.load_model",
"cv2.waitKey",
"numpy.array",
"cv2.VideoCapture",
"models.networks.LSTM.LSTM",
"time.time"
] | [((228, 253), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {}), '()\n', (251, 253), False, 'import argparse\n'), ((695, 743), 'omegaconf.OmegaConf.load', 'OmegaConf.load', (['"""./configs/project/default.yaml"""'], {}), "('./configs/project/default.yaml')\n", (709, 743), False, 'from omegaconf import Omeg... |
import requests
import pandas as pd
from bs4 import BeautifulSoup
from IPython.display import display
urls = ["http://su.edu.ph/",
"https://uz.edu.ph/",
"https://www.vsu.edu.ph/",
"https://www.msu.edu.ph/"
]
sm_sites = ['facebook.com','twitter.com']
sm_sites_present = []
columns = ['url'] + sm_sites
df ... | [
"pandas.DataFrame",
"pandas.StringDtype",
"bs4.BeautifulSoup",
"requests.get"
] | [((322, 371), 'pandas.DataFrame', 'pd.DataFrame', ([], {'data': "{'url': urls}", 'columns': 'columns'}), "(data={'url': urls}, columns=columns)\n", (334, 371), True, 'import pandas as pd\n'), ((402, 426), 'requests.get', 'requests.get', (["row['url']"], {}), "(row['url'])\n", (414, 426), False, 'import requests\n'), ((... |