code stringlengths 22 1.05M | apis listlengths 1 3.31k | extract_api stringlengths 75 3.25M |
|---|---|---|
import logging
from cliff.lister import Lister
from cliff.show import ShowOne
from atmosphere.api import AtmosphereAPI
class GroupList(Lister):
"""
List groups for a user.
"""
log = logging.getLogger(__name__)
def take_action(self, parsed_args):
column_headers = ('uuid', 'name')
... | [
"logging.getLogger",
"atmosphere.api.AtmosphereAPI"
] | [((202, 229), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (219, 229), False, 'import logging\n'), ((851, 878), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (868, 878), False, 'import logging\n'), ((327, 479), 'atmosphere.api.AtmosphereAPI', 'AtmosphereA... |
import numpy as np
from sklearn.cluster import KMeans
from splearn.cluster import SparkKMeans
from splearn.utils.testing import SplearnTestCase, assert_array_almost_equal
class TestKMeans(SplearnTestCase):
def test_same_centroids(self):
X, y, X_rdd = self.make_blobs(centers=4, n_samples=200000)
... | [
"sklearn.cluster.KMeans",
"numpy.sort",
"splearn.cluster.SparkKMeans",
"splearn.utils.testing.assert_array_almost_equal"
] | [((328, 383), 'sklearn.cluster.KMeans', 'KMeans', ([], {'n_clusters': '(4)', 'init': '"""k-means++"""', 'random_state': '(42)'}), "(n_clusters=4, init='k-means++', random_state=42)\n", (334, 383), False, 'from sklearn.cluster import KMeans\n'), ((399, 459), 'splearn.cluster.SparkKMeans', 'SparkKMeans', ([], {'n_cluster... |
""" Tests for worker """
import sys
import _pytest
import act.api
import pytest
from act.api.base import ValidationError
from act.api.helpers import handle_fact
from act.api.libs import cli
from act.types.format import object_format
from act.types.types import object_validates
from act.workers.libs import worker
de... | [
"act.workers.libs.worker.parseargs",
"pytest.raises",
"act.workers.libs.worker.init_act"
] | [((632, 653), 'act.workers.libs.worker.init_act', 'worker.init_act', (['args'], {}), '(args)\n', (647, 653), False, 'from act.workers.libs import worker\n'), ((1190, 1211), 'act.workers.libs.worker.init_act', 'worker.init_act', (['args'], {}), '(args)\n', (1205, 1211), False, 'from act.workers.libs import worker\n'), (... |
from setuptools import setup, find_packages
from codecs import open
from os import path
here = path.abspath(path.dirname(__file__))
# Get the long description from the README file
with open(path.join(here, 'README.md'), encoding='utf-8') as f:
long_description = f.read()
setup(
name='darch',
# Versions ... | [
"os.path.join",
"os.path.dirname",
"setuptools.setup"
] | [((279, 963), 'setuptools.setup', 'setup', ([], {'name': '"""darch"""', 'version': '"""0.1.0"""', 'description': '"""Deep Architect"""', 'long_description': 'long_description', 'url': '"""https://github.com/negrinho/deep_architect"""', 'author': '"""The Python Packaging Authority"""', 'author_email': '"""<EMAIL>"""', '... |
import curses
from .game import Game
def go(stdscr):
Game(stdscr).run() # Start game
def main():
# Curses convinient wrapper
curses.wrapper(go)
if __name__ == '__main__':
main() | [
"curses.wrapper"
] | [((142, 160), 'curses.wrapper', 'curses.wrapper', (['go'], {}), '(go)\n', (156, 160), False, 'import curses\n')] |
#!/usr/bin/env python
# -*- coding: utf-8 -*-
# Copyright(c) 2019 Nippon Telegraph and Telephone Corporation
# Filename: NVRDriverCLIProtocol.py
import traceback
import re
import GlobalModule
from EmCommonLog import decorater_log
from CgwshDriverCLIProtocol import CgwshDriverCLIProtocol
class NVRDriverCL... | [
"GlobalModule.EM_LOGGER.debug",
"traceback.format_exc",
"re.search"
] | [((970, 1047), 'GlobalModule.EM_LOGGER.debug', 'GlobalModule.EM_LOGGER.debug', (['("""start exit command :\n%s""" % (send_message,))'], {}), '("""start exit command :\n%s""" % (send_message,))\n', (998, 1047), False, 'import GlobalModule\n'), ((1885, 1960), 'GlobalModule.EM_LOGGER.debug', 'GlobalModule.EM_LOGGER.debug'... |
#!/usr/bin/env python
# -*- coding: utf-8 -*-
#
# Copyright (C) 2015 <NAME> (davvid -at- gmail.com)
# All rights reserved.
#
# This software is licensed as described in the file COPYING, which
# you should have received as part of this distribution.
import os
try:
import setuptools as setup_mod
except ImportError:... | [
"os.path.dirname",
"os.path.join",
"distutils.core.setup"
] | [((368, 393), 'os.path.dirname', 'os.path.dirname', (['__file__'], {}), '(__file__)\n', (383, 393), False, 'import os\n'), ((404, 456), 'os.path.join', 'os.path.join', (['here', '"""skeletor"""', '"""core"""', '"""version.py"""'], {}), "(here, 'skeletor', 'core', 'version.py')\n", (416, 456), False, 'import os\n'), ((1... |
# coding=utf-8
# Copyright 2020 HuggingFace Datasets Authors.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applica... | [
"datasets.SplitGenerator",
"datasets.features.ClassLabel",
"datasets.Version",
"datasets.logging.get_logger",
"datasets.Value"
] | [((665, 702), 'datasets.logging.get_logger', 'datasets.logging.get_logger', (['__name__'], {}), '(__name__)\n', (692, 702), False, 'import datasets\n'), ((3421, 3527), 'datasets.SplitGenerator', 'datasets.SplitGenerator', ([], {'name': 'datasets.Split.TRAIN', 'gen_kwargs': "{'filepath': downloaded_files['train']}"}), "... |
import click
import vk.config as config
import vk.utils as utils
from vk.commands.query import show_layer_state
def _add_layers(layer_str, current_layers, set_layers_func):
"""Append layers to current_layers and set to device props.
Args:
layer_str: A string in <layer1:layer2:layerN> format.
... | [
"click.echo",
"vk.utils.get_valid_app_name",
"click.BadParameter",
"vk.utils.enable_gpu_debug_layers",
"click.UsageError",
"click.option",
"click.command",
"vk.commands.query.show_layer_state",
"vk.utils.get_gpu_debug_layers",
"click.argument",
"click.confirm",
"click.prompt",
"vk.config.Set... | [((1164, 1179), 'click.command', 'click.command', ([], {}), '()\n', (1177, 1179), False, 'import click\n'), ((1181, 1295), 'click.option', 'click.option', (['"""--app"""', '"""app_name"""'], {'type': 'str', 'metavar': '"""<app_name>"""', 'help': '"""Modify per-app layer configuration."""'}), "('--app', 'app_name', type... |
#!/usr/bin/env python2
# -*- coding: utf-8 -*-
import rospy
import message_filters
from std_msgs.msg import Int32, Float32
from pololu_drv8835_rpi import motors
rospy.init_node('message_sync', anonymous=False)
speed_desired = 0.5 # desired wheel speed in rpm
angle_desired = 0.0 # desired angle - 0
k_p_angle = 4*480... | [
"rospy.init_node",
"rospy.get_rostime",
"rospy.spin",
"message_filters.Subscriber",
"message_filters.ApproximateTimeSynchronizer"
] | [((164, 212), 'rospy.init_node', 'rospy.init_node', (['"""message_sync"""'], {'anonymous': '(False)'}), "('message_sync', anonymous=False)\n", (179, 212), False, 'import rospy\n'), ((1319, 1338), 'rospy.get_rostime', 'rospy.get_rostime', ([], {}), '()\n', (1336, 1338), False, 'import rospy\n'), ((2284, 2303), 'rospy.ge... |
# -*- coding: utf-8 -*-
# Copyright (c) 2016, KOL
# All rights reserved.
# Redistribution and use in source and binary forms, with or without
# modification, are permitted provided that the following conditions are met:
# * Redistributions of source code must retain the above copyright
# notice, this list o... | [
"updater.Updater"
] | [((2445, 2476), 'updater.Updater', 'Updater', (["(PREFIX + '/update')", 'oc'], {}), "(PREFIX + '/update', oc)\n", (2452, 2476), False, 'from updater import Updater\n')] |
# !/usr/bin/env python
# coding:utf-8
# Author:XuPengTao
# Date: 2020/3/14
from ssd.layers import L2Norm
import torch.nn.functional as F
import torch.nn as nn
import torch
from ssd.modeling import registry
import os
import quantization.WqAq.dorefa.models.util_wqaq as dorefa
import quantization.WqAq.IAO.models.util_wqaq... | [
"torch.nn.ReLU",
"ssd.layers.L2Norm",
"torch.nn.Sequential",
"torch.sqrt",
"torch.nn.BatchNorm2d",
"torch.nn.ModuleList",
"torch.nn.init.zeros_",
"torch.nn.functional.relu6",
"torch.nn.AdaptiveAvgPool2d",
"torch.nn.LeakyReLU",
"ssd.modeling.registry.BACKBONES.register",
"torch.nn.init.kaiming_... | [((20852, 20895), 'ssd.modeling.registry.BACKBONES.register', 'registry.BACKBONES.register', (['"""cfg_backbone"""'], {}), "('cfg_backbone')\n", (20879, 20895), False, 'from ssd.modeling import registry\n'), ((481, 496), 'torch.no_grad', 'torch.no_grad', ([], {}), '()\n', (494, 496), False, 'import torch\n'), ((533, 72... |
import os
import shutil
def build_env(config, config_name, path):
t_path = os.path.join(path, config_name)
if config != t_path:
os.makedirs(path, exist_ok=True)
shutil.copyfile(config, os.path.join(path, config_name))
| [
"os.path.join",
"os.makedirs"
] | [((80, 111), 'os.path.join', 'os.path.join', (['path', 'config_name'], {}), '(path, config_name)\n', (92, 111), False, 'import os\n'), ((145, 177), 'os.makedirs', 'os.makedirs', (['path'], {'exist_ok': '(True)'}), '(path, exist_ok=True)\n', (156, 177), False, 'import os\n'), ((210, 241), 'os.path.join', 'os.path.join',... |
"""
Python 3.9 программа самостоятельной игры агентов текущего и предыдущего покаления
программа на Python по изучению обучения с подкреплением - Reinforcement Learning
Название файла actor.py
Version: 0.1
Author: <NAME>
Date: 2021-12-23
"""
import numpy as np
import parl
import os
from alphazero_agent import create_a... | [
"MCTS.MCTS",
"utils.win_loss_draw",
"parl.remote_class",
"alphazero_agent.create_agent",
"numpy.random.seed"
] | [((406, 435), 'parl.remote_class', 'parl.remote_class', ([], {'wait': '(False)'}), '(wait=False)\n', (423, 435), False, 'import parl\n'), ((531, 551), 'numpy.random.seed', 'np.random.seed', (['seed'], {}), '(seed)\n', (545, 551), True, 'import numpy as np\n'), ((1788, 1823), 'alphazero_agent.create_agent', 'create_agen... |
from ptkcmd import PtkCmd, Completion, complete_files
class MyPtkCmd(PtkCmd):
prompt='MyPtkCmd$ '
def __init__(self,stdin=None,stdout=None,intro=None,interactive=True,do_complete_cmd=True,default_shell=False,**psession_kwargs):
super().__init__(stdin,stdout,intro,interactive,do_complete_cmd,default_shel... | [
"ptkcmd.complete_files"
] | [((759, 783), 'ptkcmd.complete_files', 'complete_files', (['curr_arg'], {}), '(curr_arg)\n', (773, 783), False, 'from ptkcmd import PtkCmd, Completion, complete_files\n')] |
from schema.models import Member
from django.shortcuts import render
def save_userdata(backend, user, response, *args, **kwargs):
if backend.name == 'facebook':
try:
profile = Member.objects.get(user_id=user.id)
except Member.DoesNotExist:
profile = Member(user_id=user.id)
... | [
"django.shortcuts.render",
"schema.models.Member",
"schema.models.Member.objects.get"
] | [((462, 491), 'django.shortcuts.render', 'render', (['request', '"""index.html"""'], {}), "(request, 'index.html')\n", (468, 491), False, 'from django.shortcuts import render\n'), ((525, 554), 'django.shortcuts.render', 'render', (['request', '"""index.html"""'], {}), "(request, 'index.html')\n", (531, 554), False, 'fr... |
from django.http import HttpResponseRedirect, HttpResponse
from django.shortcuts import get_object_or_404, render, redirect
from django.urls import reverse
from django.views import generic
from django.views.generic import View
from .pdf import *
from .models import Campus, Usuario, Inventario, Libros,Issue,Resp,Cd
fro... | [
"django.shortcuts.render",
"django.http.HttpResponse",
"django.shortcuts.redirect"
] | [((479, 549), 'django.shortcuts.render', 'render', (['request', '"""inventario/index.html"""', "{'inventarios': inventarios}"], {}), "(request, 'inventario/index.html', {'inventarios': inventarios})\n", (485, 549), False, 'from django.shortcuts import get_object_or_404, render, redirect\n'), ((823, 919), 'django.shortc... |
import math
a = int(input('digite um numero:'))
print('O dobro do valor digitado é: {}\nO triplo é: {}\nA Raiz quadrada é: {}'.format((a*2),(a*3),(math.sqrt(a))))
| [
"math.sqrt"
] | [((150, 162), 'math.sqrt', 'math.sqrt', (['a'], {}), '(a)\n', (159, 162), False, 'import math\n')] |
#!/usr/bin/python
import os, csv
import tensorflow as tf
import numpy as np
import pandas as pd
import helpers
# fix random seed for reproducibility
np.random.seed(7)
#-------------------------- Constants --------------------------#
FLAGS = tf.flags.FLAGS
tf.flags.DEFINE_string(
"input_dir", os.path.abspath("../d... | [
"helpers.extract_hour",
"pandas.DataFrame",
"os.path.join",
"tensorflow.contrib.learn.preprocessing.VocabularyProcessor",
"helpers.create_iter_generator",
"numpy.random.seed",
"tensorflow.flags.DEFINE_integer",
"os.path.abspath",
"tensorflow.contrib.learn.preprocessing.VocabularyProcessor.restore"
] | [((151, 168), 'numpy.random.seed', 'np.random.seed', (['(7)'], {}), '(7)\n', (165, 168), True, 'import numpy as np\n'), ((549, 619), 'tensorflow.flags.DEFINE_integer', 'tf.flags.DEFINE_integer', (['"""max_vector_len"""', '(16)', '"""Maximum vector length"""'], {}), "('max_vector_len', 16, 'Maximum vector length')\n", (... |
from django.contrib import admin
from django.urls import path
# from django.contrib.auth.decorators import login_required
# from rest_framework.urlpatterns import format_suffix_patterns
from . views import *
urlpatterns = [
# path('admin/', admin.site.urls),
path('', main_view, name='main_view'),
]
| [
"django.urls.path"
] | [((267, 304), 'django.urls.path', 'path', (['""""""', 'main_view'], {'name': '"""main_view"""'}), "('', main_view, name='main_view')\n", (271, 304), False, 'from django.urls import path\n')] |
#!/usr/bin/env python
# Copyright 2013 The Chromium Authors. All rights reserved.
# Use of this source code is governed by a BSD-style license that can be
# found in the LICENSE file.
import unittest
from app_yaml_helper import AppYamlHelper
from extensions_paths import SERVER2
from host_file_system_provider import H... | [
"test_file_system.MoveTo",
"test_file_system.TestFileSystem",
"unittest.main",
"object_store_creator.ObjectStoreCreator.ForTest",
"app_yaml_helper.AppYamlHelper",
"host_file_system_provider.HostFileSystemProvider",
"test_util.DisableLogging"
] | [((2371, 2396), 'test_util.DisableLogging', 'DisableLogging', (['"""warning"""'], {}), "('warning')\n", (2385, 2396), False, 'from test_util import DisableLogging\n'), ((6137, 6152), 'unittest.main', 'unittest.main', ([], {}), '()\n', (6150, 6152), False, 'import unittest\n'), ((3295, 3323), 'object_store_creator.Objec... |
import numpy as np
import tensorflow as tf
from tensorstream.finance.supertrend import Supertrend
from tensorstream.tests import TestCase
class SupertrendSpec(TestCase):
def setUp(self):
self.sheets = self.read_ods(
self.from_test_res('supertrend.ods', __file__))
def test_supertrend(self):
sheet =... | [
"numpy.testing.assert_almost_equal",
"tensorflow.placeholder",
"tensorstream.finance.supertrend.Supertrend",
"tensorflow.Session"
] | [((364, 381), 'tensorstream.finance.supertrend.Supertrend', 'Supertrend', (['(10)', '(3)'], {}), '(10, 3)\n', (374, 381), False, 'from tensorstream.finance.supertrend import Supertrend\n'), ((401, 427), 'tensorflow.placeholder', 'tf.placeholder', (['tf.float32'], {}), '(tf.float32)\n', (415, 427), True, 'import tensorf... |
import sys
import math
sys.setrecursionlimit(999000)
def fac(x):
if x < 2:
return 1
return x * fac(x - 1)
print(math.log10( fac(int(sys.argv[1])) ))
| [
"sys.setrecursionlimit"
] | [((23, 52), 'sys.setrecursionlimit', 'sys.setrecursionlimit', (['(999000)'], {}), '(999000)\n', (44, 52), False, 'import sys\n')] |
#!/usr/bin/python
from Tkinter import *
from notifier import Notifier
from configurationReader import Configuration
class PomodoroTimer(Frame):
def __start(self):
"""
Start to work, this function is executed only the first
time
"""
self.__button.config(text = "Stop",
background = '#e21212',
ac... | [
"notifier.Notifier",
"configurationReader.Configuration"
] | [((3474, 3489), 'notifier.Notifier', 'Notifier', (['alarm'], {}), '(alarm)\n', (3482, 3489), False, 'from notifier import Notifier\n'), ((3024, 3039), 'configurationReader.Configuration', 'Configuration', ([], {}), '()\n', (3037, 3039), False, 'from configurationReader import Configuration\n')] |
"""
"""
import unittest
import arcpy
from numpy import pi
from gsfarc.test import config
class TestDatatypeDouble(unittest.TestCase):
"""Tests the double task datatype"""
@classmethod
def setUpClass(cls):
"""Class setup creates a toolbox file wrapper to GSF."""
config.setup_idl_toolbox('t... | [
"unittest.main",
"gsfarc.test.config.setup_idl_toolbox",
"arcpy.QA_IDLTaskEngine_DataType_Double_GSF"
] | [((1017, 1032), 'unittest.main', 'unittest.main', ([], {}), '()\n', (1030, 1032), False, 'import unittest\n'), ((293, 381), 'gsfarc.test.config.setup_idl_toolbox', 'config.setup_idl_toolbox', (['"""test_datatype_double"""', '"""qa_idltaskengine_datatype_double"""'], {}), "('test_datatype_double',\n 'qa_idltaskengine... |
import psutil
import os
import time
def get_function_memory_usage(function, iterations=100, max_time=None,
max_mem_usage=90, verbose=False, display_period=0.2):
'''
Cette fonction execute <iterations> fois <function> et renvoit la difference
de mémoire virtuelle utilisée pa... | [
"psutil.virtual_memory",
"time.time",
"os.getpid"
] | [((660, 671), 'time.time', 'time.time', ([], {}), '()\n', (669, 671), False, 'import time\n'), ((628, 639), 'os.getpid', 'os.getpid', ([], {}), '()\n', (637, 639), False, 'import os\n'), ((1014, 1037), 'psutil.virtual_memory', 'psutil.virtual_memory', ([], {}), '()\n', (1035, 1037), False, 'import psutil\n'), ((1051, 1... |
"""__main__ file invoked with `python -m tiatoolbox` command"""
from tiatoolbox.cli import main
main()
| [
"tiatoolbox.cli.main"
] | [((98, 104), 'tiatoolbox.cli.main', 'main', ([], {}), '()\n', (102, 104), False, 'from tiatoolbox.cli import main\n')] |
# Copyright 2020, Battelle Energy Alliance, LLC
# ALL RIGHTS RESERVED
import numpy as np
import math
import random
from scipy.integrate import quad
def timeDepLambda(t,a,b):
return a+t*b
def pdfFailure(t,a,b):
first = timeDepLambda(t,a,b)
second = math.exp(-quad(timeDepLambda, 0, t, args=(a,b))[0])
return fi... | [
"scipy.integrate.quad",
"numpy.zeros",
"numpy.ndenumerate"
] | [((452, 480), 'numpy.zeros', 'np.zeros', (["Input['time'].size"], {}), "(Input['time'].size)\n", (460, 480), True, 'import numpy as np\n'), ((503, 532), 'numpy.ndenumerate', 'np.ndenumerate', (["Input['time']"], {}), "(Input['time'])\n", (517, 532), True, 'import numpy as np\n'), ((267, 305), 'scipy.integrate.quad', 'q... |
from collections import defaultdict
from minpair import arpabet
from minpair.corpus import require as corpus_require
from nltk.corpus import brown
from nltk.corpus import cmudict
from nltk.corpus import words
import re
class Generator(object):
"""Class to generate minimal pairs.
"""
def __init__(self, do... | [
"minpair.arpabet.is_vowel",
"minpair.corpus.require",
"minpair.arpabet.has_stress",
"nltk.corpus.brown.tagged_words",
"nltk.corpus.cmudict.entries",
"collections.defaultdict",
"nltk.corpus.words.words"
] | [((2496, 2520), 'collections.defaultdict', 'defaultdict', (['(lambda : {})'], {}), '(lambda : {})\n', (2507, 2520), False, 'from collections import defaultdict\n'), ((1455, 1478), 'minpair.corpus.require', 'corpus_require', (['corpora'], {}), '(corpora)\n', (1469, 1478), True, 'from minpair.corpus import require as cor... |
import asyncio
import glob
import os
import random
import subprocess
import sys
import wave
from collections import defaultdict
from contextlib import suppress
from datetime import datetime
from functools import partial
from itertools import count, islice
from pathlib import Path
from typing import Default... | [
"discord.FFmpegPCMAudio",
"youtube_dl.YoutubeDL",
"discord.ext.commands.group",
"pathvalidate.sanitize_filename",
"discord.ext.commands.cooldown",
"discord.ext.commands.command",
"os.remove",
"pathlib.Path",
"asyncio.Queue",
"subprocess.call",
"discord.opus.load_opus",
"asyncio.get_event_loop"... | [((1927, 1946), 'youtube_dl.YoutubeDL', 'YoutubeDL', (['ytdlopts'], {}), '(ytdlopts)\n', (1936, 1946), False, 'from youtube_dl import YoutubeDL\n'), ((10776, 10808), 'discord.ext.commands.command', 'commands.command', ([], {'name': '"""players"""'}), "(name='players')\n", (10792, 10808), False, 'from discord.ext import... |
# -*- encoding: utf-8 -*-
"""
_core/queries_db/query_stats.py
"""
import re
import random
import pandas as pd
import numpy as np
from pandas.io.json import json_normalize
from log_config import log, pformat
log.debug("... _core.queries_db.query_stats.py ..." )
from bson.objectid import ObjectId
from flask_r... | [
"bson.objectid.ObjectId",
"log_config.pformat",
"log_config.log.debug",
"bson.objectid.ObjectId.is_valid"
] | [((213, 265), 'log_config.log.debug', 'log.debug', (['"""... _core.queries_db.query_stats.py ..."""'], {}), "('... _core.queries_db.query_stats.py ...')\n", (222, 265), False, 'from log_config import log, pformat\n'), ((1034, 1101), 'log_config.log.debug', 'log.debug', (['"""... Query_db_stats / document_type : %s"""',... |
# This is the file that implements a flask server to do inferences.
import os
import json
import flask
import pandas as pd
from io import StringIO, BytesIO
import mindsdb
# Define the path
prefix = '/opt/ml/'
model_path = os.path.join(prefix, 'model')
def parse_data(content_type, data):
'''
Get the request c... | [
"os.path.exists",
"mindsdb.Predictor",
"pandas.read_csv",
"flask.Flask",
"json.dumps",
"os.path.join",
"io.BytesIO",
"flask.Response",
"pandas.read_excel",
"pandas.DataFrame",
"io.StringIO"
] | [((223, 252), 'os.path.join', 'os.path.join', (['prefix', '"""model"""'], {}), "(prefix, 'model')\n", (235, 252), False, 'import os\n'), ((1000, 1021), 'flask.Flask', 'flask.Flask', (['__name__'], {}), '(__name__)\n', (1011, 1021), False, 'import flask\n'), ((1392, 1469), 'flask.Response', 'flask.Response', ([], {'resp... |
from math import floor
n = int(input())
result = 0
game_over = False
path = []
directions = {
"up": (-1, 0),
"down": (1, 0),
"left": (0, -1),
"right": (0, 1)
}
matrix = []
for row in range(n):
matrix.append(input().split())
def movement_is_valid(movement):
if movement == "up" or movement... | [
"math.floor"
] | [((2158, 2171), 'math.floor', 'floor', (['result'], {}), '(result)\n', (2163, 2171), False, 'from math import floor\n')] |
import os
import torch
import torchvision.datasets as datasets
from .imagenet import ImageNetSubsample, ImageNetSubsampleValClasses
import numpy as np
CLASS_SUBLIST = [
1, 2, 4, 6, 8, 9, 11, 13, 22, 23, 26, 29, 31, 39, 47, 63, 71, 76, 79, 84, 90, 94, 96, 97, 99, 100, 105, 107,
113, 122,
125, 130, 132, 14... | [
"os.path.join"
] | [((1596, 1637), 'os.path.join', 'os.path.join', (['self.location', '"""imagenet-r"""'], {}), "(self.location, 'imagenet-r')\n", (1608, 1637), False, 'import os\n')] |
#!/usr/bin/python
# -*- coding: utf-8 -*-
"""
This bot retreives a list of long term IP blocks on the wiki and tries to
identify if any of them are proxies. If so, it extends the block for another
year.
"""
#
# (C) User:<NAME>, 2019
# (C) User:Huji, 2019
#
# Distributed under the terms of the CC-BY-SA license.
#
from _... | [
"json.loads",
"cidr_trie.PatriciaTrie",
"pywikibot.Site",
"pywikibot.User",
"re.match",
"requests.get",
"ipwhois.IPWhois",
"pywikibot.Page",
"MySQLdb.connect",
"pywikibot.output"
] | [((583, 599), 'pywikibot.Site', 'pywikibot.Site', ([], {}), '()\n', (597, 599), False, 'import pywikibot\n'), ((940, 954), 'cidr_trie.PatriciaTrie', 'PatriciaTrie', ([], {}), '()\n', (952, 954), False, 'from cidr_trie import PatriciaTrie\n'), ((980, 994), 'cidr_trie.PatriciaTrie', 'PatriciaTrie', ([], {}), '()\n', (992... |
from backend.create_app import create_app
application = create_app()
| [
"backend.create_app.create_app"
] | [((58, 70), 'backend.create_app.create_app', 'create_app', ([], {}), '()\n', (68, 70), False, 'from backend.create_app import create_app\n')] |
# encoding: utf-8
from unittest.mock import Mock, patch
from src.client.models import utils
from src.exceptions import WalletError
from tests.unit.logging import LoggingMixin
class ClientUtilsTest(LoggingMixin):
def setUp(self):
self.data = {'test_data_key': 'test_data_value'}
def test_get_utcnow_... | [
"src.client.models.utils.serialize",
"unittest.mock.patch",
"src.client.models.utils.get_utcnow_timestamp",
"src.exceptions.WalletError"
] | [((617, 660), 'unittest.mock.patch', 'patch', (['"""src.client.models.utils.json.dumps"""'], {}), "('src.client.models.utils.json.dumps')\n", (622, 660), False, 'from unittest.mock import Mock, patch\n'), ((357, 385), 'src.client.models.utils.get_utcnow_timestamp', 'utils.get_utcnow_timestamp', ([], {}), '()\n', (383, ... |
from flask import request
from assemblyline.common.concurrency import execute_concurrently
from assemblyline.common.importing import module_attribute_by_name
from al_ui.api_base import api_login, make_api_response
from al_ui.apiv3 import core
from al_ui.config import LOGGER, config
from assemblyline.al.datasource.comm... | [
"flask.request.args.get",
"al_ui.api_base.make_api_response",
"assemblyline.common.concurrency.execute_concurrently",
"assemblyline.common.importing.module_attribute_by_name",
"al_ui.config.LOGGER.exception",
"al_ui.config.config.datasources.iteritems",
"al_ui.apiv3.core.make_subapi_blueprint",
"assem... | [((384, 419), 'al_ui.apiv3.core.make_subapi_blueprint', 'core.make_subapi_blueprint', (['SUB_API'], {}), '(SUB_API)\n', (410, 419), False, 'from al_ui.apiv3 import core\n'), ((1881, 1911), 'al_ui.api_base.api_login', 'api_login', ([], {'required_priv': "['R']"}), "(required_priv=['R'])\n", (1890, 1911), False, 'from al... |
#!/usr/bin/env python
r"""
This file contains functions useful for validating variables in robot.
"""
import gen_robot_print as grp
import gen_valid as gv
from robot.libraries.BuiltIn import BuiltIn
from robot.api import logger
def rvalid_value(var_name,
invalid_values=[],
valid_v... | [
"gen_robot_print.sprint_error_report",
"robot.libraries.BuiltIn.BuiltIn",
"gen_valid.svalid_range",
"gen_valid.svalid_integer",
"gen_valid.svalid_value"
] | [((2546, 2612), 'gen_valid.svalid_value', 'gv.svalid_value', (['var_value', 'invalid_values', 'valid_values', 'var_name'], {}), '(var_value, invalid_values, valid_values, var_name)\n', (2561, 2612), True, 'import gen_valid as gv\n'), ((2709, 2747), 'gen_robot_print.sprint_error_report', 'grp.sprint_error_report', (['er... |
import cherrypy
import os
hostIp = "127.0.0.1"
portNumber = 1337
class ImpulseServer(object):
def get_path(self, data_path):
return os.path.dirname(os.path.realpath(__file__)) + os.sep + data_path
@cherrypy.expose
def index(self):
main_path = self.get_path('main.html')
with open(main_path, "r") as f:
ret... | [
"os.path.realpath",
"os.getcwd"
] | [((509, 520), 'os.getcwd', 'os.getcwd', ([], {}), '()\n', (518, 520), False, 'import os\n'), ((153, 179), 'os.path.realpath', 'os.path.realpath', (['__file__'], {}), '(__file__)\n', (169, 179), False, 'import os\n')] |
# -*- coding: utf-8 -*-
import sys
import numpy as np
import time
from utils import print_bracketing, check_dir
import argparse
import torch
import os.path
import re
import matplotlib
import matplotlib.pyplot as plt
from matplotlib.gridspec import GridSpec
class Saver():
"""
Handles the saving of checkpoints a... | [
"time.localtime",
"re.search",
"numpy.convolve",
"numpy.ones",
"argparse.ArgumentParser",
"torch.load",
"numpy.array",
"matplotlib.pyplot.figure",
"matplotlib.gridspec.GridSpec",
"numpy.zeros",
"torch.cuda.is_available",
"time.time",
"numpy.pad",
"utils.check_dir",
"utils.print_bracketin... | [((21619, 21792), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {'description': '"""Continuous control environment for Udacity DeepRL course."""', 'formatter_class': 'argparse.ArgumentDefaultsHelpFormatter'}), "(description=\n 'Continuous control environment for Udacity DeepRL co... |
import json
from unittest.mock import ANY, MagicMock, call, patch
from pinga.events.consumer import Consumer
@patch("pinga.events.consumer.get_db_conn")
@patch("pinga.events.consumer.save_event")
@patch("pinga.events.consumer.get_logger")
@patch(
"pinga.events.consumer.KafkaConsumer",
return_value=[
... | [
"pinga.events.consumer.Consumer",
"unittest.mock.MagicMock",
"unittest.mock.call",
"json.dumps",
"unittest.mock.patch"
] | [((113, 155), 'unittest.mock.patch', 'patch', (['"""pinga.events.consumer.get_db_conn"""'], {}), "('pinga.events.consumer.get_db_conn')\n", (118, 155), False, 'from unittest.mock import ANY, MagicMock, call, patch\n'), ((157, 198), 'unittest.mock.patch', 'patch', (['"""pinga.events.consumer.save_event"""'], {}), "('pin... |
from PySide6.QtCore import Qt, QThread, Signal, QPoint, QRect, QLocale, QTranslator, QCoreApplication, QThreadPool, \
QObject, QRunnable
from PySide6.QtGui import QCursor, QPixmap, QIcon, QFont
from PySide6.QtWidgets import QApplication, QMessageBox
from gui_about import Ui_AboutWindow
from gui_main import Ui_MainW... | [
"PySide6.QtCore.QLocale",
"webbrowser.open_new",
"subprocess.STARTUPINFO",
"PySide6.QtWidgets.QMessageBox",
"os.walk",
"PySide6.QtCore.QPoint",
"PySide6.QtCore.QThreadPool",
"subprocess.Popen",
"PySide6.QtWidgets.QApplication.setHighDpiScaleFactorRoundingPolicy",
"PySide6.QtGui.QCursor",
"packag... | [((930, 947), 'os.walk', 'os.walk', (['dir_path'], {}), '(dir_path)\n', (937, 947), False, 'import os\n'), ((13008, 13019), 'PySide6.QtCore.Signal', 'Signal', (['str'], {}), '(str)\n', (13014, 13019), False, 'from PySide6.QtCore import Qt, QThread, Signal, QPoint, QRect, QLocale, QTranslator, QCoreApplication, QThreadP... |
import json_and_dictionary_constructor
class Channel:
def __init__(self, user, channel_id):
self.channel_id = channel_id
self.user = user
async def send(self, message):
http_template = dictionary_templates.http_load
message_template = dictionary_templates.new_message
... | [
"json_and_dictionary_constructor.create_http_request"
] | [((737, 888), 'json_and_dictionary_constructor.create_http_request', 'json_and_dictionary_constructor.create_http_request', (['"""GET"""', "('/channels/' + self.channel_id)", "{'Authorization': 'Bot ' + self.user.bot_data['token']}"], {}), "('GET', '/channels/' +\n self.channel_id, {'Authorization': 'Bot ' + self.us... |
# This file is just a basic example of the server comunication structure to a client program.
# The server is run on the embedded computer and passes data to a client on your local pc
# over a direct wifi connection. This can be used to display telemetry.
import time
import json
import socket
def current_time(start_... | [
"socket.socket",
"json.dumps",
"socket.bind",
"time.time",
"socket.recvfrom"
] | [((911, 922), 'time.time', 'time.time', ([], {}), '()\n', (920, 922), False, 'import time\n'), ((968, 1016), 'socket.socket', 'socket.socket', (['socket.AF_INET', 'socket.SOCK_DGRAM'], {}), '(socket.AF_INET, socket.SOCK_DGRAM)\n', (981, 1016), False, 'import socket\n'), ((1045, 1068), 'socket.bind', 'socket.bind', (['(... |
from django.db import models
from mayan.apps.testing.tests.base import BaseTestCase
from ..classes import QuerysetParametersSerializer
class QuerysetParametersSerializerTestCase(BaseTestCase):
def setUp(self):
super().setUp()
self.TestModelParent = self._create_test_model(
model_name... | [
"django.db.models.ForeignKey"
] | [((451, 546), 'django.db.models.ForeignKey', 'models.ForeignKey', ([], {'on_delete': 'models.CASCADE', 'related_name': '"""children"""', 'to': '"""TestModelParent"""'}), "(on_delete=models.CASCADE, related_name='children', to=\n 'TestModelParent')\n", (468, 546), False, 'from django.db import models\n')] |
import matplotlib.pyplot as plt
import numpy as np
def autolabel(rects):
for rect in rects:
height = rect.get_height()
plt.text(rect.get_x()+rect.get_width()/2., 1.03*height, "%s" % float(height))
text = ["10.65x", "57.62x", "54.44x"]
def autolabel_user(rects):
for i, rect in enumerate(rects... | [
"matplotlib.pyplot.xticks",
"matplotlib.pyplot.ylabel",
"matplotlib.pyplot.legend",
"matplotlib.pyplot.xlabel",
"matplotlib.pyplot.bar",
"matplotlib.pyplot.yticks",
"numpy.arange",
"matplotlib.pyplot.show"
] | [((478, 493), 'numpy.arange', 'np.arange', (['size'], {}), '(size)\n', (487, 493), True, 'import numpy as np\n'), ((760, 798), 'matplotlib.pyplot.xlabel', 'plt.xlabel', (['"""Operation"""'], {'fontsize': '(18.5)'}), "('Operation', fontsize=18.5)\n", (770, 798), True, 'import matplotlib.pyplot as plt\n'), ((799, 850), '... |
from django_tables2 import TemplateColumn
from service_catalog.models import GlobalHook
from Squest.utils.squest_table import SquestTable
class GlobalHookTable(SquestTable):
state = TemplateColumn(template_name='custom_columns/global_hook_state.html')
actions = TemplateColumn(template_name='custom_columns/gl... | [
"django_tables2.TemplateColumn"
] | [((189, 258), 'django_tables2.TemplateColumn', 'TemplateColumn', ([], {'template_name': '"""custom_columns/global_hook_state.html"""'}), "(template_name='custom_columns/global_hook_state.html')\n", (203, 258), False, 'from django_tables2 import TemplateColumn\n'), ((273, 365), 'django_tables2.TemplateColumn', 'Template... |
# Copyright 2015 The TensorFlow Authors. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applica... | [
"tensorflow.compat.v2.TensorShape",
"keras.engine.input_spec.InputSpec",
"tensorflow.python.util.tf_export.keras_export",
"keras.backend.repeat_elements"
] | [((959, 1000), 'tensorflow.python.util.tf_export.keras_export', 'keras_export', (['"""keras.layers.UpSampling1D"""'], {}), "('keras.layers.UpSampling1D')\n", (971, 1000), False, 'from tensorflow.python.util.tf_export import keras_export\n'), ((1966, 1983), 'keras.engine.input_spec.InputSpec', 'InputSpec', ([], {'ndim':... |
#!/usr/bin/env python3
# -*- coding: utf-8 -*-
"""Main script for Empty My Fridge project. Provides a CLI for the Spoonacular API.
"""
import requests
from rich.console import Console
import config
def get_missing_ingredients(available_ingredients, amount_recipes):
"""
:param available_ingredients: comma ... | [
"rich.console.Console",
"requests.get",
"rich.table.Table"
] | [((1353, 1446), 'requests.get', 'requests.get', (['"""https://api.spoonacular.com/recipes/findByIngredients/"""'], {'params': 'parameters'}), "('https://api.spoonacular.com/recipes/findByIngredients/',\n params=parameters)\n", (1365, 1446), False, 'import requests\n'), ((2262, 2271), 'rich.console.Console', 'Console... |
import json
from flask import Blueprint
from metrics import availability as avail
from metrics import utils as ut
from elasticsearch.exceptions import ConnectionError
avail_page = Blueprint('availability', __name__)
QUERY_CONTENT = '*'
@avail_page.route('/')
def hello():
return ut.json_response_formatter({'msg'... | [
"metrics.utils.es_rest",
"metrics.utils.json_response_formatter",
"metrics.availability.get_availability_per_bp_and_method",
"flask.Blueprint",
"metrics.utils.get_timestamp_timewindow"
] | [((181, 216), 'flask.Blueprint', 'Blueprint', (['"""availability"""', '__name__'], {}), "('availability', __name__)\n", (190, 216), False, 'from flask import Blueprint\n'), ((287, 352), 'metrics.utils.json_response_formatter', 'ut.json_response_formatter', (['{\'msg\': "I\'m the availability file!"}'], {}), '({\'msg\':... |
#!/usr/bin/env python
from nlputil import *
import fileinput
import re
switch_thresh_and_grid = True
errdists = dictdict()
for line in fileinput.input():
line = line.strip()
m = re.match(r'.*thresh: ([0-9.]*), grid: ([0-9.]*),.*true error distance.*\(([0-9.]*) km.*', line)
if not m:
errprint("Can't parse ... | [
"re.match",
"fileinput.input"
] | [((139, 156), 'fileinput.input', 'fileinput.input', ([], {}), '()\n', (154, 156), False, 'import fileinput\n'), ((186, 291), 're.match', 're.match', (['""".*thresh: ([0-9.]*), grid: ([0-9.]*),.*true error distance.*\\\\(([0-9.]*) km.*"""', 'line'], {}), "(\n '.*thresh: ([0-9.]*), grid: ([0-9.]*),.*true error distanc... |
import sys
import os
import re
import json
import base64
REFERENCE_BRANCH_NAME = 'master'
GRADLE_FILE_PATH = './build.gradle'
GET_GRADLE_COMMAND = f'git show {REFERENCE_BRANCH_NAME} HEAD:{GRADLE_FILE_PATH}'
class Version:
def __init__(self, version_string: str):
parts = version_string.split('.')
if le... | [
"re.findall",
"os.popen"
] | [((1443, 1511), 're.findall', 're.findall', (['"""-*\\\\+*version\\\\s*=*\\\\s*\'\\\\S+\'\\\\n"""', 'build_gradle_content'], {}), '("-*\\\\+*version\\\\s*=*\\\\s*\'\\\\S+\'\\\\n", build_gradle_content)\n', (1453, 1511), False, 'import re\n'), ((1327, 1355), 'os.popen', 'os.popen', (['GET_GRADLE_COMMAND'], {}), '(GET_GR... |
from sklearn.metrics.pairwise import cosine_similarity
import pandas as pd
import numpy as np
cs = cosine_similarity
df=pd.read_csv("df_final.csv")
def get_recommender(user_input):
""" This function produces the 5 top recommendations depending on user_input.
The user_input is the age group and skills selecte... | [
"numpy.array",
"pandas.read_csv"
] | [((121, 148), 'pandas.read_csv', 'pd.read_csv', (['"""df_final.csv"""'], {}), "('df_final.csv')\n", (132, 148), True, 'import pandas as pd\n'), ((619, 639), 'numpy.array', 'np.array', (['all_scores'], {}), '(all_scores)\n', (627, 639), True, 'import numpy as np\n')] |
# -*- coding: utf-8 -*-
#config
import configparser
#sys & os
import sys
import os.path
#set logging
import logging
logging.basicConfig(filename=os.path.join(os.path.dirname(os.path.abspath(sys.argv[0])), 'umr.log'),level=logging.DEBUG)
#Set locale
#FIXME : force French, should be in conf ?
import locale
locale.set... | [
"configparser.ConfigParser",
"logging.info",
"locale.setlocale",
"logging.debug"
] | [((310, 351), 'locale.setlocale', 'locale.setlocale', (['locale.LC_TIME', '"""fr_FR"""'], {}), "(locale.LC_TIME, 'fr_FR')\n", (326, 351), False, 'import locale\n'), ((687, 714), 'logging.info', 'logging.info', (['"""Load config"""'], {}), "('Load config')\n", (699, 714), False, 'import logging\n'), ((733, 760), 'config... |
from django.db import models
from django.conf import settings
from django.urls import reverse
# Create your models here.
class Room(models.Model):
ROOM_CATEGORIES = (
('BZS', 'BUSINESS SUITE'),
('TNS', 'TWIN SUITE'),
('EXS', 'EXECUTIVE SUITE'),
('SGB', 'SINGLE BED'),
)
room_... | [
"django.db.models.FloatField",
"django.db.models.ForeignKey",
"django.db.models.IntegerField",
"django.urls.reverse",
"django.db.models.DateTimeField",
"django.db.models.CharField"
] | [((329, 371), 'django.db.models.IntegerField', 'models.IntegerField', ([], {'null': '(True)', 'blank': '(True)'}), '(null=True, blank=True)\n', (348, 371), False, 'from django.db import models\n'), ((387, 442), 'django.db.models.CharField', 'models.CharField', ([], {'choices': 'ROOM_CATEGORIES', 'max_length': '(3)'}), ... |
from HABApp.openhab.connection_handler import http_connection
from ._plugin import on_connect, on_disconnect, setup_plugins
log = http_connection.log
def setup():
from HABApp.runtime import shutdown
# initialize callbacks
http_connection.ON_CONNECTED = on_connect
http_connection.ON_DISCONNECTED = on... | [
"HABApp.runtime.shutdown.register_func",
"HABApp.openhab.connection_handler.http_connection.start_connection"
] | [((375, 470), 'HABApp.runtime.shutdown.register_func', 'shutdown.register_func', (['http_connection.stop_connection'], {'msg': '"""Stopping openHAB connection"""'}), "(http_connection.stop_connection, msg=\n 'Stopping openHAB connection')\n", (397, 470), False, 'from HABApp.runtime import shutdown\n'), ((506, 575), ... |
#Imported to DeOXy
from userbot import bot, BOTLOG_CHATID, ALIVE_NAME, CMD_LIST
import asyncio
from telethon import events
from telethon.tl.functions.channels import EditBannedRequest
from telethon.tl.types import (PeerChat, PeerChannel,ChannelParticipantsAdmins, ChatAdminRights,ChatBannedRights, MessageEntityMentionN... | [
"telethon.tl.functions.contacts.BlockRequest",
"userbot.modules.sql_helper.gmute_sql.gmute",
"telethon.tl.functions.contacts.UnblockRequest",
"userbot.modules.sql_helper.gmute_sql.ungmute"
] | [((2568, 2582), 'userbot.modules.sql_helper.gmute_sql.gmute', 'gmute', (['user.id'], {}), '(user.id)\n', (2573, 2582), False, 'from userbot.modules.sql_helper.gmute_sql import gmute\n'), ((4763, 4779), 'userbot.modules.sql_helper.gmute_sql.ungmute', 'ungmute', (['user.id'], {}), '(user.id)\n', (4770, 4779), False, 'fro... |
import bluetooth
import time
import RPi.GPIO as GPIO
from bt_rssi import BluetoothRSSI
search_time = 10
GPIO.setmode(GPIO.BCM)
GPIO.setup(17, GPIO.OUT)
THRESHOLD = (-13, 38)
# You can hardcode the desired device ID here as a string to skip the discovery stage
addr = "F4:AF:E7:72:CB:19"
b = BluetoothRSSI(addr=addr)
pr... | [
"RPi.GPIO.cleanup",
"bt_rssi.BluetoothRSSI",
"bluetooth.discover_devices",
"RPi.GPIO.setup",
"RPi.GPIO.output",
"time.sleep",
"bluetooth.find_service",
"RPi.GPIO.setmode",
"bluetooth.lookup_name"
] | [((105, 127), 'RPi.GPIO.setmode', 'GPIO.setmode', (['GPIO.BCM'], {}), '(GPIO.BCM)\n', (117, 127), True, 'import RPi.GPIO as GPIO\n'), ((128, 152), 'RPi.GPIO.setup', 'GPIO.setup', (['(17)', 'GPIO.OUT'], {}), '(17, GPIO.OUT)\n', (138, 152), True, 'import RPi.GPIO as GPIO\n'), ((292, 316), 'bt_rssi.BluetoothRSSI', 'Blueto... |
import clockngpn.totp as totp
from clockngpn.proc_worker import Event, Broker, ProcWorkerEvent
from clockngpn.ttp import TocTocPorts, TocTocPortsWorker
from queue import Queue
import time
import os
import logging
from clockngpn.bidi import OTPBidi
import signal
import argparse
log = logging.getLogger(__name__)
def... | [
"logging.getLogger",
"logging.basicConfig",
"signal.signal",
"argparse.ArgumentParser",
"clockngpn.port_manager.PortManager",
"clockngpn.ttp.TocTocPorts",
"clockngpn.proc_worker.Broker",
"clockngpn.totp.web_secret_2_bytes",
"clockngpn.totp.gen_secret",
"clockng.firewall_manager.FirewallManager",
... | [((287, 314), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (304, 314), False, 'import logging\n'), ((1148, 1155), 'queue.Queue', 'Queue', ([], {}), '()\n', (1153, 1155), False, 'from queue import Queue\n'), ((1165, 1172), 'queue.Queue', 'Queue', ([], {}), '()\n', (1170, 1172), False, 'f... |
"""
CLI for azureml module.
"""
import json
import click
from mlflow.google.aiplatform import register_model as do_register_model
from mlflow.utils import cli_args
from mlflow.utils.annotations import experimental
@click.group("google")
def commands():
"""
Serve models on Google Cloud AI Platform. **These c... | [
"click.group",
"click.option",
"json.loads",
"mlflow.google.aiplatform.register_model"
] | [((219, 240), 'click.group', 'click.group', (['"""google"""'], {}), "('google')\n", (230, 240), False, 'import click\n'), ((722, 846), 'click.option', 'click.option', (['"""--display-name"""', '"""-n"""'], {'required': '(True)', 'help': '"""Name of the model when it\'s registered on Cloud AI Platform"""'}), '(\'--displ... |
# Copyright Amazon.com, Inc. or its affiliates. All Rights Reserved.
# SPDX-License-Identifier: Apache-2.0
import numpy as np
from music21 import midi
import pypianoroll
from pypianoroll import Multitrack
from texttable import Texttable
import os
from pprint import pprint
def play_midi(input_midi):
'''Takes path ... | [
"numpy.random.choice",
"os.path.join",
"music21.midi.MidiFile",
"music21.midi.translate.midiFileToStream",
"numpy.load",
"os.walk"
] | [((457, 472), 'music21.midi.MidiFile', 'midi.MidiFile', ([], {}), '()\n', (470, 472), False, 'from music21 import midi\n'), ((569, 613), 'music21.midi.translate.midiFileToStream', 'midi.translate.midiFileToStream', (['midi_object'], {}), '(midi_object)\n', (600, 613), False, 'from music21 import midi\n'), ((920, 933), ... |
# *****************************************************************
# Copyright 2013 MIT Lincoln Laboratory
# Project: SPAR
# Authors: SY
# Description: Section class
#
#
# Modifications:
# Date Name Modification
# ---- ---- ... | [
"spar_python.report_generation.common.latex_classes.LatexTable",
"spar_python.report_generation.ta1.ta1_analysis_input.Input",
"spar_python.report_generation.ta1.ta1_analysis_percentiles.Ta1PercentileGetter"
] | [((1833, 1950), 'spar_python.report_generation.common.latex_classes.LatexTable', 'latex_classes.LatexTable', (['caption', '"""perc_main"""', "['DBNR', 'DBRS', 'Select', 'Query Type', 'Num Passing $\\\\%$iles']"], {}), "(caption, 'perc_main', ['DBNR', 'DBRS', 'Select',\n 'Query Type', 'Num Passing $\\\\%$iles'])\n", ... |
#-*- encoding: utf-8 -*-
"""
About : Code to run as a linux daemon service, interface with it via REST POST
"""
__author__ = "<NAME>"
__email__ = "<EMAIL>"
__company__ = ""
__copyright__ = "Copyright (C) 2020 {a}".format(a=__author__)
__credits__ = ""
__license__ = "MIT"
__version__ = 0.03
__lastdate__ = "2020-04-13"... | [
"json.loads",
"uuid.uuid5",
"io.BytesIO",
"time.sleep",
"uuid.uuid4",
"http.server.HTTPServer",
"threading.Thread",
"queue.Queue",
"time.time"
] | [((868, 875), 'queue.Queue', 'Queue', ([], {}), '()\n', (873, 875), False, 'from queue import Queue\n'), ((5733, 5742), 'io.BytesIO', 'BytesIO', ([], {}), '()\n', (5740, 5742), False, 'from io import BytesIO\n'), ((5941, 5950), 'io.BytesIO', 'BytesIO', ([], {}), '()\n', (5948, 5950), False, 'from io import BytesIO\n'),... |
# Generated by Django 2.1.2 on 2018-11-08 12:48
from django.db import migrations, models
import django.db.models.deletion
class Migration(migrations.Migration):
dependencies = [
('Game', '0004_auto_20181106_1255'),
]
operations = [
migrations.CreateModel(
name='Loan',
... | [
"django.db.models.DateField",
"django.db.models.FloatField",
"django.db.models.IntegerField",
"django.db.models.ForeignKey",
"django.db.models.AutoField"
] | [((1325, 1416), 'django.db.models.ForeignKey', 'models.ForeignKey', ([], {'on_delete': 'django.db.models.deletion.DO_NOTHING', 'to': '"""Game.LoanOffer"""'}), "(on_delete=django.db.models.deletion.DO_NOTHING, to=\n 'Game.LoanOffer')\n", (1342, 1416), False, 'from django.db import migrations, models\n'), ((358, 451),... |
#!/usr/bin/env python
from __future__ import division
import rospy
from std_msgs.msg import String
from hopper_msgs.msg import ServoTelemetry, HexapodTelemetry
from sensor_msgs.msg import BatteryState
def mean(numbers):
return sum(numbers) / max(len(numbers), 1)
SERVO_COUNT = 18
MAX_VOLTAGE = 12.5
MIN_VOLTAGE ... | [
"rospy.Subscriber",
"rospy.init_node",
"sensor_msgs.msg.BatteryState",
"rospy.Duration.from_sec",
"rospy.Time.now",
"rospy.spin",
"rospy.Publisher"
] | [((403, 430), 'rospy.Duration.from_sec', 'rospy.Duration.from_sec', (['(15)'], {}), '(15)\n', (426, 430), False, 'import rospy\n'), ((517, 574), 'rospy.init_node', 'rospy.init_node', (['"""hopper_battery_monitor"""'], {'anonymous': '(True)'}), "('hopper_battery_monitor', anonymous=True)\n", (532, 574), False, 'import r... |
import pytest
from datacube.utils.geometry import box
from datacube_ows.cube_pool import cube
from datacube_ows.ogc_utils import local_solar_date_range
from datacube_ows.ows_configuration import get_config
from datacube_ows.mv_index import MVSelectOpts, mv_search_datasets
def test_full_layer():
cfg = get_config... | [
"pytest.helpers.enclosed_bbox",
"datacube_ows.mv_index.mv_search_datasets",
"datacube.utils.geometry.box",
"datacube_ows.ows_configuration.get_config",
"datacube_ows.cube_pool.cube"
] | [((310, 322), 'datacube_ows.ows_configuration.get_config', 'get_config', ([], {}), '()\n', (320, 322), False, 'from datacube_ows.ows_configuration import get_config\n'), ((693, 705), 'datacube_ows.ows_configuration.get_config', 'get_config', ([], {}), '()\n', (703, 705), False, 'from datacube_ows.ows_configuration impo... |
# coding:utf-8
# @Time : 2019-01-14 15:22
# @Author : 小贰
# @FileName: ansible_sync_hosts.py
# @function: ansible for python 3.x
import json
from collections import namedtuple
from ansible.parsing.dataloader import DataLoader
from ansible.vars.manager import VariableManager
from ansible.inventory.manager import In... | [
"collections.namedtuple",
"ansible.vars.manager.VariableManager",
"ansible.playbook.play.Play",
"json.dumps",
"ansible.inventory.manager.InventoryManager",
"ansible.parsing.dataloader.DataLoader",
"ansible.executor.task_queue_manager.TaskQueueManager"
] | [((1464, 1476), 'ansible.parsing.dataloader.DataLoader', 'DataLoader', ([], {}), '()\n', (1474, 1476), False, 'from ansible.parsing.dataloader import DataLoader\n'), ((1507, 1670), 'collections.namedtuple', 'namedtuple', (['"""Options"""', "['connection', 'module_path', 'forks', 'become', 'become_method',\n 'private... |
# -*- coding:utf-8 -*-
from django.db import models
from alipay import conf
from alipay.models import AliPayBaseModel
from alipay.create_direct_pay_by_user.dpn.signals import alipay_dpn_flagged, alipay_dpn_successful
class AliPayDPN(AliPayBaseModel):
"""
AliPay DPN
"""
gmt_close = models.DateTimeFiel... | [
"django.db.models.DateTimeField",
"django.db.models.CharField",
"alipay.create_direct_pay_by_user.dpn.signals.alipay_dpn_successful.send",
"alipay.create_direct_pay_by_user.dpn.signals.alipay_dpn_flagged.send"
] | [((301, 344), 'django.db.models.DateTimeField', 'models.DateTimeField', ([], {'blank': '(True)', 'null': '(True)'}), '(blank=True, null=True)\n', (321, 344), False, 'from django.db import models\n'), ((370, 425), 'django.db.models.CharField', 'models.CharField', ([], {'blank': '(True)', 'null': '(True)', 'max_length': ... |
import numpy as np
import pandas as pd
import time
mnist = pd.read_csv("../input/train.csv")
mnist.head()
y_train = mnist.label.values
x_train = mnist.drop('label',axis=1)
x_train = (x_train / 255.0).values
x_train = np.reshape(x_train,(42000,1,28,28))
x_train.shape
from keras.models import Sequential
from keras.l... | [
"keras.backend.set_image_data_format",
"keras.layers.core.Flatten",
"matplotlib.pyplot.grid",
"numpy.reshape",
"pandas.read_csv",
"matplotlib.pyplot.xticks",
"keras.layers.pooling.MaxPooling2D",
"time.strftime",
"numpy.argmax",
"keras.models.Sequential",
"matplotlib.pyplot.figure",
"keras.laye... | [((60, 93), 'pandas.read_csv', 'pd.read_csv', (['"""../input/train.csv"""'], {}), "('../input/train.csv')\n", (71, 93), True, 'import pandas as pd\n'), ((221, 260), 'numpy.reshape', 'np.reshape', (['x_train', '(42000, 1, 28, 28)'], {}), '(x_train, (42000, 1, 28, 28))\n', (231, 260), True, 'import numpy as np\n'), ((531... |
import subprocess
import time
from os import listdir
from os.path import isfile, join
COMMAND = ["python3", "onnx_sender.py",]
DIR = "working"
def get_file_list_from_dir(mypath):
onlyfiles = [join(mypath, f) for f in listdir(mypath) if isfile(join(mypath, f))]
return onlyfiles
file_list = get_file_list_fro... | [
"os.listdir",
"subprocess.run",
"os.path.join"
] | [((621, 652), 'subprocess.run', 'subprocess.run', (['current_command'], {}), '(current_command)\n', (635, 652), False, 'import subprocess\n'), ((199, 214), 'os.path.join', 'join', (['mypath', 'f'], {}), '(mypath, f)\n', (203, 214), False, 'from os.path import isfile, join\n'), ((224, 239), 'os.listdir', 'listdir', (['m... |
import paho.mqtt.client as mqtt
# 當地端程式連線伺服器得到回應時,要做的動作
def on_connect(client, userdata, flags, rc):
print("Connected with result code "+str(rc))
# 將訂閱主題寫在on_connet中
# 如果我們失去連線或重新連線時
# 地端程式將會重新訂閱
client.subscribe("Try/MQTT")
# 當接收到從伺服器發送的訊息時要進行的動作
def on_message(client, userdata, msg):
# 轉換編... | [
"paho.mqtt.client.Client"
] | [((414, 427), 'paho.mqtt.client.Client', 'mqtt.Client', ([], {}), '()\n', (425, 427), True, 'import paho.mqtt.client as mqtt\n')] |
import copy
import pprint
import inspect
from collections import OrderedDict
import six
from neupy.exceptions import LayerConnectionError
__all__ = ('LayerGraph',)
def filter_list(iterable, include_values):
"""
Create new list that contains only values
specified in the ``include_values`` attribute.
... | [
"collections.OrderedDict",
"inspect.ismethod",
"pprint.pformat",
"inspect.getargspec",
"copy.copy"
] | [((1278, 1291), 'collections.OrderedDict', 'OrderedDict', ([], {}), '()\n', (1289, 1291), False, 'from collections import OrderedDict\n'), ((1895, 1908), 'collections.OrderedDict', 'OrderedDict', ([], {}), '()\n', (1906, 1908), False, 'from collections import OrderedDict\n'), ((3980, 4012), 'inspect.getargspec', 'inspe... |
# Copyright (c) 2019, NVIDIA CORPORATION. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applic... | [
"torch.index_select",
"kaolin.nnsearch.nnsearch",
"torch.mul",
"kaolin.cuda.sided_distance.forward_cuda",
"torch.sum",
"kaolin.cuda.sided_distance.forward",
"torch.zeros"
] | [((976, 1001), 'torch.zeros', 'torch.zeros', (['batchsize', 'n'], {}), '(batchsize, n)\n', (987, 1001), False, 'import torch\n'), ((1017, 1059), 'torch.zeros', 'torch.zeros', (['batchsize', 'n'], {'dtype': 'torch.int'}), '(batchsize, n, dtype=torch.int)\n', (1028, 1059), False, 'import torch\n'), ((3854, 3900), 'torch.... |
import argparse
import sys
import torch
from inferno.trainers.basic import Trainer
from inferno.trainers.callbacks.logging.tensorboard import TensorboardLogger
from torch import nn
from torch.autograd import Variable
from torch.utils.data.dataloader import DataLoader
from torch.utils.data.dataset import Dataset
from t... | [
"mnist_wgangp.WGANGeneratorLoss",
"torch.nn.BatchNorm1d",
"torch.cuda.is_available",
"mnist_gan.Reshape",
"mnist_gan.format_images",
"torch.arange",
"mnist_gan.generate_video",
"torch.nn.Sigmoid",
"torch.nn.BatchNorm2d",
"argparse.ArgumentParser",
"torchvision.transforms.ToTensor",
"torch.auto... | [((7509, 7524), 'mnist_gan.save_args', 'save_args', (['args'], {}), '(args)\n', (7518, 7524), False, 'from mnist_gan import save_args, GANModel, GenerateDataCallback, GeneratorTrainingCallback\n'), ((7800, 7814), 'inferno.trainers.basic.Trainer', 'Trainer', (['model'], {}), '(model)\n', (7807, 7814), False, 'from infer... |
#!/usr/bin/env python
# coding:utf8
# Copyright (c) 2018, Tencent. All rights reserved
# This file contain DataProcessor class which can
# 1. Generate dict from train text data:
# Text format: label\t[(token )+]\t[(char )+]\t[(custom_feature )+].
# Label could be flattened or hierarchical which is separat... | [
"config.Config",
"util.Logger",
"tensorflow.TFRecordReader",
"tensorflow.cast",
"tensorflow.app.run",
"os.path.exists",
"tensorflow.train.Example",
"tensorflow.train.Coordinator",
"tensorflow.placeholder",
"json.dumps",
"tensorflow.estimator.export.ServingInputReceiver",
"tensorflow.python_io.... | [((28196, 28227), 'config.Config', 'Config', ([], {'config_file': 'sys.argv[1]'}), '(config_file=sys.argv[1])\n', (28202, 28227), False, 'from config import Config\n'), ((28348, 28360), 'tensorflow.app.run', 'tf.app.run', ([], {}), '()\n', (28358, 28360), True, 'import tensorflow as tf\n'), ((6482, 6532), 'codecs.open'... |
from nltk.corpus import stopwords
from nltk.tokenize import word_tokenize
text = "This is an example sentence for showing stop words filteration."
stop_words = set(stopwords.words("english"))
# all the stop words of englist language defined in nltk
# print(stop_words)
words = word_tokenize(text)
filtered_sentence ... | [
"nltk.corpus.stopwords.words",
"nltk.tokenize.word_tokenize"
] | [((281, 300), 'nltk.tokenize.word_tokenize', 'word_tokenize', (['text'], {}), '(text)\n', (294, 300), False, 'from nltk.tokenize import word_tokenize\n'), ((166, 192), 'nltk.corpus.stopwords.words', 'stopwords.words', (['"""english"""'], {}), "('english')\n", (181, 192), False, 'from nltk.corpus import stopwords\n')] |
import sys
import subprocess
ex = sys.executable
print(subprocess.run([ex, sys.argv[1], *sys.argv[2:]], capture_output=True).stdout.decode("UTF-8"))
| [
"subprocess.run"
] | [((57, 126), 'subprocess.run', 'subprocess.run', (['[ex, sys.argv[1], *sys.argv[2:]]'], {'capture_output': '(True)'}), '([ex, sys.argv[1], *sys.argv[2:]], capture_output=True)\n', (71, 126), False, 'import subprocess\n')] |
# -*- coding: utf-8 -*-
"""
# 3D Image Data Synthesis.
# Copyright (C) 2021 <NAME>, <NAME>, <NAME>, <NAME>, <NAME>
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the Liceense at
#
# http://www.apache.... | [
"numpy.clip",
"numpy.sqrt",
"numpy.array",
"scipy.ndimage.gaussian_filter",
"numpy.sin",
"numpy.gradient",
"numpy.arange",
"numpy.repeat",
"numpy.reshape",
"numpy.cross",
"os.path.split",
"numpy.dot",
"numpy.matmul",
"skimage.morphology.ball",
"numpy.maximum",
"pyshtools.SHCoeffs.from_... | [((2455, 2487), 'os.path.join', 'os.path.join', (['save_path', '"""masks"""'], {}), "(save_path, 'masks')\n", (2467, 2487), False, 'import os\n'), ((2580, 2621), 'numpy.random.randint', 'np.random.randint', (['min_radius', 'max_radius'], {}), '(min_radius, max_radius)\n', (2597, 2621), True, 'import numpy as np\n'), ((... |
# -*- coding: utf-8 -*-
from __future__ import division, print_function
__all__ = ["GP"]
try:
from itertools import izip
except ImportError:
izip = zip
import numpy as np
import scipy.optimize as op
from scipy.linalg import cho_factor, cho_solve, LinAlgError
from .utils import multivariate_gaussian_samples... | [
"numpy.diag_indices_from",
"scipy.linalg.cho_solve",
"numpy.atleast_2d",
"numpy.ones_like",
"scipy.linalg.cho_factor",
"scipy.optimize.minimize",
"numpy.log",
"numpy.diag",
"numpy.argsort",
"numpy.dot",
"numpy.isfinite",
"itertools.izip",
"numpy.zeros_like",
"numpy.arange",
"numpy.atleas... | [((2842, 2858), 'numpy.atleast_1d', 'np.atleast_1d', (['t'], {}), '(t)\n', (2855, 2858), True, 'import numpy as np\n'), ((3612, 3628), 'numpy.atleast_1d', 'np.atleast_1d', (['y'], {}), '(y)\n', (3625, 3628), True, 'import numpy as np\n'), ((5344, 5375), 'scipy.linalg.cho_factor', 'cho_factor', (['K'], {'overwrite_a': '... |
from surprise.model_selection import train_test_split
from surprise.model_selection import LeaveOneOut
from surprise import KNNBaseline
from surprise import Dataset, KNNBasic
from surprise import Reader
import heapq
from movies_analyzer.Movies import Movies, RATINGS, LINKS, MOVIES
from movies_recommender.utils import ... | [
"surprise.KNNBaseline",
"pandas.read_csv",
"surprise.model_selection.LeaveOneOut",
"surprise.Dataset.load_from_file",
"surprise.Dataset.load_from_df",
"surprise.model_selection.train_test_split",
"pandas.DataFrame.from_dict",
"movies_recommender.utils.get_popularity_ranking",
"surprise.Reader"
] | [((635, 679), 'pandas.read_csv', 'pd.read_csv', (['(movies.movielens_path / RATINGS)'], {}), '(movies.movielens_path / RATINGS)\n', (646, 679), True, 'import pandas as pd\n'), ((698, 769), 'surprise.Reader', 'Reader', ([], {'line_format': '"""user item rating timestamp"""', 'sep': '""","""', 'skip_lines': '(1)'}), "(li... |
# MIT License
#
# Copyright (c) 2018-2019 Red Hat, Inc.
# Permission is hereby granted, free of charge, to any person obtaining a copy
# of this software and associated documentation files (the "Software"), to deal
# in the Software without restriction, including without limitation the rights
# to use, copy, modify, m... | [
"packit_service.models.CoprBuild",
"packit.config.JobConfig",
"packit_service.config.ServiceConfig",
"packit.config.PackageConfig",
"ogr.abstract.GitService",
"packit_service.models.SRPMBuild",
"flexmock.flexmock"
] | [((2766, 2823), 'packit.config.PackageConfig', 'PackageConfig', ([], {'jobs': 'jobs', 'downstream_package_name': '"""dummy"""'}), "(jobs=jobs, downstream_package_name='dummy')\n", (2779, 2823), False, 'from packit.config import PackageConfig, JobConfig, JobType, JobConfigTriggerType\n'), ((4589, 4678), 'packit.config.J... |
# Configuration file for the Sphinx documentation builder.
#
# This file only contains a selection of the most common options. For a full
# list see the documentation:
# https://www.sphinx-doc.org/en/master/usage/configuration.html
# -- Path setup --------------------------------------------------------------
# If ex... | [
"os.getcwd"
] | [((642, 653), 'os.getcwd', 'os.getcwd', ([], {}), '()\n', (651, 653), False, 'import os\n')] |
import unittest
from collections import defaultdict
import numpy as np
import enstat.mean
class Test_mean(unittest.TestCase):
"""
tests
"""
def test_scalar(self):
"""
Basic test of "mean" and "std" using a random sample.
"""
average = enstat.scalar()
averag... | [
"numpy.mean",
"numpy.prod",
"numpy.random.random",
"numpy.logical_not",
"numpy.array",
"collections.defaultdict",
"numpy.std",
"unittest.main"
] | [((5207, 5222), 'unittest.main', 'unittest.main', ([], {}), '()\n', (5220, 5222), False, 'import unittest\n'), ((4084, 4110), 'collections.defaultdict', 'defaultdict', (['enstat.scalar'], {}), '(enstat.scalar)\n', (4095, 4110), False, 'from collections import defaultdict\n'), ((4570, 4596), 'collections.defaultdict', '... |
"""Script for testing selfies against large datasets.
"""
import argparse
import pathlib
import pandas as pd
from rdkit import Chem
from tqdm import tqdm
import selfies as sf
parser = argparse.ArgumentParser()
parser.add_argument("--data_path", type=str, default="version.smi.gz")
parser.add_argument("--col_name", t... | [
"pandas.read_csv",
"argparse.ArgumentParser",
"pathlib.Path",
"tqdm.tqdm.write",
"tqdm.tqdm",
"rdkit.Chem.MolFromSmiles",
"rdkit.Chem.CanonSmiles",
"selfies.decoder",
"selfies.set_semantic_constraints",
"selfies.encoder"
] | [((188, 213), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {}), '()\n', (211, 213), False, 'import argparse\n'), ((501, 523), 'pathlib.Path', 'pathlib.Path', (['__file__'], {}), '(__file__)\n', (513, 523), False, 'import pathlib\n'), ((708, 765), 'pandas.read_csv', 'pd.read_csv', (['TEST_SET_PATH'], {'sep... |
from typing import List, Optional # NOQA
import chainer
from chainer import cuda
from chainer import functions
from chainer import links
import numpy # NOQA
class Set2Set(chainer.Chain):
r"""MPNN subsubmodule for readout part.
See: <NAME>+, \
Order Matters: Sequence to sequence for sets. November ... | [
"chainer.functions.transpose",
"chainer.functions.stack",
"chainer.cuda.get_array_module",
"chainer.functions.sum",
"chainer.functions.concat",
"chainer.functions.reshape",
"chainer.functions.softmax",
"chainer.functions.matmul",
"chainer.links.NStepLSTM",
"chainer.functions.separate",
"chainer.... | [((1290, 1314), 'chainer.cuda.get_array_module', 'cuda.get_array_module', (['h'], {}), '(h)\n', (1311, 1314), False, 'from chainer import cuda\n'), ((1727, 1745), 'chainer.functions.stack', 'functions.stack', (['q'], {}), '(q)\n', (1742, 1745), False, 'from chainer import functions\n'), ((1777, 1815), 'chainer.function... |
"""Console script for kanbanflow_prj_selector."""
import sys
import click
from .kanbanflow_prj_selector import start
@click.command()
@click.option('-f', '--board-token-path', help='Input file with board tokens to fetch', required=True)
def main(board_token_path):
"""Console script for kanbanflow_prj_selector.""... | [
"click.option",
"click.command"
] | [((121, 136), 'click.command', 'click.command', ([], {}), '()\n', (134, 136), False, 'import click\n'), ((138, 244), 'click.option', 'click.option', (['"""-f"""', '"""--board-token-path"""'], {'help': '"""Input file with board tokens to fetch"""', 'required': '(True)'}), "('-f', '--board-token-path', help=\n 'Input ... |
#
# Copyright (C) 2014-2015
# <NAME> (<EMAIL>)
#
# This Program is free software; you can redistribute it and/or modify
# it under the terms of the GNU General Public License as published by
# the Free Software Foundation; either version 2, or (at your option)
# any later version.
#
# This Program is d... | [
"xbmcaddon.Addon"
] | [((1630, 1652), 'xbmcaddon.Addon', 'xbmcaddon.Addon', (['addon'], {}), '(addon)\n', (1645, 1652), False, 'import xbmcaddon\n'), ((1707, 1724), 'xbmcaddon.Addon', 'xbmcaddon.Addon', ([], {}), '()\n', (1722, 1724), False, 'import xbmcaddon\n')] |
import csv
import os
from os import mkdir
from os.path import isdir
from datetime import datetime
import ast
from configuration import Config
conf = Config()
traj_j = []
policy_per_plot = []
plot_policy = []
traj_j_ID = []
CSV_DIRECTORY_NAME = "Flights_trajectories"
BP_DIRECTORY_NAME = 'Best_policy/'
myfile = "./Bes... | [
"configuration.Config",
"csv.writer",
"os.path.isfile",
"datetime.datetime.now",
"ast.literal_eval",
"os.path.isdir",
"os.mkdir",
"csv.reader"
] | [((150, 158), 'configuration.Config', 'Config', ([], {}), '()\n', (156, 158), False, 'from configuration import Config\n'), ((520, 542), 'os.path.isfile', 'os.path.isfile', (['myfile'], {}), '(myfile)\n', (534, 542), False, 'import os\n'), ((3616, 3643), 'os.path.isfile', 'os.path.isfile', (['myfile_plot'], {}), '(myfi... |
#!/usr/bin/env python
# -*- coding: utf-8 -*-
from client import Sample_Client
import time
import random
# Create client
client = Sample_Client('sample.cfg')
time.sleep(5)
# Simple exemple
while(True):
print('Sending updates')
client.publish('Sensor1', 'L1_Volts', random.uniform(207,253))
client.publish... | [
"random.uniform",
"client.Sample_Client",
"time.sleep"
] | [((133, 160), 'client.Sample_Client', 'Sample_Client', (['"""sample.cfg"""'], {}), "('sample.cfg')\n", (146, 160), False, 'from client import Sample_Client\n'), ((161, 174), 'time.sleep', 'time.sleep', (['(5)'], {}), '(5)\n', (171, 174), False, 'import time\n'), ((1021, 1035), 'time.sleep', 'time.sleep', (['(60)'], {})... |
# -*- coding:utf-8 -*-
"""
外键关联相关的选项:
- model: 关联的资产Model
- field:关联的字段,这个字段必须是unique,默认可选择id
- on_delete: 当关联的外键删除的时候的操作:cascade | set_null | disable
- on_update: 当关联的外键修改的时候: cascade | disable
这些功能/约束,都是用代码逻辑来实现的,其实尽量不要使用外键:
少用的话,其实可以把约束放到业务代码中
"""
from cmdb.types.base import BaseType
# 注意别循环引用了
from cmdb.models i... | [
"cmdb.models.Field.objects.filter",
"cmdb.models.Model.objects.filter"
] | [((885, 922), 'cmdb.models.Model.objects.filter', 'Model.objects.filter', ([], {'code': 'model_code'}), '(code=model_code)\n', (905, 922), False, 'from cmdb.models import Model, Field, Instance, Value\n'), ((1481, 1531), 'cmdb.models.Field.objects.filter', 'Field.objects.filter', ([], {'model': 'model', 'code': 'field_... |
# !/usr/bin/env python
# -*- coding: utf-8 -*-
import threading
import time
import urllib2
# SELECT "bmember" as type, COUNT(*) AS bmember FROM bmember WHERE isDownPics=0
# UNION ALL SELECT "memberby " as type, COUNT(*) AS memberby FROM memberby
# UNION ALL SELECT "membercontact" as type, C... | [
"urllib2.Request",
"threading.Thread",
"urllib2.urlopen",
"time.sleep"
] | [((729, 769), 'urllib2.Request', 'urllib2.Request', ([], {'url': 'url', 'headers': 'header'}), '(url=url, headers=header)\n', (744, 769), False, 'import urllib2\n'), ((798, 822), 'urllib2.urlopen', 'urllib2.urlopen', (['request'], {}), '(request)\n', (813, 822), False, 'import urllib2\n'), ((1038, 1051), 'time.sleep', ... |
import sys, psycopg2, psycopg2.extras
def generate(user_id, db):
try:
with db.cursor(cursor_factory = psycopg2.extras.DictCursor) as cursor:
cursor.execute("""
INSERT INTO minicloud_auths (user_id) VALUES (%s) RETURNING token
""", [int(user_id)])
data = curs... | [
"sys.stderr.write",
"sys.stderr"
] | [((449, 491), 'sys.stderr', 'sys.stderr', (["('%s\\n' % 'No token generated!')"], {}), "('%s\\n' % 'No token generated!')\n", (459, 491), False, 'import sys, psycopg2, psycopg2.extras\n'), ((778, 820), 'sys.stderr.write', 'sys.stderr.write', (['"""X-Auth-Token revoked\n"""'], {}), "('X-Auth-Token revoked\\n')\n", (794,... |
import graphene
from app.models import User
class RegisterMutation(graphene.Mutation):
class Arguments(object):
email = graphene.String()
username = graphene.String()
password = graphene.String()
is_success = graphene.Boolean()
message = graphene.String()
@staticmethod
... | [
"graphene.String",
"app.models.User",
"graphene.Boolean"
] | [((247, 265), 'graphene.Boolean', 'graphene.Boolean', ([], {}), '()\n', (263, 265), False, 'import graphene\n'), ((280, 297), 'graphene.String', 'graphene.String', ([], {}), '()\n', (295, 297), False, 'import graphene\n'), ((137, 154), 'graphene.String', 'graphene.String', ([], {}), '()\n', (152, 154), False, 'import g... |
from init_repo import init_repo
def find_commit(repo, local_repo, version, branch='master'):
"""
Find commit with specified version in the DESCRIPTION file in the xcms repo
This function checkout to specified version
:param repo: git.repo.base.Repo - repository object
:param local_repo: str - path... | [
"init_repo.init_repo"
] | [((1601, 1669), 'init_repo.init_repo', 'init_repo', ([], {'repo_clone_url': 'repo_clone_url', 'path': 'path', 'version': 'version'}), '(repo_clone_url=repo_clone_url, path=path, version=version)\n', (1610, 1669), False, 'from init_repo import init_repo\n')] |
"""Author: <NAME>, Copyright 2019"""
import numpy as np
import mineral as ml
from mineral.core.samplers.sampler import Sampler
class PathSampler(Sampler):
def __init__(
self,
env,
policies,
buffers,
time_skips=(1,),
**kwargs
):
Sampl... | [
"numpy.prod",
"mineral.core.samplers.sampler.Sampler.__init__",
"mineral.nested_apply"
] | [((315, 347), 'mineral.core.samplers.sampler.Sampler.__init__', 'Sampler.__init__', (['self'], {}), '(self, **kwargs)\n', (331, 347), False, 'from mineral.core.samplers.sampler import Sampler\n'), ((1194, 1230), 'numpy.prod', 'np.prod', (['self.time_skips[:level + 1]'], {}), '(self.time_skips[:level + 1])\n', (1201, 12... |
# -*- coding: utf-8 -*-
# Generated by Django 1.10.1 on 2017-03-19 18:41
from __future__ import unicode_literals
from django.db import migrations, models
class Migration(migrations.Migration):
dependencies = [
('references', '0054_auto_20170308_2100'),
]
operations = [
migrations.Create... | [
"django.db.models.DateTimeField",
"django.db.models.AutoField",
"django.db.models.CharField"
] | [((403, 496), 'django.db.models.AutoField', 'models.AutoField', ([], {'auto_created': '(True)', 'primary_key': '(True)', 'serialize': '(False)', 'verbose_name': '"""ID"""'}), "(auto_created=True, primary_key=True, serialize=False,\n verbose_name='ID')\n", (419, 496), False, 'from django.db import migrations, models\... |
from importlib import import_module
from django.db import models
from django.db.models.signals import post_save
from django.utils.translation import gettext_lazy as _
from django.utils.module_loading import import_string
from django.conf import settings
from django_otp import devices_for_user
from .services import Us... | [
"importlib.import_module",
"django_otp.devices_for_user",
"django.db.models.signals.post_save.connect",
"django.utils.translation.gettext_lazy",
"django.utils.module_loading.import_string"
] | [((4009, 4094), 'django.db.models.signals.post_save.connect', 'post_save.connect', (['UserServicesMixin._post_save'], {'sender': 'settings.AUTH_USER_MODEL'}), '(UserServicesMixin._post_save, sender=settings.AUTH_USER_MODEL\n )\n', (4026, 4094), False, 'from django.db.models.signals import post_save\n'), ((4090, 4177... |
from __future__ import unicode_literals
import unittest
from mock import Mock
from mopidy.models import Ref
from mopidy_oe1.library import OE1LibraryProvider, OE1LibraryUri, OE1UriType
class OE1LibraryUriTest(unittest.TestCase):
def test_parse_root_uri(self):
uri = 'oe1:directory'
result = OE1... | [
"mock.Mock",
"mopidy_oe1.library.OE1LibraryUri",
"mopidy_oe1.library.OE1LibraryProvider",
"mopidy_oe1.library.OE1LibraryUri.parse"
] | [((317, 341), 'mopidy_oe1.library.OE1LibraryUri.parse', 'OE1LibraryUri.parse', (['uri'], {}), '(uri)\n', (336, 341), False, 'from mopidy_oe1.library import OE1LibraryProvider, OE1LibraryUri, OE1UriType\n'), ((479, 503), 'mopidy_oe1.library.OE1LibraryUri.parse', 'OE1LibraryUri.parse', (['uri'], {}), '(uri)\n', (498, 503... |
#! /usr/bin/env python
# -*- coding: utf-8 -*-
"""Integrate two files known as 'packages.xml'.
Usage:
python combine_packages_xml.py FPATH_1 FPATH_2 >result.xml
Description:
The script reads FPATH_1 and updates the data with FPATH_2.
Entries are sorted by 'version+language'. The file timestamp
is set... | [
"datetime.datetime.fromtimestamp",
"re.compile",
"sys.exit",
"codecs.open",
"time.time"
] | [((964, 1005), 'datetime.datetime.fromtimestamp', 'datetime.datetime.fromtimestamp', (['unixtime'], {}), '(unixtime)\n', (995, 1005), False, 'import datetime\n'), ((6808, 7021), 're.compile', 're.compile', (['"""\n <languagepack\n \\\\s+\n version="(?P<version>\\\\d+\\\\.\\\\d+\\\\.\\\\d+)"\n \\\\s+\n languag... |