blob_id
stringlengths
40
40
directory_id
stringlengths
40
40
path
stringlengths
2
616
content_id
stringlengths
40
40
detected_licenses
listlengths
0
69
license_type
stringclasses
2 values
repo_name
stringlengths
5
118
snapshot_id
stringlengths
40
40
revision_id
stringlengths
40
40
branch_name
stringlengths
4
63
visit_date
timestamp[us]
revision_date
timestamp[us]
committer_date
timestamp[us]
github_id
int64
2.91k
686M
star_events_count
int64
0
209k
fork_events_count
int64
0
110k
gha_license_id
stringclasses
23 values
gha_event_created_at
timestamp[us]
gha_created_at
timestamp[us]
gha_language
stringclasses
220 values
src_encoding
stringclasses
30 values
language
stringclasses
1 value
is_vendor
bool
2 classes
is_generated
bool
2 classes
length_bytes
int64
2
10.3M
extension
stringclasses
257 values
content
stringlengths
2
10.3M
authors
listlengths
1
1
author_id
stringlengths
0
212
a9d72cd94acdd3bc036f28fd6640c9053f57bcc8
e366e330fcf6c4f35e337056f73e65fb38bf351a
/coordenadas.py
077436ae877391917839124eec331bd88a8839a4
[]
no_license
CBGO2/Mision-02
5df5b7c67df6f2d62a24211ded1c5be67415ba9e
d5fa407d7466d8b731044d34f01cfd26b0b14202
refs/heads/master
2020-03-26T22:50:08.595899
2018-08-22T03:52:31
2018-08-22T03:52:31
145,490,598
0
0
null
2018-08-21T01:42:40
2018-08-21T01:42:39
null
UTF-8
Python
false
false
456
py
# Autor: Carlos Badillo García, A01377618 # Descripcion: Usando las dos coordenadas dadas, descubrir la distancia entre ambas. # Escribe tu programa después de esta línea. ppx = int(input("¿Cuál es el valor de x1?")) ppy = int(input("¿Cuál es el valor de y1?")) spx = int(input("¿Cuál es el valor de x2?")) spy = int(input("¿Cuál es el valor de y2?")) d = ((spx-ppx)**2+(spy-ppy)**2)**(1/2) print("La distancia entre los dos puntos es:", d)
[ "noreply@github.com" ]
CBGO2.noreply@github.com
411846624c150abad251688c80a09c1cad1dc3a9
3dcc6eaef0ca68b230ed61b9fd2bfaf78f8d1c7d
/todo_app/todos/models/__init__.py
007b0f8bc1c970fe2f9d07ff26b0dd5391d4d216
[]
no_license
ivo-bass/ToDo-App
a6f92be6ba8dcb266cd9ab58d50bafc44ce3db9f
0410fe885f729ef85e83a7779a5e971e42f74479
refs/heads/main
2023-05-14T13:28:50.219962
2021-06-18T13:14:49
2021-06-18T13:14:49
373,607,487
1
0
null
null
null
null
UTF-8
Python
false
false
85
py
from .todo import Todo from .priority import Priority from .category import Category
[ "ivailo.ignatoff@gmail.com" ]
ivailo.ignatoff@gmail.com
37e222c25edc77b749c6c4a0af3b961c6cd12de0
03dc4293ba7f7d1ab61668d1b82c88a21e50a942
/src/design/change_card.py
010acd1a2ded6f1a1dabeb37aabb8179c175f09a
[]
no_license
JabaXNT/YandexProject2
e0f333374e719c539b09c81466cfc3b6474a245c
d362b5979f37969a37ad0fd796d452f3f7f64749
refs/heads/master
2023-02-02T17:24:35.469464
2020-12-25T09:53:52
2020-12-25T09:53:52
321,351,130
0
0
null
null
null
null
UTF-8
Python
false
false
3,686
py
# -*- coding: utf-8 -*- # Form implementation generated from reading ui file 'change_card.ui' # # Created by: PyQt5 UI code generator 5.15.1 # # WARNING: Any manual changes made to this file will be lost when pyuic5 is # run again. Do not edit this file unless you know what you are doing. from PyQt5 import QtCore, QtGui, QtWidgets class Ui_Dialog_2(object): def setupUi(self, Dialog_2): Dialog_2.setObjectName("Dialog_2") Dialog_2.resize(413, 221) self.sur_new = QtWidgets.QLineEdit(Dialog_2) self.sur_new.setGeometry(QtCore.QRect(80, 30, 211, 20)) self.sur_new.setObjectName("sur_new") self.pat_new = QtWidgets.QLineEdit(Dialog_2) self.pat_new.setGeometry(QtCore.QRect(200, 90, 211, 20)) self.pat_new.setObjectName("pat_new") self.bday_lbl = QtWidgets.QLabel(Dialog_2) self.bday_lbl.setGeometry(QtCore.QRect(0, 120, 201, 21)) font = QtGui.QFont() font.setPointSize(12) self.bday_lbl.setFont(font) self.bday_lbl.setObjectName("bday_lbl") self.sur_lbl = QtWidgets.QLabel(Dialog_2) self.sur_lbl.setGeometry(QtCore.QRect(0, 30, 81, 21)) font = QtGui.QFont() font.setPointSize(12) self.sur_lbl.setFont(font) self.sur_lbl.setLineWidth(1) self.sur_lbl.setObjectName("sur_lbl") self.name_new = QtWidgets.QLineEdit(Dialog_2) self.name_new.setGeometry(QtCore.QRect(80, 60, 211, 20)) self.name_new.setObjectName("name_new") self.name_lbl = QtWidgets.QLabel(Dialog_2) self.name_lbl.setGeometry(QtCore.QRect(0, 60, 47, 21)) font = QtGui.QFont() font.setPointSize(12) self.name_lbl.setFont(font) self.name_lbl.setObjectName("name_lbl") self.bday_new = QtWidgets.QLineEdit(Dialog_2) self.bday_new.setGeometry(QtCore.QRect(120, 120, 211, 20)) self.bday_new.setObjectName("bday_new") self.Enter_lbl = QtWidgets.QLabel(Dialog_2) self.Enter_lbl.setGeometry(QtCore.QRect(0, 0, 191, 31)) font = QtGui.QFont() font.setPointSize(16) font.setBold(True) font.setWeight(75) self.Enter_lbl.setFont(font) self.Enter_lbl.setObjectName("Enter_lbl") self.fath_lbl = QtWidgets.QLabel(Dialog_2) self.fath_lbl.setGeometry(QtCore.QRect(0, 90, 201, 21)) font = QtGui.QFont() font.setPointSize(12) self.fath_lbl.setFont(font) self.fath_lbl.setObjectName("fath_lbl") self.save_btn = QtWidgets.QPushButton(Dialog_2) self.save_btn.setGeometry(QtCore.QRect(160, 150, 91, 41)) self.save_btn.setObjectName("save_btn") self.label = QtWidgets.QLabel(Dialog_2) self.label.setGeometry(QtCore.QRect(0, 190, 411, 21)) font = QtGui.QFont() font.setPointSize(12) self.label.setFont(font) self.label.setText("") self.label.setObjectName("label") self.retranslateUi(Dialog_2) QtCore.QMetaObject.connectSlotsByName(Dialog_2) def retranslateUi(self, Dialog_2): _translate = QtCore.QCoreApplication.translate Dialog_2.setWindowTitle(_translate("Dialog_2", "Dialog")) self.bday_lbl.setText(_translate("Dialog_2", "Год Рождения:")) self.sur_lbl.setText(_translate("Dialog_2", "Фамилию:")) self.name_lbl.setText(_translate("Dialog_2", "Имя:")) self.Enter_lbl.setText(_translate("Dialog_2", "Введите:")) self.fath_lbl.setText(_translate("Dialog_2", "Отчество (При наличии):")) self.save_btn.setText(_translate("Dialog_2", "Сохранить"))
[ "gluhow2015@gmail.com" ]
gluhow2015@gmail.com
636fe7f33650c3bd29921d6bf95425a2aeeaef48
d09fd96bbc931fbb8522e5c991973f064a4ded50
/baxter/devel/.private/baxter_maintenance_msgs/lib/python2.7/dist-packages/baxter_maintenance_msgs/msg/_UpdateStatus.py
dcdfcbd9e5d9bc1182afd40950d3c1c371b7df12
[]
no_license
rymonyu/EE4-Robotics
b3827ba0dff5bdfdd1e47fe07a40e955c5226f38
6cf9272abd7fe8a074dc74a032f6e0b35edb8548
refs/heads/master
2020-08-22T15:09:39.706809
2019-12-15T23:35:45
2019-12-15T23:35:45
216,420,098
6
0
null
null
null
null
UTF-8
Python
false
false
5,704
py
# This Python file uses the following encoding: utf-8 """autogenerated by genpy from baxter_maintenance_msgs/UpdateStatus.msg. Do not edit.""" import sys python3 = True if sys.hexversion > 0x03000000 else False import genpy import struct class UpdateStatus(genpy.Message): _md5sum = "74e246350421569590252c39e8aa7b85" _type = "baxter_maintenance_msgs/UpdateStatus" _has_header = False #flag to mark the presence of a Header object _full_text = """# See the class UpdateRunner() # status: One-word description of the current action being performed # long_description: Details pertaining to status if any. Used for verbose error messages. uint16 status float32 progress string long_description uint16 STS_IDLE = 0 uint16 STS_INVALID = 1 uint16 STS_BUSY = 2 uint16 STS_CANCELLED = 3 uint16 STS_ERR = 4 uint16 STS_MOUNT_UPDATE = 5 uint16 STS_VERIFY_UPDATE = 6 uint16 STS_PREP_STAGING = 7 uint16 STS_MOUNT_STAGING = 8 uint16 STS_EXTRACT_UPDATE = 9 uint16 STS_LOAD_KEXEC = 10 """ # Pseudo-constants STS_IDLE = 0 STS_INVALID = 1 STS_BUSY = 2 STS_CANCELLED = 3 STS_ERR = 4 STS_MOUNT_UPDATE = 5 STS_VERIFY_UPDATE = 6 STS_PREP_STAGING = 7 STS_MOUNT_STAGING = 8 STS_EXTRACT_UPDATE = 9 STS_LOAD_KEXEC = 10 __slots__ = ['status','progress','long_description'] _slot_types = ['uint16','float32','string'] def __init__(self, *args, **kwds): """ Constructor. Any message fields that are implicitly/explicitly set to None will be assigned a default value. The recommend use is keyword arguments as this is more robust to future message changes. You cannot mix in-order arguments and keyword arguments. The available fields are: status,progress,long_description :param args: complete set of field values, in .msg order :param kwds: use keyword arguments corresponding to message field names to set specific fields. """ if args or kwds: super(UpdateStatus, self).__init__(*args, **kwds) #message fields cannot be None, assign default values for those that are if self.status is None: self.status = 0 if self.progress is None: self.progress = 0. if self.long_description is None: self.long_description = '' else: self.status = 0 self.progress = 0. self.long_description = '' def _get_types(self): """ internal API method """ return self._slot_types def serialize(self, buff): """ serialize message into buffer :param buff: buffer, ``StringIO`` """ try: _x = self buff.write(_get_struct_Hf().pack(_x.status, _x.progress)) _x = self.long_description length = len(_x) if python3 or type(_x) == unicode: _x = _x.encode('utf-8') length = len(_x) buff.write(struct.pack('<I%ss'%length, length, _x)) except struct.error as se: self._check_types(struct.error("%s: '%s' when writing '%s'" % (type(se), str(se), str(locals().get('_x', self))))) except TypeError as te: self._check_types(ValueError("%s: '%s' when writing '%s'" % (type(te), str(te), str(locals().get('_x', self))))) def deserialize(self, str): """ unpack serialized message in str into this message instance :param str: byte array of serialized message, ``str`` """ try: end = 0 _x = self start = end end += 6 (_x.status, _x.progress,) = _get_struct_Hf().unpack(str[start:end]) start = end end += 4 (length,) = _struct_I.unpack(str[start:end]) start = end end += length if python3: self.long_description = str[start:end].decode('utf-8') else: self.long_description = str[start:end] return self except struct.error as e: raise genpy.DeserializationError(e) #most likely buffer underfill def serialize_numpy(self, buff, numpy): """ serialize message with numpy array types into buffer :param buff: buffer, ``StringIO`` :param numpy: numpy python module """ try: _x = self buff.write(_get_struct_Hf().pack(_x.status, _x.progress)) _x = self.long_description length = len(_x) if python3 or type(_x) == unicode: _x = _x.encode('utf-8') length = len(_x) buff.write(struct.pack('<I%ss'%length, length, _x)) except struct.error as se: self._check_types(struct.error("%s: '%s' when writing '%s'" % (type(se), str(se), str(locals().get('_x', self))))) except TypeError as te: self._check_types(ValueError("%s: '%s' when writing '%s'" % (type(te), str(te), str(locals().get('_x', self))))) def deserialize_numpy(self, str, numpy): """ unpack serialized message in str into this message instance using numpy for array types :param str: byte array of serialized message, ``str`` :param numpy: numpy python module """ try: end = 0 _x = self start = end end += 6 (_x.status, _x.progress,) = _get_struct_Hf().unpack(str[start:end]) start = end end += 4 (length,) = _struct_I.unpack(str[start:end]) start = end end += length if python3: self.long_description = str[start:end].decode('utf-8') else: self.long_description = str[start:end] return self except struct.error as e: raise genpy.DeserializationError(e) #most likely buffer underfill _struct_I = genpy.struct_I def _get_struct_I(): global _struct_I return _struct_I _struct_Hf = None def _get_struct_Hf(): global _struct_Hf if _struct_Hf is None: _struct_Hf = struct.Struct("<Hf") return _struct_Hf
[ "rymonyu@gmail.com" ]
rymonyu@gmail.com
5f7fbca5b07a1829d557e292fde130d12362f1c9
8a925a66777993cc0caf4085a9c309cbd59fad7d
/mdp/gridworld1d.py
4f142791692f8e6672a0d7b49fe98838f1c88406
[ "MIT" ]
permissive
yrlu/irl-imitation
96de0b435ef2ebd25f42694a90496f9e0dcbeb8a
c2f2f57d0d7cb98d4cca125579d68c411ed2cbe9
refs/heads/master
2023-05-22T16:20:35.389623
2022-07-05T07:23:56
2022-07-05T07:23:56
92,340,121
364
96
null
null
null
null
UTF-8
Python
false
false
5,163
py
# 1D Gridworld # # --- # @author Yiren Lu # @email luyiren [at] seas [dot] upenn [dot] edu # # MIT License import numpy as np from utils import * class GridWorld1D(object): """ 1D grid world environment (without terminal states) """ def __init__(self, rewards, terminals, move_rand=0.0): """ inputs: rewards 1d float array - contains rewards terminals a set of all the terminal states """ self.n_states = len(rewards) self.rewards = rewards self.terminals = terminals self.actions = [-1, 1] self.n_actions = len(self.actions) self.move_rand = move_rand def get_reward(self, state): return self.rewards[state] def get_transition_states_and_probs(self, state, action): """ inputs: state int - state action int - action returns a list of (state, probability) pair """ if action < 0 or action >= self.n_actions: # invalid input return [] if self.is_terminal(state): return [(state, 1.0)] if self.move_rand == 0: if state+self.actions[action] < 0 or state+self.actions[action] >= self.n_states: return [(state, 1.0)] return [(state+self.actions[action], 1.0)] else: mov_probs = np.zeros(3) mov_probs[1+self.actions[action]] += 1 - self.move_rand for i in range(3): mov_probs[i] += self.move_rand/3 if state == 0: mov_probs[1] += mov_probs[0] mov_probs[0] = 0 if state == self.n_states - 1: mov_probs[1] += mov_probs[2] mov_probs[2] = 0 res = [] for i in range(3): if mov_probs[i] != 0: res.append((state-1+i, mov_probs[i])) return res def is_terminal(self, state): if state in self.terminals: return True else: return False ############################################## # Stateful Functions For Model-Free Leanring # ############################################## def reset(self, start_pos): self._cur_state = start_pos def get_current_state(self): return self._cur_state def step(self, action): """ Step function for the agent to interact with gridworld inputs: action action taken by the agent returns current_state current state action input action next_state next_state reward reward on the next state is_done True/False - if the agent is already on the terminal states """ if self.is_terminal(self._cur_state): self._is_done = True return self._cur_state, action, self._cur_state, self.get_reward(self._cur_state), True st_prob = self.get_transition_states_and_probs(self._cur_state, action) rand_idx = np.random.choice(np.arange(0, len(st_prob)), p=[prob for st, prob in st_prob]) last_state = self._cur_state next_state = st_prob[rand_idx][0] reward = self.get_reward(last_state) self._cur_state = next_state return last_state, action, next_state, reward, False ####################### # Some util functions # ####################### def get_transition_mat(self): """ get transition dynamics of the gridworld return: P_a NxNxN_ACTIONS transition probabilities matrix - P_a[s0, s1, a] is the transition prob of landing at state s1 when taking action a at state s0 """ N_STATES = self.n_states N_ACTIONS = len(self.actions) P_a = np.zeros((N_STATES, N_STATES, N_ACTIONS)) for si in range(N_STATES): for a in range(N_ACTIONS): probs = self.get_transition_states_and_probs(si, a) for sj, prob in probs: # Prob of si to sj given action a P_a[si, sj, a] = prob return P_a def generate_demonstrations(self, policy, n_trajs=100, len_traj=20, rand_start=False, start_pos=0): """gatheres expert demonstrations inputs: gw Gridworld - the environment policy Nx1 matrix n_trajs int - number of trajectories to generate rand_start bool - randomly picking start position or not start_pos 2x1 list - set start position, default [0,0] returns: trajs a list of trajectories - each element in the list is a list of Steps representing an episode """ trajs = [] for i in range(n_trajs): if rand_start: # override start_pos start_pos = np.random.randint(0, self.n_states) episode = [] self.reset(start_pos) cur_state = start_pos cur_state, action, next_state, reward, is_done = self.step(int(policy[cur_state])) episode.append(Step(cur_state=cur_state, action=self.actions[action], next_state=next_state, reward=reward, done=is_done)) # while not is_done: for _ in range(1,len_traj): cur_state, action, next_state, reward, is_done = self.step(int(policy[cur_state])) episode.append(Step(cur_state=cur_state, action=self.actions[action], next_state=next_state, reward=reward, done=is_done)) if is_done: break trajs.append(episode) return trajs
[ "luyiren@seas.upenn.edu" ]
luyiren@seas.upenn.edu
e23f2edbeae333f052ce6c3b881ea6c9b93be27a
02f315c71fc653d9523607a6777ff8d89851ced2
/walky/client/common.py
175b55ad08cca8644d58ab2b39c1c48bb4966a39
[ "MIT" ]
permissive
amimoto/walky
8655d7b386a56e9f0a2e59dd2a7c45489070ed71
bf2971ccb86a8db58599aa7d8857f467a9f5fc9b
refs/heads/master
2021-01-10T22:06:32.612248
2015-07-02T02:46:44
2015-07-02T02:46:44
32,905,805
0
1
null
2015-04-06T21:23:10
2015-03-26T03:36:07
Python
UTF-8
Python
false
false
1,360
py
from __future__ import absolute_import import weakref import threading import asyncore import socket from walky.objects import * from walky.port import * from walky.engine import * class Client(object): engine = None settings = None connection = None port = None engine_class = Engine object_class = ObjectStub def __init__( self, **settings ): settings.setdefault('engine_class',self.engine_class) settings.setdefault('port_class',self.port_class) settings.setdefault('object_class',self.object_class) self.port = settings.get('port') self.settings = settings self.reset() def reset(self): if self.engine: self.engine.shutdown() self.engine = self.settings['engine_class']() def connect(self,*args,**kwargs): """ Start the engine and the asyncore """ self.engine.start() self.connection = self.engine.connection_new(*args,**kwargs) def run(self): pass def on_readline(self,line): try: pass except Exception as ex: pass def sendline(self,line): self.port().sendline(line) def object_get(self,reg_obj_id): return self.object_class(self.connection,reg_obj_id) def close(self): self.engine.shutdown()
[ "aki@zaber.com" ]
aki@zaber.com
4038aed1a3543a7584fd17318abc90b7b2ac4ea7
e3932a292695efbac57ae7adf5bd9c5554a88e39
/binding/python/setup.py
dd77ed90fa69813d54e55eef7604a48e1e43a0f5
[ "MIT" ]
permissive
makiolo/Calculate
1356e00628ef7f90423880de8190660667a407d1
81dcfcd7ce256abe1d5924329dd0ae8320fe749f
refs/heads/master
2021-07-03T07:02:44.673841
2017-09-20T22:51:18
2017-09-20T22:51:18
103,838,273
0
0
null
2017-09-17T15:10:14
2017-09-17T15:10:14
null
UTF-8
Python
false
false
1,611
py
# -*- coding: utf-8 -*- from __future__ import absolute_import import sys import os import os.path as path from platform import system from setuptools import setup from setuptools.dist import Distribution if sys.version_info < (2, 7): sys.exit('Only Python versions superior or equal than 2.7 supported') os.chdir(path.abspath(path.dirname(path.realpath(__file__)))) class BinaryDistribution(Distribution): def has_ext_modules(foo): return True def is_pure(self): return False library = 'calculate' with open('{}/__init__.py'.format(library), 'r') as file: metadata = {entry.split('=')[0].strip(' '): eval(entry.split('=')[-1]) for entry in file.read().split('\n') if '=' in entry} extensions = {'Linux': 'so', 'Darwin': 'dylib', 'Windows': 'dll'} extension = extensions.get(system(), '') library_name = 'lib' + library basedir = path.realpath(__file__).replace(path.basename(__file__), '') basedir = path.join(basedir, library) library_path = path.join(basedir, library_name + '.' + extension) if not path.lexists(library_path): raise EnvironmentError('Missing shared library') setup( name=library, distclass=BinaryDistribution, version=metadata['__version__'], license=metadata['__license__'], author=metadata['__author__'], author_email=metadata['__email__'], home_page=metadata['__site__'], description=metadata['__description__'], install_requires=[ 'cffi>=1.0.1' ], packages=[library], package_data={library: [library_name + '.' + extension]} )
[ "alorenzo.md@gmail.com" ]
alorenzo.md@gmail.com
3541be7710cfb9fd33d4c232c1bb19be30a746fe
bbc2135f8a49088686a4631dd12943c259d91c8e
/model.py
c49ad4fc039666bf55c579d55914424bcd3aec26
[]
no_license
christinababaya/hb-final-project
e264cb52b5dff889478e45610da2238982f60b37
67f466ba00377a34b40cb4ad7d53a4723745c012
refs/heads/main
2023-04-02T20:10:20.012907
2021-04-11T01:38:51
2021-04-11T01:38:51
null
0
0
null
null
null
null
UTF-8
Python
false
false
3,011
py
from datetime import datetime from flask_sqlalchemy import SQLAlchemy db = SQLAlchemy() class User(db.Model): """A user""" __tablename__ = 'users' user_id = db.Column(db.Integer, autoincrement=True, primary_key=True) user_name = db.Column(db.String, unique=True, nullable=False) password = db.Column(db.String, nullable=False) user_age = db.Column(db.Integer, nullable=True) user_weight = db.Column(db.Integer, nullable=True) user_zipcode = db.Column(db.Integer, nullable=True) def __repr__(self): return f'<User user_id={self.user_id} user_name={self.user_name}>' class Workout(db.Model): """User Workout""" __tablename__ = 'workouts' workout_id = db.Column(db.Integer, autoincrement=True, primary_key=True) user_id = db.Column(db.Integer, db.ForeignKey('users.user_id')) workout_date = db.Column(db.DateTime, nullable=True) user = db.relationship('User', backref='workouts') def __repr__(self): return f'<Workout workout_id={self.workout_id} workout_date={self.workout_date}>' class Workout_exercise(db.Model): """Exercise specific for workout""" __tablename__ = 'workout_exercises' ##TODO complete table columns/repr we_id = db.Column(db.Integer, autoincrement=True, primary_key=True) workout_id = db.Column(db.Integer, db.ForeignKey('workouts.workout_id')) exercise_id = db.Column(db.Integer, db.ForeignKey('exercises.exercise_id')) we_sets = db.Column(db.Integer, nullable=False) we_reps = db.Column(db.Integer, nullable=False) we_repunit = db.Column(db.String, nullable=True) we_weight = db.Column(db.Integer, nullable=True) we_weightunit = db.Column(db.String, nullable=True) we_equipment = db.Column(db.String, nullable=True) workout = db.relationship('Workout', backref='workout_exercises') exercise = db.relationship('Exercise', backref='workout_exercises') def __repr__(self): return f'<Workout_exercise we_id={self.we_id} we_sets={self.we_sets} we_reps={self.we_reps}>' class Exercise(db.Model): """Specific exercise details""" __tablename__ = 'exercises' exercise_id = db.Column(db.Integer, autoincrement=True, primary_key=True) exercise_name = db.Column(db.String, nullable=False) exercise_info = db.Column(db.Text, nullable=False) api_id = db.Column(db.Integer, nullable=True) def __repr__(self): return f'<Exercise exercise_id={self.exercise_id} exercise_name={self.exercise_name}>' def connect_to_db(flask_app, db_uri='postgresql:///workouts', echo=True): flask_app.config['SQLALCHEMY_DATABASE_URI'] = db_uri flask_app.config['SQLALCHEMY_ECHO'] = echo flask_app.config['SQLALCHEMY_TRACK_MODIFICATIONS'] = False db.app = flask_app db.init_app(flask_app) print('Connected to the db!') if __name__ == '__main__': from server import app connect_to_db(app) # connect_to_db(app, echo=False)
[ "jessicachap223@gmail.com" ]
jessicachap223@gmail.com
6005b320155e884dcb8bc9c7994fc6807bcf4c35
aa7c6a9276a859f75b3c5181a92f71d7c19122a5
/zvt/domain/quotes/stock/stock_1m_kdata.py
f1378710d90001bd962e69feaa23a05bf88f493e
[ "MIT" ]
permissive
Pengyuyan2/zvt
deef9c5e5bd91c65728ad9bac8c79499707519ee
9f9c77efcd34c04aaf11b12da0cf483cbe55e297
refs/heads/master
2023-07-12T16:55:15.040579
2021-08-22T09:41:33
2021-08-22T09:55:49
null
0
0
null
null
null
null
UTF-8
Python
false
false
525
py
# -*- coding: utf-8 -*- # this file is generated by gen_kdata_schema function, dont't change it from sqlalchemy.orm import declarative_base from zvt.contract.register import register_schema from zvt.domain.quotes import StockKdataCommon KdataBase = declarative_base() class Stock1mKdata(KdataBase, StockKdataCommon): __tablename__ = 'stock_1m_kdata' register_schema(providers=['joinquant'], db_name='stock_1m_kdata', schema_base=KdataBase, entity_type='stock') # the __all__ is generated __all__ = ['Stock1mKdata']
[ "5533061@qq.com" ]
5533061@qq.com
faac087b45aa6ee29a57ab65290e48b37be927c5
03143da0cf99ea92d372feca0954d22d64858d86
/Approach 4/EMNIST/EMNIST-4/utils/mnistutil.py
a7b7435b0cbd75973bfc88d7ab188e5b1eaa0596
[]
no_license
rangeetpan/moduleDecompose
ff31732a878e5f9d5e79c3bba9fd9f051c7a5d16
508c2a87e00c4e1b616bc29515fc1de2aba55c4e
refs/heads/master
2021-02-20T08:49:55.157544
2020-03-06T04:50:43
2020-03-06T04:50:43
245,332,163
0
0
null
null
null
null
UTF-8
Python
false
false
14,184
py
''' Created on Feb 8, 2019 @author: mislam ''' from keras.datasets import mnist from skimage.transform import resize import numpy as np from keras import backend as K import keras import tensorflow as tf from keras.models import Sequential from keras.layers import Dense, Dropout, Flatten from keras.layers import Conv2D, MaxPooling2D class MNISTUitl: def __init__(self): self.name = None def load(self,f): return np.load(f)['arr_0'] def getdata(self,a,b,img_rows = 28, img_cols = 28): # the data, split between train and test sets (x_train, y_train), (x_test, y_test) = mnist.load_data() x_zo = [] y_zo = [] for i in range(len(y_train)): if y_train[i] == a or y_train[i] == b: A = resize(x_train[i], (img_rows, img_cols),mode='constant') Ay = y_train[i]#resize(y_train[i], (img_rows, img_cols)) x_zo.append(A) y_zo.append(Ay) xt_zo = [] yt_zo = [] for i in range(len(y_test)): if y_test[i] == a or y_test[i] == b: A = resize(x_test[i], (img_rows, img_cols),mode='constant') Ay = y_test[i]#resize(y_train[i], (img_rows, img_cols)) xt_zo.append(A) yt_zo.append(Ay) x_zo = np.array(x_zo) y_zo = np.array(y_zo) xt_zo = np.array(xt_zo) yt_zo = np.array(yt_zo) return x_zo, y_zo, xt_zo, yt_zo def getdata2(self,a,b,img_rows = 28, img_cols = 28): # the data, split between train and test sets x_train = self.load('emnist-train-imgs.npz') x_test = self.load('emnist-test-imgs.npz') y_train = self.load('emnist-train-labels.npz') for i in range(0,len(y_train)): y_train[i]=y_train[i]-1 y_test = self.load('emnist-test-labels.npz') for i in range(0,len(y_test)): y_test[i]=y_test[i]-1 x_zo = [] y_zo = [] for i in range(len(y_train)): if y_train[i] in [0,1,2,3,4,5,6,7,8,9]: A = resize(x_train[i], (img_rows, img_cols),mode='constant') Ay = y_train[i]#resize(y_train[i], (img_rows, img_cols)) x_zo.append(A) y_zo.append(Ay) xt_zo = [] yt_zo = [] for i in range(len(y_test)): if y_test[i] in [0,1,2,3,4,5,6,7,8,9]: A = resize(x_test[i], (img_rows, img_cols),mode='constant') Ay = y_test[i]#resize(y_train[i], (img_rows, img_cols)) xt_zo.append(A) yt_zo.append(Ay) x_zo = np.array(x_zo) y_zo = np.array(y_zo) xt_zo = np.array(xt_zo) yt_zo = np.array(yt_zo) return x_zo, y_zo, xt_zo, yt_zo def train(self,x_zo,y_zo,xt_zo,yt_zo,img_rows = 28, img_cols = 28,numclass = 2): if K.image_data_format() == 'channels_first': x_zo = x_zo.reshape(x_zo.shape[0], 1, img_rows, img_cols) xt_zo = xt_zo.reshape(xt_zo.shape[0], 1, img_rows, img_cols) input_shape = (1, img_rows, img_cols) else: x_zo = x_zo.reshape(x_zo.shape[0], img_rows, img_cols, 1) xt_zo = xt_zo.reshape(xt_zo.shape[0], img_rows, img_cols, 1) input_shape = (img_rows, img_cols, 1) x_train = x_zo.astype('float32') x_test = xt_zo.astype('float32') x_train /= 255 x_test /= 255 print('x_train shape:', x_train.shape) print(x_zo.shape,x_train.shape[0], 'train samples', y_zo.shape) print(x_test.shape[0], 'test samples') y_train = y_zo#keras.utils.to_categorical(y_zo, numclass ) y_test = yt_zo#keras.utils.to_categorical(yt_zo, numclass) print(y_zo.shape,y_train.shape) nm = keras.Sequential([ keras.layers.Flatten(input_shape=(img_rows, img_cols,1), name = "Input"), keras.layers.Dense(7, activation=tf.nn.relu ,name = "H"), keras.layers.Dense(numclass, activation=tf.nn.softmax, name = "output") ]) nm.compile(optimizer='adam', loss='sparse_categorical_crossentropy', metrics=['accuracy']) nm.fit(x_train, y_train, epochs=10) return nm, x_test, y_test def train2(self,x_zo,y_zo,xt_zo,yt_zo,img_rows = 28, img_cols = 28,numclass = 10,ep = 20): if K.image_data_format() == 'channels_first': x_zo = x_zo.reshape(x_zo.shape[0], 1, img_rows, img_cols) xt_zo = xt_zo.reshape(xt_zo.shape[0], 1, img_rows, img_cols) input_shape = (1, img_rows, img_cols) else: x_zo = x_zo.reshape(x_zo.shape[0], img_rows, img_cols, 1) xt_zo = xt_zo.reshape(xt_zo.shape[0], img_rows, img_cols, 1) input_shape = (img_rows, img_cols, 1) x_train = x_zo.astype('float32') x_test = xt_zo.astype('float32') x_train /= 255 x_test /= 255 print('x_train shape:', x_train.shape) print(x_zo.shape,x_train.shape[0], 'train samples', y_zo.shape) print(x_test.shape[0], 'test samples') y_train = y_zo #keras.utils.to_categorical(y_zo, numclass ) y_test = yt_zo #keras.utils.to_categorical(yt_zo, numclass) print(y_zo.shape,y_train.shape) nm = keras.Sequential([ keras.layers.Flatten(input_shape=(img_rows, img_cols,1), name = "Input"), keras.layers.Dense(49, activation=tf.nn.relu ,name = "H"), keras.layers.Dense(numclass, activation=tf.nn.softmax, name = "output") ]) nm.compile(optimizer='adam', loss='sparse_categorical_crossentropy', metrics=['accuracy']) print(nm.summary()) nm.fit(x_train, y_train, epochs=ep) return nm, x_test, y_test def trainDense2(self,x_zo,y_zo,xt_zo,yt_zo,img_rows = 28, img_cols = 28,numclass = 10,ep = 20): if K.image_data_format() == 'channels_first': x_zo = x_zo.reshape(x_zo.shape[0], 1, img_rows, img_cols) xt_zo = xt_zo.reshape(xt_zo.shape[0], 1, img_rows, img_cols) input_shape = (1, img_rows, img_cols) else: x_zo = x_zo.reshape(x_zo.shape[0], img_rows, img_cols, 1) xt_zo = xt_zo.reshape(xt_zo.shape[0], img_rows, img_cols, 1) input_shape = (img_rows, img_cols, 1) x_train = x_zo.astype('float32') x_test = xt_zo.astype('float32') x_train /= 255 x_test /= 255 print('x_train shape:', x_train.shape) print(x_zo.shape,x_train.shape[0], 'train samples', y_zo.shape) print(x_test.shape[0], 'test samples') y_train = y_zo #keras.utils.to_categorical(y_zo, numclass ) y_test = yt_zo #keras.utils.to_categorical(yt_zo, numclass) print(y_zo.shape,y_train.shape) nm = keras.Sequential([ keras.layers.Flatten(input_shape=(img_rows, img_cols,1), name = "Input"), keras.layers.Dense(49, activation=tf.nn.relu ,name = "H1"), keras.layers.Dense(49, activation=tf.nn.relu ,name = "H2"), keras.layers.Dense(numclass, activation=tf.nn.softmax, name = "output") ]) nm.compile(optimizer='adam', loss='sparse_categorical_crossentropy', metrics=['accuracy']) print(nm.summary()) nm.fit(x_train, y_train, epochs=ep) return nm, x_test, y_test def trainDense4(self,x_zo,y_zo,xt_zo,yt_zo,img_rows = 28, img_cols = 28,numclass = 10,ep = 20): if K.image_data_format() == 'channels_first': x_zo = x_zo.reshape(x_zo.shape[0], 1, img_rows, img_cols) xt_zo = xt_zo.reshape(xt_zo.shape[0], 1, img_rows, img_cols) input_shape = (1, img_rows, img_cols) else: x_zo = x_zo.reshape(x_zo.shape[0], img_rows, img_cols, 1) xt_zo = xt_zo.reshape(xt_zo.shape[0], img_rows, img_cols, 1) input_shape = (img_rows, img_cols, 1) x_train = x_zo.astype('float32') x_test = xt_zo.astype('float32') x_train /= 255 x_test /= 255 print('x_train shape:', x_train.shape) print(x_zo.shape,x_train.shape[0], 'train samples', y_zo.shape) print(x_test.shape[0], 'test samples') y_train = y_zo #keras.utils.to_categorical(y_zo, numclass ) y_test = yt_zo #keras.utils.to_categorical(yt_zo, numclass) print(y_zo.shape,y_train.shape) nm = keras.Sequential([ keras.layers.Flatten(input_shape=(img_rows, img_cols,1), name = "Input"), keras.layers.Dense(49, activation=tf.nn.relu ,name = "H1"), keras.layers.Dense(49, activation=tf.nn.relu ,name = "H2"), keras.layers.Dense(49, activation=tf.nn.relu ,name = "H3"), keras.layers.Dense(49, activation=tf.nn.relu ,name = "H4"), keras.layers.Dense(numclass, activation=tf.nn.softmax, name = "output") ]) nm.compile(optimizer='adam', loss='sparse_categorical_crossentropy', metrics=['accuracy']) print(nm.summary()) nm.fit(x_train, y_train, epochs=ep) return nm, x_test, y_test def trainDense6(self,x_zo,y_zo,xt_zo,yt_zo,img_rows = 28, img_cols = 28,numclass = 10,ep = 20): if K.image_data_format() == 'channels_first': x_zo = x_zo.reshape(x_zo.shape[0], 1, img_rows, img_cols) xt_zo = xt_zo.reshape(xt_zo.shape[0], 1, img_rows, img_cols) input_shape = (1, img_rows, img_cols) else: x_zo = x_zo.reshape(x_zo.shape[0], img_rows, img_cols, 1) xt_zo = xt_zo.reshape(xt_zo.shape[0], img_rows, img_cols, 1) input_shape = (img_rows, img_cols, 1) x_train = x_zo.astype('float32') x_test = xt_zo.astype('float32') x_train /= 255 x_test /= 255 print('x_train shape:', x_train.shape) print(x_zo.shape,x_train.shape[0], 'train samples', y_zo.shape) print(x_test.shape[0], 'test samples') y_train = y_zo #keras.utils.to_categorical(y_zo, numclass ) y_test = yt_zo #keras.utils.to_categorical(yt_zo, numclass) print(y_zo.shape,y_train.shape) nm = keras.Sequential([ keras.layers.Flatten(input_shape=(img_rows, img_cols,1), name = "Input"), keras.layers.Dense(49, activation=tf.nn.relu ,name = "H1"), keras.layers.Dense(49, activation=tf.nn.relu ,name = "H2"), keras.layers.Dense(49, activation=tf.nn.relu ,name = "H3"), keras.layers.Dense(49, activation=tf.nn.relu ,name = "H4"), keras.layers.Dense(49, activation=tf.nn.relu ,name = "H5"), keras.layers.Dense(49, activation=tf.nn.relu ,name = "H6"), keras.layers.Dense(numclass, activation=tf.nn.softmax, name = "output") ]) nm.compile(optimizer='adam', loss='sparse_categorical_crossentropy', metrics=['accuracy']) print(nm.summary()) nm.fit(x_train, y_train, epochs=ep) return nm, x_test, y_test def trainData(self,x_zo,y_zo,xt_zo,yt_zo,img_rows = 28, img_cols = 28,numclass = 10,ep = 20): if K.image_data_format() == 'channels_first': x_zo = x_zo.reshape(x_zo.shape[0], 1, img_rows, img_cols) xt_zo = xt_zo.reshape(xt_zo.shape[0], 1, img_rows, img_cols) input_shape = (1, img_rows, img_cols) else: x_zo = x_zo.reshape(x_zo.shape[0], img_rows, img_cols, 1) xt_zo = xt_zo.reshape(xt_zo.shape[0], img_rows, img_cols, 1) input_shape = (img_rows, img_cols, 1) x_train = x_zo.astype('float32') x_test = xt_zo.astype('float32') x_train /= 255 x_test /= 255 print('x_train shape:', x_train.shape) print(x_zo.shape,x_train.shape[0], 'train samples', y_zo.shape) print(x_test.shape[0], 'test samples') y_train = y_zo #keras.utils.to_categorical(y_zo, numclass ) y_test = yt_zo #keras.utils.to_categorical(yt_zo, numclass) print(y_zo.shape,y_train.shape) # nm = keras.Sequential([ # keras.layers.Flatten(input_shape=(img_rows, img_cols,1), name = "Input"), # keras.layers.Dense(49, activation=tf.nn.relu ,name = "H"), # keras.layers.Dense(numclass, activation=tf.nn.softmax, name = "output") # ]) # nm.compile(optimizer='adam', # loss='sparse_categorical_crossentropy', # metrics=['accuracy']) # print(nm.summary()) # nm.fit(x_train, y_train, epochs=ep) return x_test, y_test,x_train, y_train def train3(self,x_zo,y_zo,xt_zo,yt_zo,img_rows = 28, img_cols = 28,numclass = 10,ep = 20): input_shape = (img_rows,img_cols,1) x_zo = x_zo.reshape(x_zo.shape[0], img_rows, img_cols, 1) xt_zo = xt_zo.reshape(xt_zo.shape[0], img_rows, img_cols, 1) x_train = x_zo.astype('float32') x_test = xt_zo.astype('float32') x_train /= 255 x_test /= 255 y_train = keras.utils.to_categorical(y_zo, numclass ) y_test = keras.utils.to_categorical(yt_zo, numclass) num_classes = 10 model = Sequential() model.add(Conv2D(32, kernel_size=(3, 3), activation='relu', input_shape=input_shape)) model.add(Conv2D(64, (3, 3), activation='relu')) model.add(MaxPooling2D(pool_size=(2, 2))) #model.add(Dropout(0.25)) model.add(Flatten()) model.add(Dense(128, activation='relu')) #model.add(Dropout(0.5)) model.add(Dense(num_classes, activation='softmax')) model.compile(loss=keras.losses.categorical_crossentropy, optimizer=keras.optimizers.Adadelta(), metrics=['accuracy']) model.fit(x_train, y_train, epochs=ep) return model, x_test, y_test
[ "rangeet@iastate.edu" ]
rangeet@iastate.edu
86118937a3c5da7d22eb06c3ed34e49f7cfa2f11
163bbb4e0920dedd5941e3edfb2d8706ba75627d
/Code/CodeRecords/2711/47774/305852.py
a8c9c3680c535404ce3caf423c50014ec1f95130
[]
no_license
AdamZhouSE/pythonHomework
a25c120b03a158d60aaa9fdc5fb203b1bb377a19
ffc5606817a666aa6241cfab27364326f5c066ff
refs/heads/master
2022-11-24T08:05:22.122011
2020-07-28T16:21:24
2020-07-28T16:21:24
259,576,640
2
1
null
null
null
null
UTF-8
Python
false
false
1,535
py
def isSimilar(s1, s2): diff, l = 0, len(s1) for i in range(l): if (s1[i] != s2[i]): diff += 1 if (diff > 2): return False return True def find(f, x): return f[x] if x == f[x] else find(f, f[x]) def merge(f, x, y): rx = find(f, f[x]) ry = find(f, f[y]) f[ry] = rx def solve(A): A = list(set(A)) l,w = len(A), len(A[0]) res = 0 f = [i for i in range(l)] if l <= w*w: for i in range(l): for j in range(i + 1, l): if (find(f, i) != find(f,j)): isS = isSimilar(A[i], A[j]) if (isS): merge(f, i, j) else: dict = {} for i in range(l): if (A[i] in dict): dict[A[i]].add(i) else: dict[A[i]] = {i} word = list(A[i]) for i0 in range(w): for j0 in range(i0+1, w): if (word[i0] != word[j0]): word[i0],word[j0] = word[j0],word[i0] neighbor = ''.join(word) if (neighbor in dict): dict[neighbor].add(i) else: dict[neighbor] = {i} word[i0],word[j0] = word[j0],word[i0] for i in range(l): for j in dict[A[i]]: merge(f,i,j) for i in range(l): if (i == f[i]): res += 1 return res s=eval(input()) print(solve(s))
[ "1069583789@qq.com" ]
1069583789@qq.com
297f0900fb895a2b9d47e22dba06bc0e49e89095
fd4b63792f0aa44acba8e656f2c71f6e4dd61377
/web/changelly.py
48c2dafe3e756c46132bdb2f0567c64e46bf7c80
[]
no_license
procaff3inator/changepark
03869a4ebe7c789bc516672ebb1fa6475cb22550
fc55a4492c4084f92031de26ae87050d045d9e94
refs/heads/master
2020-03-19T01:07:07.462694
2018-04-29T05:35:10
2018-04-29T05:35:10
135,524,080
0
0
null
null
null
null
UTF-8
Python
false
false
5,879
py
import copy import hashlib import hmac import json import requests from functools import wraps from uuid import uuid4 def api_method(f): """Decorate functions that are API methods. :param f: A function/method to be wrapped """ @wraps(f) def d(*args, **kwargs): payload = f(*args, **kwargs) print(payload) return requests.post( payload['url'], headers=payload['headers'], data=payload['payload'] ) return d class Changelly(object): def __init__(self, url, key, secret): self.url = url self.key = key self.secret = secret def _prepare_payload(self, params): json_pl = {'jsonrpc': '2.0', 'id': str(uuid4()) } json_pl.update(params) serialized_data = json.dumps(json_pl) sign = hmac.new( self.secret.encode('utf-8'), serialized_data.encode('utf-8'), hashlib.sha512 ).hexdigest() headers = { 'api-key': self.key, 'sign': sign, 'Content-type': 'application/json', } return {'url': self.url, 'payload': serialized_data, 'headers': headers} @api_method def get_currencies(self): """Fetch a list of supported currencies from the server.""" params = { 'method': 'getCurrencies', 'params': [], } return self._prepare_payload(params) @api_method def get_min_amount(self, fromcurr, tocurr): """Get min amount that can be exchanged between two different currencies. :param fromcurr: Currency to change from :param tocurr: Currency to change to """ return self._prepare_payload({ 'method': 'getMinAmount', 'params': { 'from': fromcurr, 'to': tocurr, }, }) @api_method def get_exchange_amount(self, fromcurr, tocurr, amount): """Get the exchange amount between two different currencies. :param fromcurr: Currency to change from :param tocurr: Currency to change to :param amount: Amount to be exchaned """ return self._prepare_payload({ 'method': 'getExchangeAmount', 'params': { 'from': fromcurr, 'to': tocurr, 'amount': amount, }, }) @api_method def get_status(self, transaction_id): """Get the status of a transaction. :param transaction_id: Id of the transaction """ return self._prepare_payload({ "method": "getStatus", "params": { "id": transaction_id }, }) @api_method def create_transaction(self, fromcurr, tocurr, address, amount, **kwargs): """Create a transation to convert from one currency to another. :param fromcurr: From Currency :param tocurrency: To Currency :param address: Address to send the amount to :param amount: Amount to send :param extra_id: Required for XRP, STEEM/SBD, XLM, DCT, XEM :param refund_address: Optional param, enables refund :param refund_extraid: Required for XRP, STEEM/SBD, XLM, DCT, XEM """ # raise NotImplementedError("WIP") params = { 'from': fromcurr, 'to': tocurr, 'address': address, 'amount': amount, # 'refundAddress': address, # for now let's have no refund! } if 'extraid' in kwargs: params['extraId'] = kwargs['extraid'] if 'refundextraid' in kwargs: params['refundExtraId'] = kwargs['refundextraid'] return self._prepare_payload({ 'method': 'createTransaction', 'params': params, }) @api_method def get_transactions(self, **kwargs): """Get a list of transactions according to the filter params passed. :param currency: Currency to filter from :param address: Address to filter by :param extraId: Extra id needed by some currencies :param limit: Result limit :param offset: Result offset """ return self._prepare_payload({ 'method': 'getTransactions', 'params': {}, }) @api_method def find_transactions(self, **kwargs): """Filter transations by the given params. """ return self._prepare_payload({ 'method': 'getTransactions', 'params': kwargs, }) if __name__ == '__main__': from config import api_creds c = Changelly( api_creds['url'], api_creds['key'], api_creds['secret'] ) # print("Foo: {}".format(c.get_currencies().text)) # print("Foo: {}".format(c.get_min_amount('btc', 'ltc').text)) # print("Foo: {}".format(c.get_exchange_amount("btc", "eth", "100").text)) # print("Foo: {}".format(c.get_status('f6e0c6a5bb05').text)) # print("Foo: {}".format(c.get_status('4bb51c2cca9b').text)) # print("Foo: {}".format(c.get_transactions().text)) # def create_transaction(self, fromcurr, tocurr, address, amount, **kwargs): address = 'LhNXzB2AWQ1Q2ArLPwefvrwY9cCENtDz47' print("Foo: {}".format(c.create_transaction('btc', 'ltc', address, '0.00359353', extraid=None).text))
[ "procaff3inator@gmail.com" ]
procaff3inator@gmail.com
3b02deb956c2887c303ede1abb8311eb09a6c91f
8cdcf99f1e63c380967a294653f7dc0e1d8e8291
/tests/test_reverse_polish.py
443cf4e27e21f32aa64c196dd4147db3b9d3a0ec
[]
no_license
dIgor93/Sandbox
1092f8e27dd57a4eb04676f7e445165f20df9ee7
a7ffec83ee4036a6394a8f95e30923523e1d05f8
refs/heads/master
2022-02-23T19:46:28.589014
2019-10-27T13:31:06
2019-10-27T13:31:06
null
0
0
null
null
null
null
UTF-8
Python
false
false
1,053
py
from unittest import TestCase from polish import reverse_polish class TestReverse_polish(TestCase): def test_reverse_polish_0(self): res = reverse_polish(-1, [2, 3, '+']) print(res) assert res == 5 def test_reverse_polish_1(self): res = reverse_polish(-1, [2, 2, 2, '+', '*']) print(res) assert res == 8 def test_reverse_polish_2(self): res = reverse_polish(-1, [3, 2, 1, '+', '*']) print(res) assert res == 9 def test_reverse_polish_3(self): res = reverse_polish(-1, [3, 2, '+', 1, 3, '+', '*']) print(res) assert res == 20 def test_reverse_polish_4(self): res = reverse_polish(-1, [10, 5, '<', 5, 14, '?']) print(res) assert res == 14 def test_reverse_polish_5(self): res = reverse_polish(-1, [5, 10, '<', 5, 14, '?']) print(res) assert res == 5 def test_reverse_polish_6(self): res = reverse_polish(-1, [5, 14, '-']) print(res) assert res == -9
[ "wildig1014@yandex.ru" ]
wildig1014@yandex.ru
20bcbefc778347fd1c292368763549aed1dee9e5
d51c769deeb16ea0e2f17275e269f6a9e59b1d28
/WOS + MDM/mapeo_colores.py
309e292a8fda3c29c516dfd44a5ee3badfa6cd3a
[]
no_license
MDG99/Rubiks-Cube-Solver
16cd97330987328f10d9e612e5b865acb6c13275
f17ac944e352478e5b50d9aef6f034476f8c7663
refs/heads/main
2023-02-25T20:39:05.111517
2021-02-05T11:22:39
2021-02-05T11:22:39
330,028,938
0
0
null
null
null
null
UTF-8
Python
false
false
1,782
py
import numpy as np M = [[4, 52, 81], # Azul ya [198, 51, 3], # Roj0 *** [39, 169, 43], # Verde pálido [228, 109, 6], # Naranja ya [153, 163, 30], # Amarillo *** [147, 165, 120], # Blanco [27, 55, 17] ] # Fondo def mapeo(img): w = np.shape(img)[0] h = np.shape(img)[1] contador = np.zeros(7) for i in range(w): for j in range(h): if img[i, j, 2] == M[0][0] and img[i, j, 1] == M[0][1] and img[i, j, 0] == M[0][2]: contador[0] = contador[0] + 1 if img[i, j, 2] == M[1][0] and img[i, j, 1] == M[1][1] and img[i, j, 0] == M[1][2]: contador[1] = contador[1] + 1 if img[i, j, 2] == M[2][0] and img[i, j, 1] == M[2][1] and img[i, j, 0] == M[2][2]: contador[2] = contador[2] + 1 if img[i, j, 2] == M[3][0] and img[i, j, 1] == M[3][1] and img[i, j, 0] == M[3][2]: contador[3] = contador[3] + 1 if img[i, j, 2] == M[4][0] and img[i, j, 1] == M[4][1] and img[i, j, 0] == M[4][2]: contador[4] = contador[4] + 1 if img[i, j, 2] == M[5][0] and img[i, j, 1] == M[5][1] and img[i, j, 0] == M[5][2]: contador[5] = contador[5] + 1 if img[i, j, 2] == M[6][0] and img[i, j, 1] == M[6][1] and img[i, j, 0] == M[6][2]: contador[6] = 1 moda = max(contador) indice = 0 for i in range(7): if contador[i] == moda: indice = i if indice == 0: return 'B' if indice == 1: return 'R' if indice == 2: return 'G' if indice == 3: return 'O' if indice == 4: return 'Y' if indice == 5: return 'W' if indice == 6: return 'k'
[ "ignacioisaac30@gmail.com" ]
ignacioisaac30@gmail.com
7908b1e1cfb79c9a9aed81d50ee32663279e5be3
81e14734e111a91a37888dfcc85e6d0f30dce56c
/closet palindrome number/find_palindrome.py
72b8108fbf4ce69b3de58da9c3e0f7bfa0ae2217
[]
no_license
arunvemana/pythontasks
b96ad4325c40df9c844403717e174ef2d84f7889
2d6f76bb48efcc4e718efaf88a50ec33b1a0c652
refs/heads/master
2023-05-24T17:00:27.912830
2022-11-30T12:06:32
2022-11-30T12:06:32
193,498,172
3
0
null
2023-05-23T05:16:05
2019-06-24T12:09:23
Python
UTF-8
Python
false
false
452
py
def closest_palindrome(num): num = str(num) length_number = len(num) left_shift_index = length_number/2 # print(str(num)[:left_shift_index]) if (left_shift_index % 2) == 0: return int(num[:left_shift_index]+num[:left_shift_index][::-1]) else: return int(num[:left_shift_index+1]+num[:left_shift_index][::-1]) if __name__ == '__main__': for number in [123, 1222]: print(closest_palindrome(number))
[ "avemana@loginsoft.com" ]
avemana@loginsoft.com
5245bc11bfacf34d092a6630efd1e6ec7b5948a9
32809f6f425bf5665fc19de2bc929bacc3eeb469
/src/1096-Brace-Expansion-II/1096.py
78067156acba02fd1f032327859403cee51255d5
[]
no_license
luliyucoordinate/Leetcode
9f6bf01f79aa680e2dff11e73e4d10993467f113
bcc04d49969654cb44f79218a7ef2fd5c1e5449a
refs/heads/master
2023-05-25T04:58:45.046772
2023-05-24T11:57:20
2023-05-24T11:57:20
132,753,892
1,575
569
null
2023-05-24T11:57:22
2018-05-09T12:30:59
C++
UTF-8
Python
false
false
723
py
import itertools class Solution: def braceExpansionII(self, expression): groups = [[]] level = 0 for i, c in enumerate(expression): if c == '{': if level == 0: start = i+1 level += 1 elif c == '}': level -= 1 if level == 0: groups[-1].append(self.braceExpansionII(expression[start:i])) elif level == 0: if c == ",": groups.append([]) else: groups[-1].append([c]) return sorted(set().union(*[set(map(''.join, itertools.product(*group))) for group in groups]))
[ "luliyucoordinate@outlook.com" ]
luliyucoordinate@outlook.com
658c5b53a50e59b89e284ab9c792c46081f0daeb
d20626ef3b9ae6b9a702d67333209d678e27105d
/robot_booking/freebusy/FreeBusyEvent_Builder.py
7690f3f737a404f94c5917cfcd0c9e42d122705a
[]
no_license
geleazar1000111/bebop
e3d95b641b036847e50913690b07d0eb12958693
435f134248079f369a8f0004a0e07b8d73e0ce21
refs/heads/master
2022-11-24T01:46:37.334014
2020-07-26T17:54:27
2020-07-26T17:54:27
263,472,151
0
0
null
null
null
null
UTF-8
Python
false
false
3,707
py
"""This class builds free time slots based on the specified time range and events that are booked in that time range. First, days are generated as nested dictionaries that are stored in the days in range attribute. The workday is defined as the value for the 'range' key. 'booked' is another key that gets filled in later, if there are any booked events in that specified time range. Once any booked events are appended to the 'booked' list, free time slots are then created. If there are no booked events, then the whole day is created as a free event.""" from datetime import datetime, timedelta import re class FreeBusyEventBuilder: def __init__(self, robot_id): self.robot_id = robot_id self.is_available = 0 self.days_in_range = {} self.free_events = [] def construct_free_events(self): for day in self.days_in_range: curstart = self.days_in_range[day]['range']['start'] if self.days_in_range[day]['booked']: for event in self.days_in_range[day]['booked']: curend = event['start'] self.make_free_event(curstart, curend) curstart = event['end'] curend = self.days_in_range[day]['range']['end'] self.make_free_event(curstart, curend) else: self.make_free_event(self.days_in_range[day]['range']['start'], self.days_in_range[day]['range']['end']) def initialize_free_events(self, booked, min_year, min_month, min_day, max_year, max_month, max_day): curstart = datetime(min_year, min_month, min_day, hour=9) curend = datetime(max_year, max_month, max_day, hour=17) self.fill_days(curstart, curend) for event in booked: event_start = convert_google_datetime(event['start']) event_end = convert_google_datetime(event['end']) self.days_in_range[event_start.date()]['booked'].append({'start': event_start, 'end': event_end}) self.construct_free_events() return self.free_events def fill_days(self, start_range, end_range): delta = end_range - start_range for i in range(delta.days + 1): date = start_range + timedelta(days=i) date_start = date.replace(hour=9) date_end = date.replace(hour=17) self.days_in_range[date.date()] = {'range': {'start': date_start, 'end': date_end, 'duration': 8}, 'booked': []} def calculate_gap(self, prev_event, next_event): diff = next_event - prev_event diff_hours = diff.total_seconds() // 3600 return diff_hours def make_free_event(self, start, end): if self.calculate_gap(start, end) >= 0.5: self.free_events.append({'start': start, 'end': end, 'duration': self.calculate_gap(start, end)}) # self.free_events[start] = {'start': start, 'end': end, 'duration': self.calculate_gap(start, end)} def check_if_free(self, booked): if self.free_events: self.is_available = 1 elif booked and not self.free_events: self.is_available = -1 # else: # self.is_available = 2 """Helper Functions""" def is_single_day_event(datetime_str): regex_obj = re.compile('.*-.*-.*T.*:.*:.*') if regex_obj.match(datetime_str): return True return False def convert_google_datetime(datetime_str): if not is_single_day_event(datetime_str[:datetime_str.rfind("-")]): return datetime.strptime(datetime_str, '%Y-%m-%d') else: return datetime.strptime(datetime_str[:datetime_str.rfind("-")], '%Y-%m-%dT%H:%M:%S%f')
[ "geraldine@osaro.com" ]
geraldine@osaro.com
70ba1e98be58094fefb1b3ad1735ac0b7a6c9499
80ce9b73a0447c13838de64e89fa95f24852f95a
/get_textf_from_lsi.py
fdf0f4ecb0ceed992bdde759286d4cb904a8443c
[]
no_license
thomason-jesse/synpol
17458cd05484e783d3fe84d6c951b27728286b91
2edd89d88fbab50bcaa7d77a9e83e3cfe4627a14
refs/heads/master
2021-01-20T10:00:36.555994
2017-10-20T16:39:34
2017-10-20T16:39:34
90,314,269
0
0
null
null
null
null
UTF-8
Python
false
false
5,484
py
#!/usr/bin/env python __author__ = 'jesse' ''' takes a wnid graph, a set of wnid -> text observation maps, and a serialized lsi model and calculates the textual features for the wnid observations given the observed text and outputs a map from wnid -> text features ''' import argparse import pickle import os import time def main(): # read infiles print "reading in urls, observations, lsi model, and dictionary..." f = open(FLAGS_wnid_urls, 'rb') wnid_urls = pickle.load(f) wnids = wnid_urls.keys() f.close() print "... read graph" if FLAGS_text_obs_unified > 0: f = open(FLAGS_text_obs_infile, 'rb') wnid_text = pickle.load(f) f.close() print "... read text observations" print "... done" # calculate lsi textual features from text corpus observations print "launching jobs to calculate lsi textual features from text observations..." remaining_wnid_jobs = [] for wnid_idx in range(0, len(wnids)): wnid = wnids[wnid_idx] launch_job = False if FLAGS_text_obs_unified > 0: wnid_text_obs = FLAGS_text_obs_infile if wnid in wnid_text and len(wnid_text[wnid]) > 0: launch_job = True else: wnid_text_obs = str(wnid_idx) + "_" + FLAGS_text_obs_infile try: with open(wnid_text_obs, 'rb') as pf: _ = pickle.load(pf) launch_job = True except (IOError, EOFError): print "... WARNING: missing pickle for wnid " + str(wnid_idx) + "; cannot get features for it" if launch_job: outf = str(wnid_idx) + "_lsi_temp.pickle" if FLAGS_text_obs_unified > 0 else str(wnid_idx) + "_" + FLAGS_outfile cmd = ("condorify_gpu_email python get_textf_from_lsi_for_wnid.py " + "--target_wnid " + wnid + " " + "--text_obs_infile " + wnid_text_obs + " " + "--lsi_dictionary " + FLAGS_lsi_dictionary + " " + "--lsi_dictionary " + FLAGS_lsi_dictionary + " " + "--tfidf_model " + FLAGS_tfidf_model + " " + "--lsi_model " + FLAGS_lsi_model + " " + "--lsi_fsize " + str(FLAGS_lsi_fsize) + " " + "--outfile " + outf + str(wnid_idx) + "_lsi_temp") os.system(cmd) remaining_wnid_jobs.append(wnid_idx) print "... done" # poll for jobs finished and build merged duplicates structure if FLAGS_text_obs_unified > 0: print "merging textf results into map as they become available..." wnid_textf = {} while len(remaining_wnid_jobs) > 0: time.sleep(10) # poll for finished scripts every 10 seconds newly_finished_jobs = [] for wnid_idx in remaining_wnid_jobs: log_fn = str(wnid_idx) + "_lsi_temp" lsi_fn = log_fn + ".pickle" if os.path.isfile(lsi_fn): try: with open(lsi_fn, 'rb') as pf: new_textf = pickle.load(pf) except (IOError, EOFError, ValueError, KeyError): continue newly_finished_jobs.append(wnid_idx) os.system("rm " + lsi_fn) os.system("rm err." + log_fn) os.system("rm " + log_fn) wnid_textf[wnids[wnid_idx]] = new_textf remaining_wnid_jobs = [wnid_idx for wnid_idx in remaining_wnid_jobs if wnid_idx not in newly_finished_jobs] if len(newly_finished_jobs) > 0: print ("... " + str(len(remaining_wnid_jobs)) + " wnids remain after adding wnids: " + str(newly_finished_jobs)) whether_to_continue = raw_input("continue checks(Y/n)? ") # handle this weird shit if whether_to_continue == 'n': break # write textf print "writing wnid -> textf observations to file..." with open(FLAGS_outfile, 'wb') as f: d = wnid_textf pickle.dump(d, f) print "... done" if __name__ == '__main__': parser = argparse.ArgumentParser() parser.add_argument('--wnid_urls', type=str, required=True, help="wnid urls used when getting text observations") parser.add_argument('--text_obs_infile', type=str, required=True, help="wnid text observations file") parser.add_argument('--text_obs_unified', type=int, required=True, help="whether wnid text observations are in one file or one per wnid (1 for one file)") parser.add_argument('--lsi_dictionary', type=str, required=True, help="dictionary of words used in lsi model") parser.add_argument('--tfidf_model', type=str, required=True, help="tfidf model used by lsi") parser.add_argument('--lsi_model', type=str, required=True, help="serialized lsi model") parser.add_argument('--lsi_fsize', type=int, required=True, help="number of features in lsi") parser.add_argument('--outfile', type=str, required=True, help="output text features from w2v") args = parser.parse_args() for k, v in vars(args).items(): globals()['FLAGS_%s' % k] = v main()
[ "jesse@cs.utexas.edu" ]
jesse@cs.utexas.edu
f6e1dbcd885565e82d2661115d0d44d5701a04c0
1dff43a4fd4a8e84f8ce104dbb3a0d79185d9744
/jekyde/tests/test_jekyde.py
50005df25c5e0d44d9e7104a73ee31ab9e2912fb
[ "MIT" ]
permissive
devilicecream/jekyde
090134b060baffc4d75e7248b5113d4ee2a445b5
2440c183db00e97bf67ead6a00974a2bd9bc8fac
refs/heads/master
2020-05-19T07:56:36.774930
2019-05-04T17:01:20
2019-05-04T17:01:20
184,909,369
2
0
null
null
null
null
UTF-8
Python
false
false
2,061
py
import sqlalchemy as sa from ming import schema as s from ming.odm import MappedClass from jekyde.drivers import Driver from jekyde.meta import JekydeModel from sqlalchemy.orm import configure_mappers from .conftest import BaseModel def test_change_type(sql_session, ming_session): class AMingModel(MappedClass): class __mongometa__: session = ming_session name = "amodel" _id = s.ObjectId() value = s.String() class ASQLModel(BaseModel): __tablename__ = "amodel" id = sa.Column(sa.Integer(), primary_key=True) value = sa.Column(sa.String(length=32)) class AModel(JekydeModel): _driver = Driver.Ming _models = {Driver.Ming: AMingModel, Driver.SQLAlchemy: ASQLModel} configure_mappers() AModel(value="test") ming_session.flush() assert AModel.query.find({}).first() AModel.use(Driver.SQLAlchemy) with sql_session() as session: new_obj = AModel(value="test2") session.add(new_obj) session.flush() created = session.query(AModel).get(new_obj.id) assert created def test_migration(sql_session, ming_session): class BMingModel(MappedClass): class __mongometa__: session = ming_session name = "bmodel" _id = s.ObjectId() value = s.String() class BSQLModel(BaseModel): __tablename__ = "bmodel" id = sa.Column(sa.Integer(), primary_key=True) value = sa.Column(sa.String(length=32)) class BModel(JekydeModel): _driver = Driver.Ming _models = {Driver.Ming: BMingModel, Driver.SQLAlchemy: BSQLModel} configure_mappers() BModel(value="test") ming_session.flush() doc = BModel.query.find({}).first() assert doc with sql_session() as session: new_obj = BModel.migrate_to(doc, Driver.SQLAlchemy) session.add(new_obj) session.flush() obj_id = new_obj.id with sql_session() as session: assert session.query(BSQLModel).get(obj_id)
[ "walterdangalante@gmail.com" ]
walterdangalante@gmail.com
e50e159516d02e4f28151196e8e7cffa1b2abc36
454f5318d68aded03b7ce43371a68f02d51b5e5e
/playlist_website/playlist_website/wsgi.py
7de9f26408b4609b7b162d9966286c5d1a8c81af
[]
no_license
issaitorres/Portfolio
5ffd6c3e9c9e59af010a51960e3e116c57557cf9
b3324b28a0d1c8b8b5f4a1b5fc10428530f18336
refs/heads/master
2023-06-18T07:02:43.400662
2021-07-15T19:31:13
2021-07-15T19:31:13
306,566,451
0
0
null
null
null
null
UTF-8
Python
false
false
409
py
""" WSGI config for playlist_website project. It exposes the WSGI callable as a module-level variable named ``application``. For more information on this file, see https://docs.djangoproject.com/en/3.1/howto/deployment/wsgi/ """ import os from django.core.wsgi import get_wsgi_application os.environ.setdefault('DJANGO_SETTINGS_MODULE', 'playlist_website.settings') application = get_wsgi_application()
[ "issaitorres@gmail.com" ]
issaitorres@gmail.com
1c2030fad87f3a2f73634809687e3293afdb9848
79991f71b68590de24a5f8bb904054e30d38e77f
/main.py
cc71bc69feec84cef3cb93add31bf5d67ab407b7
[]
no_license
ottinger/housing-price-prediction-web
7d745d335b5253d27396ffe3a31c148bd0e11423
12d574159274d8afc0345be5bc97868fc7c753ca
refs/heads/master
2020-09-08T13:33:16.744118
2019-11-28T07:57:38
2019-11-28T07:57:38
221,148,235
0
0
null
null
null
null
UTF-8
Python
false
false
1,467
py
from flask import Flask, send_from_directory, request import pandas as pd import pickle import json app = Flask(__name__) @app.route('/predict', methods=['POST','GET']) def predict(): json_ = request.get_json() print(json_) predict_df = pd.DataFrame(columns=field_names) # empty df for req data predict_df = predict_df.append(pd.Series(), ignore_index=True) if json_: predict_data = json_ else: with open("sample_item.json", "rb") as file: predict_data = json.loads(file.read()) for item in predict_data.items(): if type(item[1]) != str: predict_df[item[0]][0] = item[1] else: # it's a string - onehot encoded for col_name, col_val in predict_df.iteritems(): if(col_name.startswith(item[0])): if col_name == item[0] + "-" + item[1]: predict_df[col_name][0] = 1 else: predict_df[col_name][0] = 0 prediction = model.predict(predict_df) return_dict = {"prediction": prediction[0]} return(json.dumps(return_dict)) @app.route('/<path:path>') def send_static(path): return send_from_directory('static', path) @app.route('/') def send_root(): return app.send_static_file('index.html') if __name__ == '__main__': field_names = pickle.load(open("column_names.pkl", "rb")) model = pickle.load(open("model.pkl", "rb")) app.run(port=8000)
[ "michael@ottinger.net" ]
michael@ottinger.net
00b4d530809478ba8d5e470aab29216b7d93a01e
6e781205dfb2aa7cf43709b4a9c208f4bb7117b7
/victory.py
45dca785110c60f4f79650b494f1ec19b5518a3f
[]
no_license
timofeyegorov/Python-developer-5-Console_file_manager
d04458e6c9750dbb5c98e039facc362b3c6e41a6
a83bc9d79b6345aef14e54f6a983f9447264950b
refs/heads/master
2023-06-16T23:23:44.563669
2021-07-14T07:32:22
2021-07-14T07:32:22
384,894,594
0
0
null
2021-07-14T07:32:23
2021-07-11T08:18:08
Python
UTF-8
Python
false
false
1,912
py
import random def victory_game(): while True: data = {'Рубен Диаш': ['14.05.1997', 'Четырнадцатое мая 1997 года'], 'Мохаммед Салах': ['15.06.1992', 'Пятнадцатое июня 1992 года'], 'Кевин Де Брюйне': ['28.06.1991', 'Двадцать восьмое июня 1991 года'], 'Неймар': ['05.02.1992', 'Пятое февраля 1992 года'], 'Эрлинг Холланд': ['21.07.2000', 'Двадцать первое июля 2000 года'], 'Килиан Мбаппе': ['20.12.1998', 'Двадцатое декабря 1998 года'], 'Бруну Фернандеш': ['08.09.1994', 'Восьмое сентября 1994 года'], 'Лионель Месси': ['24.06.1987', 'Двадцать четвертое июня 1987 года'], 'Криштиану Роналду': ['05.02.1985', 'Пятое февраля 1985 года'], 'Роберт Левандовски': ['21.08.1988', 'Двадцать первое августа 1988 года'] } filtered_data = random.sample(list(data.keys()), 5) wrong, right = 0, 0 for name in filtered_data: answer = input(f'Введите дату рождения {name} в формате dd.mm.yyyy: ') if answer != data[name][0]: print(f'Неверно, дата рождения {name} - {data[name][1]}') wrong += 1 else: right += 1 print('Правильных ответов: ', right) print('Ошибок: ', wrong) game = input('Сыграть еще? (введите нет, чтобы завершить игру): ') if game == 'нет': break
[ "timofeyegorov48@gmail.com" ]
timofeyegorov48@gmail.com
3ecdd02380aca40b278797ea18ccfd30972aae5d
2d768a41df277b92ca8a15407f24ff01eafb3de6
/4_Database_Flask_FastAPI/2_FastAPI/main.py
23468e2a732f7a0ab686277e55b99de58d1a6f3e
[]
no_license
morganpartee/docker_class
1a2de1e252373f71efa548fe8cccff8911f4b2cd
b9b5977a91133533b144dbcbd889cc25bb4d6070
refs/heads/master
2023-03-08T01:41:54.892533
2021-02-21T23:46:50
2021-02-21T23:46:50
284,179,958
5
1
null
null
null
null
UTF-8
Python
false
false
526
py
from fastapi import FastAPI from pickle import load import numpy as np import uvicorn app = FastAPI() with open("model.pkl", "rb") as f: clf = load(f) @app.get("/") def root(): return {"Hello": "World"} @app.get("/predict") def predict(sep_len: int, sep_wid: int, ped_len: int, ped_wid: int): return { "result": clf.predict( np.array([sep_len, sep_wid, ped_len, ped_wid]).reshape(1, -1) )[0] } if __name__ == "__main__": uvicorn.run("main:app", host="0.0.0.0", port=80)
[ "morganpartee@gmail.com" ]
morganpartee@gmail.com
6b19da70918b7711aee9f2fda10eb6fbec50ba0d
de24f83a5e3768a2638ebcf13cbe717e75740168
/moodledata/vpl_data/126/usersdata/191/29517/submittedfiles/ap2.py
c8f2da701341911eecf630c83018954555844586
[]
no_license
rafaelperazzo/programacao-web
95643423a35c44613b0f64bed05bd34780fe2436
170dd5440afb9ee68a973f3de13a99aa4c735d79
refs/heads/master
2021-01-12T14:06:25.773146
2017-12-22T16:05:45
2017-12-22T16:05:45
69,566,344
0
0
null
null
null
null
UTF-8
Python
false
false
423
py
# -*- coding: utf-8 -*- a=float(input('digite a:')) b=float(input('digite b:')) c=float(input('digite c:')) d=float(input('digite d:')) if a>=b and b>=c and a>=d: print(a) elif b>=a and b>=c and b>=d: print(b) elif c>=a and c>=b and c>=d: print(c) else: print(d) if a<=b and a<=c and a<=d: print(a) elif b<=a and b<=c and c<=d: print(b) elif c<=a and c<=b and c<=d: print(c) else: print(d)
[ "rafael.mota@ufca.edu.br" ]
rafael.mota@ufca.edu.br
05926ba3ca184ff6f9aeaa8c26c96fecba2df8dd
14148656b8c28ea4a28f45b00b0792004a2904db
/message/message_adapter.py
0645d332ff3f3a6c8321b001ee1a775bf77acf33
[]
no_license
CaveMike/mercury
7c2d2bbb1e1352db1faa5ad049fab018ac3410d4
eedaa52c1e49e91897533d93f2bf85654f80f423
refs/heads/master
2021-01-21T12:26:55.861250
2011-08-03T15:58:20
2011-08-03T15:58:20
2,149,384
0
0
null
null
null
null
UTF-8
Python
false
false
4,147
py
#!/usr/bin/env python from iron.dispatcher import Dispatcher from iron.event import Event from mercury.core import SipException from mercury.header.header import SIP_CRLF from mercury.message.message import Message from mercury.message.message import MessageEvent from mercury.message.message_assembler import DatagramReassembler from mercury.message.message_assembler import StreamReassembler from mercury.message.message_coder import MessageCoder from mercury.network.netevent import NetError from mercury.network.netevent import NetEvent from mercury.network.network import Network import logging class MessageAdapter(object): """Adapts between NetEvents and MessageEvents. Routes outgoing MessageEvents to the appropriate destination using the the appropriate source. """ def __init__( self, name, parent ): #super( MessageAdapter, self ).__init__( name, parent ) self.log = logging.getLogger( self.__class__.__name__ ) self.network = Network( 'net', self ) self.network.addListener( self ) self.coder = MessageCoder( self.query( 'network.encoding' ) ) self.default = DatagramReassembler() self.connections = {} def identifyEvent( self, event ): self.log.info( str(event) ) if isinstance( event, MessageEvent ): return event.id elif isinstance( event, NetEvent ): return event.id elif isinstance( event, NetError ): return event.id raise SipException( '[' + str(self.name) + '] ' + 'Ignoring event ' + str(event) + '.' ) def onBind( self, event ): # Pass through to the underlying network implementation. self.send( event, self.network, queued=False ) def onUnbind( self, event ): # Pass through to the underlying network implementation. self.send( event, self.network, queued=False ) def onRxPacket( self, event ): # Decode the message and, if the decoding succeeded, pass the MessageEvent up. text = self.coder.decode( event.packet ) if not event.connection: message = self.default.parse( text ) else: #FIXME: handle KeyError. message = self.connections[event.connection].parse( text ) if message != None: newEvent = MessageEvent( MessageEvent.EVENT_RX, message, transport=event.transport, localAddress=event.localAddress, localPort=event.localPort, remoteAddress=event.remoteAddress, remotePort=event.remotePort, useragent=self ) self.notify( newEvent, queued=False ) event.handled = True def __onTxPacket( self, event ): # Determine the transport, addresses, and ports to use and adjust the # SIP message as necessary. self.routeMessage( event ) # Encode the message, and if the encoding succeeded, pass the NetEvent down. text = self.coder.encode( event.message ) newEvent = NetEvent( NetEvent.EVENT_TX_PACKET, event.transport, event.localAddress, event.localPort, event.remoteAddress, event.remotePort, packet=text ) if newEvent: self.send( newEvent, self.network, queued=False ) event.handled = True def onTxRequest( self, event ): self.__onTxPacket( event ) def onTxResponse( self, event ): self.__onTxPacket( event ) def onConnected( self, event ): print 'ccc', str(event.connection) self.connections[event.connection] = StreamReassembler() event.handled = True def onDisconnected( self, event ): #FIXME: handle KeyError. print 'ddd', str(event.connection) #print self.connections del self.connections[event.connection] event.handled = True def onNetError( self, event ): self.log.error( str(event) ) #FIXME: Not sure what to do with these events. Should they be sent up as-is? # Or converted to MessageEvents? self.notify( event, queued=False ) event.handled = True def routeMessage( self, event ): """This function determines the address and port to send the message to. For requests, this function also determines the transport, address, and port to send the message from. Request: Look up Request-URI host and get remote transport, address and port. Modify/set Contact. Modify/set Via to local transport, address, and port. Response: Get the destination from the Via. """ #FIXME:IMPLEMENT. pass
[ "corrigan@gmail.com" ]
corrigan@gmail.com
69e96d91f1e97b1e4777741ed5926f0e3ffe5d96
d37ab0fa7dd0026425fc15a13288847ae0954f48
/src/helixweb/billing/forms_filters.py
dd3d23578a55a833025b34264b3fabe186615716
[]
no_license
sand8080/helixweb
4fd84e3df8add42996684a288c16148f8582297b
5f08b4cc41d6bd72f54382ebe5e9b45c428fac4b
refs/heads/master
2020-12-24T15:23:16.944216
2014-02-17T10:56:45
2014-02-17T10:56:45
1,048,085
0
0
null
null
null
null
UTF-8
Python
false
false
8,846
py
from django import forms from django.utils.translation import ugettext_lazy as _ from helixweb.core.widgets import ConstInput from helixweb.core.forms_filters import (FilterForm, AbstractFilterActionLogsForm, AbstractFilterAllActionLogsForm, AbstractFilterSelfActionLogsForm, AbstractFilterUserActionLogsForm) from helixweb.billing.forms import BillingForm class FilterBillingForm(FilterForm, BillingForm): pass class AbstractBillingFilterActionLogsForm(AbstractFilterActionLogsForm, FilterBillingForm): action = 'get_action_logs' def __init__(self, *args, **kwargs): kwargs['choices'] = (('', ''), ('add_balance', _('add balance')), ('modify_balance', _('modify balance')), ('add_receipt', _('add receipt')), ('add_bounus', _('add bonus')), ('lock', _('lock')), ('unlock', _('unlock')), ('charge_off', _('charge off')), ('modify_used_currencies', _('modify currencies')), ) super(AbstractBillingFilterActionLogsForm, self).__init__(*args, **kwargs) class FilterAllActionLogsForm(AbstractBillingFilterActionLogsForm, AbstractFilterAllActionLogsForm): pass class FilterSelfActionLogsForm(AbstractBillingFilterActionLogsForm, AbstractFilterSelfActionLogsForm): pass class FilterUserActionLogsForm(AbstractBillingFilterActionLogsForm, AbstractFilterUserActionLogsForm): pass class FilterCurrenciesForm(FilterBillingForm): action = 'get_currencies' ordering_param = '-code' class FilterUsedCurrenciesForm(FilterBillingForm): action = 'get_used_currencies' ordering_param = '-code' class FilterBalanceForm(FilterBillingForm): action = 'get_balances' def __init__(self, *args, **kwargs): currencies = kwargs.pop('currencies', []) super(FilterBalanceForm, self).__init__(*args, **kwargs) self.fields['id'] = forms.IntegerField(label=_('balance id'), required=False) self.fields['user_id'] = forms.IntegerField(label=_('user id'), required=False) self.fields['currency_code'] = self._gen_currency_code(currencies, required=False) self.fields['from_real_amount'] = forms.DecimalField(label=_('real amount from'), required=False) self.fields['to_real_amount'] = forms.DecimalField(label=_('real amount to'), required=False) self.fields['from_virtual_amount'] = forms.DecimalField(label=_('virtual amount from'), required=False) self.fields['to_virtual_amount'] = forms.DecimalField(label=_('virtual amount to'), required=False) self.fields['from_overdraft_limit'] = forms.DecimalField(label=_('overdraft limit from'), required=False) self.fields['to_overdraft_limit'] = forms.DecimalField(label=_('overdraft limit to'), required=False) self.fields['from_locked_amount'] = forms.DecimalField(label=_('locked amount from'), required=False) self.fields['to_locked_amount'] = forms.DecimalField(label=_('locked amount to'), required=False) self.fields['is_active'] = forms.ChoiceField(label=_('is active'), required=False, widget=forms.widgets.RadioSelect(), choices=(('all', _('all')), ('1', _('active')), ('0', _('inactive'))), initial='all') def as_helix_request(self): d = super(FilterBalanceForm, self).as_helix_request() self._strip_filter_param(d, 'id') self._strip_filter_param(d, 'user_id') self._strip_filter_param(d, 'currency_code') self._strip_filter_param(d, 'from_real_amount') self._strip_filter_param(d, 'to_real_amount') self._strip_filter_param(d, 'from_virtual_amount') self._strip_filter_param(d, 'to_virtual_amount') self._strip_filter_param(d, 'from_overdraft_limit') self._strip_filter_param(d, 'to_overdraft_limit') self._strip_filter_param(d, 'from_locked_amount') self._strip_filter_param(d, 'to_locked_amount') if (not d['filter_params']['is_active'] or d['filter_params']['is_active'] == 'all'): d['filter_params'].pop('is_active') else: val = bool(int(d['filter_params']['is_active'])) d['filter_params']['is_active'] = val return d class AbstractFilterLocksForm(FilterBillingForm): action = 'get_locks' def _add_common_fields(self): self.fields['order_id'] = forms.CharField(label=_('order id'), max_length=64, required=False) self.fields['from_creation_date'] = forms.DateField(label=_('from'), required=False) self.fields['to_creation_date'] = forms.DateField(label=_('to'), required=False) def as_helix_request(self): d = super(AbstractFilterLocksForm, self).as_helix_request() self._strip_filter_param(d, 'user_id') self._strip_filter_param(d, 'order_id') self._strip_filter_param(d, 'balance_id') self._strip_from_date_param(d, 'from_creation_date') self._strip_to_date_param(d, 'to_creation_date') return d class FilterLocksForm(AbstractFilterLocksForm): def __init__(self, *args, **kwargs): super(FilterLocksForm, self).__init__(*args, **kwargs) self.fields['user_id'] = forms.IntegerField(label=_('user id'), required=False) self.fields['balance_id'] = forms.IntegerField(label=_('balance id'), required=False) self._add_common_fields() class FilterUserBalanceLocksForm(AbstractFilterLocksForm): def __init__(self, *args, **kwargs): super(FilterUserBalanceLocksForm, self).__init__(*args, **kwargs) self.fields['user_id'] = forms.IntegerField(label=_('user id'), widget=ConstInput, required=False) self.fields['balance_id'] = forms.IntegerField(label=_('balance id'), widget=ConstInput, required=False) self._add_common_fields() class FilterSelfLocksForm(AbstractFilterLocksForm): action = 'get_locks_self' def __init__(self, *args, **kwargs): super(FilterSelfLocksForm, self).__init__(*args, **kwargs) self._add_common_fields() class AbstractFilterTransactionsForm(FilterBillingForm): action = 'get_transactions' def _add_common_fields(self): self.fields['order_id'] = forms.CharField(label=_('order id'), max_length=64, required=False) self.fields['type'] = forms.ChoiceField(label=_('type'), required=False, widget=forms.widgets.Select(), choices=((None, _('all')), ('receipt', _('receipt')), ('bonus', _('bonus')), ('lock', _('lock')), ('unlock', _('unlock')), ('charge_off', _('charge off'))), initial='all') self.fields['from_creation_date'] = forms.DateField(label=_('from'), required=False) self.fields['to_creation_date'] = forms.DateField(label=_('to'), required=False) def as_helix_request(self): d = super(AbstractFilterTransactionsForm, self).as_helix_request() self._strip_filter_param(d, 'id') self._strip_filter_param(d, 'user_id') self._strip_filter_param(d, 'order_id') self._strip_filter_param(d, 'type') self._strip_filter_param(d, 'balance_id') self._strip_from_date_param(d, 'from_creation_date') self._strip_to_date_param(d, 'to_creation_date') return d class FilterTransactionsForm(AbstractFilterTransactionsForm): def __init__(self, *args, **kwargs): super(FilterTransactionsForm, self).__init__(*args, **kwargs) self.fields['user_id'] = forms.IntegerField(label=_('user id'), required=False) self.fields['balance_id'] = forms.IntegerField(label=_('balance id'), required=False) self.fields['id'] = forms.IntegerField(label=_('id'), required=False) self._add_common_fields() class FilterUserTransactionsForm(AbstractFilterTransactionsForm): def __init__(self, *args, **kwargs): super(FilterUserTransactionsForm, self).__init__(*args, **kwargs) self.fields['user_id'] = forms.IntegerField(label=_('user id'), widget=ConstInput, required=False) self.fields['balance_id'] = forms.IntegerField(label=_('balance id'), widget=ConstInput, required=False) self.fields['id'] = forms.IntegerField(label=_('id'), required=False) self._add_common_fields() class FilterSelfTransactionsForm(AbstractFilterTransactionsForm): action = 'get_transactions_self' def __init__(self, *args, **kwargs): super(FilterSelfTransactionsForm, self).__init__(*args, **kwargs) self.fields['id'] = forms.IntegerField(label=_('id'), required=False) self._add_common_fields()
[ "sand8080@gmail.com" ]
sand8080@gmail.com
406e7aae33727bb43762bd69aaa12abd355c309f
fd85b477888a3061233d9ef4814ce45953145618
/demo/SimplePlot.py
b679ed0e3a0dee702055abc760d3149382a038d9
[ "MIT", "LicenseRef-scancode-unknown-license-reference" ]
permissive
anondroid5/Tkinter
799f318006c886b592dc340d1d08f34870b5e998
3556224ec2e30b31da9f4a472dfe978c639c2c9f
refs/heads/master
2021-01-17T07:43:32.166211
2016-07-18T09:10:12
2016-07-18T09:10:12
37,201,096
2
0
null
null
null
null
UTF-8
Python
false
false
1,202
py
#coding: utf-8 from Tkinter import * def main(): root = Tk() root.title('Simple Plot') try: canvas = Canvas(root, width=450, height=300, bg = 'white') canvas.pack() Button(root, text='Quit', command=root.quit).pack() canvas.create_line(100,250,400,250, width=2) canvas.create_line(100,250,100,50, width=2) for i in range(11): x = 100 + (i * 30) canvas.create_line(x,250,x,245, width=2) canvas.create_text(x,254, text='%d'% (10*i), anchor=N) for i in range(6): x = 250 - (i + 40) canvas.create_line(100,y,105,y, width=2) canvas.create_text(96,y, text='%5.1f'% (50.*i), anchor=E) scaled = [] for x,y in [(12, 56), (20, 94), (33, 98), (45, 120), (61, 180), (75, 160), (98, 223)]: scaled.append(100 + 3*x, 250 - (4*y)/5) canvas.create_line(scaled, fill='black', smooth=1) for xs,ys in scaled: canvas.create_oval(x-6,y-6,x+6,y+6, width=1, outline='black', fill='SkyBlue2') except: print 'An error has occured!' root.mainloop() main()
[ "boy11922960shooping@gmail.com" ]
boy11922960shooping@gmail.com
a4f91d131845e9d72b46556179c3d49e831e15f0
c7bb2f583148c3851720e1f535520cfb5f3df32e
/lib/models/removal_model.py
8ded2eec4bfcab124b17f5aacaa9ac33ddc74fd1
[]
no_license
PeterZhouSZ/Shadow-aware-portrait-relight
e248d8afee96a812997044b7f51d0a0616fce070
119fb975de77c58a1049074195bc009660700575
refs/heads/main
2023-03-24T06:53:08.161122
2021-03-28T10:55:35
2021-03-28T10:55:35
374,818,414
1
0
null
2021-06-07T22:43:55
2021-06-07T22:43:55
null
UTF-8
Python
false
false
4,347
py
from lib.networks.base_network import MsImageDis from lib.networks.removal_network import Gen from lib.utils.utils import weights_init, get_model_list, vgg_preprocess, load_vgg16, get_scheduler,ssim from lib.models.base_model import BaseModels from torch.autograd import Variable import torch import torch.nn as nn import os import sys # x y is subject # a b is illumination class Models(BaseModels): def __init__(self, hyperparameters): super(Models, self).__init__() lr = hyperparameters['lr'] self.model_name = hyperparameters['models_name'] # Initiate the networks if(self.model_name=='removal'): self.gen = Gen(hyperparameters['input_dim_a'], hyperparameters['gen']) self.dis = MsImageDis(hyperparameters['input_dim_a'], hyperparameters['dis']) # discriminator for domain a else: sys.exit('error on models') self.instancenorm = nn.InstanceNorm2d(512, affine=False) self.style_dim = hyperparameters['gen']['style_dim'] self.gen = nn.DataParallel(self.gen) self.dis = nn.DataParallel(self.dis) # fix the noise used in sampling display_size = int(hyperparameters['display_size']) # Setup the optimizers beta1 = hyperparameters['beta1'] beta2 = hyperparameters['beta2'] gen_params = list(self.gen.parameters()) self.gen_opt = torch.optim.Adam([p for p in gen_params if p.requires_grad], lr=lr, betas=(beta1, beta2), weight_decay=hyperparameters['weight_decay']) self.gen_scheduler = get_scheduler(self.gen_opt, hyperparameters) dis_params = list(self.dis.parameters()) self.dis_opt = torch.optim.Adam([p for p in dis_params if p.requires_grad], lr=lr, betas=(beta1, beta2), weight_decay=hyperparameters['weight_decay']) self.dis_scheduler = get_scheduler(self.dis_opt, hyperparameters) # Network weight initialization self.apply(weights_init(hyperparameters['init'])) self.dis.apply(weights_init('gaussian')) # Load VGG model if needed if 'vgg_w' in hyperparameters.keys() and hyperparameters['vgg_w'] > 0: self.vgg = load_vgg16(hyperparameters['vgg_model_path'] + '/models') self.vgg.eval() for param in self.vgg.parameters(): param.requires_grad = False self.vgg = nn.DataParallel(self.vgg) def gen_update(self, out_data,hyperparameters): Xa_out, X_removal, mask, depth=out_data self.gen_opt.zero_grad() # encode out_x,p_x = self.gen.forward(torch.mul(Xa_out, mask)) # main relight loss self.loss_gen_prime_x_b = self.recon_criterion_mask(out_x, X_removal, mask) # main p loss self.loss_gen_prime_x_p = self.recon_criterion_mask(p_x, depth, mask[:,0:1,:,:]) # GAN loss self.loss_gen_adv = self.calc_gen_loss(self.dis.forward(torch.mul(out_x, mask))) self.loss_gen_total = hyperparameters['relight'] * self.loss_gen_prime_x_b + \ hyperparameters['x_p'] * self.loss_gen_prime_x_p + hyperparameters['gan_w']* self.loss_gen_adv self.loss_gen_total.backward() self.gen_opt.step() image_anchor = Xa_out[0:1].detach().cpu()[:3] image_recons = torch.mul(out_x, mask)[0:1].detach().cpu()[:3] image_gt = X_removal[0:1].detach().cpu()[:3] depth_gt = (depth[0:1].detach().cpu()[:3]).repeat(1,3,1,1) depth = (p_x[0:1].detach().cpu()[:3]).repeat(1,3,1,1) self.image_display = torch.cat((image_anchor, image_recons, image_gt,depth_gt,depth),dim=3) def dis_update(self, x_a,gt_xb,x_mask,hyperparameters): self.dis_opt.zero_grad() out_x,_ = self.gen.forward(torch.mul(x_a, x_mask)) # D loss out_x = torch.mul(out_x, x_mask) gt_xb = torch.mul(gt_xb, x_mask) self.loss_dis = self.calc_dis_loss(self.dis.forward(out_x.detach()),self.dis.forward(gt_xb)) #self.loss_dis = self.dis.calc_dis_loss(out_x.detach(), gt_xb) self.loss_dis_total = hyperparameters['gan_w'] * self.loss_dis self.loss_dis_total.backward() self.dis_opt.step()
[ "guoxian.song@bytedance.com" ]
guoxian.song@bytedance.com
2daf9cca1251a8f201f5a3437dc55c28c447f373
c54896ba0703fc4de01d19ed9e7eecfdfc4aa810
/Python_Projects/Python_MySQL/Foreign_Trade_Data_Pipe_Delimination/scripts/merchandise_trade_exports.py
e5ba7c2739bc71535561f9073a09b9bc30fd063e
[ "MIT" ]
permissive
YangLiu928/NDP_Projects
f4d1c3d45161ff353f89dff2ebc36fabde28db62
d2ebfa7b95a0003481dde1361c6ab563ac94f2e6
refs/heads/master
2021-01-10T14:39:17.141520
2016-04-28T19:57:48
2016-04-28T19:57:48
45,988,930
0
0
null
null
null
null
UTF-8
Python
false
false
841
py
import process_data data_folder = '../data/merchandise_trade_exports/cdromtxt/' output_folder = '../output/merchandise_trade_exports/' process_data.process_concord(data_folder, output_folder) process_data.process_country(data_folder, output_folder) process_data.process_district(data_folder, output_folder) process_data.process_enduse(data_folder, output_folder) process_data.process_exp_comm(data_folder, output_folder) process_data.process_exp_cty(data_folder, output_folder) process_data.process_exp_detl(data_folder, output_folder) process_data.process_exp_dist(data_folder, output_folder) process_data.process_hitech(data_folder, output_folder) process_data.process_hsdesc(data_folder, output_folder) process_data.process_naics(data_folder, output_folder) process_data.process_sitc(data_folder, output_folder) print 'work completed'
[ "yangliu1989@gwu.edu" ]
yangliu1989@gwu.edu
ca7114b9431d0b12bef71d9fd7e5c95a095770dd
a1c08cca122d580994e6a495095fbfeb3e7120d6
/bert4keras/__init__.py
e559963fc6776e35aaaee73347b221f421fde74d
[ "Apache-2.0" ]
permissive
meissenzheng/bert4keras
cc4c4b25a52521612b9e460e3356e38ad4158d93
93c1b8c78a8efac7843e634ee7d5b3b1ea631f5d
refs/heads/master
2023-04-17T14:14:57.103530
2021-04-25T05:10:56
2021-04-25T05:10:56
null
0
0
null
null
null
null
UTF-8
Python
false
false
49
py
#! -*- coding: utf-8 -*- __version__ = '0.10.5'
[ "noreply@github.com" ]
meissenzheng.noreply@github.com
e26f787cdac7bccc03eebca71b634c965ab28d9b
8bf4243abe4f5f66d0068d96dc09e5cf975430d5
/Selenium_lessons/neskolko_brauserov/auth_data.py
92e50613b48ae3cde7e5b5defba74be75c6c227c
[]
no_license
Azhdar1990/Parsers
899776388b4dbe257eb5b6d08dce3f0ef5ea8190
aa39fd3f651b92405c12caf1bd784ed3afa4f3fd
refs/heads/master
2023-09-05T12:45:03.604726
2021-10-31T10:51:56
2021-10-31T10:51:56
404,690,752
0
0
null
null
null
null
UTF-8
Python
false
false
56
py
login = "maniyev@rambler.ru" password = "@kmNom@k!k12@"
[ "amaniyev@gmail.com" ]
amaniyev@gmail.com
25a31c75346e71626fe953790437618a840aa0b7
3a4860c05baa2791986747bf1d73f6b8da0d7fa6
/list_append.py
7387fee0899f55dac4231dc81b7576f89d3b8ca3
[]
no_license
Raun551/pythontraining
dcdb0acf966d555b7a7ab7c242031a9cff1ac354
65e9b32bd92c5e096f8c11fadfca21ed6ea38553
refs/heads/master
2020-03-26T14:21:03.761316
2018-08-30T14:38:31
2018-08-30T14:38:31
144,983,611
0
0
null
null
null
null
UTF-8
Python
false
false
134
py
#Append a list to the second list list1 = [1, 2, 3, 0] list2 = ['Red', 'Green', 'Black'] final_list = list1 + list2 print(final_list)
[ "raunaq.malik23@gmail.com" ]
raunaq.malik23@gmail.com
33ca51a5d85d8352fdbdea328351d34be314446f
c7b1282a1d66eb2756a63909320f515d03917deb
/day00/ex05/kata03.py
5c2e7ccc39d7df276537cc565f9161a3e8a4fd7a
[]
no_license
AdrianWR/PythonBootcamp
5603f2108f4ff80292e8964f75af44223fe0febc
9dc6ef196b3bc51c079efcd19ea421561ceb8a6d
refs/heads/master
2022-12-23T00:26:22.833866
2020-09-23T12:36:14
2020-09-23T12:36:14
259,933,367
1
2
null
2020-04-29T13:35:31
2020-04-29T13:31:00
Python
UTF-8
Python
false
false
107
py
phrase = "The right format" if __name__ == "__main__": s = phrase.rjust(42, '-') print(s, end='')
[ "adrian.w.roque@gmail.com" ]
adrian.w.roque@gmail.com
d0fd73ad51f76303fa3a3baf786e05f599ca263c
59f8e9d5c273974adcac14ccbd410c1e2689c4e1
/setup.py
a58a482d26006affe9310b191c0521ccebe38755
[ "MIT" ]
permissive
andrewzeneski/pylodge
74a9f8c33af4e7d509ca0dc91013ad966ca3b685
f5fdc470ef986228f8a44ad1c9d2d45ea221a95b
refs/heads/master
2020-12-27T03:52:11.185337
2016-06-02T16:50:55
2016-06-02T16:50:55
59,770,088
1
1
null
2016-05-26T17:33:06
2016-05-26T17:33:06
null
UTF-8
Python
false
false
833
py
__author__ = 'ashwin' """A Test Lodge based pylodge module. """ from setuptools import setup setup( name='pylodge', version='0.2.8', description='Test Automation framework for TestLodge', url='https://github.com/gettalent/pylodge', # Author details author='Ashwin Kondapalli', author_email='ashwin@gettalent.com', license='MIT', # See https://pypi.python.org/pypi?%3Aaction=list_classifiers classifiers=[ 'Development Status :: 5 - Production/Stable', 'Intended Audience :: Developers', 'Topic :: Software Development :: Quality Assurance', 'Programming Language :: Python', 'License :: OSI Approved :: MIT License', ], keywords='TestLodge, test automation, pylodge', packages=['pylodge'], install_requires=['requests'], )
[ "ashwin@gettalent.com" ]
ashwin@gettalent.com
1ca7d40ffa698cb26fa3c02af75ca221a34029a0
8112bdea83d0a51c12720f751fb7759ff6741d9d
/tests/test_friends.py
7d1912acdcf1a75a28e20aa396d280b9cc08ffcf
[]
no_license
dushyantpatel/smap_api
e5ea67a1c28645a7ebfa74576a9be7f848fd9769
263e44efd4a976c80ad71413fa05279b4a4efbe0
refs/heads/master
2020-03-17T07:39:57.776872
2018-06-10T02:37:15
2018-06-10T02:37:15
133,407,881
1
0
null
null
null
null
UTF-8
Python
false
false
3,116
py
import unittest import pymysql import rds_config import sys from tests.event import Event from gateway import main_handler # rds settings rds_host = rds_config.db_host name = rds_config.db_username password = rds_config.db_password db_name = rds_config.db_name connection = None # resource settings path = 'friends' context = None class TestUsers(unittest.TestCase): # @classmethod # def setUpClass(cls): # global connection # # connect to database # try: # connection = pymysql.connect(rds_host, user=name, passwd=password, db=db_name, connect_timeout=5) # except pymysql.err.Error as ex: # template = "ERROR: {0} - Could not connect to MySql instance \n{1!r}" # message = template.format(type(ex).__name__, ex.args) # print(message) # sys.exit() # # # setup test users in the database # add_usr_template = 'INSERT INTO user (display_name, email, first_name, last_name, missionCurator) ' \ # 'VALUES ("{0}", "{1}", "{2}", "{3}", {4});' # add_usr1 = add_usr_template.format('test_user', 'test1.email@smap.com', 'Test1', 'Case1', 0) # add_usr2 = add_usr_template.format('test_user', 'test2.email@smap.com', 'Test2', 'Case2', 0) # add_usr3 = add_usr_template.format('test_user', 'test3.email@smap.com', 'Test3', 'Case3', 0) # add_usr4 = add_usr_template.format('test_user', 'test4.email@smap.com', 'Test4', 'Case4', 0) # with connection.cursor() as cur: # cur.execute(add_usr1) # cur.execute(add_usr2) # cur.execute(add_usr3) # cur.execute(add_usr4) # connection.commit() # # @classmethod # def tearDownClass(cls): # # clean up the database # with connection.cursor() as cur: # cur.execute('SELECT * FROM user WHERE display_name="test_user"') # li = cur.fetchall() # for row in li: # cur.execute('DELETE FROM user WHERE user_id=' + str(row[0])) # connection.commit() # connection.close() # # def setUp(self): # self.event = Event() # self.event.setPath(path) # self.req_body = dict() # # def tearDown(self): # return # # def test_add_new_user(self): # self.event.setHttpMethod('POST') # # # user 1 sends friend request to user 2 # self.req_body['user_1'] = 'test1.email@smap.com' # self.req_body['user_2'] = 'test2.email@smap.com' # self.event.setBody(str(self.req_body)) # # response = main_handler(self.event.getEvent(), context) # resp_body = response['body'] # status_code = response['statusCode'] # headers = response['headers'] # print(headers['message']) # print(headers['details']) # # # check for correct status code # self.assertEqual(201, status_code) # # # check for correct body # self.assertEqual(resp_body, str(None)) def test_dummy(self): return
[ "dushyantpatel_r@outlook.com" ]
dushyantpatel_r@outlook.com
210f7ba427b184848d5fb3c985f326bb4dde8314
d3ac52556da9f8a8c0dddbd4b30a81d67579ed5f
/options.py
34091e25f6dd787bd663ad5ffd4c92ef9d6de41a
[]
no_license
mustaphaasbbar/Silaty
59b06eaa614e8d202089a13088fe84c522a4fb79
f8a4ddc2e7e351b51030759b72bfe9f6b9d3fc77
refs/heads/master
2020-04-29T08:16:19.144501
2015-04-22T17:59:19
2015-04-22T17:59:19
null
0
0
null
null
null
null
UTF-8
Python
false
false
10,474
py
import configparser import os import datetime class Calendar(object): UmmAlQuraUniv, \ EgyptianGeneralAuthorityOfSurvey,\ UnivOfIslamicSciencesKarachi,\ IslamicSocietyOfNorthAmerica,\ MuslimWorldLeague = range(5) class Madhab(object): Default, Hanafi = 0, 1 class Options: def __init__(self): print ("DEBUG: Initializing the Options module @", (str(datetime.datetime.now()))) cparse = configparser.ConfigParser() cparse.read([os.path.expanduser('~/.silaty')]) try: self._city = cparse.get('DEFAULT', 'city') self._country = cparse.get('DEFAULT', 'country') self._calcmethodname = cparse.get('DEFAULT', 'calculation-method') self._madhab = cparse.get('DEFAULT', 'madhab') self._clockformat = cparse.get('DEFAULT', 'clock-format') self._latitude = cparse.get('DEFAULT', 'latitude') self._longitude = cparse.get('DEFAULT', 'longitude') self._timezone = cparse.get('DEFAULT', 'timezone') self._notif = cparse.get('DEFAULT', 'notif') self._iconlabel = cparse.get('DEFAULT', 'iconlabel') self._startminimized = cparse.get('DEFAULT', 'minimized') self._fajradhan = cparse.get('DEFAULT', 'fajr-adhan') self._normaladhan = cparse.get('DEFAULT', 'normal-adhan') self._audionotifications = cparse.get('DEFAULT', 'audio-notifications') except configparser.NoOptionError: print ("DEBUG: No configration file using default settings") self._city = 'Makkah' self._country = 'Saudi Arabia' self._latitude = '21.25' self._longitude = '39.49' self._timezone = '3' self._calcmethodname = 'Makkah' self._madhab = 'Default' self._clockformat = '24h' self._notif = '10' self._iconlabel = '1' self._startminimized = '1' self._fajradhan = (self.get_fajr_adhans())[0] self._normaladhan = (self.get_normal_adhans())[0] self._audionotifications = '1' self.save_options() except ValueError: print ("DEBUG: Problem while reading setting file, using the default settings") os.system("rm ~/.silaty") self._city = 'Makkah' self._country = 'Saudi Arabia' self._latitude = '21.25' self._longitude = '39.49' self._timezone = '3' self._calcmethodname = 'Makkah' self._madhab = 'Default' self._clockformat = '24h' self._notif = '10' self._iconlabel = '1' self._startminimized = '1' self._fajradhan = (self.get_fajr_adhans())[0] self._normaladhan = (self.get_normal_adhans())[0] self._audionotifications = '1' self.save_options() ##Functions with lists for the Buttons def get_cal_methods(self): return ['Makkah', 'Egypt', 'Karachi', 'ISNA', 'MWL'] def get_madhahed(self): return ['Hanafi','Default'] def get_clock_formats(self): return ['12h', '24h'] def get_fajr_adhans(self): dirfiles = os.listdir( os.path.dirname(os.path.realpath(__file__))+"/audio/Fajr/") wavfiles = filter(lambda song: song.endswith(".ogg"), dirfiles) adhans = list(map(lambda x: os.path.splitext(x)[0], wavfiles)) return adhans def get_normal_adhans(self): dirfiles = os.listdir( os.path.dirname(os.path.realpath(__file__))+"/audio/Normal/") wavfiles = filter(lambda song: song.endswith(".ogg"), dirfiles) adhans = list(map(lambda x: os.path.splitext(x)[0], wavfiles)) return adhans ##Functions to get and set settings @property def audio_notifications_num(self): return self._audionotifications @audio_notifications_num.setter def audio_notifications_num(self, value): self._audionotifications = value @property def audio_notifications(self): print ("DEBUG: getting icon label settings @", (str(datetime.datetime.now()))) if self.audio_notifications_num == '1': return True else: return False @audio_notifications.setter def audio_notifications(self, data): print ("DEBUG: setting icon label settings @", (str(datetime.datetime.now()))) if data == True: self.audio_notifications_num = '1' else: self.audio_notifications_num = '0' @property def fajr_adhan(self): return self._fajradhan @fajr_adhan.setter def fajr_adhan(self, value): self._fajradhan = value @property def normal_adhan(self): return self._normaladhan @normal_adhan.setter def normal_adhan(self, value): self._normaladhan = value @property def city(self): print ("DEBUG: getting city settings @", (str(datetime.datetime.now()))) return self._city @city.setter def city(self, data): print ("DEBUG: setting city settings @", (str(datetime.datetime.now()))) self._city = data @property def country(self): print ("DEBUG: getting country settings @", (str(datetime.datetime.now()))) return self._country @country.setter def country(self, value): print ("DEBUG: setting country settings @", (str(datetime.datetime.now()))) self._country = value @property def calculation_method_name(self): return self._calcmethodname @calculation_method_name.setter def calculation_method_name(self, value): self._calcmethodname = value @property def calculation_method(self): print ("DEBUG: getting calculation method settings @", (str(datetime.datetime.now()))) if self.calculation_method_name == 'Makkah': return Calendar.UmmAlQuraUniv elif self.calculation_method_name == 'Egypt': return Calendar.EgyptianGeneralAuthorityOfSurvey elif self.calculation_method_name == 'Karachi': return Calendar.UnivOfIslamicSciencesKarachi elif self.calculation_method_name == 'ISNA': return Calendar.IslamicSocietyOfNorthAmerica elif self.calculation_method_name == 'MWL': return Calendar.MuslimWorldLeague @calculation_method.setter def calculation_method(self, data): print ("DEBUG: setting calculation method settings @", (str(datetime.datetime.now()))) self.calculation_method_name = data @property def madhab_name(self): return self._madhab @madhab_name.setter def madhab_name(self, value): self._madhab = value @property def madhab(self): print ("DEBUG: getting madhab settings @", (str(datetime.datetime.now()))) if self.madhab_name == 'Default': return Madhab.Default if self.madhab_name == 'Hanafi': return Madhab.Hanafi @madhab.setter def madhab(self, data): print ("DEBUG: setting madhab settings @", (str(datetime.datetime.now()))) self._madhab = data @property def latitude(self): print ("DEBUG: getting latitude settings @", (str(datetime.datetime.now()))) return float(self._latitude) @latitude.setter def latitude(self, data): print ("DEBUG: setting latitude settings @", (str(datetime.datetime.now()))) self._latitude = str(data) @property def longitude(self): print ("DEBUG: getting longitude settings @", (str(datetime.datetime.now()))) return float(self._longitude) @longitude.setter def longitude(self, data): print ("DEBUG: setting longitude settings @", (str(datetime.datetime.now()))) self._longitude = str(data) @property def timezone(self): print ("DEBUG: getting timezone settings @", (str(datetime.datetime.now()))) return float(self._timezone) @timezone.setter def timezone(self, data): print ("DEBUG: setting timezone settings @", (str(datetime.datetime.now()))) self._timezone = str(data) @property def notification_time(self): print ("DEBUG: getting notification time settings @", (str(datetime.datetime.now()))) return float(self._notif) @notification_time.setter def notification_time(self, data): print ("DEBUG: setting notification time settings @", (str(datetime.datetime.now()))) self._notif = str(data) @property def iconlabel_num(self): return self._iconlabel @iconlabel_num.setter def iconlabel_num(self, value): self._iconlabel = value @property def iconlabel(self): print ("DEBUG: getting icon label settings @", (str(datetime.datetime.now()))) if self.iconlabel_num == '1': return True else: return False @iconlabel.setter def iconlabel(self, data): print ("DEBUG: setting icon label settings @", (str(datetime.datetime.now()))) if data == True: self.iconlabel_num = '1' else: self.iconlabel_num = '0' @property def start_minimized_num(self): return self._startminimized @start_minimized_num.setter def start_minimized_num(self, value): self._startminimized = value @property def start_minimized(self): print ("DEBUG: getting icon label settings @", (str(datetime.datetime.now()))) if self.start_minimized_num == '1': return True else: return False @start_minimized.setter def start_minimized(self, data): print ("DEBUG: setting icon label settings @", (str(datetime.datetime.now()))) if data == True: self.start_minimized_num = '1' else: self.start_minimized_num = '0' @property def clock_format(self): print ("DEBUG: getting clock format settings @", (str(datetime.datetime.now()))) return self._clockformat @clock_format.setter def clock_format(self, data): print ("DEBUG: setting clock format settings @", (str(datetime.datetime.now()))) self._clockformat = data ## Function to save the options def save_options(self): print ("DEBUG: saving settings file @", (str(datetime.datetime.now()))) config = open(os.path.expanduser('~/.silaty'), 'w') Text='''# Silaty Settings File [DEFAULT] # Location Information city = %s country = %s latitude = %s longitude = %s timezone = %s # Possible Values for Calculation Methods # Makkah # Egypt # Karachi # ISNA # MWL calculation-method = %s # Possible Values for Madhaheb # Default # Hanafi madhab = %s # Possible Values for Clock Format # 24h # 12h clock-format = %s # Time before prayer for notification notif = %s # Display Time by the indicator icon iconlabel = %s # Should audio notifications be enabled audio-notifications = %s # Should the application state minimized minimized = %s # Paths to the audio files fajr-adhan = %s normal-adhan = %s ''' % (self.city, self.country, self.latitude, self.longitude, self.timezone,\ self.calculation_method_name, self.madhab_name, self.clock_format, \ self.notification_time, self.iconlabel_num, self.audio_notifications_num,self.start_minimized_num, \ self.fajr_adhan, self.normal_adhan) config.write(Text) config.close()
[ "www.jwb@gmail.com" ]
www.jwb@gmail.com
1c3d00acafd76a610342ab1ef712ad250ee8870c
b2bdd5997ac84b0e19071c1ddc1c1a4d2f4fab58
/catkin_ws/devel/.private/p2/lib/python2.7/dist-packages/p2/msg/_Ackermann.py
0dff4e208b8c08e4de290b065cd192a52bee173e
[]
no_license
hbtslys01/RosCodingProject
860d18531dabe4a969278deff5dbad8a8703ea83
226feda08724e92fd94191e123b9442c028283dd
refs/heads/master
2020-04-11T09:16:17.808626
2018-12-13T17:30:08
2018-12-13T17:30:08
161,671,560
0
0
null
null
null
null
UTF-8
Python
false
false
3,768
py
# This Python file uses the following encoding: utf-8 """autogenerated by genpy from p2/Ackermann.msg. Do not edit.""" import sys python3 = True if sys.hexversion > 0x03000000 else False import genpy import struct class Ackermann(genpy.Message): _md5sum = "61c7e29a36f91d9c196a9722234d7472" _type = "p2/Ackermann" _has_header = False #flag to mark the presence of a Header object _full_text = """float64 steering_angle float64 vel """ __slots__ = ['steering_angle','vel'] _slot_types = ['float64','float64'] def __init__(self, *args, **kwds): """ Constructor. Any message fields that are implicitly/explicitly set to None will be assigned a default value. The recommend use is keyword arguments as this is more robust to future message changes. You cannot mix in-order arguments and keyword arguments. The available fields are: steering_angle,vel :param args: complete set of field values, in .msg order :param kwds: use keyword arguments corresponding to message field names to set specific fields. """ if args or kwds: super(Ackermann, self).__init__(*args, **kwds) #message fields cannot be None, assign default values for those that are if self.steering_angle is None: self.steering_angle = 0. if self.vel is None: self.vel = 0. else: self.steering_angle = 0. self.vel = 0. def _get_types(self): """ internal API method """ return self._slot_types def serialize(self, buff): """ serialize message into buffer :param buff: buffer, ``StringIO`` """ try: _x = self buff.write(_get_struct_2d().pack(_x.steering_angle, _x.vel)) except struct.error as se: self._check_types(struct.error("%s: '%s' when writing '%s'" % (type(se), str(se), str(locals().get('_x', self))))) except TypeError as te: self._check_types(ValueError("%s: '%s' when writing '%s'" % (type(te), str(te), str(locals().get('_x', self))))) def deserialize(self, str): """ unpack serialized message in str into this message instance :param str: byte array of serialized message, ``str`` """ try: end = 0 _x = self start = end end += 16 (_x.steering_angle, _x.vel,) = _get_struct_2d().unpack(str[start:end]) return self except struct.error as e: raise genpy.DeserializationError(e) #most likely buffer underfill def serialize_numpy(self, buff, numpy): """ serialize message with numpy array types into buffer :param buff: buffer, ``StringIO`` :param numpy: numpy python module """ try: _x = self buff.write(_get_struct_2d().pack(_x.steering_angle, _x.vel)) except struct.error as se: self._check_types(struct.error("%s: '%s' when writing '%s'" % (type(se), str(se), str(locals().get('_x', self))))) except TypeError as te: self._check_types(ValueError("%s: '%s' when writing '%s'" % (type(te), str(te), str(locals().get('_x', self))))) def deserialize_numpy(self, str, numpy): """ unpack serialized message in str into this message instance using numpy for array types :param str: byte array of serialized message, ``str`` :param numpy: numpy python module """ try: end = 0 _x = self start = end end += 16 (_x.steering_angle, _x.vel,) = _get_struct_2d().unpack(str[start:end]) return self except struct.error as e: raise genpy.DeserializationError(e) #most likely buffer underfill _struct_I = genpy.struct_I def _get_struct_I(): global _struct_I return _struct_I _struct_2d = None def _get_struct_2d(): global _struct_2d if _struct_2d is None: _struct_2d = struct.Struct("<2d") return _struct_2d
[ "907098316@qq.com" ]
907098316@qq.com
b2f59b138ab429063808e0cfaee6e1c099c3be50
7313698d76e4606e290702463464fc8eeeff32c0
/Intermo/migrations/versions/20210512_184835_.py
d140577d9f0f34af9c85ed83b0e26419b21e7c29
[]
no_license
NathanFrank285/Intermo
53e55b576b4fcc2741b06f387e706fa6e8a43702
69ca818376dbeb22b100cc4a1e5873210fe52ce4
refs/heads/master
2023-06-23T05:54:07.452573
2021-07-20T21:14:27
2021-07-20T21:14:27
364,291,740
0
0
null
2021-07-20T21:14:27
2021-05-04T14:54:36
Python
UTF-8
Python
false
false
701
py
"""empty message Revision ID: 271520156ed4 Revises: 591c56ed2dac Create Date: 2021-05-12 18:48:35.174597 """ from alembic import op import sqlalchemy as sa # revision identifiers, used by Alembic. revision = '271520156ed4' down_revision = '591c56ed2dac' branch_labels = None depends_on = None def upgrade(): # ### commands auto generated by Alembic - please adjust! ### op.drop_constraint('trades_uniqueTradeId_key', 'trades', type_='unique') # ### end Alembic commands ### def downgrade(): # ### commands auto generated by Alembic - please adjust! ### op.create_unique_constraint('trades_uniqueTradeId_key', 'trades', ['uniqueTradeId']) # ### end Alembic commands ###
[ "nathansfrank@yahoo.com" ]
nathansfrank@yahoo.com
ae12d6fd85aad0110b05be6b293b0a3b5ce40ede
c9fea70a3dc5efafc718b5ede8e9971804cd5265
/venv/Scripts/pip3-script.py
1e4799b5e7995b569d60518ccbce4aa11d7c1567
[]
no_license
zekrihicham/creche
b4de8df3b92d9080e6d5a8ecf2dd4bdfa5759d64
c23ca4095777cbb9c7d5ccc25377eff63af460f4
refs/heads/master
2020-04-22T01:38:33.513119
2019-02-10T20:13:48
2019-02-10T20:13:48
170,020,607
0
0
null
null
null
null
UTF-8
Python
false
false
409
py
#!C:\Users\zekri\PycharmProjects\Creche\venv\Scripts\python.exe # EASY-INSTALL-ENTRY-SCRIPT: 'pip==9.0.1','console_scripts','pip3' __requires__ = 'pip==9.0.1' import re import sys from pkg_resources import load_entry_point if __name__ == '__main__': sys.argv[0] = re.sub(r'(-script\.pyw?|\.exe)?$', '', sys.argv[0]) sys.exit( load_entry_point('pip==9.0.1', 'console_scripts', 'pip3')() )
[ "36763072+zekrihicham@users.noreply.github.com" ]
36763072+zekrihicham@users.noreply.github.com
a14e1188bdfc65d5b6c7835a865851cf2b468dce
e3365bc8fa7da2753c248c2b8a5c5e16aef84d9f
/indices/srt.py
fe6f0681dbf56a25b61e42d88d931b804baf7678
[]
no_license
psdh/WhatsintheVector
e8aabacc054a88b4cb25303548980af9a10c12a8
a24168d068d9c69dc7a0fd13f606c080ae82e2a6
refs/heads/master
2021-01-25T10:34:22.651619
2015-09-23T11:54:06
2015-09-23T11:54:06
42,749,205
2
3
null
2015-09-23T11:54:07
2015-09-18T22:06:38
Python
UTF-8
Python
false
false
82
py
ii = [('ShawHDE.py', 1), ('AubePRP.py', 1), ('FerrSDO2.py', 1), ('ClarGE3.py', 1)]
[ "prabhjyotsingh95@gmail.com" ]
prabhjyotsingh95@gmail.com
294e27c16b37a8cdf7816d7381f0bd726197257e
86a85d81ae484939694b76bda29057e744022016
/src/tf_transformers/models/roberta.py
9ea34f9f5a0da63614f2727587e8f64514ed8c91
[ "Apache-2.0" ]
permissive
Vibha111094/tf-transformers
903f16224434d4ded9174f6fe2de8bb6bead1fd9
f26d440a4de0557e0e481279bfd70a732aaa8825
refs/heads/main
2023-07-27T16:11:22.101410
2021-07-23T15:33:39
2021-07-23T15:33:39
405,870,934
1
0
Apache-2.0
2021-09-13T07:15:31
2021-09-13T07:15:31
null
UTF-8
Python
false
false
49,343
py
from __future__ import absolute_import, division, print_function import tensorflow as tf from absl import logging from tf_transformers.activations import get_activation from tf_transformers.core import LegacyLayer from tf_transformers.layers import MLMLayer, OnDeviceEmbedding, SimplePositionEmbedding from tf_transformers.layers.mask import CausalMask, CrossAttentionMask, SelfAttentionMask, prefix_mask from tf_transformers.layers.transformer import TransformerBERT logging.set_verbosity("INFO") class ROBERTAEncoder(LegacyLayer): """RoBERTa based encoder / Decoder . RoBERTa: A Robustly Optimized BERT Pretraining Approach Authors: Yinhan Liu, Myle Ott, Naman Goyal, Jingfei Du, Mandar Joshi, Danqi Chen, Omer Levy, Mike Lewis, Luke Zettlemoyer, Veselin Stoyanov Implementation of Roberta in TF2.0 Paper: https://arxiv.org/abs/1907.11692 Official Code: https://github.com/pytorch/fairseq/tree/master/examples/roberta """ def __init__( self, config, mask_mode="user_defined", name=None, use_dropout=False, is_training=None, batch_size=None, sequence_length=None, use_type_embeddings=True, use_positonal_embeddings=True, pipeline_mode=None, is_decoder=False, cross_attention_inside_encoder=False, share_attention_layers=True, share_encoder_embeddings=False, encoder_embedding_layer=None, encoder_type_embedding_layer=None, encoder_positional_embedding_layer=None, use_mlm_layer=False, return_all_layer_token_embeddings=True, attention_type="full_attention", **kwargs, ): """ Args: config: dict mask_mode: str, `user_defined` BERT by default uses masking for PADDED or MLM. But can be overridden . # noqa name: str, Name of the model use_dropout: bool, It is strictly optional. Sometimes, while training you can set `use_dropout` to False. If `is_training` is False, `use_dropout` will be automatically set to False. # noqa batch_size: int, `batch_size` can be None or any int sequence_length: int, `sequence_length` can be None or any int use_type_embeddings: bool, By default BERT has type_embeddings, GPT2 don't. use_positonal_embeddings: bool, T5 don't have postional embeddings bidirectional: use in relative postional embedding (we can infer it based on mask_mode) is_decoder: bool, if True it will become decoder mode (as in Seq2Seq) use_mlm_layer: bool ( To use MLM layer or not ) share_encoder_embeddings: bool, When is_decoder = True, most cases, it will re-use the embedding layer from Encoder. So. if you still want to initialize , set this to False. If True, share embedding layers from encoder (word_embeddings, positional_embeddings, type_embeddings) cross_attention_inside_encoder: bool, Encoder Decoder Cross attention in each layer """ # Because saved_model causes some serialization problems here # self.config = config self.vocab_size = config["vocab_size"] self.type_vocab_size = config["type_vocab_size"] self.num_hidden_layers = config["num_hidden_layers"] self.num_attention_heads = config["num_attention_heads"] self.attention_head_size = config["attention_head_size"] self.max_position_embeddings = config["max_position_embeddings"] self.intermediate_size = config["intermediate_size"] self.embedding_size = config["embedding_size"] self.initializer_range = config["initializer_range"] self.hidden_act = config["hidden_act"] self.hidden_dropout_prob = config["hidden_dropout_prob"] self.attention_probs_dropout_prob = config["attention_probs_dropout_prob"] self.intermediate_act = config["intermediate_act"] self.layer_norm_epsilon = config["layer_norm_epsilon"] # Get activation and initiliazers self.activation = get_activation(self.hidden_act) self.intermediate_activation = get_activation(self.intermediate_act) initializer = tf.keras.initializers.TruncatedNormal(stddev=self.initializer_range) self.initializer = tf.keras.initializers.get(initializer) self.mask_mode = mask_mode # If we use self.name , its a conflict with keras property self.model_name = name self.pipeline_mode = pipeline_mode self.is_decoder = is_decoder # self._self_setattr_tracking = False self.mask_mode = mask_mode self.use_dropout = use_dropout self.is_training = is_training self.batch_size = batch_size self.sequence_length = sequence_length self.use_type_embeddings = use_type_embeddings self.use_positonal_embeddings = use_positonal_embeddings self.share_encoder_embeddings = share_encoder_embeddings self.share_attention_layers = share_attention_layers self.use_mlm_layer = use_mlm_layer self.cross_attention_inside_encoder = cross_attention_inside_encoder self.return_all_layer_token_embeddings = return_all_layer_token_embeddings self.attention_type = attention_type if not name.startswith("tf_transformers"): kwargs["name"] = "tf_transformers/" + self.model_name else: kwargs["name"] = self.model_name self.validate_and_set_inputs() super(ROBERTAEncoder, self).__init__(is_training=self.is_training, use_dropout=self.use_dropout, **kwargs) self._config_dict = { "initializer": tf.keras.initializers.serialize(initializer), "is_training": self.is_training, "use_dropout": self.use_dropout, "batch_size": self.batch_size, "sequence_length": self.sequence_length, "name": kwargs["name"], "use_type_embeddings": self.use_type_embeddings, "use_positonal_embeddings": self.use_positonal_embeddings, "is_decoder": self.is_decoder, "share_encoder_embeddings": self.share_encoder_embeddings, "share_attention_layers": self.share_attention_layers, "cross_attention_inside_encoder": cross_attention_inside_encoder, "return_all_layer_token_embeddings": self.return_all_layer_token_embeddings, } # Update config dict with passed config self._config_dict.update(config) # Call embedding layers self._embedding_layer, self._type_embeddings, self._position_embedding_layer = self.get_embedding_layers() if self.is_decoder: # If embedding has to shared from the encoder if self.share_encoder_embeddings: self._embedding_layer = encoder_embedding_layer self._type_embeddings = encoder_type_embedding_layer self._position_embedding_layer = encoder_positional_embedding_layer # Embedding Norm self._embedding_norm = tf.keras.layers.LayerNormalization( name="embeddings/layer_norm", axis=-1, epsilon=self.layer_norm_epsilon, dtype=tf.float32, ) # Embedding dropout Layer self._embedding_dropout = tf.keras.layers.Dropout(rate=self.hidden_dropout_prob) # Transformer Layer self._transformer_layers = [] for i in range(self.num_hidden_layers): layer = TransformerBERT( num_attention_heads=self.num_attention_heads, intermediate_size=self.intermediate_size, intermediate_activation=self.activation, dropout_rate=self.hidden_dropout_prob, attention_dropout_rate=self.attention_probs_dropout_prob, kernel_initializer=self.initializer, is_training=self.is_training, use_dropout=self.use_dropout, is_decoder=is_decoder, share_attention_layers=share_attention_layers, layer_norm_epsilon=self.layer_norm_epsilon, cross_attention_inside_encoder=self.cross_attention_inside_encoder, attention_type=self.attention_type, name="transformer/layer_%d" % i, ) self._transformer_layers.append(layer) self._pooler_layer = tf.keras.layers.Dense( units=self.embedding_size, activation="tanh", kernel_initializer=self.initializer, name="pooler_transform", ) if self.use_mlm_layer: self.mlm_layer = MLMLayer( self.embedding_size, self.layer_norm_epsilon, self.hidden_act, name="mlm_layer", ) self._last_logits_bias = self.add_weight( "tf_transformers/last_logits_bias", shape=(self.vocab_size,), dtype=tf.float32, trainable=True, ) self.call_fn = self.get_call_method() # Initialize model self.model_inputs, self.model_outputs = self.get_model(initialize_only=True) logging.info("Initialized Variables") def call_predict(self, inputs): """Inputs will be pass to this method, when is_training = False. The need to cache the past `key` and `value` tensors are necessary while predicting, to make the inference/NLG faster in case of AutoRegressive Decoding. """ input_ids_mod = inputs["input_ids"] all_cache_key = inputs["all_cache_key"] all_cache_value = inputs["all_cache_value"] past_length = inputs["past_length"] # Come from kwargs if self.mask_mode in ["user_defined", "prefix"]: input_mask = inputs["input_mask"] if self.use_type_embeddings: input_type_ids = inputs["input_type_ids"] # Convert past_length 2D to 1D past_length = tf.squeeze(past_length, 0) # In case of variable batch decoding, we will pad the inputs with -1 # So, we will replace -1 with 0, because -1 is not a valid index in word embeddings # >> input_ids_mod = [[ 1, 5, 7, 8, 10], # 2, 3, -1, -1, -1]] # # >> input_ids = [[1, 5, 7, 8,10], # 2, 3, 0, 0, 0]] input_ids = input_ids_mod * tf.cast(tf.not_equal(input_ids_mod, -1), tf.int32) sequence_length = tf.shape(input_ids)[1] # Asserting tf.assert_equal(tf.shape(all_cache_value)[0], self.num_hidden_layers) # Step 0 of inference. For step0, we do not have valid cache. We pass zero tensor def step_0(input_ids): sequence_length = tf.shape(input_ids)[1] position_embeddings = self._position_embedding_layer(tf.range(sequence_length)) return sequence_length, position_embeddings # From step_1 (autoregressive mode starts) onwards, we need to account for # `past_length` of previous words (inputs + generated) . Due to our logic, # we need to take a transpose of `position_embeddings` in this specific setting def step_other(input_ids): sequence_length = tf.shape(input_ids)[1] # Because past_length varies with batch position_embeddings = self._position_embedding_layer(past_length + sequence_length) position_embeddings = tf.transpose(position_embeddings, [1, 0, 2]) return sequence_length, position_embeddings # Condition to switch functions # if `sum(past_length) = 0` , means no outputs has been generated. # the given inputs is the first input sequence_length, positional_embeddings = tf.cond( tf.equal(tf.reduce_sum(past_length), 0), lambda: step_0(input_ids), lambda: step_other(input_ids), ) all_cache_key = [ tf.squeeze(item, axis=0) for item in tf.split(all_cache_key, num_or_size_splits=self.num_hidden_layers, axis=0) ] all_cache_value = [ tf.squeeze(item, axis=0) for item in tf.split(all_cache_value, num_or_size_splits=self.num_hidden_layers, axis=0) ] word_embeddings = self._embedding_layer(input_ids) embeddings = word_embeddings # Add word_embeddings + position_embeddings + type_embeddings if self.use_type_embeddings: type_embeddings = self._type_embeddings(input_type_ids) embeddings = embeddings + type_embeddings if self.use_positonal_embeddings: embeddings = embeddings + positional_embeddings # Norm + dropout embeddings = self._embedding_norm(embeddings) embeddings = self._embedding_dropout(embeddings, training=self.use_dropout) # Initialize `attention_mask` as empty list attention_mask = [] if self.mask_mode == "user_defined": attention_mask = SelfAttentionMask()([embeddings, input_mask]) if self.mask_mode == "prefix": attention_mask = tf.map_fn(prefix_mask, input_mask, fn_output_signature=tf.float32) if self.mask_mode == "causal": attention_mask = CausalMask()(embeddings) encoder_outputs = [] # Make all -1 positions to 0 (as -1 represents padding in the input) mask_values = tf.cast(tf.not_equal(input_ids_mod, -1), tf.float32) # We want zero values , where embeddings inputs where 0 (by replacing PAD -1) # So we use the mask and multiply it with embeddings embeddings = embeddings * tf.expand_dims(mask_values, -1) for i in range(self.num_hidden_layers): layer = self._transformer_layers[i] # Fetching cache_value = all_cache_value[i] cache_key = all_cache_key[i] embeddings, cache_key, cache_value = layer( [embeddings, attention_mask], cache_key=cache_key, cache_value=cache_value, ) # Updating all_cache_key[i] = cache_key all_cache_value[i] = cache_value # Mask next layer embedding (PAD positions to 0) embeddings = tf.identity( embeddings * tf.expand_dims(mask_values, -1), name="encoder_outputs_{}".format(i), ) encoder_outputs.append(embeddings) # First word of last layer outputs [CLS] cls_token_tensor = tf.keras.layers.Lambda(lambda x: tf.squeeze(x[:, 0:1, :], axis=1))(encoder_outputs[-1]) # batch_size x embedding_size cls_output = self._pooler_layer(cls_token_tensor) # batch_size x sequence_length x embedding_size token_embeddings = encoder_outputs[-1] # MLM Projection if self.use_mlm_layer: token_embeddings = self.mlm_layer(token_embeddings) # token --> vocab ( batch_size x sequence_length x vocab_size) token_logits = ( tf.matmul( token_embeddings, self.get_embedding_table(), transpose_b=True, name="token_logits", ) + self._last_logits_bias ) else: # token --> vocab ( batch_size x sequence_length x vocab_size) token_logits = tf.matmul( token_embeddings, self.get_embedding_table(), transpose_b=True, name="token_logits", ) def step_0_gather(past_length, token_embeddings): cache_length = tf.reduce_sum(tf.cast(tf.not_equal(input_ids_mod, -1), tf.int32), axis=1) - 1 # Getting corresponding last token tensor and last token logits last_token_tensor = tf.gather_nd(token_embeddings, tf.expand_dims(cache_length, axis=1), batch_dims=1) past_length = past_length + cache_length return past_length, last_token_tensor def step_other_gather(past_length, token_embeddings): past_length = past_length + sequence_length last_token_tensor = tf.keras.layers.Lambda(lambda x: x[:, -1, :])(token_embeddings) return past_length, last_token_tensor # Condition to switch functionsn (When batch_size > 1, # past_length will be different for each entry) # if `sum(past_length) = 0` , means no outputs has been generated. # the given inputs is the first input past_length, last_token_tensor = tf.cond( tf.equal(tf.reduce_sum(past_length), 0), lambda: step_0_gather(past_length, token_embeddings), lambda: step_other_gather(past_length, token_embeddings), ) # token --> vocab ( batch_size x sequence_length x vocab_size) last_token_logits = tf.matmul( last_token_tensor, self.get_embedding_table(), transpose_b=True, name="token_logits", ) # Expand dims of past_length back to 2D past_length = tf.expand_dims(past_length, 0, name="past_length") # Stack all layers key and value together # num_layers x batch_size x num_heads x sequence_length x (hidden_dimension/num_heads) all_cache_key = tf.stack(all_cache_key, axis=0, name="all_cache_key") all_cache_value = tf.stack(all_cache_value, axis=0, name="all_cache_value") return { "cls_output": cls_output, "token_logits": token_logits, "token_embeddings": token_embeddings, "last_token_logits": last_token_logits, "past_length": past_length, "all_cache_key": all_cache_key, "all_cache_value": all_cache_value, } def call_training(self, inputs): """Forward Pass for BERT Args: inputs: dict inputs is a dict with keys [`input_ids` , `input_mask`, `input_type_ids`]. These keys might or might not be present based on `mask_mode` and other criterias """ input_ids = inputs["input_ids"] # When `mask_mode` is `causal` , input_mask is not required if self.mask_mode in ["user_defined", "prefix"]: input_mask = inputs["input_mask"] # Default True in BERT if self.use_type_embeddings: input_type_ids = inputs["input_type_ids"] sequence_length = tf.shape(input_ids)[1] word_embeddings = self._embedding_layer(input_ids) embeddings = word_embeddings # Add word_embeddings + position_embeddings + type_embeddings if self.use_type_embeddings: type_embeddings = self._type_embeddings(input_type_ids) embeddings = embeddings + type_embeddings if self.use_positonal_embeddings: positional_embeddings = self._position_embedding_layer(tf.range(sequence_length)) embeddings = embeddings + positional_embeddings # Norm + dropout embeddings = self._embedding_norm(embeddings) embeddings = self._embedding_dropout(embeddings, training=self.use_dropout) if self.attention_type == "block_attention" or self.attention_type == "bigbird": attention_mask = input_mask else: # Initialize `attention_mask` as empty list attention_mask = [] if self.mask_mode == "user_defined": attention_mask = SelfAttentionMask()([embeddings, input_mask]) if self.mask_mode == "prefix": attention_mask = tf.map_fn(prefix_mask, input_mask, dtype=tf.float32) if self.mask_mode == "causal": attention_mask = CausalMask()(embeddings) encoder_outputs = [] for i in range(self.num_hidden_layers): layer = self._transformer_layers[i] embeddings, _, _ = layer([embeddings, attention_mask]) encoder_outputs.append(embeddings) # First word of last layer outputs [CLS] cls_token_tensor = tf.keras.layers.Lambda(lambda x: tf.squeeze(x[:, 0:1, :], axis=1))(encoder_outputs[-1]) # batch_size x embedding_size cls_output = self._pooler_layer(cls_token_tensor) # batch_size x sequence_length x embedding_size token_embeddings = encoder_outputs[-1] all_cls_output = [] for per_layer_token_embeddings in encoder_outputs: per_cls_token_tensor = tf.keras.layers.Lambda(lambda x: tf.squeeze(x[:, 0:1, :], axis=1))( per_layer_token_embeddings ) all_cls_output.append(self._pooler_layer(per_cls_token_tensor)) # MLM Projection if self.use_mlm_layer: token_embeddings = self.mlm_layer(token_embeddings) # token --> vocab ( batch_size x sequence_length x vocab_size) token_logits = ( tf.matmul( token_embeddings, self.get_embedding_table(), transpose_b=True, name="token_logits", ) + self._last_logits_bias ) else: # token --> vocab ( batch_size x sequence_length x vocab_size) token_logits = tf.matmul( token_embeddings, self.get_embedding_table(), transpose_b=True, name="token_logits", ) last_token_logits = tf.keras.layers.Lambda(lambda x: x[:, -1, :])(token_logits) result = { "cls_output": cls_output, "token_embeddings": token_embeddings, "token_logits": token_logits, "last_token_logits": last_token_logits, } if self.return_all_layer_token_embeddings: result["all_layer_token_embeddings"] = encoder_outputs result["all_layer_cls_output"] = all_cls_output return result def call_cross_attention_encoder(self, inputs): """[summary] Args: inputs ([type]): [description] """ encoder_input_ids = inputs["encoder_input_ids"] decoder_input_ids = inputs["decoder_input_ids"] encoder_input_type_ids = None decoder_input_type_ids = None if self.use_type_embeddings: encoder_input_type_ids = inputs["encoder_input_type_ids"] decoder_input_type_ids = inputs["decoder_input_type_ids"] encoder_input_mask = None if self.mask_mode in ["user_defined", "prefix"]: encoder_input_mask = inputs["encoder_input_mask"] def get_embeddings(input_ids, input_type_ids): """Get embedding for encoder as well as decoder Args: input_ids ([type]): [description] input_type_ids ([type]): [description] """ embeddings = self._embedding_layer(input_ids) sequence_length = tf.shape(input_ids)[1] # Add word_embeddings + position_embeddings + type_embeddings if self.use_type_embeddings: type_embeddings = self._type_embeddings(input_type_ids) embeddings = embeddings + type_embeddings if self.use_positonal_embeddings: positional_embeddings = self._position_embedding_layer(tf.range(sequence_length)) embeddings = embeddings + positional_embeddings # Norm + dropout embeddings = self._embedding_norm(embeddings) embeddings = self._embedding_dropout(embeddings, training=self.use_dropout) return embeddings encoder_embeddings = get_embeddings(encoder_input_ids, encoder_input_type_ids) decoder_embeddings = get_embeddings(decoder_input_ids, decoder_input_type_ids) # Initialize `encoder_attention_mask` as empty list encoder_attention_mask = [] if self.mask_mode == "user_defined": encoder_attention_mask = SelfAttentionMask()([encoder_embeddings, encoder_input_mask]) if self.mask_mode == "prefix": encoder_attention_mask = tf.map_fn(prefix_mask, encoder_input_mask, dtype=tf.float32) if self.mask_mode == "causal": encoder_attention_mask = CausalMask()(encoder_embeddings) # Decoder mask is always None decoder_attention_mask = CausalMask()(decoder_embeddings) decoder_encoder_mask = CrossAttentionMask()([decoder_input_ids, encoder_input_mask]) decoder_outputs = [] encoder_outputs = [] # Encoder Layer for i in range(self.num_hidden_layers): layer = self._transformer_layers[i] encoder_embeddings, _, _ = layer( [ encoder_embeddings, encoder_attention_mask, decoder_encoder_mask, # dummy decoder_encoder_mask encoder_embeddings, # dummy encoder_hidden_states ], mode="encoder", ) encoder_outputs.append(encoder_embeddings) # Decoder Layer encoder_hidden_states = encoder_outputs[-1] for i in range(self.num_hidden_layers): layer = self._transformer_layers[i] decoder_embeddings, _, _ = layer( [decoder_embeddings, decoder_attention_mask, decoder_encoder_mask, encoder_hidden_states], mode="decoder", ) decoder_outputs.append(decoder_embeddings) # First word of last layer outputs [CLS] cls_token_tensor = tf.keras.layers.Lambda(lambda x: tf.squeeze(x[:, 0:1, :], axis=1))(decoder_outputs[-1]) # batch_size x embedding_size cls_output = self._pooler_layer(cls_token_tensor) # batch_size x sequence_length x embedding_size token_embeddings = decoder_outputs[-1] # MLM Projection if self.use_mlm_layer: token_embeddings = self.mlm_layer(token_embeddings) # token --> vocab ( batch_size x sequence_length x vocab_size) token_logits = ( tf.matmul( token_embeddings, self.get_embedding_table(), transpose_b=True, name="token_logits", ) + self._last_logits_bias ) else: # token --> vocab ( batch_size x sequence_length x vocab_size) token_logits = tf.matmul( token_embeddings, self.get_embedding_table(), transpose_b=True, name="token_logits", ) last_token_logits = tf.keras.layers.Lambda(lambda x: x[:, -1, :])(token_logits) result = { "cls_output": cls_output, "token_embeddings": token_embeddings, "token_logits": token_logits, "last_token_logits": last_token_logits, } if self.return_all_layer_token_embeddings: result["all_layer_token_embeddings"] = decoder_outputs return result def call_cross_attention_encoder_predict(self, inputs): """[summary] Args: inputs ([type]): [description] """ encoder_input_ids = inputs["encoder_input_ids"] decoder_input_ids = inputs["decoder_input_ids"] encoder_input_type_ids = None decoder_input_type_ids = None if self.use_type_embeddings: encoder_input_type_ids = inputs["encoder_input_type_ids"] decoder_input_type_ids = inputs["decoder_input_type_ids"] encoder_input_mask = None if self.mask_mode in ["user_defined", "prefix"]: encoder_input_mask = inputs["encoder_input_mask"] # self.num_hidden_layers, batch_size, sequence_length, embeddingd_imension encoder_hidden_states = inputs["encoder_hidden_states"] all_cache_key = inputs["decoder_all_cache_key"] all_cache_value = inputs["decoder_all_cache_value"] def get_encoder_embeddings(input_ids, input_type_ids): """Get embedding for encoder as well as decoder Args: input_ids ([type]): [description] input_type_ids ([type]): [description] """ embeddings = self._embedding_layer(input_ids) sequence_length = tf.shape(input_ids)[1] # Add word_embeddings + position_embeddings + type_embeddings if self.use_type_embeddings: type_embeddings = self._type_embeddings(input_type_ids) embeddings = embeddings + type_embeddings if self.use_positonal_embeddings: positional_embeddings = self._position_embedding_layer(tf.range(sequence_length)) embeddings = embeddings + positional_embeddings # Norm + dropout embeddings = self._embedding_norm(embeddings) embeddings = self._embedding_dropout(embeddings, training=self.use_dropout) return embeddings # this function is slightly different from the other function # because, we do not need tf.range(sequence_length) # we need it for (one word) from, step 1 onwards, as we decode # word by word. So we use all_cache_key for getting the past_length def get_decoder_embeddings_step_other(input_ids, input_type_ids): """Get embedding for encoder as well as decoder Args: input_ids ([type]): [description] input_type_ids ([type]): [description] """ def step_0_cache_length(_): return tf.constant(0, dtype=tf.int32) def step_other_cache_length(all_cache_key): past_length = tf.shape(all_cache_key)[3] # Why -1, because When iter 2 (our positional # embedding should be 1 not 2 and so on) sequence_length = tf.shape(input_ids)[1] + past_length - 1 return sequence_length sequence_length = tf.cond( tf.equal(tf.reduce_sum(all_cache_key), 0), lambda: step_0_cache_length(all_cache_key), lambda: step_other_cache_length(all_cache_key), ) embeddings = self._embedding_layer(input_ids) # Add word_embeddings + position_embeddings + type_embeddings if self.use_type_embeddings: type_embeddings = self._type_embeddings(input_type_ids) embeddings = embeddings + type_embeddings if self.use_positonal_embeddings: positional_embeddings = self._position_embedding_layer(sequence_length) # Make it 3D for sum ( For decoder we decode one at a time) positional_embeddings = tf.expand_dims(positional_embeddings, 0) embeddings = embeddings + positional_embeddings # Norm + dropout embeddings = self._embedding_norm(embeddings) embeddings = self._embedding_dropout(embeddings, training=self.use_dropout) return embeddings # Encoder embeddings remains same throughout the decoding process # so we have to calculate it only once # So , we check if cache_key == 0, if its 0 its step 0 # else, pass a dummy encoder_embeddings, as we dont have to use it from step1 # because, what we need from encoder is encoder_hidden_states_batch encoder_embeddings = tf.cond( tf.equal(tf.reduce_sum(all_cache_key), 0.0), lambda: get_encoder_embeddings(encoder_input_ids, encoder_input_type_ids), lambda: tf.zeros_like(encoder_hidden_states), # dummy ) decoder_embeddings = tf.cond( tf.equal(tf.reduce_sum(all_cache_key), 0.0), lambda: get_encoder_embeddings(decoder_input_ids, decoder_input_type_ids), lambda: get_decoder_embeddings_step_other(decoder_input_ids, decoder_input_type_ids), ) # Initialize `encoder_attention_mask` as empty list encoder_attention_mask = [] if self.mask_mode == "user_defined": encoder_attention_mask = SelfAttentionMask()([encoder_embeddings, encoder_input_mask]) if self.mask_mode == "prefix": encoder_attention_mask = tf.map_fn(prefix_mask, encoder_input_mask, dtype=tf.float32) if self.mask_mode == "causal": encoder_attention_mask = CausalMask()(encoder_embeddings) # Decoder mask is always None decoder_attention_mask = CausalMask()(decoder_embeddings) decoder_encoder_mask = CrossAttentionMask()([decoder_input_ids, encoder_input_mask]) all_cache_key = [ tf.squeeze(item, axis=0) for item in tf.split(all_cache_key, num_or_size_splits=self.num_hidden_layers, axis=0) ] all_cache_value = [ tf.squeeze(item, axis=0) for item in tf.split(all_cache_value, num_or_size_splits=self.num_hidden_layers, axis=0) ] def calculate_encoder_hidden_state(encoder_embeddings): # Encoder Layer encoder_outputs = [] for i in range(self.num_hidden_layers): layer = self._transformer_layers[i] cache_key = all_cache_key[i] cache_value = all_cache_value[i] encoder_embeddings, _, _ = layer( [ encoder_embeddings, encoder_attention_mask, decoder_encoder_mask, # decoder_encoder_mask encoder_embeddings, ], mode="encoder", cache_key=cache_key, cache_value=cache_value, ) encoder_outputs.append(encoder_embeddings) encoder_hidden_states = encoder_outputs[-1] return encoder_hidden_states # While decoding we have to calculate it only once def use_cache_encoder(): return tf.identity(inputs["encoder_hidden_states"]) encoder_hidden_states = tf.cond( tf.equal(tf.reduce_sum(inputs["encoder_hidden_states"]), 0.0), lambda: calculate_encoder_hidden_state(encoder_embeddings), lambda: use_cache_encoder(), ) # Decoder layer decoder_outputs = [] for i in range(self.num_hidden_layers): layer = self._transformer_layers[i] # Fetching cache_value = all_cache_value[i] cache_key = all_cache_key[i] decoder_embeddings, cache_key, cache_value = layer( [ decoder_embeddings, decoder_attention_mask, decoder_encoder_mask, encoder_hidden_states, ], mode="decoder", cache_key=cache_key, cache_value=cache_value, ) # Updating all_cache_key[i] = cache_key all_cache_value[i] = cache_value decoder_outputs.append(decoder_embeddings) # Stack all layers key and value together # num_layers x batch_size x num_heads x sequence_length x # (hidden_dimension/num_heads) # noqa all_cache_key = tf.stack(all_cache_key, axis=0, name="decoder_all_cache_key") all_cache_value = tf.stack(all_cache_value, axis=0, name="decoder_all_cache_value") # First word of last layer outputs [CLS] cls_token_tensor = tf.keras.layers.Lambda(lambda x: tf.squeeze(x[:, 0:1, :], axis=1))(decoder_outputs[-1]) # batch_size x embedding_size cls_output = self._pooler_layer(cls_token_tensor) # batch_size x sequence_length x embedding_size token_embeddings = decoder_outputs[-1] # MLM Projection if self.use_mlm_layer: token_embeddings = self.mlm_layer(token_embeddings) # token --> vocab ( batch_size x sequence_length x vocab_size) token_logits = ( tf.matmul( token_embeddings, self.get_embedding_table(), transpose_b=True, name="token_logits", ) + self._last_logits_bias ) else: # token --> vocab ( batch_size x sequence_length x vocab_size) token_logits = tf.matmul( token_embeddings, self.get_embedding_table(), transpose_b=True, name="token_logits", ) last_token_logits = tf.keras.layers.Lambda(lambda x: x[:, -1, :])(token_logits) return { "encoder_hidden_states": encoder_hidden_states, "decoder_all_cache_key": all_cache_key, "decoder_all_cache_value": all_cache_value, "cls_output": cls_output, "token_embeddings": token_embeddings, "token_logits": token_logits, "last_token_logits": last_token_logits, } def call_decoder_predict(self, inputs): """Inputs will be pass to this method, when is_training = False and is_decoder = True. # noqa The need to cache the past `key` and `value` tensors for decoders necessary while predicting, to make the inference/NLG faster in case of AutoRegressive Decoding. """ input_ids = inputs["input_ids"] encoder_hidden_state = inputs["encoder_hidden_states"] decoder_encoder_mask = inputs["decoder_encoder_mask"] all_cache_key = inputs["all_cache_key"] all_cache_value = inputs["all_cache_value"] # Decoder don't need this # # When `mask_mode` is `causal` , input_mask is not required # if self.mask_mode in ['user_defined']: # input_mask = inputs['input_mask'] if self.use_type_embeddings: input_type_ids = inputs["input_type_ids"] # cache_length = tf.constant(0, dtype=tf.int32) def step_0_cache_length(_): return tf.constant(0, dtype=tf.int32) def step_other_cache_length(all_cache_key): past_length = tf.shape(all_cache_key)[3] # Why -1, because When iter 2 (our positional embedding should be 1 not 2 and so on) sequence_length = tf.shape(input_ids)[1] + past_length - 1 return sequence_length sequence_length = tf.cond( tf.equal(tf.reduce_sum(all_cache_key), 0), lambda: step_0_cache_length(all_cache_key), lambda: step_other_cache_length(all_cache_key), ) all_cache_key = [ tf.squeeze(item, axis=0) for item in tf.split(all_cache_key, num_or_size_splits=self.num_hidden_layers, axis=0) ] all_cache_value = [ tf.squeeze(item, axis=0) for item in tf.split(all_cache_value, num_or_size_splits=self.num_hidden_layers, axis=0) ] # If decoder is not sharing embeddings word_embeddings = self._embedding_layer(input_ids) embeddings = word_embeddings # Add word_embeddings + position_embeddings + type_embeddings if self.use_type_embeddings: type_embeddings = self._type_embeddings(input_type_ids) embeddings = embeddings + type_embeddings if self.use_positonal_embeddings: positional_embeddings = self._position_embedding_layer(sequence_length) # Make it 3D for sum ( For decoder we decode one at a time) positional_embeddings = tf.expand_dims(positional_embeddings, 0) embeddings = embeddings + positional_embeddings # Norm + dropout embeddings = self._embedding_norm(embeddings) embeddings = self._embedding_dropout(embeddings, training=self.use_dropout) # Initialize `attention_mask` as empty list attention_mask = [] if self.mask_mode == "causal": attention_mask = CausalMask()(embeddings) decoder_outputs = [] for i in range(self.num_hidden_layers): layer = self._transformer_layers[i] # Fetching cache_value = all_cache_value[i] cache_key = all_cache_key[i] embeddings, cache_key, cache_value = layer( [ embeddings, attention_mask, encoder_hidden_state, decoder_encoder_mask, ], cache_key=cache_key, cache_value=cache_value, ) # Updating all_cache_key[i] = cache_key all_cache_value[i] = cache_value decoder_outputs.append(embeddings) # Stack all layers key and value together # num_layers x batch_size x num_heads x sequence_length x (hidden_dimension/num_heads) all_cache_key = tf.stack(all_cache_key, axis=0, name="all_cache_key") all_cache_value = tf.stack(all_cache_value, axis=0, name="all_cache_value") # batch_size x sequence_length x embedding_size token_embeddings = decoder_outputs[-1] # MLM Projection if self.use_mlm_layer: token_embeddings = self.mlm_layer(token_embeddings) # token --> vocab ( batch_size x sequence_length x vocab_size) token_logits = ( tf.matmul( token_embeddings, self.get_embedding_table(), transpose_b=True, name="token_logits", ) + self._last_logits_bias ) else: # token --> vocab ( batch_size x sequence_length x vocab_size) token_logits = tf.matmul( token_embeddings, self.get_embedding_table(), transpose_b=True, name="token_logits", ) last_token_logits = tf.keras.layers.Lambda(lambda x: x[:, -1, :])(token_logits) return { "all_cache_key": all_cache_key, "all_cache_value": all_cache_value, "token_embeddings": token_embeddings, "token_logits": token_logits, "last_token_logits": last_token_logits, } def call_decoder(self, inputs): """Forward Pass for Decoder Args: inputs: dict inputs is a dict with keys [`input_ids` , `input_mask`, `input_type_ids`, `encoder_hidden_states`, `decoder_encoder_mask`]. These keys might or might not be present based on `mask_mode` and other criterias """ input_ids = inputs["input_ids"] encoder_output = inputs["encoder_hidden_states"] decoder_encoder_mask = inputs["decoder_encoder_mask"] if self.mask_mode in ["user_defined"]: input_mask = inputs["input_mask"] if self.use_type_embeddings: input_type_ids = inputs["input_type_ids"] sequence_length = tf.shape(input_ids)[1] # If decoder is not sharing embeddings word_embeddings = self._embedding_layer(input_ids) embeddings = word_embeddings # Add word_embeddings + position_embeddings + type_embeddings if self.use_type_embeddings: type_embeddings = self._type_embeddings(input_type_ids) embeddings = embeddings + type_embeddings if self.use_positonal_embeddings: positional_embeddings = self._position_embedding_layer(tf.range(sequence_length)) # positional_embeddings = self._position_embedding_layer(sequence_length) # # Make it 3D for sum ( For decoder we decode one at a time) # positional_embeddings = tf.expand_dims(positional_embeddings, 0) embeddings = embeddings + positional_embeddings # Norm + dropout embeddings = self._embedding_norm(embeddings) embeddings = self._embedding_dropout(embeddings, training=self.use_dropout) # Initialize `attention_mask` as empty list attention_mask = [] if self.mask_mode == "user_defined": attention_mask = SelfAttentionMask()([embeddings, input_mask]) if self.mask_mode == "causal": attention_mask = CausalMask()(embeddings) decoder_outputs = [] for i in range(self.num_hidden_layers): layer = self._transformer_layers[i] embeddings, _key, _value = layer([embeddings, attention_mask, encoder_output, decoder_encoder_mask]) decoder_outputs.append(embeddings) # batch_size x sequence_length x embedding_size token_embeddings = decoder_outputs[-1] # MLM Projection if self.use_mlm_layer: token_embeddings = self.mlm_layer(token_embeddings) # token --> vocab ( batch_size x sequence_length x vocab_size) token_logits = ( tf.matmul( token_embeddings, self.get_embedding_table(), transpose_b=True, name="token_logits", ) + self._last_logits_bias ) else: # token --> vocab ( batch_size x sequence_length x vocab_size) token_logits = tf.matmul( token_embeddings, self.get_embedding_table(), transpose_b=True, name="token_logits", ) last_token_logits = tf.keras.layers.Lambda(lambda x: x[:, -1, :])(token_logits) result = { "token_embeddings": token_embeddings, "token_logits": token_logits, "last_token_logits": last_token_logits, } if self.return_all_layer_token_embeddings: result["all_layer_token_embeddings"] = decoder_outputs return result def call(self, inputs): """Forward Pass. We have 2 pipelines . Training pipeline is relatively simpler Testing pipeline has few changes to accomodate caching of `key` and `value` for Transformer. Caching is significant for AutoRegressive modeling. Also, minor changes to make use of variable batch decoding Args: inputs, dict if self.is_training: self.call_training(inputs) else: self.call_predict(inputs) """ outputs = self.call_fn(inputs) return outputs def extend_positional_embeddings(self, factor): """Extends positional embeddings, by a factor. If factor = 2, we replicate the positional embeddings. If matrix is 512 x 768 , we convert it into 1024 x 768. Args: factor: int Returns: a new object of the class method """ if not isinstance(factor, int): raise ValueError(" `factor` must be an int with value > 1") # Squeeze is used to convert 3D to 2D updated_pos_embeddings = tf.squeeze(tf.repeat(self._position_embedding_layer.variables, factor, axis=1), 0) self.config["max_position_embeddings"] = 2 * self.config["max_position_embeddings"] tf.keras.backend.clear_session() new_layer = self.__class__( config=self.config, mask_mode=self.mask_mode, name=self.model_name, use_dropout=self.use_dropout, is_training=self.is_training, batch_size=self.batch_size, sequence_length=self.sequence_length, use_type_embeddings=self.use_type_embeddings, pipeline_mode=self.pipeline_mode, ) # layer to model to instantiate variables new_model = new_layer.get_model() del new_model model_new_dict = {} for var in self.variables: if "positional_embedding" in var.name: # Add the replicated previous embeddings to this embeddings model_new_dict[var.name] = updated_pos_embeddings else: model_new_dict[var.name] = var # Re assign it to model_new for var in new_layer.variables: var.assign(model_new_dict[var.name]) # Release the memory del model_new_dict logging.info("Succesfully changed position_embeddings to {}".format(updated_pos_embeddings.shape)) return new_layer def get_embedding_table(self): return self._embedding_layer.embeddings def get_config(self): return self._config_dict @property def transformer_layers(self): """List of Transformer layers in the encoder.""" return self._transformer_layers @classmethod def from_config(cls, config, custom_objects=None): return cls(**config)
[ "legacyai.org@gmail.com" ]
legacyai.org@gmail.com
3a3429686a4f84cd642ada4255092bcd590d5cb0
9a0704ef8a2ccd3aff58af8a9623578c5d4f7353
/app.py
86919f4ec581d3240bd567d28ff58e685ff4fefc
[]
no_license
rahulagrawal01/KNN-and-Naive-base-
d29396d6d886455ce9d62d844ac9c74d31e0c5f0
d62318410106fa0a707b39785b5514f083319cb5
refs/heads/main
2023-04-14T16:53:53.494660
2021-04-24T10:42:59
2021-04-24T10:42:59
361,135,119
0
0
null
null
null
null
UTF-8
Python
false
false
2,929
py
import streamlit as st from PIL import Image import pickle import numpy as np import matplotlib.pyplot as plt import pandas as pd st.set_option('deprecation.showfileUploaderEncoding', False) # Load the pickled model model = pickle.load(open('Knearestneighborclassifier.pkl', 'rb')) model_naive = pickle.load(open('naivebayesclassifier.pkl', 'rb')) dataset= pd.read_csv('titanic.csv') X=dataset[["Age","SibSp","Parch","Fare","Sex","Pclass"]] from sklearn.preprocessing import LabelEncoder labelencoder_X = LabelEncoder() X["Sex"] = labelencoder_X.fit_transform(X["Sex"]) from sklearn.preprocessing import StandardScaler sc = StandardScaler() X = sc.fit_transform(X) def predict_note_authentication(Age,SibSp,Parch,Sex,Fare,Pclass): output= model.predict(sc.transform([[Age,SibSp,Parch,Sex,Fare,Pclass]])) print("Passenger will die =", output) if output==[1]: prediction="Passanger will survive" else: prediction="Passanger will die" print(prediction) return prediction def predict_naive(Age,SibSp,Parch,Fare,Sex,Pclass): output= model_naive.predict(sc.transform([[Age,SibSp,Parch,Fare,Sex,Pclass]])) print("Passenger will die =", output) if output==[1]: prediction="Passanger will survive" else: prediction="Passanger will die" print(prediction) return prediction def main(): html_temp = """ <div class="" style="background-color:green;" > <div class="clearfix"> <div class="col-md-12"> <center><p style="font-size:35px;color:white;margin-top:10px;">Poornima Institute of Engineering & Technology</p></center> <center><p style="font-size:29px;color:white;margin-top:10px;">Department of Computer Engineering</p></center> <center><p style="font-size:26px;color:white;margin-top:10px;"Machine Learning Lab Experiment 4 KNN and Naive base Algo By Rahul Kr. Agrawal PIET18CS116 section 1</p></center> </div> </div> </div> """ st.markdown(html_temp,unsafe_allow_html=True) st.header("Passenger Survived Prediction using KNN And NB By Rahul 116") Sex = st.number_input('Insert 1 for Male 2 for Female 3 Others',1,3) Age = st.number_input('Insert a Age',18,60) SibSp = st.number_input('Insert a SibSp',0,10) Parch = st.number_input('Insert a Parch',1,10) Pclass = st.number_input('Insert a Pclass',1,8) Fare = st.number_input("Insert Fare",1,15000) resul="" if st.button("Predict by KNN "): result=predict_note_authentication(Age,SibSp,Parch,Fare,Sex,Pclass) st.success('KNN Model has predicted {}'.format(result)) if st.button("Predict by Naive Bayes "): result=predict_naive(Age,SibSp,Parch,Fare,Sex,Pclass) st.success('Naive Bayes Model has predicted {}'.format(result)) if st.button("About"): st.subheader("Developed by Rahul kumar agrawal PIET18CS116") st.subheader("Department of Computer Engineering Section C1") if __name__=='__main__': main()
[ "noreply@github.com" ]
rahulagrawal01.noreply@github.com
cf0e83ceacc341a163c92016f970b462250aa926
4d58c58165792bf22986c3cf53990ba17b28085a
/maya/python/lib/mpcJob.py
4e27b28d3f9522420d08f0b94fcfcab4e47d8368
[]
no_license
esernaalonso/dev
aabb54827db0f8507a0abbc24de5e1f7d27fe6c9
4fa41b8007e44dc0b574e372617ae65d198e27e9
refs/heads/master
2021-01-21T14:02:00.113420
2016-06-04T11:56:41
2016-06-04T11:56:41
47,589,157
1
0
null
null
null
null
UTF-8
Python
false
false
1,041
py
####################################### # imports import maya.cmds as cmds import maya.OpenMayaUI as apiUI import sys import time from mpc import jobtools as jobTls ####################################### # functionality def getJob(): """ Return the job from the environment. Returns: (str or None): Job, or None if no job is found. """ return jobTls.jobName() def getScene(): """ Returns the scene from the environment. Returns: (str or None): Scene, or None if no scene is found. """ return jobTls.sceneName() def getShot(): """ Returns the shot from the environment. Returns: (str or None): Shot, or None if no shot is found. """ return jobTls.shotName() def getPlayblastPath(): jobName = jobTls.jobName() sceneName = jobTls.sceneName() shotName = jobTls.shotName() if jobName and sceneName and shotName: return ("/jobs/" + jobName + "/" + sceneName + "/" + shotName + "/maya/playblasts/") else: return None ####################################### # execution if __name__ == "__main__": pass
[ "edusernalonso@gmail.com" ]
edusernalonso@gmail.com
2159710ed19f6e6b65f4a46e2509a0dbadb81e3b
119c716206804aef3eb99c5ca24e8e16eed3473a
/openaddr/tests/coverage.py
8693f0645fb3ed0e3777f5f06456c426eb04111b
[ "ISC" ]
permissive
enterstudio/machine-1
d190db7cd8fceb409e25232b65507ec21c1b7009
43d4d3b41e1ad3410c2442b8220e6e1a9fe2255f
refs/heads/master
2021-08-30T08:06:05.055323
2017-12-16T23:30:56
2017-12-16T23:30:56
null
0
0
null
null
null
null
UTF-8
Python
false
false
35,950
py
import unittest import unittest.mock import os import psycopg2 from httmock import HTTMock, response DATABASE_URL = os.environ.get('DATABASE_URL', 'postgres:///hooked_on_sources') from ..ci import recreate_db from ..ci.coverage import calculate class TestCalculate (unittest.TestCase): def setUp(self): ''' ''' recreate_db.recreate(DATABASE_URL) with psycopg2.connect(DATABASE_URL) as conn: with conn.cursor() as db: db.execute("insert into cb_2013_us_state_20m (gid, name, usps_code, geom) values (1, 'Kansas', 'KS', ST_SetSRID('MULTIPOLYGON(((-102.0472 40.0033, -94.6143 40.0033, -94.6143 36.9985, -102.0472 36.9985, -102.0472 40.0033)))'::geometry, 4326))") db.execute("insert into ne_50m_admin_0_countries (gid, name, name_long, iso_a2, iso_a3, geom) values (1, 'Null Is.', 'Null Island', 'XX', 'XXX', '0106000020E610000002000000010300000001000000270000008EB1135E82533FBF691D554D1075EF3E90BE49D3A0683EBF664CC11A67D3F13EA208A9DBD9573EBF3FABCC94D6DFE23EEC3026FDBD143EBF8DEDB5A0F7C6E03E3659A31EA2D13DBF664CC11A67D3F13E13D21A834E083DBF54E41071732AE93EDCF3FC69A33A3DBF8DEDB5A0F7C6D03E0188BB7A15193DBF0000000000000000CCB6D3D688603CBF54E41071732AC93E0395F1EF332E3CBFF168E388B5F8E4BE7044F7AC6BB43CBF05A227655243EBBE7044F7AC6BB43CBF8DEDB5A0F7C6F0BE2829B000A60C3CBF664CC11A67D3F1BE16DF50F86C1D3CBF180AD80E46EC03BF7044F7AC6BB43CBF180AD80E46EC03BFA72215C616823CBF3677F4BF5C8B06BFEE3D5C72DC293DBF05A2276552430BBFA515DF50F86C3DBFC093162EABB009BF3659A31EA2D13DBF2D431CEBE2360ABF3659A31EA2D13DBF240F441669E20DBF344C6DA983BC3EBF240F441669E20DBF344C6DA983BC3EBFC093162EABB009BFD7D9907F66103FBFC093162EABB009BF691D554D10753FBF4BB0389CF9D50CBF61C1FD80070640BFE7340BB43BA408BFD5CC5A0A48FB3FBFA226FA7C941107BF213EB0E3BF4040BF664CC11A67D301BF4F779E78CE1640BFB75F3E59315CFDBE4F779E78CE1640BF05A227655243FBBEFCA9F1D24D6240BF54E41071732AF9BE213EB0E3BF4040BFF168E388B5F8E4BE46D26EF4311F40BFF168E388B5F8E4BE46D26EF4311F40BF54E41071732AD9BE2AE3DF675C3840BF00000000000000801FF5D72B2CB83FBF3FABCC94D6DFE23E56D3F544D7853FBF54E41071732AD93EA0FB7266BB423FBFB75F3E59315CDD3EA0FB7266BB423FBFA226FA7C9411E73E8EB1135E82533FBF691D554D1075EF3E010300000003000000C7010000C9C7EE022505163F7BA35698BED7303F2829B000A60C1C3F7689EAAD81AD323F1FF5D72B2CB81F3F41B8020AF5F4313F86C613419C87233F0ADAE4F04927323FA019C40776FC273FD2FBC6D79E59323F7D923B6C2233273F7689EAAD81AD323F9E0C8E9257E7283F613255302AA9333FE527D53E1D8F293FAB5AD2510E66333F99F221A81ABD2A3F3D9E961FB8CA333F72512D228AC92B3F2A5437177FDB333F261C7A8B87F72C3F3D9E961FB8CA333F240F441669E22D3FE02BBAF59A1E343FFA60191BBAD92F3FF5824F73F222333F8EB1135E82532F3F3FABCC94D6DF323FFA60191BBAD92F3FF78F85E81038323FFA60191BBAD92F3FE7525C55F65D313F0EF450DB8651303F1F317A6EA12B313FB345D26EF4312F3FD5CC5A0A48FB2F3F6B2A8BC22E8A2E3F46D26EF4311F303F92CB7F48BF7D2D3F4489963C9E962F3FBC79AA436E862B3F691D554D10752F3FE31A9FC9FE792A3F1FF5D72B2CB82F3FC093162EABB0293FD7D9907F66102F3FA019C40776FC273F33880FECF82F303F315D88D51F61283FE95F92CA1473303F6859F78F85E8203F46D26EF4311F303F664CC11A67D3213F581CCEFC6A0E303FF5824F73F222233F6B2A8BC22E8A2E3F3FABCC94D6DF223F48A30227DBC02D3FAB5AD2510E66233F92CB7F48BF7D2D3F3D9E961FB8CA233F95D8B5BDDD922C3FA94D9CDCEF50243F95D8B5BDDD922C3FF168E388B5F8243FDCF3FC69A33A2D3FEC4E779E78CE263FB96C74CE4F712C3FEA4141295AB9273F261C7A8B87F72C3F56F146E6913F283F261C7A8B87F72C3F0CC9C9C4AD82283FDE0033DFC14F2C3F97E5EB32FCA72B3FB96C74CE4F712C3FDE0033DFC14F2C3F4DBD6E1118EB2B3F240F441669E22D3FBE86E0B88C9B2A3F8EB1135E82532F3FBE86E0B88C9B2A3F213EB0E3BF40303F52D7DAFB54152A3F6859F78F85E8303F52D7DAFB54152A3F560F98874CF9303FC093162EABB0293F6859F78F85E8303F9E0C8E9257E7283F317BD976DA1A313F56F146E6913F283F1F317A6EA12B313F13F06B2409C2253F6859F78F85E8303F3A9160AA99B5243F8EB1135E82532F3F3A9160AA99B5243FB345D26EF4312F3F3FABCC94D6DF223F1FF5D72B2CB82F3FAE6708C72C7B223F46D26EF4311F303F1A170E846401233FE95F92CA1473303F3FABCC94D6DF223FC5CBD3B9A294303FF78F85E81038223F7BA35698BED7303FAE6708C72C7B223F1F317A6EA12B313FF78F85E81038223F6859F78F85E8303FFA9CBB5D2F4D213F560F98874CF9303FB28174B169A5203FE7525C55F65D313FFCA9F1D24D62203F8BE07F2BD9B1313F1FF5D72B2CB81F3F41B8020AF5F4313F213EB0E3BF40203FC0B167CF656A323FFCA9F1D24D62203F7689EAAD81AD323F691D554D10751F3FF5824F73F222333F691D554D10751F3FAB5AD2510E66333F90BE49D3A0681E3FAB5AD2510E66333F2829B000A60C1C3F4FE8F527F1B9333F2829B000A60C1C3F180AD80E46EC333F4BB0389CF9D51C3FBB97FBE42840343F4BB0389CF9D51C3F4DDBBFB2D2A4343F4FCAA48636001B3F3A9160AA99B5343F9E0C8E9257E7183FB98AC56F0A2B353FC5AD8218E8DA173F5D18E945ED7E353FE7340BB43BA4183F01A60C1CD0D2353FC5AD8218E8DA173FB77D8FFAEB15363F3677F4BF5C8B163F3677F4BF5C8B363F5D18E945ED7E153FEC4E779E78CE363F5D18E945ED7E153FC7BAB88D06F0363F3A9160AA99B5143FB5705985CD00373FAB5AD2510E66133F58FE7C5BB054373FF5824F73F222133F58FE7C5BB054373FCEE15AED612F143F21205F420587373FCEE15AED612F143FD7F7E12021CA373F3FABCC94D6DF123FD7F7E12021CA373FD2FBC6D79E59123F1F1329CDE671383FD2FBC6D79E59123FB056ED9A90D6383F664CC11A67D3113FB056ED9A90D6383F8DEDB5A0F7C6103F693BA6EECA2E383F05A2276552430B3F315D88D51F61383FC093162EABB0093F315D88D51F61383F3677F4BF5C8B063F7B8505F7031E383F180AD80E46EC033FD7F7E12021CA373F664CC11A67D3F13E6B48DC63E943373F180AD80E46ECF33EFE98D6A6B1BD363F8DEDB5A0F7C6E03EDC114E0B5EF4353F691D554D1075EF3E263ACB2C42B1353F54E41071732AE93E263ACB2C42B1353F54E41071732AD93E94F6065F984C353F54E41071732AD93E4ACE893DB48F353F8DEDB5A0F7C6E0BECCD42478431A353FA226FA7C9411E7BECCD42478431A353F180AD80E46ECF3BE15FDA19927D7343FF168E388B5F8F4BE15FDA19927D7343FB75F3E59315CFDBE84B9DDCB7D72343F90BE49D3A068FEBEF37519FED30D343F54E41071732AF9BE2A5437177FDB333F2D431CEBE236FABE05C078060DFD333FAB5AD2510E6603BF99107349D576333F180AD80E46EC03BF99107349D576333FA226FA7C941107BF7689EAAD81AD323FA226FA7C941107BF7689EAAD81AD323F54E41071732A09BFE54526E0D748323FC093162EABB009BF1C2444F98216323F4BB0389CF9D50CBFF78F85E81038323F691D554D10750FBFF78F85E81038323FB0743E3C4B9011BFC0B167CF656A323FFA9CBB5D2F4D11BF7689EAAD81AD323FD2FBC6D79E5912BFBEA4315A4755333F8DEDB5A0F7C610BF86C613419C87333F89D349B6BA9C12BF180AD80E46EC333FD2FBC6D79E5912BFF37519FED30D343FAB5AD2510E6613BF180AD80E46EC333F3A9160AA99B514BFCEE15AED612F343F3A9160AA99B514BFA94D9CDCEF50343F315D88D51F6118BF05C078060DFD333FE7340BB43BA418BF05C078060DFD333F766B990CC7F319BF86C613419C87333FBC79AA436E861BBF86C613419C87333F4BB0389CF9D51CBF613255302AA9333F4BB0389CF9D51CBF180AD80E46EC333F72512D228AC91BBFA94D9CDCEF50343F95D8B5BDDD921CBFF37519FED30D343F90BE49D3A0681EBFA94D9CDCEF50343FFCA9F1D24D6220BF97033DD4B661343F6859F78F85E820BFBB97FBE42840343FD508FD4CBD6E21BF4DDBBFB2D2A4343FD2FBC6D79E5922BFF168E388B5F8343FD2FBC6D79E5922BFCCD42478431A353F3FABCC94D6DF22BFF168E388B5F8343F86C613419C8723BF94F6065F984C353F5F251FBB0B9424BFC9C7EE022505363F5F251FBB0B9424BFB77D8FFAEB15363FF168E388B5F824BF01A60C1CD0D2353F38842A357BA025BF01A60C1CD0D2353FA43330F2B22626BFC9C7EE022505363F11E335AFEAAC26BFDC114E0B5EF4353FA226FA7C941127BF13F06B2409C2353F58FE7C5BB05427BF13F06B2409C2353FC5AD8218E8DA27BF4ACE893DB48F353FA019C40776FC27BF5D18E945ED7E353F0CC9C9C4AD8228BFEE5BAD1397E3353F2F505260014C29BFEE5BAD1397E3353F745E6397A8DE2ABF263ACB2C42B1353F4FCAA48636002BBF38842A357BA0353F7044F7AC6BB42CBF03B34291EEE7343F4BB0389CF9D52CBFF168E388B5F8343FDAE6C6F484252EBF284701A260C6343FB55208E412472EBF3A9160AA99B5343FD7D9907F66102FBF84B9DDCB7D72343FD7D9907F66102FBF97033DD4B661343F0EF450DB865130BFCEE15AED612F343FA03715A930B630BF3D9E961FB8CA333FC5CBD3B9A29430BF99107349D576333F560F98874CF930BF643F8BA548BE323F43C5387F130A31BFAE6708C72C7B323FC5CBD3B9A29430BF41B8020AF5F4313FC5CBD3B9A29430BF8BE07F2BD9B1313F46D26EF4311F30BF317BD976DA1A313FA03715A930B630BF6859F78F85E8303FB28174B169A530BF8DEDB5A0F7C6303F213EB0E3BF4030BFC5CBD3B9A294303F213EB0E3BF4030BFC5CBD3B9A294303FB28174B169A530BF0EF450DB8651303F317BD976DA1A31BFFCA9F1D24D62303FB0743E3C4B9031BF581CCEFC6A0E303F664CC11A67D331BFB345D26EF4312F3F41B8020AF5F431BFD7D9907F66102F3FD2FBC6D79E5932BF6B2A8BC22E8A2E3FD2FBC6D79E5932BF240F441669E22D3F1C2444F9821632BF4BB0389CF9D52C3FE54526E0D74832BF7044F7AC6BB42C3F78962023A0C231BF0395F1EF332E2C3F9D2ADF3312A131BF0395F1EF332E2C3F41B8020AF5F431BFE31A9FC9FE792A3F41B8020AF5F431BF766B990CC7F3293FE54526E0D74832BF54E41071732A293F1C2444F9821632BFE7340BB43BA4283FE54526E0D74832BF7B8505F7031E283F2F6EA301BC0532BFEA4141295AB9273F540262122EE431BFEA4141295AB9273FB0743E3C4B9031BFEA4141295AB9273F43C5387F130A31BFEA4141295AB9273F8DEDB5A0F7C630BFEC4E779E78CE263F7BA35698BED730BF5D18E945ED7E253F33880FECF82F30BF5D18E945ED7E253F1FF5D72B2CB82FBF15FDA19927D7243FD5CC5A0A48FB2FBF84B9DDCB7D72243FD5CC5A0A48FB2FBFA94D9CDCEF50243FB345D26EF4312FBF180AD80E46EC233FD7D9907F66102FBF3D9E961FB8CA233FB55208E412472EBF3A9160AA99B5243F240F441669E22DBF3A9160AA99B5243F4BB0389CF9D52CBFA7406667D13B253F4DBD6E1118EB2BBFCCD42478431A253FE10D6954E0642BBFCCD42478431A253F745E6397A8DE2ABF180AD80E46EC233FE10D6954E0642BBF3D9E961FB8CA233F99F221A81ABD2ABF1A170E846401233F2A36E675C4212BBF643F8BA548BE223F4FCAA48636002BBFD2FBC6D79E59223FE527D53E1D8F29BF664CC11A67D3213FE527D53E1D8F29BF1F317A6EA12B213FE7340BB43BA428BF1C2444F98216223F336ABE4A3E7627BFF78F85E81038223F809F71E1404826BF89D349B6BA9C223FC9C7EE02250526BF89D349B6BA9C223F38842A357BA025BF8BE07F2BD9B1213FC9C7EE02250526BFD508FD4CBD6E213FEE5BAD1397E325BF8DEDB5A0F7C6203F5B0BB3D0CE6926BF1FF5D72B2CB81F3F82ACA7565F5D25BF240F441669E21D3F38842A357BA025BF240F441669E21D3F15FDA19927D724BF72512D228AC91B3FF168E388B5F824BF72512D228AC91B3FA94D9CDCEF5024BF05A2276552431B3F3D9E961FB8CA23BFBC79AA436E861B3FD0EE9062804423BFDE0033DFC14F1C3F3FABCC94D6DF22BFDE0033DFC14F1C3FF78F85E8103822BFE31A9FC9FE791A3F89D349B6BA9C22BF2D431CEBE2361A3F5F251FBB0B9424BFE31A9FC9FE791A3FA7406667D13B25BFC093162EABB0193F38842A357BA025BF315D88D51F61183F82ACA7565F5D25BFC5AD8218E8DA173F3A9160AA99B524BF7B8505F7031E083F3A9160AA99B524BF7B8505F7031E083F82ACA7565F5D25BF05A2276552430B3F38842A357BA025BFFA9CBB5D2F4D113FA43330F2B22626BF84B9DDCB7D72143FA226FA7C941127BF58FE7C5BB054173F58FE7C5BB05427BF54E41071732A193F0FD6FF39CC9727BF0ABC934F8F6D193F7B8505F7031E28BF0FD6FF39CC97173F7978CF81E50829BF13F06B2409C2153F2F505260014C29BF3FABCC94D6DF123F99F221A81ABD2ABF43C5387F130A113FE10D6954E0642BBFC093162EABB0093FDE0033DFC14F2CBF213EB0E3BF40003FDCF3FC69A33A2DBF54E41071732AC93E240F441669E22DBFA226FA7C9411F7BEDAE6C6F484252EBF7B8505F7031E08BFDAE6C6F484252EBFCEE15AED612F14BFDAE6C6F484252EBFA7406667D13B15BFBE86E0B88C9B2ABFA7406667D13B15BF7B8505F7031E28BF13F06B2409C215BF315D88D51F6128BF0FD6FF39CC9717BF56F146E6913F28BFE7340BB43BA418BFE7340BB43BA428BFE7340BB43BA418BF2F505260014C29BF0FD6FF39CC9717BF9BFF571D39D229BFC5AD8218E8DA17BF2D431CEBE2362ABF9E0C8E9257E718BF08AF5DDA70582ABF0ABC934F8F6D19BFE10D6954E0642BBF2D431CEBE2361ABF0395F1EF332E2CBF2829B000A60C1CBF4DBD6E1118EB2BBF95D8B5BDDD921CBF95D8B5BDDD922CBF4696CCB1BCAB1EBF7044F7AC6BB42CBFD5CC5A0A48FB1FBF0395F1EF332E2CBFB28174B169A520BFDE0033DFC14F2CBFD508FD4CBD6E21BF7044F7AC6BB42CBF41B8020AF5F421BFB96C74CE4F712CBF1C2444F9821622BF4DBD6E1118EB2BBF3FABCC94D6DF22BF4DBD6E1118EB2BBF1A170E84640123BFE10D6954E0642BBF613255302AA923BF2A36E675C4212BBFCEE15AED612F24BFBC79AA436E862BBF5F251FBB0B9424BF4FCAA48636002BBF82ACA7565F5D25BF4FCAA48636002BBF13F06B2409C225BFBC79AA436E862BBF5B0BB3D0CE6926BFBC79AA436E862BBF3677F4BF5C8B26BF99F221A81ABD2ABF11E335AFEAAC26BF2D431CEBE2362ABF7D923B6C223327BFE527D53E1D8F29BFA019C40776FC27BFC093162EABB029BF0CC9C9C4AD8228BF54E41071732A29BF54E41071732A29BF54E41071732A29BFE527D53E1D8F29BFC3A04CA3C9C528BFE527D53E1D8F29BFA019C40776FC27BF766B990CC7F329BFC5AD8218E8DA27BF766B990CC7F329BFC7BAB88D06F026BFE527D53E1D8F29BFEE5BAD1397E325BF766B990CC7F329BF13F06B2409C225BF08AF5DDA70582ABFC7BAB88D06F026BF4FCAA48636002BBFA226FA7C941127BF05A2276552432BBF336ABE4A3E7627BF97E5EB32FCA72BBF0FD6FF39CC9727BF0395F1EF332E2CBF7D923B6C223327BF261C7A8B87F72CBFA226FA7C941127BF0188BB7A15192DBF809F71E1404826BF6D37C1374D9F2DBFC9C7EE02250526BF6D37C1374D9F2DBF84B9DDCB7D7224BFDAE6C6F484252EBF84B9DDCB7D7224BFDAE6C6F484252EBF613255302AA923BFB345D26EF4312FBFAE6708C72C7B22BFB345D26EF4312FBFB0743E3C4B9021BFFC6D4F90D8EE2EBF1F317A6EA12B21BFB345D26EF4312FBF6859F78F85E820BFB345D26EF4312FBF46D26EF4311F20BFFA60191BBAD92FBF46D26EF4311F20BF46D26EF4311F30BF43C5387F130A21BFFCA9F1D24D6230BF43C5387F130A21BFD71533C2DB8330BFB28174B169A520BF8DEDB5A0F7C630BFB28174B169A520BF6859F78F85E830BF46D26EF4311F20BF7BA35698BED730BFDAE6C6F484251EBFFA9CBB5D2F4D31BF95D8B5BDDD921CBF78962023A0C231BF240F441669E21DBF41B8020AF5F431BF0188BB7A15191DBF540262122EE431BFBC79AA436E861BBF0ADAE4F0492732BF4FCAA48636001BBF1C2444F9821632BF766B990CC7F319BF643F8BA548BE32BF2D431CEBE2361ABF1A170E84640133BF05A2276552431BBFBEA4315A475533BFC093162EABB019BFBEA4315A475533BF9E0C8E9257E718BF613255302AA933BFE7340BB43BA418BFBB97FBE4284034BF0FD6FF39CC9717BF726F7EC3448334BF54E41071732A19BFB98AC56F0A2B35BF315D88D51F6118BF6F62484E266E35BFA226FA7C941117BFC9C7EE02250536BF315D88D51F6118BF92E9D0E9793736BF7B8505F7031E18BF92E9D0E9793736BF13F06B2409C215BF48C153C8957A36BF5D18E945ED7E15BF48C153C8957A36BF180AD80E46EC13BF5B0BB3D0CE6936BFAB5AD2510E6613BF5B0BB3D0CE6936BF89D349B6BA9C12BFFE98D6A6B1BD36BF89D349B6BA9C12BF232D95B7239C36BF90BE49D3A0680EBFB5705985CD0037BF99F221A81ABD0ABFC7BAB88D06F036BF3677F4BF5C8B06BF7D923B6C223337BFF168E388B5F804BF21205F42058737BF3677F4BF5C8B06BFEA4141295AB937BFAB5AD2510E6603BF7B8505F7031E38BF180AD80E46EC03BF44A7E7DD585038BFD2FBC6D79E5902BF44A7E7DD585038BF05A227655243FBBE0CC9C9C4AD8238BF2D431CEBE236FABEB056ED9A90D638BF2D431CEBE236FABEC3A04CA3C9C538BF05A227655243EBBE419AB1683A3B39BFA226FA7C9411E7BE419AB1683A3B39BF00000000000000807978CF81E50839BF8DEDB5A0F7C6C03E7978CF81E50839BF54E41071732AE93EF8713447567E39BF54E41071732AE93ED3DD7536E49F39BF691D554D1075EF3EC093162EABB039BFC9C7EE022505F63E662E7079AC1939BFFA9CBB5D2F4D013F54E41071732A39BF180AD80E46EC033FFA7E6ABC749338BF05A2276552430B3F7B8505F7031E38BFDE0033DFC14F0C3F693BA6EECA2E38BF691D554D10750F3FC5AD8218E8DA37BF8DEDB5A0F7C6103F0FD6FF39CC9737BF8DEDB5A0F7C6103F90DC9A745B2237BF8DEDB5A0F7C6103F48C153C8957A36BF1C2444F98216123FA43330F2B22636BF1C2444F98216123F13F06B2409C235BFD2FBC6D79E59123F6F62484E266E35BFCEE15AED612F143F6F62484E266E35BFC9C7EE022505163F15FDA19927D734BF58FE7C5BB054173F97033DD4B66134BF0FD6FF39CC97173FF37519FED30D34BFE7340BB43BA4183F99107349D57633BFE7340BB43BA4183F86C613419C8733BF05A2276552431B3F2C616D8C9DF032BF4BB0389CF9D51C3F9B1DA9BEF38B32BF72512D228AC91B3F2F6EA301BC0532BF4FCAA48636001B3F1C2444F9821632BF90BE49D3A0681E3F7689EAAD81AD32BF4696CCB1BCAB1E3F7689EAAD81AD32BF213EB0E3BF40203FF78F85E8103832BF8DEDB5A0F7C6203FF78F85E8103832BF643F8BA548BE223F1C2444F9821632BF1A170E846401233F8BE07F2BD9B131BFF5824F73F222233FFA9CBB5D2F4D31BFF168E388B5F8243F1F317A6EA12B31BF38842A357BA0253FD71533C2DB8330BFA7406667D13B253F46D26EF4311F30BF11E335AFEAAC263F691D554D10752FBF3677F4BF5C8B263FFF7A8505F7032EBF0FD6FF39CC97273FDCF3FC69A33A2DBF0FD6FF39CC97273F4BB0389CF9D52CBF7978CF81E508293F0395F1EF332E2CBF2F505260014C293F97E5EB32FCA72BBFBE86E0B88C9B2A3F9BFF571D39D229BF0395F1EF332E2C3FE527D53E1D8F29BF0188BB7A15192D3FEA4141295AB927BFB96C74CE4F712C3FC5AD8218E8DA27BF72512D228AC92B3FE7340BB43BA428BFE31A9FC9FE792A3F315D88D51F6128BF9BFF571D39D2293F315D88D51F6128BF2F505260014C293FE7340BB43BA428BFA019C40776FC273FA019C40776FC27BFA019C40776FC273FA226FA7C941127BFEC4E779E78CE263F3677F4BF5C8B26BFEC4E779E78CE263F82ACA7565F5D25BF56F146E6913F283F15FDA19927D724BF315D88D51F61283FCEE15AED612F24BF9E0C8E9257E7283FAB5AD2510E6623BF7B8505F7031E283F1C2444F9821622BF7978CF81E508293FD71533C2DB8320BFC3A04CA3C9C5283FFCA9F1D24D6220BFC5AD8218E8DA273FFC6D4F90D8EE1EBFEA4141295AB9273F4696CCB1BCAB1EBF9E0C8E9257E7283FBC79AA436E861BBF7978CF81E508293FBC79AA436E861BBF97E5EB32FCA72B3FDE0033DFC14F1CBF95D8B5BDDD922C3F4FCAA48636001BBF4BB0389CF9D52C3F99F221A81ABD1ABFDAE6C6F484252E3FC5AD8218E8DA17BF21020EA14ACD2E3FEC4E779E78CE16BF6B2A8BC22E8A2E3FEC4E779E78CE16BF691D554D10752F3F13F06B2409C215BF581CCEFC6A0E303F613255302AA913BF33880FECF82F303F43C5387F130A11BFB345D26EF4312F3FFC6D4F90D8EE0EBFD7D9907F66102F3F90BE49D3A0680EBF4696CCB1BCAB2E3FF168E388B5F804BF6B2A8BC22E8A2E3F84B9DDCB7D7204BF4489963C9E962F3F691D554D1075FFBEFA60191BBAD92F3F691D554D1075FFBEFC6D4F90D8EE2E3F05A227655243FBBEFC6D4F90D8EE2E3F3FABCC94D6DFF2BE0188BB7A15192D3FF168E388B5F8E4BE0188BB7A15192D3FB75F3E59315CDDBEB96C74CE4F712C3F54E41071732AC9BEB96C74CE4F712C3F54E41071732AC93E97E5EB32FCA72B3FF168E388B5F8E43E0395F1EF332E2C3F54E41071732AE93E0188BB7A15192D3F664CC11A67D3F13E0188BB7A15192D3F691D554D1075FF3E6B2A8BC22E8A2E3F664CC11A67D3013F90BE49D3A0682E3F3FABCC94D6DF023F4489963C9E962F3F43C5387F130A113F8EB1135E82532F3F1C2444F98216123FFA60191BBAD92F3FD2FBC6D79E59123F213EB0E3BF40303F8DEDB5A0F7C6103F213EB0E3BF40303FDE0033DFC14F0C3FA03715A930B6303F240F441669E20D3F43C5387F130A313F240F441669E20D3F2F6EA301BC05323F691D554D10750F3F0ADAE4F04927323F8DEDB5A0F7C6103F0CE71A66683C313F1C2444F98216123F1F317A6EA12B313FF5824F73F222133F7BA35698BED7303FC9C7EE022505163F7BA35698BED7303F10000000F5824F73F22213BFC7BAB88D06F026BFFA9CBB5D2F4D11BF11E335AFEAAC26BF3677F4BF5C8B06BFEE5BAD1397E325BF664CC11A67D3F1BE5D18E945ED7E25BF8DEDB5A0F7C6F0BE5F251FBB0B9424BF240F441669E20DBF5F251FBB0B9424BF90BE49D3A0680EBF5D18E945ED7E25BF213EB0E3BF4010BF13F06B2409C225BFFA9CBB5D2F4D11BF5D18E945ED7E25BFB0743E3C4B9011BF180AD80E46EC23BF3FABCC94D6DF12BF180AD80E46EC23BF3FABCC94D6DF12BF5F251FBB0B9424BF664CC11A67D311BFF168E388B5F824BFD2FBC6D79E5912BF38842A357BA025BFD2FBC6D79E5912BF5B0BB3D0CE6926BFF5824F73F22213BFC7BAB88D06F026BF06000000B75F3E59315CDDBE5F251FBB0B9424BFB75F3E59315CDDBEA7406667D13B25BF691D554D1075EF3ECCD42478431A25BFAB5AD2510E66033FA7406667D13B25BFAB5AD2510E66033F3A9160AA99B524BFB75F3E59315CDDBE5F251FBB0B9424BF')") db.execute("insert into ne_50m_admin_0_countries (gid, name, name_long, iso_a2, iso_a3, geom) values (2, 'USA', 'United States', 'US', 'USA', ST_SetSRID('MULTIPOLYGON(((-123.6 49.6, -65.3 49.6, -65.3 24.0, -123.6 24.0, -123.6 49.6)))'::geometry, 4326))") db.execute("insert into boxes (id, lon, lat, size, geom) values (1, 0, 0, 1, st_setsrid('polygon(( 0 0, 0 1, 1 1, 1 0, 0 0))'::geometry, 4326))") db.execute("insert into boxes (id, lon, lat, size, geom) values (2, 0, -1, 1, st_setsrid('polygon(( 0 -1, 0 0, 1 0, 1 -1, 0 -1))'::geometry, 4326))") db.execute("insert into boxes (id, lon, lat, size, geom) values (3, -1, -1, 1, st_setsrid('polygon((-1 -1, -1 0, 0 0, 0 -1, -1 -1))'::geometry, 4326))") db.execute("insert into boxes (id, lon, lat, size, geom) values (4, -1, 0, 1, st_setsrid('polygon((-1 0, -1 1, 0 1, 0 0, -1 0))'::geometry, 4326))") db.execute("insert into boxes (id, lon, lat, size, geom) values (5, -99, 39, 1, st_setsrid('polygon((-99 38, -99 39, -98 39, -98 38, -99 38))'::geometry, 4326))") db.execute("insert into gpwv4_2015 (iso_a2, box_id, population, area) values ('XX', 1, 2000, 800)") db.execute("insert into gpwv4_2015 (iso_a2, box_id, population, area) values ('XX', 2, 4000, 600)") db.execute("insert into gpwv4_2015 (iso_a2, box_id, population, area) values ('XX', 3, 6000, 400)") db.execute("insert into gpwv4_2015 (iso_a2, box_id, population, area) values ('XX', 4, 8000, 200)") db.execute("insert into gpwv4_2015 (iso_a2, box_id, population, area) values ('US', 5, 17907, 9540)") db.execute("insert into acs5yr_2015 (usps_code, box_id, population, area) values ('KS', 5, 17907, 9540)") def test_guess_iso_a2(self): get_iso3166 = lambda n: 'XX' if (n == 'ISO 3166') else None get_iso3166_2 = lambda n: 'YY-YY' if (n == 'ISO 3166-2') else None get_us_census = lambda n: '06001' if (n == 'US Census GEOID') else None get_intl_src_path = lambda n: 'sources/xx/yy.json' if (n == 'source paths') else None get_us_src_path = lambda n: 'sources/us/ca/oakland.json' if (n == 'source paths') else None feature = unittest.mock.Mock() feature.GetField = get_iso3166 self.assertEqual(calculate.guess_iso_a2(feature), 'XX') feature.GetField = get_iso3166_2 self.assertEqual(calculate.guess_iso_a2(feature), 'YY') feature.GetField = get_us_census self.assertEqual(calculate.guess_iso_a2(feature), 'US') feature.GetField = get_intl_src_path self.assertEqual(calculate.guess_iso_a2(feature), 'XX') feature.GetField = get_us_src_path self.assertEqual(calculate.guess_iso_a2(feature), 'US') def test_guess_state_abbrev(self): get_us_census = lambda n: '06001' if (n == 'US Census GEOID') else None get_intl_src_path = lambda n: 'sources/xx/yy.json' if (n == 'source paths') else None get_us_src_path = lambda n: 'sources/us/ca/oakland.json' if (n == 'source paths') else None feature = unittest.mock.Mock() feature.GetField = get_us_census self.assertEqual(calculate.guess_state_abbrev(feature), 'CA') feature.GetField = get_intl_src_path self.assertIsNone(calculate.guess_state_abbrev(feature)) feature.GetField = get_us_src_path self.assertEqual(calculate.guess_state_abbrev(feature), 'CA') def test_calculate(self): def response_geojson(url, request): if (request.method, url.hostname, url.path) == ('GET', 'results.openaddresses.io', '/index.json'): return response(200, b'{"render_geojson_url": "http://data.openaddresses.io/render-world.geojson"}', headers={'Content-Type': 'application/json'}) if (request.method, url.hostname, url.path) == ('GET', 'data.openaddresses.io', '/render-world.geojson'): null_geojson = '''{\n"type": "FeatureCollection",\n"features": [\n{ "type": "Feature", "properties": {"source count": 1, "name": "Null Island", "source dates": "2017-03-12 21:54:49.107291+00:00", "source paths": "sources/xx/countrywide.json", "ISO 3166": "XX", "ISO 3166-2": null, "US Census GEOID": null, "status": "good", "address count": 9990}, "geometry": { "type": "MultiPolygon", "coordinates": [ [ [ [ -0.000478, 0.000015 ], [ -0.000464, 0.000017 ], [ -0.000463, 0.000009 ], [ -0.000459, 0.000008 ], [ -0.000455, 0.000017 ], [ -0.000443, 0.000012 ], [ -0.000446, 0.000004 ], [ -0.000444, 0.0 ], [ -0.000433, 0.000003 ], [ -0.00043, -0.00001 ], [ -0.000438, -0.000013 ], [ -0.000438, -0.000016 ], [ -0.000428, -0.000017 ], [ -0.000429, -0.000038 ], [ -0.000438, -0.000038 ], [ -0.000435, -0.000043 ], [ -0.000445, -0.000052 ], [ -0.000449, -0.000049 ], [ -0.000455, -0.00005 ], [ -0.000455, -0.000057 ], [ -0.000469, -0.000057 ], [ -0.000469, -0.000049 ], [ -0.000474, -0.000049 ], [ -0.00048, -0.000055 ], [ -0.000489, -0.000047 ], [ -0.000488, -0.000044 ], [ -0.000496, -0.000034 ], [ -0.000491, -0.000028 ], [ -0.000491, -0.000026 ], [ -0.0005, -0.000024 ], [ -0.000496, -0.00001 ], [ -0.000492, -0.00001 ], [ -0.000492, -0.000006 ], [ -0.000495, -0.0 ], [ -0.000484, 0.000009 ], [ -0.000481, 0.000006 ], [ -0.000477, 0.000007 ], [ -0.000477, 0.000011 ], [ -0.000478, 0.000015 ] ] ], [ [ [ 0.000084, 0.000257 ], [ 0.000107, 0.000285 ], [ 0.000121, 0.000274 ], [ 0.000149, 0.000277 ], [ 0.000183, 0.00028 ], [ 0.000177, 0.000285 ], [ 0.00019, 0.0003 ], [ 0.000195, 0.000296 ], [ 0.000204, 0.000302 ], [ 0.000212, 0.000303 ], [ 0.000221, 0.000302 ], [ 0.000228, 0.000307 ], [ 0.000243, 0.000292 ], [ 0.000239, 0.000288 ], [ 0.000243, 0.000278 ], [ 0.000243, 0.000265 ], [ 0.000249, 0.000262 ], [ 0.000238, 0.000244 ], [ 0.000233, 0.000246 ], [ 0.000225, 0.000241 ], [ 0.00021, 0.00024 ], [ 0.000202, 0.000242 ], [ 0.000196, 0.000237 ], [ 0.000183, 0.000247 ], [ 0.000186, 0.000251 ], [ 0.000129, 0.000246 ], [ 0.000136, 0.000245 ], [ 0.000146, 0.000233 ], [ 0.000144, 0.000227 ], [ 0.000148, 0.000225 ], [ 0.000151, 0.000218 ], [ 0.000155, 0.000218 ], [ 0.00016, 0.000223 ], [ 0.000174, 0.000217 ], [ 0.000181, 0.000221 ], [ 0.000185, 0.000221 ], [ 0.000187, 0.000216 ], [ 0.000211, 0.000217 ], [ 0.000216, 0.000213 ], [ 0.000228, 0.000203 ], [ 0.000239, 0.000203 ], [ 0.000248, 0.000199 ], [ 0.000258, 0.000199 ], [ 0.000259, 0.000196 ], [ 0.000258, 0.00019 ], [ 0.000261, 0.000185 ], [ 0.000262, 0.000166 ], [ 0.000258, 0.000158 ], [ 0.000239, 0.000158 ], [ 0.000238, 0.000144 ], [ 0.000242, 0.000141 ], [ 0.000246, 0.000145 ], [ 0.000251, 0.000144 ], [ 0.000253, 0.000139 ], [ 0.000257, 0.000141 ], [ 0.000262, 0.000139 ], [ 0.000258, 0.000132 ], [ 0.000259, 0.000127 ], [ 0.000265, 0.000125 ], [ 0.00027, 0.000121 ], [ 0.000274, 0.000124 ], [ 0.000281, 0.000125 ], [ 0.000285, 0.00012 ], [ 0.000292, 0.00012 ], [ 0.000296, 0.000116 ], [ 0.000296, 0.000107 ], [ 0.000301, 0.000107 ], [ 0.000304, 0.00011 ], [ 0.000309, 0.00011 ], [ 0.000315, 0.000103 ], [ 0.000316, 0.000095 ], [ 0.000323, 0.000091 ], [ 0.000328, 0.000094 ], [ 0.000333, 0.000091 ], [ 0.000337, 0.000086 ], [ 0.000344, 0.000082 ], [ 0.000348, 0.000082 ], [ 0.00035, 0.000079 ], [ 0.000351, 0.000074 ], [ 0.000356, 0.000073 ], [ 0.000356, 0.000077 ], [ 0.000359, 0.000077 ], [ 0.000363, 0.000072 ], [ 0.000363, 0.00007 ], [ 0.000373, 0.00007 ], [ 0.000379, 0.000068 ], [ 0.000379, 0.000064 ], [ 0.000369, 0.000052 ], [ 0.000372, 0.000049 ], [ 0.000372, 0.000043 ], [ 0.000368, 0.000038 ], [ 0.000363, 0.000017 ], [ 0.000355, 0.000019 ], [ 0.000347, 0.000008 ], [ 0.000335, 0.000015 ], [ 0.000331, 0.000012 ], [ 0.000331, 0.000006 ], [ 0.000325, 0.000006 ], [ 0.000329, -0.000008 ], [ 0.000322, -0.000011 ], [ 0.000322, -0.000019 ], [ 0.000318, -0.00002 ], [ 0.000318, -0.000028 ], [ 0.000312, -0.000029 ], [ 0.000306, -0.000024 ], [ 0.000303, -0.000025 ], [ 0.000305, -0.000037 ], [ 0.000297, -0.000038 ], [ 0.000297, -0.000044 ], [ 0.000285, -0.000044 ], [ 0.000285, -0.000048 ], [ 0.000279, -0.000049 ], [ 0.000276, -0.000055 ], [ 0.000278, -0.00006 ], [ 0.000278, -0.000067 ], [ 0.000281, -0.000066 ], [ 0.000285, -0.00007 ], [ 0.000295, -0.000064 ], [ 0.000298, -0.000071 ], [ 0.000304, -0.00007 ], [ 0.000306, -0.000074 ], [ 0.000304, -0.000079 ], [ 0.000308, -0.000079 ], [ 0.00031, -0.000093 ], [ 0.000305, -0.000094 ], [ 0.000305, -0.000099 ], [ 0.000298, -0.000105 ], [ 0.000298, -0.00011 ], [ 0.0003, -0.00011 ], [ 0.000304, -0.000106 ], [ 0.00031, -0.000109 ], [ 0.000306, -0.000116 ], [ 0.00031, -0.000125 ], [ 0.000311, -0.000129 ], [ 0.000309, -0.000133 ], [ 0.000315, -0.00014 ], [ 0.00032, -0.00014 ], [ 0.000322, -0.000144 ], [ 0.00032, -0.000149 ], [ 0.000325, -0.000157 ], [ 0.000336, -0.000157 ], [ 0.000337, -0.00016 ], [ 0.000333, -0.000165 ], [ 0.000333, -0.000169 ], [ 0.000336, -0.000173 ], [ 0.000335, -0.000176 ], [ 0.000332, -0.000178 ], [ 0.000332, -0.000182 ], [ 0.000329, -0.000183 ], [ 0.000328, -0.000187 ], [ 0.000334, -0.000193 ], [ 0.000334, -0.000205 ], [ 0.000331, -0.000206 ], [ 0.00033, -0.000219 ], [ 0.000319, -0.00022 ], [ 0.00032, -0.00023 ], [ 0.000317, -0.000231 ], [ 0.000316, -0.000237 ], [ 0.000312, -0.000237 ], [ 0.000311, -0.000249 ], [ 0.000308, -0.000255 ], [ 0.000302, -0.000253 ], [ 0.000297, -0.000259 ], [ 0.000286, -0.00026 ], [ 0.000282, -0.000253 ], [ 0.000274, -0.000253 ], [ 0.00027, -0.000246 ], [ 0.000261, -0.000255 ], [ 0.000258, -0.000254 ], [ 0.000256, -0.000248 ], [ 0.000253, -0.000248 ], [ 0.000253, -0.000254 ], [ 0.000249, -0.000261 ], [ 0.00025, -0.000268 ], [ 0.000245, -0.000272 ], [ 0.000238, -0.000274 ], [ 0.000237, -0.00028 ], [ 0.000233, -0.00028 ], [ 0.000228, -0.000276 ], [ 0.00022, -0.000279 ], [ 0.000219, -0.000271 ], [ 0.000215, -0.000269 ], [ 0.000215, -0.000274 ], [ 0.000202, -0.000274 ], [ 0.000198, -0.000279 ], [ 0.000192, -0.000276 ], [ 0.000188, -0.000279 ], [ 0.000184, -0.000275 ], [ 0.000181, -0.000273 ], [ 0.000181, -0.000268 ], [ 0.000181, -0.00026 ], [ 0.000181, -0.000256 ], [ 0.000174, -0.000257 ], [ 0.000164, -0.000247 ], [ 0.000164, -0.000242 ], [ 0.000159, -0.000244 ], [ 0.000156, -0.000244 ], [ 0.000155, -0.000238 ], [ 0.000152, -0.000237 ], [ 0.000151, -0.000231 ], [ 0.000158, -0.000228 ], [ 0.000158, -0.00022 ], [ 0.000162, -0.000213 ], [ 0.000161, -0.000209 ], [ 0.000161, -0.000205 ], [ 0.000152, -0.000209 ], [ 0.000151, -0.000204 ], [ 0.000145, -0.000207 ], [ 0.000143, -0.000206 ], [ 0.00014, -0.000195 ], [ 0.000136, -0.000195 ], [ 0.000131, -0.000188 ], [ 0.000138, -0.000179 ], [ 0.000139, -0.00017 ], [ 0.000142, -0.000168 ], [ 0.000142, -0.000165 ], [ 0.000135, -0.000168 ], [ 0.000133, -0.000167 ], [ 0.000128, -0.000171 ], [ 0.000121, -0.000163 ], [ 0.000114, -0.000165 ], [ 0.000114, -0.000159 ], [ 0.000106, -0.00016 ], [ 0.000106, -0.000155 ], [ 0.000104, -0.000151 ], [ 0.000105, -0.000147 ], [ 0.000108, -0.000144 ], [ 0.000108, -0.000139 ], [ 0.000101, -0.000142 ], [ 0.0001, -0.000157 ], [ 0.000101, -0.000162 ], [ 0.000098, -0.000165 ], [ 0.000093, -0.000163 ], [ 0.000091, -0.000158 ], [ 0.000046, -0.000158 ], [ 0.000046, -0.000163 ], [ 0.000052, -0.000165 ], [ 0.000066, -0.000169 ], [ 0.000078, -0.000176 ], [ 0.000089, -0.000178 ], [ 0.000096, -0.00018 ], [ 0.000097, -0.000184 ], [ 0.00009, -0.000191 ], [ 0.000083, -0.000193 ], [ 0.000072, -0.000204 ], [ 0.000065, -0.000209 ], [ 0.000049, -0.000216 ], [ 0.000031, -0.000223 ], [ 0.000003, -0.000228 ], [ -0.000022, -0.00023 ], [ -0.000046, -0.00023 ], [ -0.000077, -0.00023 ], [ -0.000081, -0.000203 ], [ -0.000081, -0.000184 ], [ -0.000083, -0.000186 ], [ -0.00009, -0.000185 ], [ -0.000094, -0.000188 ], [ -0.000094, -0.000193 ], [ -0.00009, -0.000197 ], [ -0.000091, -0.0002 ], [ -0.000095, -0.000201 ], [ -0.000097, -0.000209 ], [ -0.0001, -0.000215 ], [ -0.000107, -0.000213 ], [ -0.000109, -0.000218 ], [ -0.000117, -0.000219 ], [ -0.000122, -0.000215 ], [ -0.000127, -0.000216 ], [ -0.000133, -0.000219 ], [ -0.000137, -0.000217 ], [ -0.000138, -0.000213 ], [ -0.000144, -0.000213 ], [ -0.000145, -0.000209 ], [ -0.00015, -0.000207 ], [ -0.000154, -0.00021 ], [ -0.000157, -0.000206 ], [ -0.000163, -0.000206 ], [ -0.000166, -0.00021 ], [ -0.000171, -0.00021 ], [ -0.000172, -0.000204 ], [ -0.000173, -0.0002 ], [ -0.000177, -0.000195 ], [ -0.000183, -0.000196 ], [ -0.000187, -0.000192 ], [ -0.000192, -0.000192 ], [ -0.000195, -0.000189 ], [ -0.000195, -0.000183 ], [ -0.000198, -0.000182 ], [ -0.000198, -0.000175 ], [ -0.000195, -0.000167 ], [ -0.000198, -0.000166 ], [ -0.000201, -0.000175 ], [ -0.000206, -0.000176 ], [ -0.000208, -0.000179 ], [ -0.000211, -0.00018 ], [ -0.000215, -0.000177 ], [ -0.000221, -0.000176 ], [ -0.000222, -0.00017 ], [ -0.000226, -0.000168 ], [ -0.000226, -0.000156 ], [ -0.00023, -0.000156 ], [ -0.00023, -0.00015 ], [ -0.000238, -0.000141 ], [ -0.000238, -0.000134 ], [ -0.000236, -0.000131 ], [ -0.000238, -0.000129 ], [ -0.000238, -0.000123 ], [ -0.000243, -0.000123 ], [ -0.000246, -0.00013 ], [ -0.00025, -0.00013 ], [ -0.000252, -0.000127 ], [ -0.000256, -0.000127 ], [ -0.000258, -0.000123 ], [ -0.000257, -0.000115 ], [ -0.000264, -0.000109 ], [ -0.000271, -0.000114 ], [ -0.000274, -0.000111 ], [ -0.000273, -0.000105 ], [ -0.000277, -0.000103 ], [ -0.000276, -0.000099 ], [ -0.000286, -0.0001 ], [ -0.00029, -0.000104 ], [ -0.000295, -0.000098 ], [ -0.000295, -0.000095 ], [ -0.0003, -0.000094 ], [ -0.000309, -0.00009 ], [ -0.000313, -0.000096 ], [ -0.000323, -0.000093 ], [ -0.000327, -0.000088 ], [ -0.000336, -0.000093 ], [ -0.000339, -0.000092 ], [ -0.000339, -0.000083 ], [ -0.000343, -0.000082 ], [ -0.000343, -0.000076 ], [ -0.000342, -0.000074 ], [ -0.000342, -0.000071 ], [ -0.000347, -0.000071 ], [ -0.000345, -0.000058 ], [ -0.000351, -0.000051 ], [ -0.00035, -0.000043 ], [ -0.000354, -0.00004 ], [ -0.000359, -0.000043 ], [ -0.000362, -0.000037 ], [ -0.000368, -0.000038 ], [ -0.000371, -0.000035 ], [ -0.000371, -0.000026 ], [ -0.000374, -0.000025 ], [ -0.000379, -0.000025 ], [ -0.000378, -0.000013 ], [ -0.000385, -0.000011 ], [ -0.000385, -0.0 ], [ -0.000382, 0.000002 ], [ -0.000382, 0.000012 ], [ -0.000389, 0.000012 ], [ -0.000391, 0.000015 ], [ -0.000392, 0.000021 ], [ -0.000383, 0.000033 ], [ -0.000384, 0.000038 ], [ -0.000375, 0.000052 ], [ -0.000368, 0.000054 ], [ -0.000369, 0.00006 ], [ -0.000364, 0.000064 ], [ -0.00036, 0.000064 ], [ -0.000353, 0.000064 ], [ -0.000343, 0.000069 ], [ -0.000338, 0.000069 ], [ -0.000332, 0.00007 ], [ -0.000327, 0.000077 ], [ -0.000327, 0.000084 ], [ -0.000318, 0.000089 ], [ -0.000311, 0.00009 ], [ -0.000306, 0.000094 ], [ -0.000297, 0.000094 ], [ -0.000298, 0.000104 ], [ -0.000289, 0.00011 ], [ -0.000283, 0.000106 ], [ -0.000275, 0.000103 ], [ -0.000276, 0.000116 ], [ -0.000285, 0.000117 ], [ -0.000285, 0.000124 ], [ -0.000278, 0.000128 ], [ -0.000278, 0.000143 ], [ -0.000276, 0.000145 ], [ -0.00027, 0.000146 ], [ -0.000264, 0.00016 ], [ -0.000262, 0.000165 ], [ -0.000252, 0.000162 ], [ -0.000246, 0.000173 ], [ -0.00024, 0.000172 ], [ -0.000229, 0.00018 ], [ -0.000223, 0.00018 ], [ -0.00022, 0.000191 ], [ -0.000215, 0.000193 ], [ -0.000211, 0.000203 ], [ -0.000197, 0.000215 ], [ -0.000195, 0.000222 ], [ -0.000181, 0.000217 ], [ -0.000182, 0.000212 ], [ -0.000188, 0.000202 ], [ -0.000186, 0.000197 ], [ -0.000186, 0.000193 ], [ -0.000188, 0.000183 ], [ -0.000183, 0.000183 ], [ -0.000176, 0.000174 ], [ -0.000172, 0.000174 ], [ -0.000163, 0.000185 ], [ -0.000159, 0.000186 ], [ -0.000154, 0.00019 ], [ -0.000148, 0.000184 ], [ -0.000138, 0.000191 ], [ -0.000126, 0.000189 ], [ -0.000125, 0.000182 ], [ -0.000118, 0.000181 ], [ -0.000117, 0.00019 ], [ -0.000105, 0.000191 ], [ -0.000105, 0.000211 ], [ -0.000108, 0.000218 ], [ -0.000103, 0.00022 ], [ -0.000102, 0.00023 ], [ -0.000091, 0.000235 ], [ -0.000087, 0.000233 ], [ -0.000087, 0.00024 ], [ -0.000083, 0.000245 ], [ -0.000075, 0.000247 ], [ -0.000065, 0.000238 ], [ -0.000059, 0.000237 ], [ -0.000058, 0.000234 ], [ -0.00004, 0.000233 ], [ -0.000039, 0.000241 ], [ -0.00003, 0.000243 ], [ -0.00003, 0.000236 ], [ -0.000026, 0.000236 ], [ -0.000018, 0.000222 ], [ -0.00001, 0.000222 ], [ -0.000007, 0.000217 ], [ -0.000003, 0.000217 ], [ 0.000003, 0.000211 ], [ 0.00001, 0.000215 ], [ 0.000012, 0.000222 ], [ 0.000017, 0.000222 ], [ 0.00003, 0.000233 ], [ 0.000034, 0.000232 ], [ 0.000036, 0.000241 ], [ 0.000065, 0.000239 ], [ 0.000069, 0.000243 ], [ 0.00007, 0.000248 ], [ 0.000064, 0.000248 ], [ 0.000054, 0.000255 ], [ 0.000057, 0.00026 ], [ 0.000057, 0.000275 ], [ 0.00006, 0.000277 ], [ 0.000064, 0.000263 ], [ 0.000069, 0.000262 ], [ 0.000073, 0.000257 ], [ 0.000084, 0.000257 ] ], [ [ -0.000073, -0.000175 ], [ -0.000066, -0.000173 ], [ -0.000043, -0.000167 ], [ -0.000017, -0.000164 ], [ -0.000016, -0.000157 ], [ -0.000057, -0.000157 ], [ -0.000058, -0.000164 ], [ -0.000062, -0.000166 ], [ -0.000066, -0.000164 ], [ -0.000067, -0.000152 ], [ -0.000072, -0.000152 ], [ -0.000072, -0.000157 ], [ -0.000068, -0.00016 ], [ -0.00007, -0.000165 ], [ -0.00007, -0.000171 ], [ -0.000073, -0.000175 ] ], [ [ -0.000007, -0.000157 ], [ -0.000007, -0.000162 ], [ 0.000015, -0.000161 ], [ 0.000037, -0.000162 ], [ 0.000037, -0.000158 ], [ -0.000007, -0.000157 ] ] ] ] } }, { "type": "Feature", "properties": {"source count": 1, "name": "Null Ranch", "source dates": "2017-03-12 21:54:49.107291+00:00", "source paths": "sources/us/ks/null-ranch.json", "ISO 3166": null, "ISO 3166-2": null, "US Census GEOID": null, "status": "good", "address count": 9}, "geometry": { "type": "Polygon", "coordinates": [[[-99, 38], [-99, 39], [-98, 39], [-98, 38], [-99, 38]]] } }\n]\n}\n''' return response(200, null_geojson.encode('utf8'), headers={'Content-Type': 'application/json'}) raise Exception() with HTTMock(response_geojson): calculate.calculate(DATABASE_URL) with psycopg2.connect(DATABASE_URL) as conn: with conn.cursor() as db: db.execute('select iso_a2, addr_count, area_total, area_pct, pop_total, pop_pct from areas order by iso_a2') (row1, row2) = db.fetchall() self.assertEqual(row1, ('US', 9, 9540, 1.0, 17907, 1.0)) self.assertEqual(row2, ('XX', 9990, 2000, 1.0, 20000, 1.0))
[ "mike@teczno.com" ]
mike@teczno.com
b26444ad2d6f2216e041816a9cd9a0238f7491e6
6d493d09085d4d398132204925078a179774f138
/melgan_vocoder.py
2ec8f713892afcce0d01ff4faa4f26ebc87935ea
[ "MIT" ]
permissive
zongxiangli/CycleGAN-VC3
6a41f843b430fd307d9ea0b43aa5910816fba450
431b332fa17638391ca913e6821b526456fd874f
refs/heads/main
2023-02-21T02:19:39.058010
2021-01-25T09:49:00
2021-01-25T09:49:00
null
0
0
null
null
null
null
UTF-8
Python
false
false
4,448
py
#!python # -*- coding: utf-8 -*- import os import yaml from pathlib import Path import torch import torch.nn as nn from torch.nn.utils import weight_norm from feature_utils import Audio2Mel def weights_init(m): classname = m.__class__.__name__ if classname.find("Conv") != -1: m.weight.data.normal_(0.0, 0.02) elif classname.find("BatchNorm2d") != -1: m.weight.data.normal_(1.0, 0.02) m.bias.data.fill_(0) def WNConv1d(*args, **kwargs): return weight_norm(nn.Conv1d(*args, **kwargs)) def WNConvTranspose1d(*args, **kwargs): return weight_norm(nn.ConvTranspose1d(*args, **kwargs)) class ResnetBlock(nn.Module): def __init__(self, dim, dilation=1): super().__init__() self.block = nn.Sequential( nn.LeakyReLU(0.2), nn.ReflectionPad1d(dilation), WNConv1d(dim, dim, kernel_size=3, dilation=dilation), nn.LeakyReLU(0.2), WNConv1d(dim, dim, kernel_size=1), ) self.shortcut = WNConv1d(dim, dim, kernel_size=1) def forward(self, x): return self.shortcut(x) + self.block(x) class Generator(nn.Module): def __init__(self, input_size, ngf, n_residual_layers): super().__init__() ratios = [8, 8, 2, 2] self.hop_length = np.prod(ratios) mult = int(2 ** len(ratios)) model = [ nn.ReflectionPad1d(3), WNConv1d(input_size, mult * ngf, kernel_size=7, padding=0), ] # Upsample to raw audio scale for i, r in enumerate(ratios): model += [ nn.LeakyReLU(0.2), WNConvTranspose1d( mult * ngf, mult * ngf // 2, kernel_size=r * 2, stride=r, padding=r // 2 + r % 2, output_padding=r % 2, ), ] for j in range(n_residual_layers): model += [ResnetBlock(mult * ngf // 2, dilation=3 ** j)] mult //= 2 model += [ nn.LeakyReLU(0.2), nn.ReflectionPad1d(3), WNConv1d(ngf, 1, kernel_size=7, padding=0), nn.Tanh(), ] self.model = nn.Sequential(*model) self.apply(weights_init) def forward(self, x): return self.model(x) def get_default_device(): if torch.cuda.is_available(): return "cuda" else: return "cpu" def load_model(mel2wav_path, device=get_default_device()): """ Args: mel2wav_path (str or Path): path to the root folder of dumped text2mel device (str or torch.device): device to load the model """ root = Path(mel2wav_path) with open(root / "args.yml", "r") as f: args = yaml.load(f, Loader=yaml.FullLoader) netG = Generator(args.n_mel_channels, args.ngf, args.n_residual_layers).to(device) netG.load_state_dict(torch.load(root / "best_netG.pt", map_location=device)) return netG class MelVocoder: def __init__( self, path, device=get_default_device(), github=False, model_name="multi_speaker", ): self.fft = Audio2Mel().to(device) if github: netG = Generator(80, 32, 3).to(device) root = Path(os.path.dirname(__file__)).parent netG.load_state_dict( torch.load(root / f"models/{model_name}.pt", map_location=device) ) self.mel2wav = netG else: self.mel2wav = load_model(path, device) self.device = device def __call__(self, audio): """ Performs audio to mel conversion (See Audio2Mel in mel2wav/modules.py) Args: audio (torch.tensor): PyTorch tensor containing audio (batch_size, timesteps) Returns: torch.tensor: log-mel-spectrogram computed on input audio (batch_size, 80, timesteps) """ return self.fft(audio.unsqueeze(1).to(self.device)) def inverse(self, mel): """ Performs mel2audio conversion Args: mel (torch.tensor): PyTorch tensor containing log-mel spectrograms (batch_size, 80, timesteps) Returns: torch.tensor: Inverted raw audio (batch_size, timesteps) """ with torch.no_grad(): return self.mel2wav(mel.to(self.device)).squeeze(1)
[ "jackaduma@gmail.com" ]
jackaduma@gmail.com
5cc1e828fb014c43f7ef00b0a93ab7e27a5e0eb4
5ee6858e60d1065c797a105710e9d6c835f3b7ad
/app/user/migrations/0008_alter_customuser_avatar.py
0d17298ad2e47e51857959c13560f1c37e7deca3
[]
no_license
kiyoshion/django-rest-framework
19260d1b864939a69fe6668062fdea12a1b1b1d5
8f81064511b9db066ef2475fad78629a5839c213
refs/heads/main
2023-07-08T08:29:32.483651
2021-08-09T11:36:13
2021-08-09T11:36:13
394,266,414
0
0
null
null
null
null
UTF-8
Python
false
false
547
py
# Generated by Django 3.2.4 on 2021-06-10 02:09 from django.db import migrations import imagekit.models.fields import user.models class Migration(migrations.Migration): dependencies = [ ('user', '0007_alter_customuser_avatar'), ] operations = [ migrations.AlterField( model_name='customuser', name='avatar', field=imagekit.models.fields.ProcessedImageField(blank=True, default='img/avatar.svg', null=True, upload_to=user.models.CustomUser.user_directory_path), ), ]
[ "kiyoshion@gmail.com" ]
kiyoshion@gmail.com
c98f587d3f300afa98366ee38ee38743ef8a8905
025ad48264afd44ccc9a3ec820a74711631a77d7
/ludo.py
5195e20168f60670e6bcdfa1552868a8d865f136
[]
no_license
buffosens/LudoRGB
1939f6588b624162500ee74dadef7efbd22a5fd9
72a940c4e309e3ea128d332528303f246d3bc2de
refs/heads/master
2023-03-27T20:33:08.731595
2021-03-29T06:50:27
2021-03-29T06:50:27
347,163,268
0
0
null
null
null
null
UTF-8
Python
false
false
4,273
py
# Simple demo of of the WS2801/SPI-like addressable RGB LED lights. import time import RPi.GPIO as GPIO # Import the WS2801 module. import Adafruit_WS2801 import Adafruit_GPIO.SPI as SPI # Configure the count of pixels: PIXEL_COUNT = 100 # Alternatively specify a hardware SPI connection on /dev/spidev0.0: SPI_PORT = 0 SPI_DEVICE = 0 pixels = Adafruit_WS2801.WS2801Pixels(PIXEL_COUNT, spi=SPI.SpiDev(SPI_PORT, SPI_DEVICE), gpio=GPIO) # Define the wheel function to interpolate between different hues. def wheel(pos): if pos < 85: return Adafruit_WS2801.RGB_to_color(pos * 3, 255 - pos * 3, 0) elif pos < 170: pos -= 85 return Adafruit_WS2801.RGB_to_color(255 - pos * 3, 0, pos * 3) else: pos -= 170 return Adafruit_WS2801.RGB_to_color(0, pos * 3, 255 - pos * 3) # Define rainbow cycle function to do a cycle of all hues. def rainbow_cycle_successive(pixels, wait=0.1): for i in range(pixels.count()): # tricky math! we use each pixel as a fraction of the full 96-color wheel # (thats the i / strip.numPixels() part) # Then add in j which makes the colors go around per pixel # the % 96 is to make the wheel cycle around pixels.set_pixel(i, wheel(((i * 256 // pixels.count())) % 256) ) pixels.show() if wait > 0: time.sleep(wait) def rainbow_cycle(pixels, wait=0.005): for j in range(256): # one cycle of all 256 colors in the wheel for i in range(pixels.count()): pixels.set_pixel(i, wheel(((i * 256 // pixels.count()) + j) % 256) ) pixels.show() if wait > 0: time.sleep(wait) def rainbow_colors(pixels, wait=0.05): for j in range(256): # one cycle of all 256 colors in the wheel for i in range(pixels.count()): pixels.set_pixel(i, wheel(((256 // pixels.count() + j)) % 256) ) pixels.show() if wait > 0: time.sleep(wait) def brightness_decrease(pixels, wait=0.01, step=1): for j in range(int(256 // step)): for i in range(pixels.count()): r, g, b = pixels.get_pixel_rgb(i) r = int(max(0, r - step)) g = int(max(0, g - step)) b = int(max(0, b - step)) pixels.set_pixel(i, Adafruit_WS2801.RGB_to_color( r, g, b )) pixels.show() if wait > 0: time.sleep(wait) def blink_color(pixels, blink_times=5, wait=0.5, color=(255,0,0)): for i in range(blink_times): # blink two times, then wait pixels.clear() for j in range(2): for k in range(pixels.count()): pixels.set_pixel(k, Adafruit_WS2801.RGB_to_color( color[0], color[1], color[2] )) pixels.show() time.sleep(0.08) pixels.clear() pixels.show() time.sleep(0.08) time.sleep(wait) def appear_from_back(pixels, color=(255, 0, 0)): pos = 0 for i in range(pixels.count()): for j in reversed(range(i, pixels.count())): pixels.clear() # first set all pixels at the begin for k in range(i): pixels.set_pixel(k, Adafruit_WS2801.RGB_to_color( color[0], color[1], color[2] )) # set then the pixel at position j pixels.set_pixel(j, Adafruit_WS2801.RGB_to_color( color[0], color[1], color[2] )) pixels.show() time.sleep(0.02) def test_led(pixels): for i in range(pixels.count()): pixels.set_pixel(i, Adafruit_WS2801.RGB_to_color( 255,255,255 )) pixels.show() if __name__ == "__main__": # Clear all the pixels to turn them off. pixels.clear() pixels.show() # Make sure to call show() after changing any pixels! #rainbow_cycle_successive(pixels, wait=1.0) #rainbow_cycle(pixels, wait=0.01) #brightness_decrease(pixels) #appear_from_back(pixels) #for i in range(3): #blink_color(pixels, blink_times = 1, color=(255, 0, 0)) #blink_color(pixels, blink_times = 1, color=(0, 255, 0)) #blink_color(pixels, blink_times = 1, color=(0, 0, 255)) #rainbow_colors(pixels) #brightness_decrease(pixels) test_led(pixels)
[ "volker.weber@swarco.de" ]
volker.weber@swarco.de
673bb5c0289c9335d604b2d187d96c5573ad2619
6ecf6e0545709592996acbe7d7752870a7e7d179
/GUI/gui1studentinfo.py
5e0203cf72c384c5532a7ad01a3639ddf854c513
[]
no_license
namujagtap/Namrata
9bf152882703ca6d6819c6ddf915d88d47c9a3e2
f90faabc142238e305511fdd44011e5db9767903
refs/heads/main
2023-08-01T22:44:22.360108
2021-09-16T05:05:00
2021-09-16T05:05:00
396,077,357
0
0
null
2021-08-14T17:26:58
2021-08-14T17:19:37
null
UTF-8
Python
false
false
1,661
py
#create student information from using Tkinter #name,address,email,schooltype,stding in year ,dob,,gender,schooltype,marathi mediunm,english medium,english,convent,semi import tkinter as tk from tkinter import * win=tk.Tk(className="StudentInfoForm......") win.geometry("1000x1000") label=tk.Label(win,text="***** STUDENT INFORMATION FORM *****").place(x=380,y=10) Name=tk.Label(win,text="NAME").place(x=35,y=60) e1=tk.Entry(win).place(x=170,y=60) Address=tk.Label(win,text="ADDRESS").place(x=35,y=90) e2=tk.Entry(win).place(x=170,y=90) Email=tk.Label(win,text="EMAIL").place(x=35,y=120) e3=tk.Entry(win).place(x=170,y=120) Schoolname=tk.Label(win,text="SCHOOL NAME ").place(x=35,y=150) e4=tk.Entry(win).place(x=170,y=150) Stding_in_year=tk.Label(win,text="STUDING IN YEAR ").place(x=35,y=180) e5=tk.Entry(win).place(x=170,y=180) dob=tk.Label(win,text="DATE OF BIRTH ").place(x=35,y=210) e6=tk.Entry(win).place(x=170,y=210) Gender=tk.Label(win,text=" 1] SELECT GENDER ").place(x=35,y=280) radio1=tk.Radiobutton(win,text="Male",value=0).place(x=120,y=310) radio2=tk.Radiobutton(win,text="Female",value=1).place(x=320,y=310) schooltype=tk.Label(win,text=" 2] SCHOOL TYPE ").place(x=35,y=360) radio3=tk.Radiobutton(win,text="Marathi Medium",value=0).place(x=120,y=390) radio4=tk.Radiobutton(win,text="English Medium",value=0).place(x=320,y=390) radio5=tk.Radiobutton(win,text="Convent Medium",value=0).place(x=520,y=390) radio6=tk.Radiobutton(win,text="Semi-English Medium",value=1).place(x=720,y=390) submit=tk.Button(win,text="SUBMIT",activebackground="pink",activeforeground="purple").place(x=420,y=500) win.mainloop()
[ "noreply@github.com" ]
namujagtap.noreply@github.com
19c873ebd90585623d912f2794ba9b06bb70f3d7
0634aed371b1c08888bd0b69be98c02779a49889
/Chapters/code/appendices3.py
3917a26bbd1e5e3bb99cbfa0dfeba4f77b3a73f8
[]
no_license
samiarja/Thesis2
e3918c3ea2b5ad15412e79368f7a1c6f82d544e8
4954b36c9f24881b4a724fdefcd29cd03aa0ec1b
refs/heads/master
2020-06-01T10:39:41.384465
2019-12-21T03:01:52
2019-12-21T03:01:52
190,751,932
0
0
null
null
null
null
UTF-8
Python
false
false
1,251
py
#Make the code running on both python2 and python3 from __future__ import unicode_literals from __future__ import print_function from __future__ import division # Imports Libraries and dependencies import os import cv2 from sklearn.datasets import load_files from sklearn.model_selection import train_test_split raw_data = load_files(os.getcwd() + r'/Data', shuffle=False) files = raw_data['filenames'] targets = raw_data['target'] train_files, test_files, train_targets, test_targets = train_test_split(files, targets, test_size=1/3, random_state=191) # Taking ~25% of the training data for validation valid_files = train_files[300:] valid_targets = train_targets[300:] # Remaining data will be used for training the model train_files = train_files[:300] train_targets = train_targets[:300] # Generic details about the data print('Total number of videos:', len(files)) print('\nNumber of videos in training data:', train_files.shape[0]) print('Number of videos in validation data:', valid_files.shape[0]) print('Number of videos in test data:', test_files.shape[0])
[ "sami18040571@outlook.com" ]
sami18040571@outlook.com
01d0564258a79bdc181836896a6e65794b1dbcee
6b94e0aba8e1bd3daf4e6ca4ab472007ab13bf97
/Py 16.py
8dc56e4adaa822d846e38227617dfeb958671ec6
[]
no_license
Dre-AsiliVentures/Python-Programming-scripts
92ccdb29fb302387386bc3f1714918ba2f97c07f
6ea0408965276afee5b5a1603e56703853075023
refs/heads/master
2023-05-10T10:09:56.460191
2020-03-02T09:46:52
2020-03-02T09:46:52
null
0
0
null
null
null
null
UTF-8
Python
false
false
125
py
numbers=[12,13,14,15,18] numbers[0]=numbers[2]*2-6 if 20 in numbers: print(numbers[3]) else: print(numbers[4])
[ "noreply@github.com" ]
Dre-AsiliVentures.noreply@github.com
6f5709bc731865f387e0060495401727486d9ca7
16fe56a5fd3beea997bea716f103f4165f2f9dfc
/reviewsMapper.py
e9073ec498b7cb4b7ab9568ba1f0b9a46bdf53f0
[]
no_license
jgnguy/yelp_dataset_project
2299b1054a0e284762ef3b9c9ab49cc0c6213bd8
a4099978374053cae2b7702675121ab4051fc16f
refs/heads/master
2021-01-08T15:12:37.079944
2019-12-11T19:55:00
2019-12-11T19:55:00
null
0
0
null
null
null
null
UTF-8
Python
false
false
1,234
py
#!/usr/bin/python #REVIEWS mapper.py #Evan Yao, Jonathan Nguyen, Richard Pham #get all the stars and businessID of all businesses within a specified number of years ago import datetime import json import sys #get today's date and convert to YYYY-MM-DD format numYearsAgo = 2 today = datetime.datetime.now() #get a year ago from today yearAgo = today - datetime.timedelta(numYearsAgo * 365) reformatYear = yearAgo.strftime('%Y-%m-%d') pastYear, pastMonth, pastDay = reformatYear.split('-') pastYearDate = datetime.datetime(int(pastYear), int(pastMonth), int(pastDay)) #adds each review to a list for line in sys.stdin: line = line.strip() review = json.loads(line) #check if review exists if review is not None: try: reviewDate = review['date'] reviewYear, reviewMonth, reviewDay = reviewDate.split(' ')[0].split('-') refReviewDate = datetime.datetime(int(reviewYear), int(reviewMonth), int(reviewDay)) #check if review is relevant and current if pastYearDate < refReviewDate: #print the business ID the review is for and the amount of stars they gave print(review['business_id'] + '\t' + str(review['stars'])) except: continue
[ "noreply@github.com" ]
jgnguy.noreply@github.com
502d4b86495a11eb9b80fc09726abe7faa363290
93de0276028ee893ef25e269f25c0492fec6d0d8
/Client-bl.py
c540da59914e76373be612207e2c7ccaf3d50520
[]
no_license
antarasargam/Lab2
dcbdee7cf8c460773df558f39dc1b1541150cfc4
3854a175dff5ffde136d213db54c01b2e93c2a37
refs/heads/master
2021-08-24T13:19:56.899452
2017-11-21T07:32:31
2017-11-21T07:32:31
105,047,853
0
3
null
2017-11-21T07:32:31
2017-09-27T17:24:39
Python
UTF-8
Python
false
false
19,464
py
#Client import asyncio import playground import random, logging from .Servera import PEEPpacket from playground.network.packet import PacketType from playground.network.packet.fieldtypes import UINT32, STRING, UINT16, UINT8, BUFFER from playground.network.packet.fieldtypes.attributes import Optional from playground.network.common.Protocol import StackingProtocol, StackingProtocolFactory, StackingTransport import zlib '''class PEEPpacket(PacketType): DEFINITION_IDENTIFIER = "PEEP.Packet" DEFINITION_VERSION = "1.0" FIELDS = [ ("Type", UINT8), ("SequenceNumber", UINT32({Optional: True})), ("Checksum", UINT16), ("Acknowledgement", UINT32({Optional: True})), ("Data", BUFFER({Optional: True})) ]''' class PeepClientTransport(StackingTransport): def __init__(self,protocol,transport): self.protocol = protocol self.transport = transport self.exc = None super().__init__(self.transport) def write(self, data): self.protocol.write(data) def close(self): self.protocol.close() def connection_lost(self): self.protocol.connection_lost(self.exc) class PEEPClient(StackingProtocol): global_number_seq = 0 global_number_ack = 0 count_of_function_call = 0 first_data_seq_number = 0 count_of_function_call_ack = 0 global_packet_size = 0 number_of_packs = 0 recv_window = {} prev_sequence_number = 0 expected_ackno = 0 sending_window = {} sending_window_count = 0 global_pig = 0 keylist1= [] t = {} n = 0 global_received_ack = 0 prev_ack_number = 0 backlog_window = [] def __init__(self, loop): self.transport = None self.loop = loop self._state = 0 def calculateChecksum(self, c): self.c = c self.c.Checksum = 0 checkbytes = self.c.__serialize__() return zlib.adler32(checkbytes) & 0xffff def checkChecksum(self, instance): self.instance = instance pullChecksum = self.instance.Checksum instance.Checksum = 0 bytes = self.instance.__serialize__() if pullChecksum == zlib.adler32(bytes) & 0xffff : return True else: return False async def syn_timeout(self): while self._state < 2: await asyncio.sleep(1) if self._state < 2: self.transport.write(self.syn) async def ack_timeout(self): while self._state < 3: await asyncio.sleep(0.9) if self._state < 3: self.transport.write(self.clientpacketbytes) '''async def data_timeout(self): packets = list(self.t.values()) for each_packet in packets: while each_packet.packet.Acknowledgement < each_packet.packet.SequenceNumber: await asyncio.sleep(0.3) for each_packet in packets: if each_packet.packet.SequenceNumber == each_packet.packet.Acknowledgement: print("Inside Data Timer") self.transport.write(each_packet.packet.__serialize__())''' async def data_timeout(self): print("Inside Data Timer") packets = list(self.t.values()) while self.global_received_ack < self.global_number_seq: await asyncio.sleep(0.1) for each_packet in packets: await asyncio.sleep(0.1) if self.global_received_ack < self.global_number_seq: if each_packet.packet.SequenceNumber == self.global_received_ack and each_packet.flag<10: self.transport.write(each_packet.packet.__serialize__()) each_packet.flag += 1 print("Packet Retransmitted.") def connection_made(self, transport): print("=============== PEEP Client Connection_made CALLED =========\n") self.transport = transport self.protocol = self if self._state == 0: packet = PEEPpacket() packet.Type = 0 packet.SequenceNumber = random.randrange(1, 1000, 1) packet.Acknowledgement = 0 packet.Data = b"Piggy" self._state += 1 print("Value of actual state is",self._state) print("=============== Sending SYN packet ==================\n") packet.Checksum = self.calculateChecksum(packet) self.syn = packet.__serialize__() self.transport.write(self.syn) self.ta = Timerx(0.1, self.syn_timeout, self.syn) def data_received(self, data): print("=============== PEEP Client Data_Received CALLED =============\n") self.deserializer = PEEPpacket.Deserializer() self.deserializer.update(data) for packet in self.deserializer.nextPackets(): checkvalue = self.checkChecksum(packet) if self._state == 1 and packet.Type == 1: if checkvalue: print("SYN-ACK Received. Seqno= ", packet.SequenceNumber, " Ackno=", packet.Acknowledgement) self.ta.cancel() #Sending ACK if packet.Data == b"Piggy": self.global_pig = 56 print(self.global_pig) print("Choosing Piggybacking") else: print ("Choosing Selective") ack = PEEPpacket() ack.Type = 2 ack.SequenceNumber = packet.Acknowledgement self.global_number_seq = ack.SequenceNumber ack.Acknowledgement = packet.SequenceNumber + 1 if self.global_pig == 56: ack.Data = b"Piggy" self.global_number_ack = ack.Acknowledgement self._state += 1 ack.Checksum = self.calculateChecksum(ack) self.clientpacketbytes = ack.__serialize__() print ("\n=================== Sending ACK =================\n") self.transport.write(self.clientpacketbytes) self.tb = Timerx(0.1, self.ack_timeout, self.clientpacketbytes) peeptransport = PeepClientTransport(self, self.transport) self.higherProtocol().connection_made(peeptransport) else: print("Corrupt SYN-ACK packet received. Please check on server end.") elif packet.Type == 5: if checkvalue: if self._state == 2: self.tb.cancel() print("====================Got Encapasulated Packet and Deserialized==================") #print(packet.Data) self._state +=1 self.global_received_ack = packet.Acknowledgement self.global_packet_size = len(packet.Data) print("The size of packet is:", self.global_packet_size) print("Seq number of incoming packet", packet.SequenceNumber) print("Ack Number of incoming packet", packet.Acknowledgement) self.receive_window(packet) #if self.global_pig != 56: # self.sendack(self.update_ack(packet.SequenceNumber, self.global_packet_size)) #self.higherProtocol().data_received(packet.Data) else: print("Corrupt Data packet received. Please check on server end.") elif packet.Type == 2: if checkvalue: '''self.return_value = self.check_if_ack_received_before(packet) if self.return_value == 1: self.prev_ack_number = 0 else:''' self.prev_ack_number = packet.Acknowledgement self.pop_sending_window(packet.Acknowledgement) #self.prev_ack_number = packet.Acknowledgement print("ACK Received from the server. Removing data from buffer.", packet.Acknowledgement) self.global_received_ack = packet.Acknowledgement #self.pop_sending_window(packet.Acknowledgement) elif packet.Type == 3: if checkvalue: print("RIP Received from Server. Sending RIP-ACK") # RIPack ripack = PEEPpacket() self.exc = 0 ripack.Type = 4 ripack.Acknowledgement = packet.SequenceNumber + 1 ripack.SequenceNumber = 5555 calcChecksum = PEEPClient(self.loop) ripack.Checksum = calcChecksum.calculateChecksum(ripack) ripz = ripack.__serialize__() self.transport.write(ripz) else: print("Corrupt RIP packet received. Please check on server end.") elif packet.Type == 4: if checkvalue: print("RIP-ACK Received from Server. Closing down the connection.") self.exc = 0 self.connection_lost(self.exc) else: print("Corrupt RIP-ACK packet received. Please check on server end.") else: print("======== Incorrect packet received. Closing connection!=========\n") self.transport.close() def sendack(self, ackno): print ("================== Sending ACK ================\n") ack = PEEPpacket() calcChecksum = PEEPClient(self.loop) ack.Type = 2 ack.Acknowledgement = ackno print ("ACK No:" + str(ack.Acknowledgement)) # For debugging ack.Checksum = calcChecksum.calculateChecksum(ack) #print(ack.Checksum) bytes = ack.__serialize__() self.transport.write(bytes) '''def check_if_ack_received_before(self, packet): keylist = list(self.sending_window) self.keylist1 = sorted(keylist) if self.prev_ack_number == packet.Acknowledgement: print ("REceived two acks of the same value") print ("33333333333",self.keylist1) for key in self.keylist1: if key == packet.Acknowledgement: print ("found a key that equals the acknow received") packet_to_be_retrans = self.sending_window[self.keylist1[0]] print("So far so goood!") packet_to_be_retrans.Acknowledgment = self.global_number_ack bytes_retrans = packet_to_be_retrans.__serialize__() self.transport.write(bytes_retrans) print ("ready to return") return 1''' def write(self,data): print ("=================== Writing Data down to wire from Client ================\n") self.backlog_window.append(data) print("Post appending BL window in client", self.backlog_window) if self.sending_window_count <= 100: print("About to pop backlog in client") data_from_BL = self.backlog_window.pop(0) self.encapsulating_packet(data_from_BL) def encapsulating_packet(self,data_from_BL_1): udata = data_from_BL_1 self.data_from_BL = data_from_BL_1 i = 0 l = 1 while i < len(udata): # print("Chunk {}". format(l)) chunk, self.data_from_BL = self.data_from_BL[:1024], self.data_from_BL[1024:] self.Cencap = PEEPpacket() self.n += 1 calcChecksum = PEEPClient(self.loop) self.Cencap.Type = 5 self.Cencap.SequenceNumber = self.update_sequence(chunk) self.prev_sequence_number = self.Cencap.SequenceNumber # prev_sequence_number is the seq number of the packet sent by client print("SEQ No:" + str(self.Cencap.SequenceNumber)) self.Cencap.Acknowledgement = self.global_number_ack # print("ACK No:" + str(self.Cencap.Acknowledgement)) self.Cencap.Data = chunk # print ("Data is", chunk) print("Size of data", len(chunk)) self.Cencap.Checksum = calcChecksum.calculateChecksum(self.Cencap) if self.sending_window_count <= 100: # print (" Entered count ") self.Cencap = self.update_sending_window(self.Cencap) self.bytes = self.Cencap.__serialize__() i += 1024 l += 1 self.transport.write(self.bytes) # Creating timer for each data packet self.timer = PEEPClient(loop) self.tx = Timerx(0.1, self.data_timeout, self.Cencap) self.chabi = self.global_number_seq self.t[self.chabi] = self.tx else: print(" Sorry, window is full. ") i += 1024 #### Put some return statement to handle this exception. Code shouldn't hang. ### def receive_window(self, pkt): self.number_of_packs += 1 self.packet = pkt if self.packet.SequenceNumber == self.global_number_ack: self.global_number_ack = self.update_ack(self.packet.SequenceNumber, self.global_packet_size) #It's actually updating the expected Seq Number self.sendack(self.update_ack(self.packet.SequenceNumber, self.global_packet_size)) self.higherProtocol().data_received(self.packet.Data) self.check_receive_window() elif self.number_of_packs <= 100: #and self.packet.SequenceNumber <= self.global_number_ack + (1024*1000): self.recv_window[self.packet.SequenceNumber] = self.packet.Data self.sendack(self.global_number_ack) else: print ("Receive window is full or the packet has already been received!") def check_receive_window(self): sorted_list = [] sorted_list = self.recv_window.keys() for k in sorted_list: if k == self.global_number_ack: self.packet_to_be_popped = self.recv_window[k] self.sendack(self.update_ack(self.packet_to_be_popped.SequenceNumber, self.global_packet_size)) self.higherProtocol().data_received(self.packet_to_be_popped.Data) else: return prev_packet_size = 0 def calculate_length(self, data): self.prev_packet_size = len(data) def update_sequence(self, data): if self.count_of_function_call == 0: self.count_of_function_call = 1 self.calculate_length(data) return self.global_number_seq #for first packet this is equal to synack.ackno else: self.global_number_seq = self.prev_sequence_number + self.prev_packet_size self.calculate_length(data) return self.global_number_seq def update_ack(self, received_seq_number, size): self.received_seq_number = received_seq_number self.global_number_ack = self.received_seq_number + size return self.global_number_ack def update_sending_window(self, packet): self.packet = packet self.sending_window_count += 1 self.key = self.global_number_seq #self.key = self.prev_sequence_number + self.prev_packet_size #removed this because it is redundant to the previous line. self.sending_window[self.key] = self.packet #for k,v in self.sending_window.items(): #print ("Key is: ",k, "Packet is: ", v) #self.sending_window = (sorted(self.sending_window.items())) keylist = self.sending_window.keys() self.keylist1 = sorted(keylist) print("###########################################", self.keylist1) #print("Sorted keys list is", keylist) #print("dic type is", type(self.sending_window)) return self.packet def pop_sending_window(self, AckNum): #print (" Entered Popping Values ") self.AckNum = AckNum print (" Ack Number is: ", self.AckNum) #self.sending_window = OrderedDict(sorted(self.sending_window.items())) #print("Keylist1 is", self.keylist1) for key in self.keylist1: #print ("Key is: ", key) if (self.AckNum > key): #print("Inside Acknum loo.") #print("The current Dictionary is", self.sending_window) #Finishing off timers for the packets with ACKs received ''' seqs = list(self.t.keys()) for chabi in seqs: if self.AckNum > chabi: (self.t[chabi]).cancel() self.t.pop(chabi) ''' #print("Key value to pop is", key) self.sending_window.pop(key) print ("sending window count is",self.sending_window_count) self.sending_window_count = self.sending_window_count - 1 if self.sending_window_count <= 100: print("About to pop backlog") data_from_BL = self.backlog_window.pop(0) self.encapsulating_packet(data_from_BL) #else: #print (" Popped all packets ") #self.k self.keylist1 = [] return def close(self): #RIPpacket rip = PEEPpacket() rip.Type = 3 rip.Acknowledgement = 0 rip.SequenceNumber = 9999 calcChecksum = PEEPClient(self.loop) rip.Checksum = calcChecksum.calculateChecksum(rip) ripz = rip.__serialize__() self.transport.write(ripz) def connection_lost(self,exc): print ("============== PEEPClient Closing connection ===========\n") self.transport.close() self.loop.stop() #Timer Function code block starts here class Timerx(): def __init__(self, timeout, callback, packet): self._timeout = timeout self._callback = callback self.packet = packet self.flag = 0 self._task = asyncio.ensure_future(self._job()) async def _job(self): await asyncio.sleep(self._timeout) await self._callback() def cancel(self): self._task.cancel() loop = asyncio.get_event_loop() #logging.getLogger().setLevel(logging.NOTSET) # this logs *everything* #logging.getLogger().addHandler(logging.StreamHandler()) # logs to stderr Clientfactory = StackingProtocolFactory(lambda: PEEPClient(loop)) '''if __name__ == "__main__": loop = asyncio.get_event_loop() logging.getLogger().setLevel(logging.NOTSET) # this logs *everything* logging.getLogger().addHandler(logging.StreamHandler()) # logs to stderr Clientfactory = StackingProtocolFactory(lambda: PEEPClient(loop)) ptConnector = playground.Connector(protocolStack=Clientfactory) playground.setConnector("passthrough", ptConnector) go = initiate(loop) coro = playground.getConnector('passthrough').create_playground_connection(go.send_first_packet, '20174.1.1.1', 8888) loop.run_until_complete(coro) loop.run_forever() loop.close()'''
[ "noreply@github.com" ]
antarasargam.noreply@github.com
e4c55d2046bfa5e009793218e4f1236cc147abc2
b58bf20946b2f0e43ccacd4d8206d513ba5ce6dd
/vecap/Churches/models.py
b57041b56fd66a90dc10f7f4f9bfb22b3af229de
[]
no_license
ifeoluwaDeterminas/vecap
fb5540838ef170fbd91b2075e66a1d19a7df6ade
cc7bc941d82ee3c8852da6fd946756703629b62e
refs/heads/main
2023-08-16T01:55:45.949424
2021-10-11T08:20:19
2021-10-11T08:20:19
null
0
0
null
null
null
null
UTF-8
Python
false
false
394
py
from django.db import models from django_tenants.models import TenantMixin, DomainMixin # Create your models here. class Client(TenantMixin): name = models.CharField(max_length=50) description = models.CharField(max_length=255) on_trial = models.BooleanField() created_on = models.DateField(auto_now_add=True) auto_create_schema = True class Domain(DomainMixin): pass
[ "jotunbade@gmail.com" ]
jotunbade@gmail.com
f4eb52622028a08e0dec011b2b776b1650007f4e
00c33337f4023c8d257a7da1c47db4bf36441b94
/events/migrations/0002_auto_20190807_1713.py
c16f45334c7935886774babcf5dea0f26385fdd3
[]
no_license
ECellNitrr/EcellWeb2k19
694d733737b1ac26c136994ee631f36904ca942d
75d53aad788d6deac8c5dda72cabf34b8d4bcb15
refs/heads/master
2021-06-24T06:23:09.097667
2021-01-12T08:37:48
2021-01-12T08:37:48
189,839,122
4
23
null
2021-01-12T08:24:49
2019-06-02T11:39:44
JavaScript
UTF-8
Python
false
false
828
py
# Generated by Django 2.2.2 on 2019-08-07 17:13 from django.conf import settings from django.db import migrations, models import django.db.models.deletion class Migration(migrations.Migration): initial = True dependencies = [ ('events', '0001_initial'), migrations.swappable_dependency(settings.AUTH_USER_MODEL), ] operations = [ migrations.AddField( model_name='eventregister', name='user', field=models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to=settings.AUTH_USER_MODEL), ), migrations.AddField( model_name='event', name='ecell_user', field=models.ForeignKey(blank=True, null=True, on_delete=django.db.models.deletion.SET_NULL, to=settings.AUTH_USER_MODEL), ), ]
[ "naveennvrgup@gmail.com" ]
naveennvrgup@gmail.com
b1bfb061db1148311e785c410fd9730250806673
cf954be6c93a3dcd81f5b094faf005af371336f0
/src/utils/__init__.py
43d3f48b46546235749fa29c39d05ced1e1c54ac
[]
no_license
urielsinger/text2Mol
819e20ea2638ddc97ecf7d57944ee1faaa8c97ff
2d8b0cfb17414f43904856944e3d6f0a11fd2b96
refs/heads/master
2020-07-05T18:53:33.064530
2019-08-16T14:04:05
2019-08-16T14:04:05
202,737,161
1
0
null
null
null
null
UTF-8
Python
false
false
202
py
from utils.datetime_utils import * from utils.general_utils import * from utils.graph_utils import * from utils.visualize_utils import * from utils.text_utils import * from utils.molecule_utils import *
[ "urielsinger@gmail.com" ]
urielsinger@gmail.com
befaa9ced886f7bb04ea7ea29f60e173dddb39b8
3c660571c9c2b028a88b844bd8c376970bea9f4b
/loginRegister/models.py
5098e0f9afd5240a604226763323970cbcd8771a
[]
no_license
felipesma/almacen
c1914b30b63d43ca2ed7bc87db4d63a8c0abfd44
ba2ddbff635db0ae1be7becc1a4ef6520cdaebbf
refs/heads/main
2023-08-18T14:00:12.886588
2021-10-26T15:52:00
2021-10-26T15:52:00
420,873,047
1
0
null
null
null
null
UTF-8
Python
false
false
1,388
py
from django.db import models # Create your models here. class UserManager(models.Manager): def basic_validator(self, postData): errors = {} if postData['password'] != postData['password_confirmation']: errors['password_match'] = "Las constraseñas no coinciden, favor reintente." if len(postData['password']) < 8: errors['len_password'] = "La contraseña debe tener al menos 8 carácteres." return errors class Usuario(models.Model): nombre = models.CharField(max_length=255) email = models.EmailField(max_length=255) direccion = models.CharField(max_length=255) telefono = models.CharField(max_length=15) password = models.CharField(max_length=255) objects = UserManager() nivel = models.IntegerField(default=1) class Categoria(models.Model): nombre = models.CharField(max_length=50) class Producto(models.Model): producto = models.CharField(max_length=50) precio = models.IntegerField() categoria = models.ForeignKey(Categoria, related_name="productos", on_delete=models.CASCADE) class Pedido(models.Model): cliente = models.ForeignKey(Usuario, related_name="pedidos", on_delete=models.CASCADE) productos = models.TextField(default='null') total = models.IntegerField(default=0) estado = models.IntegerField(default=1) pago = models.IntegerField(default=1)
[ "felipem39@gmail.com" ]
felipem39@gmail.com
434fad0eaa4c453385b4cd6adfaddfc88e0bb4e4
4db7e83f27a07c7838b80ab5cb25a01da4db7199
/main.py
acea4071a2692d7355b771ff79fc5c632c99b704
[ "MIT" ]
permissive
knowmetoowell/PUBG-API
8f01a0bd65134f98e3d619bb25a27cdcba8a7022
d26c4ebc054750cb2a28eba8feff00674e34f263
refs/heads/main
2023-06-09T18:26:04.271997
2021-07-02T13:35:34
2021-07-02T13:35:34
null
0
0
null
null
null
null
UTF-8
Python
false
false
19,388
py
import asyncio import pymysql import os import sys import json import aiohttp import importlib import datetime from math import trunc from sanic import Sanic from pytz import timezone import sanic.response as response from log_config import LOGGING_CONFIG app = Sanic(__name__, log_config=LOGGING_CONFIG) platform_name = ["Steam","Kakao","XBOX","PS","Stadia"] platform_site = ["steam","kakao","xbox","psn","stadia"] DB_platform = ["Steam","Kakao","XBOX","PSN","Stadia"] directory = os.path.dirname(os.path.abspath(__file__)).replace("\\","/") db_f = open(f"{directory}/data/database.json",mode='r') db = db_f.read() db_f.close() db_json = json.loads(db) db_ip = db_json["mysql"]["ip"] db_user = db_json["mysql"]["user"] db_pw = db_json["mysql"]["password"] db_name = db_json["mysql"]["database"] connect = pymysql.connect(host=db_ip, user=db_user, password=db_pw,db=db_name, charset='utf8') #클라이언트 API키 불러오기. cur = connect.cursor() cur.execute("SELECT * from PUBG_BOT") client_list = cur.fetchall() pubg_token = client_list[0][2] connect.close() sys.path.append(directory + "/modules") #다른 파일내 함수추가 p_info = importlib.import_module("player") s_info = importlib.import_module("status") header = { "Authorization": "Bearer " + pubg_token, "Accept": "application/vnd.api+json" } sample_f = open(f"{directory}/data/last_update_sample.json",mode='r') sample1 = json.loads(sample_f.read()) sample_f.close() async def get_season(pubg_platform): connect = pymysql.connect(host=db_ip, user=db_user, password=db_pw, db='PUBG_BOT', charset='utf8') cur = connect.cursor() sql = f"SELECT {DB_platform[pubg_platform]} FROM SEASON_STATUS" cur.execute(sql) cache = cur.fetchone() html = cache[0] data_json = json.loads(html)['data'] for i in data_json: if i['attributes']['isCurrentSeason']: least_season = i return least_season['id'] def time_num(f_playtime): playtime = datetime.datetime.fromtimestamp(f_playtime, timezone('UTC')) if playtime.month == 1: if playtime.day == 1: if playtime.hour == 0: if playtime.minute == 0: return f"{playtime.second}초" return f"{playtime.minute}분 {playtime.second}초" return f"{playtime.hour}시간 {playtime.minute}분 {playtime.second}초" return f"{playtime.day-1}일 {playtime.hour}시간 {playtime.minute}분 {playtime.second}초" return f"{playtime.month-1}달 {playtime.day-1}일 {playtime.hour}시간 {playtime.minute}분 {playtime.second}초" @app.route("/api/PUBG/") async def main(request): return response.redirect("https://github.com/team-alpha-kr/PUBG-API") @app.route("/api/PUBG/player") async def player(request): args = request.get_args(keep_blank_values=True) if not ("nickname" in args): return response.json({'code':'01', 'msg':"Please write your nickname."}, status=400) else: nickname = args['nickname'][0] connect = pymysql.connect(host=db_ip, user=db_user, password=db_pw,db=db_name, charset='utf8') cur = connect.cursor() exist_nickname = pymysql.escape_string("SELECT EXISTS (SELECT name FROM player WHERE name=%s) as succees;") cur.execute(exist_nickname,(nickname)) exist = cur.fetchone() if exist[0]: command = pymysql.escape_string("SELECT id, name, platform, last_update FROM player WHERE name=%s") cur.execute(command,(nickname)) fetch = cur.fetchone() connect.close() data = { "id":fetch[0], "nickname":fetch[1], "platform":fetch[2], "lastupdate":json.loads(fetch[3]) } return response.json(data,status=200) else: if not ("platform" in args): return response.json({"code":"02","msg":"The value is not stored in DB, so you need to create a platform."},status=400) else: try: platform = int(args['platform'][0]) except ValueError: return response.json({'code':'06', 'msg':"Platform values can only contain numbers."}, status=400) if not (platform >= 0 and platform < 5): return response.json({'code':'07', 'msg':"Platform values can contain only 0-4 values."}, status=400) async with aiohttp.ClientSession() as session: async with session.get(f"https://api.pubg.com/shards/{platform_site[platform]}/players?filter[playerNames]={nickname}", headers=header) as resp: if resp.status == 200: json_data = await resp.json() else: e_resp = s_info.response_num(resp.status) print(await resp.json(),resp.status) return response.json({'code': e_resp[1], 'msg': e_resp[2]}, status=e_resp[0]) data = { "id":json_data["data"][0]["id"], "nickname":json_data["data"][0]["attributes"]["name"], "platform":platform, "lastupdate":sample1 } command = pymysql.escape_string("insert into player(id,name,last_update,platform) value(%s,%s,%s,%s)") cur.execute(command,(json_data["data"][0]["id"],json_data["data"][0]["attributes"]["name"],json.dumps(sample1),platform)) connect.commit() connect.close() return response.json(data,status=200) @app.route("/api/PUBG/normal") async def normal_status(request): args = request.get_args(keep_blank_values=True) if not ("id" in args): return response.json({'code':'01', 'msg':"Please write your id."}, status=400) else: pubg_id = args['id'][0] connect = pymysql.connect(host=db_ip, user=db_user, password=db_pw,db=db_name, charset='utf8') cur = connect.cursor() exist_nickname = pymysql.escape_string("SELECT EXISTS (SELECT name FROM player WHERE id=%s) as succees;") cur.execute(exist_nickname,(pubg_id)) fetch1 = cur.fetchone() if fetch1[0]: command = pymysql.escape_string("SELECT platform FROM player WHERE id=%s") cur.execute(command, (pubg_id)) platform_info = cur.fetchone()[0] else: return response.json({'code':'05', 'msg':"No information about the user was found. Please proceed with \"/PUBG/player\" first."}, status=400) if ("season" in args): try: season = int(args['season'][0]) except ValueError: return response.json({'code':'08', 'msg':"Season values can only contain numbers."}, status=400) if platform_info >= 0 and platform_info <= 1: type_season = "pc-2018" else: type_season = "console" if len(str(season)) < 2: season = f"division.bro.official.{type_season}-0{season}" else: season = f"division.bro.official.{type_season}-{season}" else: season = await get_season(platform_info) status, html = await s_info.season_status(pubg_id,platform_info,season) if not status: return response.json({'code': html[1], 'msg': html[2]}, status=html[0]) else: data = { "id":pubg_id, "gameMode":{} } gamestat = html['data']['attributes']['gameModeStats'] for i in ['solo','solo-fpp','duo','duo-fpp','squad','squad-fpp']: modestat = gamestat[i] losses = modestat['losses'] if losses == 0: losses = 1 KDA_point = round((modestat['assists'] + modestat['kills']) / losses,2) KD_point = round(modestat['kills'] / losses,2) i_data = { i:{ "assists":modestat['assists'], "boosts": modestat['boosts'], "dBNOs": modestat['dBNOs'], "dailyKills": modestat['dailyKills'], "dailyWins": modestat['dailyWins'], "damageDealt": modestat['damageDealt'], "days": modestat['days'], "headshotKills": modestat['headshotKills'], "heals": modestat['heals'], "KDA_point": KDA_point, "KD_point": KD_point, "kills": modestat['kills'], "longestKill": modestat['longestKill'], "longestTimeSurvived": modestat['longestTimeSurvived'], "longestTimeSurvivedAnswer": time_num(modestat['longestTimeSurvived']), "losses": modestat['losses'], "maxKillStreaks": modestat['maxKillStreaks'], "mostSurvivalTime": modestat['mostSurvivalTime'], "revives": modestat['revives'], "rideDistance": modestat['rideDistance'], "roadKills": modestat['roadKills'], "roundMostKills": modestat['roundMostKills'], "roundsPlayed": modestat['roundsPlayed'], "suicides": modestat['suicides'], "swimDistance": modestat['swimDistance'], "teamKills": modestat['teamKills'], "timeSurvived": modestat['timeSurvived'], "timeSurvivedAnswer": time_num(modestat['timeSurvived']), "top10s": modestat['top10s'], "vehicleDestroys": modestat['vehicleDestroys'], "walkDistance": modestat['walkDistance'], "weaponsAcquired": modestat['weaponsAcquired'], "weeklyKills": modestat['weeklyKills'], "weeklyWins": modestat['weeklyWins'], "wins": modestat['wins'] } } data['gameMode'].update(i_data) return response.json(data, status=200) @app.route("/api/PUBG/normal/update") async def update_normal_status(request): args = request.get_args(keep_blank_values=True) if not ("id" in args): return response.json({'code':'01', 'msg':"Please write your id."}, status=400) else: pubg_id = args['id'][0] connect = pymysql.connect(host=db_ip, user=db_user, password=db_pw,db=db_name, charset='utf8') cur = connect.cursor() exist_nickname = pymysql.escape_string("SELECT EXISTS (SELECT name FROM player WHERE id=%s) as succees;") cur.execute(exist_nickname,(pubg_id)) fetch1 = cur.fetchone() if fetch1[0]: command = pymysql.escape_string("SELECT platform FROM player WHERE id=%s") cur.execute(command, (pubg_id)) platform_info = cur.fetchone()[0] else: return response.json({'code':'05', 'msg':"No information about the user was found. Please proceed with \"/PUBG/player\" first."}, status=400) if ("season" in args): try: season = int(args['season'][0]) except ValueError: return response.json({'code':'08', 'msg':"Season values can only contain numbers."}, status=400) if platform_info >= 0 and platform_info <= 1: type_season = "pc-2018" else: type_season = "console" if len(str(season)) < 2: season = f"division.bro.official.{type_season}-0{season}" else: season = f"division.bro.official.{type_season}-{season}" else: season = await get_season(platform_info) await s_info.season_status_update(pubg_id, platform_info, season) return response.json({ "code":"00", "msg":"Updated successfully." },status=200) @app.route("/api/PUBG/ranked") async def ranked_status(request): args = request.get_args(keep_blank_values=True) if not ("id" in args): return response.json({'code':'01', 'msg':"Please write your id."}, status=400) else: pubg_id = args['id'][0] connect = pymysql.connect(host=db_ip, user=db_user, password=db_pw,db=db_name, charset='utf8') cur = connect.cursor() exist_nickname = pymysql.escape_string("SELECT EXISTS (SELECT name FROM player WHERE id=%s) as succees;") cur.execute(exist_nickname,(pubg_id)) fetch1 = cur.fetchone() if fetch1[0]: command = pymysql.escape_string("SELECT platform FROM player WHERE id=%s") cur.execute(command, (pubg_id)) platform_info = cur.fetchone()[0] else: return response.json({'code':'05', 'msg':"No information about the user was found. Please proceed with \"/PUBG/player\" first."}, status=400) if ("season" in args): try: season = int(args['season'][0]) except ValueError: return response.json({'code':'08', 'msg':"Season values can only contain numbers."}, status=400) if platform_info >= 0 and platform_info <= 1: type_season = "pc-2018" else: type_season = "console" if len(str(season)) < 2: season = f"division.bro.official.{type_season}-0{season}" else: season = f"division.bro.official.{type_season}-{season}" else: season = await get_season(platform_info) status, html = await s_info.ranked_status(pubg_id,platform_info,season) if not status: return response.json({'code': html[1], 'msg': html[2]}, status=html[0]) else: data = { "id":pubg_id, "gameMode":{} } gamestat = html['data']['attributes']['rankedGameModeStats'] for i in ['solo','solo-fpp','squad','squad-fpp']: if not (i in gamestat): i_data = { i: { "assists": 0, "avgRank": 0, "currentRank":{ "tier":"Unranked", "subTier":"1" }, "currentRankAnswer":"Unranked", "currentRankPoint":0, "bestRank":{ "tier":"Unranked", "subTier":"1" }, "bestRankAnswer":"Unranked", "bestRankPoint": 0, "damageDealt": 0, "deaths": 0, "dBNOs": 0, "KDA_point": 0, "KD_point": 0, "kills": 0, "top10s": 0, "top10_point": 0, "wins": 0, "win_point": 0 } } data['gameMode'].update(i_data) continue modestat = gamestat[i] losses = modestat['deaths'] if losses == 0: losses = 1 KD_point = round(modestat['kills'] / losses,2) currentTier1 = modestat["currentTier"]["tier"] currentTier2 = modestat["currentTier"]["subTier"] bestTier1 = modestat["bestTier"]["tier"] bestTier2 = modestat["bestTier"]["subTier"] if currentTier1 == "Unranked" or currentTier1 == "Master": tier_name1 = currentTier1 else: tier_name1 = f"{currentTier1} {currentTier2}" if bestTier1 == "Unranked" or bestTier1 == "Master": tier_name2 = bestTier1 else: tier_name2 = f"{bestTier1} {bestTier2}" i_data = { i:{ "assists": modestat['assists'], "avgRank": modestat['avgRank'], "currentRank":modestat['currentTier'], "currentRankAnswer":tier_name1, "currentRankPoint":modestat['currentRankPoint'], "bestRank":modestat['bestTier'], "bestRankAnswer":tier_name2, "bestRankPoint": modestat['bestRankPoint'], "damageDealt": modestat['damageDealt'], "deaths": modestat['deaths'], "dBNOs": modestat['dBNOs'], "KDA_point": modestat['kda'], "KD_point": KD_point, "kills": modestat['kills'], "roundsPlayed": modestat['roundsPlayed'], "top10s": trunc(modestat['top10Ratio'] * modestat['roundsPlayed']), "top10_point": modestat['top10Ratio'], "wins": modestat['wins'], "win_point": modestat['winRatio'] } } data['gameMode'].update(i_data) return response.json(data, status=200) @app.route("/api/PUBG/ranked/update") async def update_ranked_status(request): args = request.get_args(keep_blank_values=True) if not ("id" in args): return response.json({'code':'01', 'msg':"Please write your id."}, status=400) else: pubg_id = args['id'][0] connect = pymysql.connect(host=db_ip, user=db_user, password=db_pw,db=db_name, charset='utf8') cur = connect.cursor() exist_nickname = pymysql.escape_string("SELECT EXISTS (SELECT name FROM player WHERE id=%s) as succees;") cur.execute(exist_nickname,(pubg_id)) fetch1 = cur.fetchone() if fetch1[0]: command = pymysql.escape_string("SELECT platform FROM player WHERE id=%s") cur.execute(command, (pubg_id)) platform_info = cur.fetchone()[0] else: return response.json({'code':'05', 'msg':"No information about the user was found. Please proceed with \"/PUBG/player\" first."}, status=400) if ("season" in args): try: season = int(args['season'][0]) except ValueError: return response.json({'code':'08', 'msg':"Season values can only contain numbers."}, status=400) if platform_info >= 0 and platform_info <= 1: type_season = "pc-2018" else: type_season = "console" if len(str(season)) < 2: season = f"division.bro.official.{type_season}-0{season}" else: season = f"division.bro.official.{type_season}-{season}" else: season = await get_season(platform_info) await s_info.ranked_status_update(pubg_id, platform_info, season) return response.json({ "code":"00", "msg":"Updated successfully." },status=200) @app.route("/api/PUBG/player/change_platform") async def change_platform(request): args = request.get_args(keep_blank_values=True) if not ("nickname" in args): return response.json({'code':'01', 'msg':"Please write your nickname."}, status=400) else: nickname = args['nickname'][0] connect = pymysql.connect(host=db_ip, user=db_user, password=db_pw,db=db_name, charset='utf8') cur = connect.cursor() exist_nickname = pymysql.escape_string("SELECT EXISTS (SELECT name FROM player WHERE name=%s) as succees;") cur.execute(exist_nickname,(nickname)) exist = cur.fetchone() if exist[0]: if not ("platform" in args): return response.json({'code':'02', 'msg':"Please write your platform."}, status=400) else: try: platform = int(args['platform'][0]) except ValueError: return response.json({'code':'06', 'msg':"Platform values can only contain numbers."}, status=400) if not (platform >= 0 and platform < 5): return response.json({'code':'07', 'msg':"Platform values can contain only 0-4 values."}, status=400) command = pymysql.escape_string("UPDATE player SET platform=%s WHERE name=%s") cur.execute(command,(platform,nickname)) connect.commit() connect.close() return response.json({ "code":"00", "msg":"Updated successfully." },status=200) else: connect.close() return response.json({'code': '05','msg': "No information about the user was found. Please proceed with \"/PUBG/player\" first."},status=400) app.run('127.0.0.1', 3200)
[ "gunyu1019@gmail.com" ]
gunyu1019@gmail.com
09e9bd7c4d26cbef87fec0324296b7b0e260ef3a
5453a23a6c59bd354e2b07f87336b8fb3a618741
/Growth curve/doubleing_time.py
31edccbc56f2420eaf0340889d88101573cd5f5b
[]
no_license
MinTTT/pycode
bfe7d311f7bd14c9775a8d8c5b6a4969b0d1ec57
7658b1460d0ca7bef50937bbc18f6bd4bcdb1b4e
refs/heads/master
2020-04-01T04:09:32.787824
2018-10-13T08:54:41
2018-10-13T08:54:41
152,852,024
0
0
null
null
null
null
UTF-8
Python
false
false
552
py
# -*- coding: utf-8 -*- ''' This module need three parameters x: time sequence, y: ABS or OD max: logmarthtic midpoint ''' import numpy as np from scipy import optimize def doutime(x,y,max): logtime=x[np.arange(max-38,max-10,1).tolist()] logod=y.iloc[np.arange(max-38,max-10,1).tolist()] def residuals(p): k,b=p return np.log(logod) - (k*logtime+b) r=optimize.leastsq(residuals,[0,-10]) k,b=r[0] k=np.log(2)/k b=np.e**b print("Doubling time =",k,"min.") fity=b*(2**(logtime/k)) return k,logtime,fity
[ "34529833+MinTTT@users.noreply.github.com" ]
34529833+MinTTT@users.noreply.github.com
979b4aba56fd5ea83870647b5611677ae5844785
ce1b0adade9b0cab3bfa5299f985a3d0d59a40cd
/session_key_test.py
9a478a1e37167f80b5b494705f84137d584b036a
[ "Apache-2.0" ]
permissive
xyhlk520/GPG-Decrypt
5e5f8522429df2aaf6026b603ef848849f439396
43efec3890b5ff9c905180eef6778a94f10bb6b0
refs/heads/main
2023-03-17T20:55:19.215318
2021-03-14T09:37:10
2021-03-14T09:37:10
347,595,523
0
0
null
null
null
null
UTF-8
Python
false
false
158
py
passphrase = "test" session_key = b"\x09\x8f\x6b\xcd\x46\x21\xd3\x73\xca\xde\x4e\x83" + \ b"\x26\x27\xb4\xf6\x5f\x8f\x8e\x05\xef\xdc\x22\xe8"
[ "noreply@github.com" ]
xyhlk520.noreply@github.com
04a6497aa4fb2f2caa9082dd44f7be71664ea415
1f96e1101cf253a94efa824a0406d807fd93acc1
/archive/MyBot_213.py
a4963ca3dc631aa58ab5ddfd54bfe1a6dc0445eb
[]
no_license
sp00/google_ai_bot
228f48302e556271ed6714ebad9c0e449e992f25
1f3f77a00dc8289ad8e5ae9378b748fb099bc756
refs/heads/master
2020-12-25T00:49:59.968563
2010-11-30T03:14:16
2010-11-30T03:14:16
null
0
0
null
null
null
null
UTF-8
Python
false
false
35,352
py
from planetwars import BaseBot, Game from planetwars.universe3 import Universe3 from planetwars.planet import Planet from planetwars.player import PLAYER1, PLAYER2, NOBODY from planetwars.planet2 import Planet2, getLogger from planetwars.universe import player, Fleet from logging import getLogger import planetwars.planet from math import ceil from copy import copy import random HORIZON_FIRST = 40 HORIZON = 40 ATTACK_SCORE_THRESHOLD_FIRST = 0 ATTACK_SCORE_THRESHOLD = 140 ATTACK_SCORE_ENEMY_MULTIPLIER = 2 log = getLogger(__name__) def zeros(rows,cols): row = [] data = [] for i in range(cols): row.append(0) for i in range(rows): data.append(row[:]) return data # v = list of item values or profit # w = list of item weight or cost # W = max weight or max cost for the knapsack def zeroOneKnapsack(v, w, W): # c is the cost matrix c = [] n = len(v) c = zeros(n,W+1) for i in range(0,n): #for ever possible weight for j in range(0,W+1): #can we add this item to this? if (w[i] > j): c[i][j] = c[i-1][j] else: c[i][j] = max(c[i-1][j],v[i] +c[i-1][j-w[i]]) return [c[n-1][W], getUsedItems(w,c)] # w = list of item weight or cost # c = the cost matrix created by the dynamic programming solution def getUsedItems(w,c): # item count i = len(c)-1 currentW = len(c[0])-1 # set everything to not marked marked = [] for i in range(i+1): marked.append(0) while (i >= 0 and currentW >=0): if (i==0 and c[i][currentW] >0 )or c[i][currentW] != c[i-1][currentW]: marked[i] =1 currentW = currentW-w[i] i = i-1 return marked class Move(object): def __init__(self, source, target, turn, ship_count): self.source = source self.target = target self.turn = turn self.ship_count = int(ship_count) def __repr__(self): return "Move from %s to %s at turn %s with %s ships" % (self.source, self.target, self.turn, self.ship_count) class MyBot(BaseBot): def __init__(self, universe): self.universe = universe self.scheduled_moves_at_turn= {} def total_fleet_ship_count(self, owner): return sum( [ fleet.ship_count for fleet in self.universe.find_fleets(owner) ] ) def get_neutrals_under_player_attack(self, player): result = [] for planet in self.nobodies_planets: if sum( [ 1 for fleet in planet.attacking_fleets if fleet.owner == player ] ) > 0: result.append(planet) return result def get_available_ships_within_distance(self, planet_to_attack, player, distance): result = 0 for planet in (list(self.universe.find_planets(player)) + self.get_neutrals_under_player_attack(player)): if planet.id != planet_to_attack.id and planet.distance(planet_to_attack) <= distance and self.ships_needed[planet] == 0: ships_avail = self.ships_available_at_turn[planet][distance-planet.distance(planet_to_attack)] # if planet_to_attack.id == 0: # log.info("get avail from %s = %s" % (planet, ships_avail)) result += ships_avail return result def get_attack_score(self, planet_to_attack, future_owner, distance): turns = self.max_distance_between_planets - distance + HORIZON attack_score = turns * planet_to_attack.growth_rate if future_owner in player.ENEMIES: attack_score *= ATTACK_SCORE_ENEMY_MULTIPLIER return attack_score def get_attack_score_200(self, planet_to_attack, future_owner, distance): profit_turns = max(200 - self.current_turn - distance, 0) attack_score = profit_turns * planet_to_attack.growth_rate if future_owner in player.ENEMIES: attack_score *= 2 return attack_score def get_scheduled_fleets_to(self, planet): result = [] for moves in self.scheduled_moves_at_turn.values(): for move in moves: if move.target == planet: distance = move.source.distance(move.target) turns_remaining = distance + (move.turn - self.universe.game.turn_count) fleet = Fleet(self.universe,random.randint(1,1000000),1, move.ship_count, move.source.id, move.target.id, distance, turns_remaining) result.append(fleet) return result def get_scheduled_fleets_from(self, planet): result = [] for moves in self.scheduled_moves_at_turn.values(): for move in moves: if move.source == planet: turns_remaining = move.turn - self.universe.game.turn_count fleet = Fleet(self.universe,random.randint(1,1000000),1, move.ship_count, move.source.id, move.target.id, turns_remaining, turns_remaining) result.append(fleet) return result def get_scheduled_fleets_shipcount_from_within_distance(self, planet, turns): result = 0 for moves in self.scheduled_moves_at_turn.values(): for move in moves: if move.source == planet: turns_remaining = move.turn - self.universe.game.turn_count if turns_remaining == turns: result += move.ship_count return result def get_attack_ship_count_first_turn(self, planet_to_attack, my_home, enemy_home): my_dist = my_home.distance(planet_to_attack) enemy_dist = enemy_home.distance(planet_to_attack) if my_dist < enemy_dist: return planet_to_attack.ship_count+1 if my_dist == enemy_dist and planet_to_attack.ship_count <= planet_to_attack.growth_rate: return planet_to_attack.ship_count+1 return 1000000 def closest_enemy_planet(self, p): if len(self.enemy_planets) == 0: return None sorted_planets = sorted(self.enemy_planets, key=lambda ep : p.distance(ep) + ep.id/1000000.0) return sorted_planets[0] def closest_enemy_planet_distance(self, p): return min((lambda ep:ep.distance(p))(ep) for ep in self.enemy_planets) def my_fleets_attacking(self, planet): return sum( [ 1 for fleet in planet.attacking_fleets if fleet.owner == player.ME] ) def closest_to_enemy_neutral_under_my_attack(self): best_distance = 1000000 result_planet = None for planet in self.nobodies_planets: if self.my_fleets_attacking(planet) > 0: distance = self.enemy_com.distance(planet) if distance < best_distance: best_distance = distance result_planet = planet return result_planet def decrease_ships_available(self, planet, start_turn, ship_count): for turn in range(start_turn, self.max_distance_between_planets + 21): self.ships_available_at_turn[planet][turn] -= ship_count def send_fleet(self, source, target, ship_count): if source.owner == PLAYER1 and ship_count > 0 and ship_count <= source.ship_count: source.send_fleet(target, ship_count) else: log.info("Error sending fleet from %s to %s with % ships" % (source, target, ship_count)) def doScheduled(self): log.info("Scheduled move phase") # execute delayed moves first if self.scheduled_moves_at_turn.has_key(self.current_turn): for move in self.scheduled_moves_at_turn[self.current_turn]: #if move.ship_count <= move.source.ship_count and move.ship_count > 0 and move.source.owner == PLAYER1 and self.ships_available_at_turn[move.source][0] >= move.ship_count: #if move.ship_count <= move.source.ship_count and move.ship_count > 0 and move.source.owner == PLAYER1 and move.source.ship_count >= move.ship_count: if move.ship_count <= move.source.ship_count and move.ship_count > 0 and move.source.owner == PLAYER1 and move.source.ship_count >= move.ship_count and self.ships_available_at_turn[move.source][0] >= move.ship_count: self.send_fleet(move.source, move.target, move.ship_count) self.decrease_ships_available(move.source, 0, move.ship_count) #self.cumulative_ships_sent += move.ship_count #self.ships_available[move.source] -= move.ship_count else: log.info("Can't execute move: %s, ships avail: %s" % (move, self.ships_available_at_turn[move.source][0])) del self.scheduled_moves_at_turn[self.current_turn] def doPrep(self): log.info("Prep phase") if self.current_turn == 1: self.my_home = list(self.my_planets)[0] self.enemy_home = list(self.enemy_planets)[0] self.max_distance_between_planets = 0 for p1 in self.all_planets: for p2 in self.all_planets: self.max_distance_between_planets = max(self.max_distance_between_planets, p1.distance(p2)) #log.info("Max distance: %s" % self.max_distance_between_planets) # calculate current high level metrics self.total_ships = {PLAYER1:0, PLAYER2:0} self.total_growth_rate = {PLAYER1:0, PLAYER2:0} self.ships_available_at_turn = {} self.ships_needed = {} self.ships_needed_at_turn = {} self.ships_needed_timeline = {} self.planet_timeline = {} for planet in self.all_planets: self.ships_available_at_turn[planet] = {} scheduled_fleets_to_planet = self.get_scheduled_fleets_to(planet) scheduled_fleets_from_planet = self.get_scheduled_fleets_from(planet) self.planet_timeline[planet] = planet.in_future_timeline(self.max_distance_between_planets + 20, scheduled_fleets_to_planet, scheduled_fleets_from_planet) need_help = False # if planet.id == 7: # log.info("timeline for %s: %s" % (planet, self.planet_timeline[planet])) #log.info("attacking fleets by me: %s" % (self.universe.find_fleets(PLAYER1, destination=planet))) prev_owner = planet.owner for step in self.planet_timeline[planet]: owner = step[0] ship_count = step[1] if owner != prev_owner and prev_owner == planet.owner and prev_owner != NOBODY and not need_help: self.ships_needed[planet] = ship_count self.ships_needed_at_turn[planet] = self.planet_timeline[planet].index(step) + 1 need_help = True self.ships_needed_timeline[planet] = [ship_count] #log.info("Planet %s needs help %s at %s" % (planet, ship_count, self.ships_needed_at_turn[planet])) if need_help and owner == prev_owner: delta = self.planet_timeline[planet].index(step) + 1 - self.ships_needed_at_turn[planet] ships_needed_delta = ship_count - delta * 2 * planet.growth_rate self.ships_needed_timeline[planet].append(ships_needed_delta) prev_owner = owner if not need_help: self.ships_needed[planet] = 0 min_available = 1000000 step_index = len(self.planet_timeline[planet]) for step in reversed(self.planet_timeline[planet]): ship_count = step[1] min_available = min(min_available, ship_count) if step[0] == NOBODY: min_available = 0 if min_available < 0: log.info("Negative min_available: %s for %s" % (min_available, planet)) min_available = 0 self.ships_available_at_turn[planet][step_index] = min_available #log.info("avail for %s at %s: %s" % (planet, step_index, min_available)) step_index -= 1 self.ships_available_at_turn[planet][0] = max(0,min(planet.ship_count, self.ships_available_at_turn[planet][1] - planet.growth_rate)) else: for step_index in range(0, len(self.planet_timeline[planet])+1): self.ships_available_at_turn[planet][step_index] = 0 if planet.owner != NOBODY: self.total_ships[planet.owner] += planet.ship_count self.total_growth_rate[planet.owner] += planet.growth_rate # if planet.id == 14: # log.info("avail timeline for %s is: %s" % (planet, self.ships_available_at_turn[planet])) self.total_ships[PLAYER1] += self.total_fleet_ship_count(PLAYER1) self.total_ships[PLAYER2] += self.total_fleet_ship_count(PLAYER2) for my_planet in [self.my_home]: for enemy_planet in [self.enemy_home]: # if self.ships_available_at_turn[enemy_planet][0] < self.ships_available_at_turn[my_planet][0]: # continue if my_planet.owner != PLAYER1 or enemy_planet.owner != PLAYER2: continue max_enemy_fleet = self.ships_available_at_turn[enemy_planet][0] distance = my_planet.distance(enemy_planet) ships_needed_for_safety = max_enemy_fleet-(self.planet_timeline[my_planet][distance-1][1] - my_planet.ship_count) - enemy_planet.growth_rate #ships_needed_for_safety = max_enemy_fleet-(self.planet_timeline[my_planet][distance-1][1] - my_planet.ship_count) if ships_needed_for_safety > (my_planet.ship_count - self.ships_available_at_turn[my_planet][0]): deficit = ships_needed_for_safety - (my_planet.ship_count - self.ships_available_at_turn[my_planet][0]) #log.info("deficit for %s: %s, max enemy fleet %s" % (my_planet, deficit, max_enemy_fleet)) if deficit > self.ships_available_at_turn[my_planet][0]: deficit = self.ships_available_at_turn[my_planet][0] self.decrease_ships_available(my_planet, 0, deficit) # calculate enemy's center of mass weighted_x = 0 weighted_y = 0 div = 0 for planet in self.enemy_planets: weighted_x += planet.position.x * (self.ships_available_at_turn[planet][0] + planet.growth_rate) weighted_y += planet.position.y * (self.ships_available_at_turn[planet][0] + planet.growth_rate) div += self.ships_available_at_turn[planet][0] + planet.growth_rate if div == 0: div = 1 self.enemy_com = Planet(self.universe, 666, weighted_x/div, weighted_y/div, 2, 0, 0) # For every planet, and every turn, calculate how many ships each player can send to it # TODO should we use ships_available_at_turn here? self.max_aid_at_turn = {PLAYER1:{}, PLAYER2:{}} for player in (PLAYER1 | PLAYER2): source_planets = list(self.universe.find_planets(player)) + self.get_neutrals_under_player_attack(player) for planet in self.all_planets: self.max_aid_at_turn[player][planet] = {} for turn in range(1, self.max_distance_between_planets+21): max_aid = 0 for source_planet in source_planets: if source_planet.id != planet.id and planet.distance(source_planet) < turn: source_planet_time_step = self.planet_timeline[source_planet][turn - planet.distance(source_planet) - 1] if (source_planet_time_step[0] == player): #log.info("Max aid by %s for %s from %s at %s: %s" % (player.id, planet.id, source_planet.id, turn, source_planet_time_step[1])) max_aid += source_planet_time_step[1] else: if source_planet.id != planet.id and planet.distance(source_planet) == turn: if (source_planet.owner == player): max_aid += source_planet.ship_count self.max_aid_at_turn[player][planet][turn] = max_aid #log.info("Max aid by %s for %s at %s: %s" % (player.id, planet.id, turn, self.max_aid_at_turn[player][planet][turn])) log.info("MY STATUS: %s/%s" % (self.total_ships[PLAYER1], self.total_growth_rate[PLAYER1])) log.info("ENEMY STATUS: %s/%s" % (self.total_ships[PLAYER2], self.total_growth_rate[PLAYER2])) def doDefense(self): log.info("Defense phase") for planet_to_defend in sorted(self.my_planets, key=lambda p: p.growth_rate + p.id/1000000.0, reverse=True): ships_to_send = self.ships_needed[planet_to_defend] if ships_to_send <= 0: continue min_distance = self.max_distance_between_planets max_distance = self.ships_needed_at_turn[planet_to_defend] for my_planet in self.my_planets: distance = my_planet.distance(planet_to_defend) min_distance = min(min_distance, distance) min_distance = max(min_distance, 1) timeline = [elem for elem in self.ships_needed_timeline[planet_to_defend] if elem > 0] ship_counts_to_attempt = sorted(list(set(timeline)), key=lambda p : p, reverse=True) #log.info("evaluating defense for %s needed %s" % (planet_to_defend, ship_counts_to_attempt)) defended = False avail_ships_within_distance = {} for ships_to_send in ship_counts_to_attempt: for distance in range(min_distance, max_distance+1): # calculate if we can get enough ships from my planets to planet_to_defend within 'distance' turns ships_avail_to_defend = 0 if avail_ships_within_distance.has_key((planet_to_defend, distance)): ships_avail_to_defend = avail_ships_within_distance[(planet_to_defend, distance)] else: ships_avail_to_defend = self.get_available_ships_within_distance(planet_to_defend, PLAYER1, distance) avail_ships_within_distance[(planet_to_defend, distance)] = ships_avail_to_defend #log.info("Ships avail to defend %s within %s dist: %s" % (planet_to_defend, distance, ships_avail_to_defend)) if ships_avail_to_defend >= ships_to_send: ships_left_to_send = ships_to_send for source_planet in sorted(list(self.my_planets) + self.get_neutrals_under_player_attack(PLAYER1), key=lambda p : p.distance(planet_to_defend) + p.id/1000000.0): if self.ships_needed[source_planet] > 0: continue #log.info("evaluating for D: %s" % (source_planet)) current_distance = source_planet.distance(planet_to_defend) ships_avail = self.ships_available_at_turn[source_planet][distance-current_distance] if source_planet.id != planet_to_defend.id and ships_avail > 0: #log.info("Ships avail from %s: %s at dist %s, dist = %s" % (source_planet, ships_avail, current_distance, distance)) ships_to_send = min(ships_left_to_send, ships_avail) if current_distance == distance: #log.info("defending avail from %s: %s at dist %s" % (source_planet, ships_to_send, current_distance)) self.send_fleet(source_planet, planet_to_defend, ships_to_send) #self.cumulative_ships_sent += ships_to_send if current_distance < distance: future_turn = self.current_turn + (distance - current_distance) future_move = Move(source_planet, planet_to_defend, future_turn, ships_to_send) log.info("Scheduled move: %s" % future_move) if not self.scheduled_moves_at_turn.has_key(future_turn): self.scheduled_moves_at_turn[future_turn] = [] self.scheduled_moves_at_turn[future_turn].append(future_move) ships_left_to_send -= ships_to_send self.decrease_ships_available(source_planet, 0, ships_to_send) if ships_left_to_send == 0: defended = True break if defended: break if defended: break def doFirstTurnOffense(self): candidates = [] candidate_map = {} home_planet_distance = self.my_home.distance(self.enemy_home) ships_available = min(self.my_home.ship_count, self.my_home.growth_rate * (home_planet_distance+0)) i = 0 max_attack_distance=0 for p in sorted(self.nobodies_planets, key=lambda p : self.get_attack_ship_count_first_turn(p, self.my_home, self.enemy_home) + p.id/1000000.0): if p.distance(self.my_home) < p.distance(self.enemy_home) or p.distance(self.my_home) == p.distance(self.enemy_home): if p.distance(self.my_home) == p.distance(self.enemy_home) and p.ship_count > 10: continue candidates.append(p) candidate_map[i] = p max_attack_distance = max(max_attack_distance, p.distance(self.my_home)) i += 1 weights = [] profits = [] for c in candidates: weight = self.get_attack_ship_count_first_turn(c, self.my_home, self.enemy_home) attack_score = (self.max_distance_between_planets - c.distance(self.my_home) + HORIZON_FIRST) * c.growth_rate - (weight - 1) if attack_score < ATTACK_SCORE_THRESHOLD_FIRST: attack_score = 0 weights.append(weight) profits.append(attack_score) #log.info("candidate %s: score %s, weight %s" % (c, attack_score, weight)) best_planets_to_attack = zeroOneKnapsack(profits,weights,ships_available) #log.info("best planets: %s, ships_avail: %s" % (best_planets_to_attack,ships_available)) sorted_moves = [] for i in range(len(best_planets_to_attack[1])): if (best_planets_to_attack[1][i] != 0): planet_to_attack = candidate_map[i] self.send_fleet(self.my_home, planet_to_attack, planet_to_attack.ship_count+1) def doOffense(self): log.info("Offense phase") if self.current_turn == 1: self.doFirstTurnOffense() return planets_attacked = [] best_planet_to_attack = None while True: best_planet_to_attack = None best_planet_to_attack_score = 0 best_planet_to_attack_distance = 0 best_planet_to_attack_ships_to_send = 0 for planet_to_attack in self.all_planets: if planet_to_attack in planets_attacked: continue min_distance = self.max_distance_between_planets max_distance = 0 for my_planet in self.my_planets: distance = my_planet.distance(planet_to_attack) min_distance = min(min_distance, distance) max_distance = max(max_distance, distance) for fleet in self.universe.find_fleets(owner=PLAYER2, destination=planet_to_attack): max_distance = max(max_distance, fleet.turns_remaining) #log.info("Max distance for %s: %s" % (planet_to_attack, max_distance)) min_distance = max(min_distance, 1) for distance in range(min_distance, max_distance+1): # calculate how many ships we need to get from my planets to planet_to_attack within 'distance' turns planet_to_attack_future = self.planet_timeline[planet_to_attack][distance-1] planet_to_attack_future_owner = planet_to_attack_future[0] if planet_to_attack_future_owner == PLAYER1: break cost_to_conquer = 0 if planet_to_attack_future_owner == PLAYER2 else -1 time_to_profit = 0 if planet_to_attack_future_owner == player.NOBODY: cost_to_conquer = planet_to_attack_future[1] time_to_profit = int(ceil((cost_to_conquer+0.001)/planet_to_attack.growth_rate)) if planet_to_attack.growth_rate > 0 else 1000000 if planet_to_attack_future_owner == NOBODY and self.enemy_com.distance(planet_to_attack) < distance: break #log.info("Time to profit for %s is %s" % (planet_to_attack, time_to_profit)) # if (distance+time_to_profit) >= self.max_distance_between_planets: # break can_hold = True for turn in range(distance, min(distance+time_to_profit+1, self.max_distance_between_planets + 20)): enemy_max_aid = self.max_aid_at_turn[PLAYER2][planet_to_attack][turn] if planet_to_attack_future_owner == player.PLAYER2: enemy_max_aid += self.planet_timeline[planet_to_attack][turn+time_to_profit-1][1] my_max_aid = self.max_aid_at_turn[PLAYER1][planet_to_attack][turn] - cost_to_conquer + planet_to_attack.growth_rate * (turn-distance) - self.cumulative_ships_sent if enemy_max_aid > my_max_aid: can_hold = False #log.info("can't hold %s at turn %s, enemy %s, me %s" % (planet_to_attack, turn, enemy_max_aid, my_max_aid)) break if not can_hold: continue simulation_distance = min(distance+time_to_profit, self.max_distance_between_planets + 20) if simulation_distance <= 0: continue enemy_max_aid = self.max_aid_at_turn[PLAYER2][planet_to_attack][simulation_distance] if planet_to_attack_future_owner == player.PLAYER2: enemy_max_aid += self.planet_timeline[planet_to_attack][simulation_distance-1][1] my_max_aid = self.max_aid_at_turn[PLAYER1][planet_to_attack][simulation_distance] - (cost_to_conquer + 1) - self.cumulative_ships_sent if planet_to_attack_future_owner == NOBODY else 0 ships_to_send = cost_to_conquer + max(enemy_max_aid - my_max_aid, 0) + 1 #log.info("aids for %s at distance %s: enemy %s , me %s, cost %s" % (planet_to_attack, distance, enemy_max_aid, my_max_aid, cost_to_conquer)) # calculate if we can get enough ships from my planets to planet_to_attack within 'distance' turns ships_avail_to_attack = self.get_available_ships_within_distance(planet_to_attack, PLAYER1, distance) #log.info("avail to attack: %s, need to send %s" % (ships_avail_to_attack, ships_to_send)) if ships_avail_to_attack >= ships_to_send: if self.planet_timeline[planet_to_attack][distance-1][0] in player.ENEMIES and self.planet_timeline[planet_to_attack][distance-2][0] == player.NOBODY: continue attack_score = self.get_attack_score(planet_to_attack, planet_to_attack_future_owner, distance) log.info("Attack score of %s at dist %s is: %s - %s ships, cost %s" % (planet_to_attack, distance, attack_score, ships_to_send, cost_to_conquer)) if planet_to_attack_future_owner in player.ENEMIES or (attack_score-cost_to_conquer) >= ATTACK_SCORE_THRESHOLD: if attack_score > best_planet_to_attack_score: best_planet_to_attack_score = attack_score best_planet_to_attack = planet_to_attack best_planet_to_attack_distance = distance best_planet_to_attack_ships_to_send = ships_to_send break if best_planet_to_attack is None: return log.info("Best planet to attack: %s at dist %s with score %s" % (best_planet_to_attack, best_planet_to_attack_distance, best_planet_to_attack_score)) ships_left_to_send = best_planet_to_attack_ships_to_send source_planets = list(self.my_planets) + self.get_neutrals_under_player_attack(PLAYER1) for source_planet in sorted(source_planets, key=lambda p : p.distance(best_planet_to_attack) + p.id/1000000.0): distance = source_planet.distance(best_planet_to_attack) if distance > best_planet_to_attack_distance: continue ships_avail = self.ships_available_at_turn[source_planet][best_planet_to_attack_distance-distance] #log.info("ships avail to attack from %s at dist %s: %s" % (source_planet, best_planet_to_attack_distance-distance, ships_avail)) if self.ships_needed[source_planet] > 0: ships_avail = 0 if source_planet.id != best_planet_to_attack.id and ships_avail > 0: ships_to_send = min(ships_left_to_send, ships_avail) #log.info("ships to send from %s: %s" % (source_planet, ships_to_send)) if distance == best_planet_to_attack_distance and source_planet.owner == PLAYER1: self.send_fleet(source_planet, best_planet_to_attack, ships_to_send) #self.cumulative_ships_sent += ships_to_send if distance < best_planet_to_attack_distance: future_turn = self.current_turn + (best_planet_to_attack_distance - distance) future_move = Move(source_planet, best_planet_to_attack, future_turn, ships_to_send) log.info("Scheduled move: %s" % future_move) if not self.scheduled_moves_at_turn.has_key(future_turn): self.scheduled_moves_at_turn[future_turn] = [] self.scheduled_moves_at_turn[future_turn].append(future_move) ships_left_to_send -= ships_to_send self.decrease_ships_available(source_planet, 0, ships_to_send) if ships_left_to_send == 0: break planets_attacked.append(best_planet_to_attack) def doPostOffense2(self): log.info("Post-Offense phase") if len(self.enemy_planets) == 0: return planets_to_send_to = copy(self.my_planets) neutral_candidate = self.closest_to_enemy_neutral_under_my_attack() if neutral_candidate is not None: planets_to_send_to = planets_to_send_to | neutral_candidate for source_planet in self.my_planets: closest_enemy_planet = self.closest_enemy_planet(source_planet) #log.info("Eval Post-Offense for %s: closest enemy is %s" % (source_planet, closest_enemy_planet)) min_distance_to_enemy = 1000000 dest_planet = None for planet_to_send_to in sorted(planets_to_send_to, key=lambda p : p.id if p.id != source_planet.id else 1000000): if source_planet.distance(planet_to_send_to) < source_planet.distance(closest_enemy_planet) \ and planet_to_send_to.distance(closest_enemy_planet) < min_distance_to_enemy: min_distance_to_enemy = planet_to_send_to.distance(closest_enemy_planet) dest_planet = planet_to_send_to if dest_planet is not None and source_planet.id != dest_planet.id and self.ships_available_at_turn[source_planet][0] > 0: ships_to_send = min(self.ships_available_at_turn[source_planet][0], source_planet.ship_count) self.send_fleet(source_planet, dest_planet, ships_to_send) self.decrease_ships_available(source_planet, 0, ships_to_send) def doPostOffense(self): log.info("Post-Offense phase") if len(self.enemy_planets) == 0: return planets_to_send_to = copy(self.my_planets) neutral_candidate = self.closest_to_enemy_neutral_under_my_attack() if neutral_candidate is not None: planets_to_send_to = planets_to_send_to | neutral_candidate # cache closest and com enemy planet distances closest_enemy_planet_distance_map = {} com_enemy_planet_distance_map = {} for planet in planets_to_send_to: closest_enemy_planet_distance_map[planet] = self.closest_enemy_planet_distance(planet) com_enemy_planet_distance_map[planet] = self.enemy_com.distance(planet) my_nearest_to_enemy_planets = sorted(planets_to_send_to, key=lambda p : p.distance(self.enemy_com) + p.id/1000000.0) for source_planet in self.my_planets: if self.ships_needed[source_planet] == 0 and self.ships_available_at_turn[source_planet][0] > 0: #log.info("Post-Offense for %s" % source_planet) for dest_planet in my_nearest_to_enemy_planets: distance = source_planet.distance(dest_planet) if distance > 0 and distance < com_enemy_planet_distance_map[source_planet]: if com_enemy_planet_distance_map[dest_planet] < com_enemy_planet_distance_map[source_planet] and \ closest_enemy_planet_distance_map[dest_planet] <= closest_enemy_planet_distance_map[source_planet]: self.send_fleet(source_planet, dest_planet, self.ships_available_at_turn[source_planet][0]) self.decrease_ships_available(source_planet, 0, self.ships_available_at_turn[source_planet][0]) break def do_turn(self): self.all_planets = self.universe.all_planets self.my_planets = self.universe.my_planets self.enemy_planets = self.universe.enemy_planets self.nobodies_planets = self.universe.nobodies_planets self.not_my_planets = self.universe.not_my_planets self.current_turn = self.universe.game.turn_count if len(self.my_planets) == 0: return self.cumulative_ships_sent = 0 self.doPrep() self.doScheduled() self.doDefense() self.doOffense() self.doPostOffense2() Game(MyBot, universe_class=Universe3, planet_class=Planet2)
[ "apinkin@apinkin-lappy.(none)" ]
apinkin@apinkin-lappy.(none)
650c945ba471ee76b6ba900217fb8ee31eea82ae
ca98a533d53da95249df7aed710a2db9424cdcdc
/pyspark_proxy/sql/udf.py
f95753ccc8abdd05912b9ce41ea60fcbe0d2eeec
[ "Apache-2.0" ]
permissive
abronte/PysparkProxy
9b3b0cc5dd9c826ca01b1562c639e9ebfc163c84
cc28bacb0d4ee6fb87ced763a73e9ea791612414
refs/heads/master
2021-08-06T12:46:55.562589
2018-12-12T21:57:01
2018-12-12T21:57:01
147,398,292
4
0
NOASSERTION
2018-12-12T21:57:03
2018-09-04T19:07:00
Python
UTF-8
Python
false
false
1,101
py
import base64 import imp import sys from pyspark_proxy.proxy import Proxy from pyspark_proxy.sql.types import DataType from pyspark_proxy.sql.column import Column class UDFRegistration(Proxy): def __init__(self, context_id): self._context_id = context_id def register(self, name, f, returnType=None): if returnType != None: returnType = {'_PROXY_ID': returnType._id} self._call( self._context_id, 'udf.register', [(name, f), {'returnType': returnType}]) class UserDefinedFunction(Proxy): def __init__(self, f, returnType=None): if isinstance(returnType, DataType): returnType = {'_PROXY_ID': returnType._id} result = self._call( 'pyspark', 'sql.functions.udf', [(f, returnType), {}]) self._id = result['id'] def __call__(self, *args, **kwargs): result = self._call( self._id, None, [args, kwargs]) result._name = args[0] return result
[ "adam.bronte@coupa.com" ]
adam.bronte@coupa.com
c439a364452d8c084ee3f038cccde1007cb39e21
560d0c8a59b7933d91e4e37ed3302abfd195395c
/test/player_p2.py
b8fb82f201cb18d9905a079ae85572c4034d03bd
[ "Apache-2.0" ]
permissive
kakao/pycon2016apac-gawibawibo
b56043ca0cdfcfe0375a1d86a85a4534f762420d
b74862562c51f0d027f20030346a8a3e8cb081b6
refs/heads/master
2020-09-26T22:01:45.387401
2016-08-22T09:01:09
2016-08-22T09:01:09
66,256,293
9
3
null
null
null
null
UTF-8
Python
false
false
49
py
def show_me_the_hand(records): return 'gawi'
[ "iolothebard@gmail.com" ]
iolothebard@gmail.com
3cb54764d75b5f50045410295526124deb9a2ad6
eda865165f43083ea3281c0b55e8118147caa604
/mysite/views.py
654659a8306ec411683d4eaf29189841974542b2
[]
no_license
MrPathak21/TextMagic
d86aa75c23881b2d28b1b3fb1efa9b708cf413c5
92b9a6950d7c36eedb48c6751f7e4dff75ea9f63
refs/heads/main
2023-05-08T19:02:02.339862
2021-05-28T21:01:11
2021-05-28T21:01:11
371,816,881
0
0
null
null
null
null
UTF-8
Python
false
false
2,412
py
# I have created this file - Shikhar from django.http import HttpResponse from django.shortcuts import render ''' def index(request): return HttpResponse('<a href = "http://127.0.0.1:8000/navigation"> Click here to go to navigation page</a>') def about(request): return HttpResponse('This is about page') def navigation(request): return HttpResponse('<h1>Navigation</h1><a href = "http://www.google.com"> Google</a> <a href = "http://www.facebook.com"> Facebook</a> <a href = "http://www.twitter.com">Twitter</a>') ''' def index(request): #para = {'name':'Shikhar','place':'Etawah'} return render(request, 'index.html') def analyze(request): mytext = request.POST.get('text', 'default') punc = request.POST.get('removepunc', 'off') upper = request.POST.get('uppercase', 'off') lineremove = request.POST.get('newlineremover', 'off') spaceremover = request.POST.get('spaceremover', 'off') charcount = request.POST.get('charcount', 'off') text = mytext c = "Option Didn't selected " cn = "Option Didn't selected " if punc == "on": mytext = removepunc(mytext) if upper=="on": mytext = uppercase(mytext) if lineremove == 'on': mytext = lineremover(mytext) if spaceremover== 'on': mytext = spacerem(mytext) if charcount == 'on': c = charcounter(text) cn = charcounter(mytext) params = {'purpose': 'Removed Punctuations', 'analyzedtext': mytext, 'charcountold': c, 'charcountnew': cn} return render(request, 'analyze.html', params) #return HttpResponse('''<h1> </h1> <a href = 'http://127.0.0.1:8000'>Back</a>''') def about(request): return render(request, 'AboutUs.html') def contact(request): return render(request, 'ContactUs.html') def removepunc(s): result = '' p = '''`~!@#$%^&*()[{}]|:;"'<,>.?/''' for i in s: if i not in p: result += i return result def uppercase(s): return s.upper() def lineremover(s): result = '' for i in s: if i != '\n' and i != '\r': result += i return result def spacerem(s): result = '' if s[0] != '': result += s[0] for i in range(1, len(s)): if s[i] == ' ' and s[i-1] == ' ': continue result += s[i] return result def charcounter(s): count = 0 for i in s: count += 1 return count
[ "66026814+MrPathak21@users.noreply.github.com" ]
66026814+MrPathak21@users.noreply.github.com
2961d204760a2520fcea0110dbca3da1a4547114
19e263ce69087e386b5bb92a534f69dc39f46877
/main.py
d961541aca3d03c63940ce20ef6e66755e29c7f0
[]
no_license
Jyuukun/PS5-Alerts
055b855bde75540b8534d7c713c0edf04dbcb807
f13deb7353f11ea486c204e52e7e4a38a1544161
refs/heads/main
2023-04-04T16:51:57.826709
2020-11-26T15:44:20
2021-04-12T11:36:30
316,273,706
3
0
null
null
null
null
UTF-8
Python
false
false
3,507
py
# -*- coding: utf-8 -*- import logging import os import sys import signal import smtplib import tempfile import time from configparser import ConfigParser from weboob.tools.log import createColoredFormatter, getLogger from amazon import AmazonBrowser from auchan import AuchanBrowser from boulanger import BoulangerBrowser from cdiscount import CdiscountBrowser from carrefour import CarrefourBrowser from cultura import CulturaBrowser from darty import DartyBrowser from fnac import FnacBrowser from leclerc import LeclercBrowser from micromania import MicromaniaBrowser def get_config(): config = ConfigParser() config.read(os.path.dirname(os.path.abspath(sys.argv[0])) + '/config') return config def send_mail(subject, text): config = get_config() try: with smtplib.SMTP('smtp.gmail.com', 587) as server: server.starttls() server.login( self.config['mail']['login'], self.config['mail']['password'] ) message = 'Subject: {}\n\n{}'.format(subject, text) server.sendmail( self.config['mail']['sender'], self.config['mail']['receiver'], message.encode('utf-8') ) except Exception as e: logger = getLogger('send-mail') logger.error("Something went wrong while sending mail : %s" % str(e)) def create_colored_handler(): # stderr logger format = '%(asctime)s:%(levelname)s:%(lineno)d:%(funcName)s %(message)s' handler = logging.StreamHandler(sys.stderr) handler.setFormatter(createColoredFormatter(sys.stderr, format)) return handler def signal_handler(signal, frame): sys.exit(0) def main(): signal.signal(signal.SIGINT, signal_handler) # create colored logger logging.root.setLevel(logging.DEBUG) logging.root.addHandler(create_colored_handler()) logger = getLogger('ps5-availability') # create temporary directory responses_dirname = tempfile.mkdtemp(prefix='ps5_availability_session_') logger.info('Debug data will be saved in this directory: %s' % responses_dirname) browsers = ( AmazonBrowser, AuchanBrowser, BoulangerBrowser, CarrefourBrowser, CdiscountBrowser, CulturaBrowser, DartyBrowser, FnacBrowser, LeclercBrowser, MicromaniaBrowser, ) while True: waiting_time = 60 * 30 for browser in browsers: browser = browser(logger=logger, responses_dirname=responses_dirname) logger.warning("Now trying on %s" % browser.BASEURL) try: is_available = browser.is_available except Exception as e: logger.error("Something went wrong : %s" % str(e)) continue if is_available: logger.warning("Playstation 5 is AVAILABLE !!") send_mail( "Playstation 5 Available !", "Vite mec ! La PS5 est disponible sur %s ! Va dépenser toute ta tune ! :')" % browser.BASEURL ) # we found one, no need to check to much now waiting_time = 60 * 60 break else: logger.warning("Playstation 5 is not available on %s, so sad. :(" % browser.BASEURL) else: logger.critical("No PS5 found on any providers") logger.critical("waiting %s seconds to check again" % waiting_time) time.sleep(waiting_time) if __name__ == '__main__': main()
[ "mail@elambert.me" ]
mail@elambert.me
4eb1da69c4b7f74b682db604e23967cb55b2f970
e62606922d4820e90f5f8e59de1a3812cb05f14f
/Centralized BlockChain/secondClient.py
578a7f8f85479375e6e0685023b8e976e51310f0
[]
no_license
ozgunozerk/DistributedApplications
8d0d647020a2856ede5938507d8a5b571d1df0a6
f9d21174c62e037242a53ca5a6ba66e0d82d8193
refs/heads/main
2023-03-04T22:13:15.943790
2021-02-12T14:43:54
2021-02-12T14:43:54
338,344,208
0
1
null
null
null
null
UTF-8
Python
false
false
434
py
import Pyro4 BTC = Pyro4.Proxy("PYRONAME:BTC") acc1 = BTC.createAccount(100) print(acc1) bal = BTC.calculateBalance(acc1) if bal > 20: BTC.transfer(acc1, 1, -60) BTC.printChain() ETH = Pyro4.Proxy("PYRONAME:ETH") e1 = ETH.createAccount(30) print(e1) bal = ETH.calculateBalance(e1) if bal > 20: ETH.transfer(e1, 1, -20) ETH.printChain() BTC.exchange(acc1, e1, ETH, 50) BTC.printChain() ETH.printChain()
[ "noreply@github.com" ]
ozgunozerk.noreply@github.com
81e8a9bdb7e89305a5eb4d8e265ea9bbcb1ec8e1
d96764c85f86c580029232f226f8cc6bc5fde1e8
/EyeQ1_33/utils/trainer.py
c2223c6bf323127c7378fecb83c0e624bddf8b1d
[]
no_license
QuantumZhang/FIQA_training_module
b93f775487b11ad172f4d1df57d59a8ad189d035
789c11e23d0a120d2f04006b3187f47d4300415d
refs/heads/main
2023-02-09T14:01:51.707809
2021-01-04T06:31:03
2021-01-04T06:31:03
326,200,579
0
0
null
null
null
null
UTF-8
Python
false
false
9,724
py
import time import torch import numpy as np import pandas as pd from sklearn.metrics import cohen_kappa_score, confusion_matrix, accuracy_score, recall_score from tqdm import tqdm import torch.nn as nn import csv Label_Merge = ['fundus_image_qualified', 'fi_unqualified_disc-position', 'fi_unqualified_macular-position', 'fi_unqualified_focus-clearness', 'fi_unqualified_readable-range', 'fi_unqualified_others'] def show_message(y_tru, y_p, model='train'): message = [] y_p[y_p > 0.5] = 1 y_p[y_p <= 0.5] = 0 acc = 0 kappa = 0 kappa_weight = [1.0, 1.0, 0.8, 1.2, 0.8, 1.2] for i in range(y_p.shape[1]): kappa_t = cohen_kappa_score(y_tru[:, i], y_p[:, i]) kappa += kappa_weight[i] * kappa_t acc_t = accuracy_score(y_tru[:, i], y_p[:, i]) acc += acc_t confu_t = confusion_matrix(y_tru[:, i], y_p[:, i]) recall_t = recall_score(y_tru[:, i], y_p[:, i]) recall0_t = recall_score(y_tru[:, i], y_p[:, i], pos_label=0) message.append(kappa_t) message.append(acc_t) message.append(recall_t) message.append(recall0_t) print('label==>%s<==>%s<==的信息:'%(Label_Merge[i], model)) print('kappa:', kappa_t, '-----acc:', acc_t, '-----recall:', recall_t, '---recall0', recall0_t) print('confusion_matrix:\n', confu_t) return acc / y_p.shape[1], kappa / y_p.shape[1], message # 定义训练类 def train_step(train_loader, model, epoch, optimizer, criterion, epochs, log_csv, cycle_scheduler= None): # switch to train mode model.train() epoch_loss = 0.0 iters_per_epoch = len(train_loader) y_tru = None y_p = None for step, (imagesA, imagesB, imagesC, labels) in enumerate(train_loader): imagesA = imagesA.cuda() imagesB = imagesB.cuda() imagesC = imagesC.cuda() if y_tru is None: y_tru = np.array(labels) else: y_tru = np.vstack((y_tru, np.array(labels))) labels = labels.float().cuda() # labels = labels.cuda().long() # labels = torch.tensor(labels).reshape(4,-1) # labels = labels.reshape(labels.shape[0],1) # labels = torch.zeros(labels.shape[0], 2).scatter_(1, labels, 1).cuda() combine = model(imagesA) combine = torch.sigmoid(combine) # out_A, out_B, out_C, out_F, combine = model(imagesA, imagesB, imagesC) # loss_x = criterion(out_A, labels) # loss_y = criterion(out_B, labels) # loss_z = criterion(out_C, labels) # loss_c = criterion(out_F, labels) loss_f = criterion(combine, labels) lossValue = loss_f # lossValue = loss_w[0]*loss_x+loss_w[1]*loss_y+loss_w[2]*loss_z+loss_w[3]*loss_c+loss_w[4]*loss_f # writer.add_scalar('/epoch_loss', lossValue, step) # pre = torch.cat((pre, combine), 0) # tru = torch.cat((tru, labels.float()), 0) y_pre = combine.detach().cpu().numpy() if y_p is None: y_p = np.array(y_pre) else: y_p = np.vstack((y_p, np.array(y_pre))) optimizer.zero_grad() lossValue.backward() optimizer.step() if cycle_scheduler is not None: cycle_scheduler.batch_step() epoch_loss += lossValue.item() acc, kappa, message = show_message(y_tru, y_p, model='train') with open(log_csv, 'a+') as f: if epoch == 0: csv_write = csv.writer(f) data_row = ['epoch', 'qua_kappa', 'qua_acc', 'qua_recall', 'qua_recall0', 'disc_kappa', 'disc_acc', 'disc_recall', 'disc_recall0', 'macular_kappa', 'macular_acc', 'macular_recall', 'macular_recall0', 'clear_kappa', 'clear_acc', 'clear_recall', 'clear_recall0', 'read_kappa', 'read_acc', 'read_recall', 'read_recall0', 'others_kappa', 'others_acc', 'others_recall', 'others_recall0', ] csv_write.writerow(data_row) csv_write = csv.writer(f) data_row = [epoch] + message csv_write.writerow(data_row) epoch_loss = epoch_loss / iters_per_epoch return epoch_loss, acc, kappa def validation_step(train_loader, model, epoch, optimizer, criterion, epochs, log_csv, cycle_scheduler=None): # switch to train mode model.eval() epoch_loss = 0.0 iters_per_epoch = len(train_loader) y_tru = None y_p = None for step, (imagesA, imagesB, imagesC, labels) in enumerate(train_loader): imagesA = imagesA.cuda() imagesB = imagesB.cuda() imagesC = imagesC.cuda() if y_tru is None: y_tru = np.array(labels) else: y_tru = np.vstack((y_tru, np.array(labels))) labels = labels.float().cuda() # labels = labels.cuda().long() # labels = torch.tensor(labels).reshape(4,-1) # labels = labels.reshape(labels.shape[0],1) # labels = torch.zeros(labels.shape[0], 2).scatter_(1, labels, 1).cuda() with torch.no_grad(): combine = model(imagesA) combine = torch.sigmoid(combine) # out_A, out_B, out_C, out_F, combine = model(imagesA, imagesB, imagesC) # loss_x = criterion(out_A, labels) # loss_y = criterion(out_B, labels) # loss_z = criterion(out_C, labels) # loss_c = criterion(out_F, labels) loss_f = criterion(combine, labels) lossValue = loss_f # lossValue = loss_w[0]*loss_x+loss_w[1]*loss_y+loss_w[2]*loss_z+loss_w[3]*loss_c+loss_w[4]*loss_f # writer.add_scalar('/epoch_loss', lossValue, step) # pre = torch.cat((pre, combine), 0) # tru = torch.cat((tru, labels.float()), 0) y_pre = combine.detach().cpu().numpy() if y_p is None: y_p = np.array(y_pre) else: y_p = np.vstack((y_p, np.array(y_pre))) # optimizer.zero_grad() # lossValue.backward() # optimizer.step() # cycle_scheduler.batch_step() epoch_loss += lossValue.item() acc, kappa, message = show_message(y_tru, y_p, model='val') with open(log_csv, 'a+') as f: if epoch == 0: csv_write = csv.writer(f) data_row = ['epoch', 'qua_kappa', 'qua_acc', 'qua_recall', 'qua_recall0', 'disc_kappa', 'disc_acc', 'disc_recall', 'disc_recall0', 'macular_kappa', 'macular_acc', 'macular_recall', 'macular_recall0', 'clear_kappa', 'clear_acc', 'clear_recall', 'clear_recall0', 'read_kappa', 'read_acc', 'read_recall', 'read_recall0', 'others_kappa', 'others_acc', 'others_recall', 'others_recall0', ] csv_write.writerow(data_row) csv_write = csv.writer(f) data_row = [epoch] + message csv_write.writerow(data_row) epoch_loss = epoch_loss / iters_per_epoch return epoch_loss, acc, kappa # # def validation_step(val_loader, model, criterion): # # # switch to train mode # model.eval() # epoch_loss = 0 # iters_per_epoch = len(val_loader) # y_tru = None # y_p = None # for step, (imagesA, imagesB, imagesC, labels) in enumerate(val_loader): # imagesA = imagesA.cuda() # imagesB = imagesB.cuda() # imagesC = imagesC.cuda() # # if y_tru is None: # y_tru = np.array(labels) # else: # y_tru = np.vstack((y_tru, np.array(labels))) # # labels = labels.float().cuda() # # _, _, _, _, outputs = model(imagesA, imagesB, imagesC) # combine = model(imagesA) # combine = torch.sigmoid(combine) # with torch.no_grad(): # loss = criterion(combine, labels) # epoch_loss += loss.item() # # # y_pre = combine.detach().cpu().numpy() # if y_p is None: # y_p = np.array(y_pre) # else: # y_p = np.vstack((y_p, np.array(y_pre))) # # y_tru = y_tru.reshape((-1)) # y_p = y_p.reshape((-1)) # acc = show_message(y_tru, y_p) # epoch_loss = epoch_loss / iters_per_epoch # return epoch_loss, acc def save_output(label_test_file, dataPRED, label_idx, save_file): label_list = label_idx n_class = len(label_list) datanpPRED = np.squeeze(dataPRED.cpu().numpy()) df_tmp = pd.read_csv(label_test_file) image_names = df_tmp["image"].tolist() result = {label_list[i]: datanpPRED[:, i] for i in range(n_class)} result['image_name'] = image_names out_df = pd.DataFrame(result) name_older = ['image_name'] for i in range(n_class): name_older.append(label_list[i]) out_df.to_csv(save_file, columns=name_older) def acc_mol(val_loader, model): model.eval() iters_per_epoch = len(val_loader) pre_all = [] y_all = [] for step, (imagesA, imagesB, imagesC, labels) in enumerate(val_loader): imagesA = imagesA.cuda() imagesB = imagesB.cuda() imagesC = imagesC.cuda() labels = labels.cuda() _, _, _, _, outputs = model(imagesA, imagesB, imagesC) pre = outputs.argmax(dim=1) pre_all.append(pre) y_all.append(labels) return pre_all, y_all def save_file(model,loss, kappa, acc, recall,recall0,model_save_file, epoch): torch.save({'state_dict': model.state_dict(), 'loss':loss,'kappa':kappa, 'acc': acc, 'recall':recall,'recall0':recall0,'epoch': epoch + 1}, model_save_file) print('已保存模型至:',model_save_file)
[ "noreply@github.com" ]
QuantumZhang.noreply@github.com
f725e3ee7f1631ff7c08050264fac3d3468ddd14
859951d9fc8ec1377c9fdadbe5e21128879d89b0
/unit21/a214_simple_window1_AC.py
288aa54aa9b7e8dd1c0ab955b32a339df182ac98
[]
no_license
Axis-Denied/csp_files
d61ff240306de8114e40e030b92582b962850b73
579a57e56bc616f42eac529de106f603aebc427a
refs/heads/master
2020-12-06T08:01:10.765190
2020-01-09T19:13:03
2020-01-09T19:13:03
232,401,232
0
0
null
null
null
null
UTF-8
Python
false
false
343
py
############################################################################## # a113_TR_simple_window1.py # Example solution: Change its size to 200 by 100 pixels. ############################################################################## import tkinter as tk # main window root = tk.Tk() root.wm_geometry("200x100") root.mainloop()
[ "cokera01@student.psdr3.org" ]
cokera01@student.psdr3.org
b2751e339d9962f161e087abd2904dce3feee3f8
029fe3c82ace900aeeb6b5624aea9f02cccee571
/dtlearn/nbayes/gaussian.py
e258251abd78c8921142c223696dca2fb7342347
[ "MIT" ]
permissive
i-agrawal/dtlearn
5b86bad7d67d84ee5f0ccfb130e142ce6a099748
5e925aece613a0402cecf9ee869dfd41e0e18bce
refs/heads/master
2020-03-19T23:55:28.603406
2018-12-27T00:58:37
2018-12-27T00:58:37
137,024,986
2
0
null
null
null
null
UTF-8
Python
false
false
668
py
import numpy as np from .. import Model class Gaussian(Model): def train(self, x, y): self.labels = np.unique(y) groups = [x[y == label] for label in self.labels] self.means = np.array([np.mean(group, axis=0) for group in groups]) self.varis = np.array([np.var(group, axis=0) for group in groups]) self.fracs = np.array([len(group) for group in groups]) / len(y) def predict(self, x): diff = x[:, np.newaxis] - self.means[np.newaxis] prob = np.exp(-np.multiply(diff, diff) / (2 * self.varis)) chosen = np.argmax(np.prod(prob, axis=2) * self.fracs, axis=1) return self.labels[chosen]
[ "ishan.agrawal97@gmail.com" ]
ishan.agrawal97@gmail.com
03d238f1b9ccba256f82bd150d054bb5982e40d6
c4362b27e969f5ffb1b39c85fc212198de91b646
/extractLegalOpinionsCstmMetadata.py
048f239378095a6344d1b548accb795462af0859
[ "MIT" ]
permissive
OpenLawsGR/judgments2AKN
af8be80f931571b1c67fa957dff6c0ae76c67f9f
0c6217349cde36058d5599800e289fdf0d3eaf23
refs/heads/master
2020-09-05T00:34:53.860129
2020-01-08T15:35:49
2020-01-08T15:35:49
219,934,916
5
0
null
null
null
null
UTF-8
Python
false
false
6,896
py
# -*- coding: utf-8 -*- import os import sys import fnmatch import argparse from functions import extract_data_from_nsk from variables import NSK_METADATA, NSK_CSTM_METADATA, NSK, TXT_EXT from variables import LEGAL_TEXTS from lxml import etree program_description = 'A module for downloading specific data (e.g. keywords) ' program_description += 'from Legal Council of State official website. ' program_description += 'Extracted data may be used later to build some ' program_description += 'appropriate Akoma Ntoso metadata nodes.' parser = argparse.ArgumentParser( description = program_description ) parser.add_argument( '-fn', metavar = 'FILENAME', help = 'choose a specific legal opinion to extract data' ) # create a namespace object args = parser.parse_args() if __name__ == '__main__': source_path = os.path.join( os.getcwd(), os.path.join( LEGAL_TEXTS, NSK ) ) if args.fn is not None: file_pattern = '*' + args.fn else: file_pattern = '*' + TXT_EXT #print(source_path) # Create custom metadata folder if it does not exist if not os.path.exists(source_path.replace(NSK, NSK_CSTM_METADATA)): os.makedirs(source_path.replace(NSK, NSK_CSTM_METADATA)) for root, dirs, files in os.walk(source_path): for name in files: #print name if fnmatch.fnmatch(name, file_pattern): print name # check metadata folder if meta_file exists # open and get post parameters meta_file_exists = os.path.isfile( os.path.join( source_path.replace(NSK, NSK_METADATA), name ) ) if meta_file_exists: with open( os.path.join( source_path.replace(NSK, NSK_METADATA), name ), 'r') as fin: XML = etree.parse(fin) #print XML.getroot().nsmap XML_root = XML.getroot() #print list(root.nsmap.values())[0] try: ns = list(XML_root.nsmap.values())[0] protocolNumber = XML.findtext( '//ns:protocolNumber', namespaces = {'ns' : ns} ) issueDate = XML.findtext( '//ns:issueDate', namespaces = {'ns' : ns} ) except IndexError: protocolNumber = XML.findtext( '//protocolNumber' ) issueDate = XML.findtext( '//issueDate' ) try: issueYear = protocolNumber.split('/')[1] except IndexError: issueYear = issueDate.split('-')[0] decisionNumber = protocolNumber.split('/')[0] #print issueYear #print decisionNumber # Create POST url (based in NSK search form) and POST data post_url ='http://www.nsk.gr/web/nsk/' post_url +='anazitisi-gnomodoteseon' post_url +='?p_p_id=nskconsulatories_WAR_nskplatformportlet' post_url +='&p_p_lifecycle=0&p_p_state=normal&p_p_mode=view' post_url +='&p_p_col_id=column-4&p_p_col_pos=2' post_url +='&p_p_col_count=3' #print post_url post_data = { "_nskconsulatories_WAR_nskplatformportlet_isSearch" : "1", "_nskconsulatories_WAR_nskplatformportlet_inputSuggestionNo" : decisionNumber, "_nskconsulatories_WAR_nskplatformportlet_inputDatefrom" : issueYear, "_nskconsulatories_WAR_nskplatformportlet_consulState":"null" } extracted_data = extract_data_from_nsk(post_url, post_data) #print extracted_data if extracted_data: # Create a custom element that will hold extracted data custom_metadata = etree.Element("customMetadata") keywords = etree.SubElement(custom_metadata, 'keywords') cnt = 0 for keyword in extracted_data['keywords']: # If its is not empty string if keyword: cnt += 1 keyword_elem = etree.SubElement( keywords, 'keyword_' + str(cnt) ) keyword_elem.text = keyword.strip() chairman = etree.SubElement(custom_metadata, 'chairman') chairman.text = extracted_data['chairman'] rapporteur = etree.SubElement(custom_metadata, 'rapporteur') rapporteur.text = extracted_data['rapporteur'] status = etree.SubElement(custom_metadata, 'status') status.text = extracted_data['status'] #print etree.tostring( # custom_metadata, # pretty_print=True, # encoding="UTF-8" # ) XmlTree = etree.ElementTree(custom_metadata) # Write ElementTree to file with open( os.path.join( source_path.replace(NSK, NSK_CSTM_METADATA), name), 'w') as fin: fin.write( etree.tostring( XmlTree, pretty_print = True, encoding = "UTF-8", xml_declaration = True ) )
[ "plessas@ceid.upatras.gr" ]
plessas@ceid.upatras.gr
61181ee6c9b296d449fba0d588a715d3307c4d76
5fca0fc6fcada38c91227ee8b10bfe5ae39d374d
/manage.py
c2404b6994bc86aba007914f31ff62ad12e5e8b2
[]
no_license
Lex3528/Polling
4d1712993afa8a2a0516f524862455900135bf3a
bdfa8da16367de518162ca87d8d9aa14951b2368
refs/heads/master
2022-12-16T04:51:46.528666
2020-09-24T18:13:39
2020-09-24T18:13:39
298,347,378
0
0
null
null
null
null
UTF-8
Python
false
false
629
py
#!/usr/bin/env python """Django's command-line utility for administrative tasks.""" import os import sys def main(): os.environ.setdefault('DJANGO_SETTINGS_MODULE', 'apps_conf.settings') try: from django.core.management import execute_from_command_line except ImportError as exc: raise ImportError( "Couldn't import Django. Are you sure it's installed and " "available on your PYTHONPATH environment variable? Did you " "forget to activate a virtual environment?" ) from exc execute_from_command_line(sys.argv) if __name__ == '__main__': main()
[ "mikhaylov611@mail.ru" ]
mikhaylov611@mail.ru
c72d9299bc10665a4db3242dbdca70d84cf13520
68ea05d0d276441cb2d1e39c620d5991e0211b94
/2714.py
c816933a2eed56ec8282d45061a5d42bbd7766f2
[]
no_license
mcavalca/uri-python
286bc43aa157d3a6880dc222e0136c80cf079565
e22875d2609fe7e215f9f3ed3ca73a1bc2cf67be
refs/heads/master
2021-11-23T08:35:17.614443
2021-10-05T13:26:03
2021-10-05T13:26:03
131,339,175
50
27
null
2021-11-22T12:21:59
2018-04-27T19:54:09
Python
UTF-8
Python
false
false
221
py
n = int(input()) while n > 0: n -= 1 ra = input() saida = 'INVALID DATA' if len(ra) == 20: if ra[0:2] == 'RA': if ra[2:].isdigit(): saida = int(ra[2:]) print(saida)
[ "m.cavalca@hotmail.com" ]
m.cavalca@hotmail.com
1116093189d5bbdecb2913592b931a0012e07283
4501c310488b7d019098858e8577b020b6346d79
/Code/PythagoreGUI.spec
0e1160a49ef72111528452fa78fd7e07d45c8cdb
[]
no_license
nathanbegin/Pythaface
13d4b030fd9809dd59715f1b0a6cb79db499e012
0660421f62b67653a833f31d0fcd8864de7508ae
refs/heads/master
2020-04-06T04:27:56.778429
2017-02-23T01:35:50
2017-02-23T01:35:50
82,733,602
0
0
null
null
null
null
UTF-8
Python
false
false
767
spec
# -*- mode: python -*- block_cipher = None a = Analysis(['PythagoreGUI.py'], pathex=['C:\\Users\\natha\\Desktop\\Pythagore\\Code'], binaries=[], datas=[], hiddenimports=[], hookspath=[], runtime_hooks=[], excludes=[], win_no_prefer_redirects=False, win_private_assemblies=False, cipher=block_cipher) pyz = PYZ(a.pure, a.zipped_data, cipher=block_cipher) exe = EXE(pyz, a.scripts, a.binaries, a.zipfiles, a.datas, name='PythagoreGUI', debug=False, strip=False, upx=True, console=False , icon='Logo.ico')
[ "nathanbegin@gmail.com" ]
nathanbegin@gmail.com
18b774f735198bcefe1ca5a4e8ffe56b83d8dfd2
120622dd09db3aa677e0c0100ea6380b04883aa7
/.envPy3/lib/python3.6/site-packages/pip/_vendor/distlib/wheel.py
9a09509ac907b10ca6b57882f7e41d100c2bf407
[]
no_license
jeffonmac/MacGyver
bf72beacb14a5ddc2aed5787b616d85871b7f3ba
9b9db6b79164bad7874ba134695174792e2d9a28
refs/heads/master
2021-01-18T13:26:01.335442
2017-09-13T13:35:37
2017-09-13T13:35:37
100,375,402
0
0
null
null
null
null
UTF-8
Python
false
false
39,119
py
# -*- coding: utf-8 -*- # # Copyright (C) 2013-2016 Vinay Sajip. # Licensed to the Python Software Foundation under a contributor agreement. # See LICENSE.txt and CONTRIBUTORS.txt. # from __future__ import unicode_literals import base64 import codecs import datetime import distutils.util from email import message_from_file import hashlib import imp import json import logging import os import posixpath import re import shutil import sys import tempfile import zipfile from . import __version__, DistlibException from .compat import sysconfig, ZipFile, fsdecode, text_type, filter from .database import InstalledDistribution from .metadata import Metadata, METADATA_FILENAME from .util import (FileOperator, convert_path, CSVReader, CSVWriter, Cache, cached_property, get_cache_base, read_exports, tempdir) from .version import NormalizedVersion, UnsupportedVersionError logger = logging.getLogger(__name__) cache = None # created when needed if hasattr(sys, 'pypy_version_info'): IMP_PREFIX = 'pp' elif sys.platform.startswith('java'): IMP_PREFIX = 'jy' elif sys.platform == 'cli': IMP_PREFIX = 'ip' else: IMP_PREFIX = 'cp' VER_SUFFIX = sysconfig.get_config_var('py_version_nodot') if not VER_SUFFIX: # pragma: no cover VER_SUFFIX = '%s%s' % sys.version_info[:2] PYVER = 'py' + VER_SUFFIX IMPVER = IMP_PREFIX + VER_SUFFIX ARCH = distutils.util.get_platform().replace('-', '_').replace('.', '_') ABI = sysconfig.get_config_var('SOABI') if ABI and ABI.startswith('cpython-'): ABI = ABI.replace('cpython-', 'cp') else: def _derive_abi(): parts = ['cp', VER_SUFFIX] if sysconfig.get_config_var('Py_DEBUG'): parts.append('d') if sysconfig.get_config_var('WITH_PYMALLOC'): parts.append('m') if sysconfig.get_config_var('Py_UNICODE_SIZE') == 4: parts.append('u') return ''.join(parts) ABI = _derive_abi() del _derive_abi FILENAME_RE = re.compile(r''' (?P<nm>[^-]+) -(?P<vn>\d+[^-]*) (-(?P<bn>\d+[^-]*))? -(?P<py>\w+\d+(\.\w+\d+)*) -(?P<bi>\w+) -(?P<ar>\w+(\.\w+)*) \.whl$ ''', re.IGNORECASE | re.VERBOSE) NAME_VERSION_RE = re.compile(r''' (?P<nm>[^-]+) -(?P<vn>\d+[^-]*) (-(?P<bn>\d+[^-]*))?$ ''', re.IGNORECASE | re.VERBOSE) SHEBANG_RE = re.compile(br'\s*#![^\r\n]*') SHEBANG_DETAIL_RE = re.compile(br'^(\s*#!("[^"]+"|\S+))\s+(.*)$') SHEBANG_PYTHON = b'#!python' SHEBANG_PYTHONW = b'#!pythonw' if os.sep == '/': to_posix = lambda o: o else: to_posix = lambda o: o.replace(os.sep, '/') class Mounter(object): def __init__(self): self.impure_wheels = {} self.libs = {} def add(self, pathname, extensions): self.impure_wheels[pathname] = extensions self.libs.update(extensions) def remove(self, pathname): extensions = self.impure_wheels.pop(pathname) for k, v in extensions: if k in self.libs: del self.libs[k] def find_module(self, fullname, path=None): if fullname in self.libs: result = self else: result = None return result def load_module(self, fullname): if fullname in sys.modules: result = sys.modules[fullname] else: if fullname not in self.libs: raise ImportError('unable to find extension for %s' % fullname) result = imp.load_dynamic(fullname, self.libs[fullname]) result.__loader__ = self parts = fullname.rsplit('.', 1) if len(parts) > 1: result.__package__ = parts[0] return result _hook = Mounter() class Wheel(object): """ Class to build and install from Wheel files (PEP 427). """ wheel_version = (1, 1) hash_kind = 'sha256' def __init__(self, filename=None, sign=False, verify=False): """ Initialise an instance using a (valid) filename. """ self.sign = sign self.should_verify = verify self.buildver = '' self.pyver = [PYVER] self.abi = ['none'] self.arch = ['any'] self.dirname = os.getcwd() if filename is None: self.name = 'dummy' self.version = '0.1' self._filename = self.filename else: m = NAME_VERSION_RE.match(filename) if m: info = m.groupdict('') self.name = info['nm'] # Reinstate the local version separator self.version = info['vn'].replace('_', '-') self.buildver = info['bn'] self._filename = self.filename else: dirname, filename = os.path.split(filename) m = FILENAME_RE.match(filename) if not m: raise DistlibException('Invalid name or ' 'filename: %r' % filename) if dirname: self.dirname = os.path.abspath(dirname) self._filename = filename info = m.groupdict('') self.name = info['nm'] self.version = info['vn'] self.buildver = info['bn'] self.pyver = info['py'].split('.') self.abi = info['bi'].split('.') self.arch = info['ar'].split('.') @property def filename(self): """ Build and return a filename from the various components. """ if self.buildver: buildver = '-' + self.buildver else: buildver = '' pyver = '.'.join(self.pyver) abi = '.'.join(self.abi) arch = '.'.join(self.arch) # replace - with _ as a local version separator version = self.version.replace('-', '_') return '%s-%s%s-%s-%s-%s.whl' % (self.name, version, buildver, pyver, abi, arch) @property def exists(self): path = os.path.join(self.dirname, self.filename) return os.path.isfile(path) @property def tags(self): for pyver in self.pyver: for abi in self.abi: for arch in self.arch: yield pyver, abi, arch @cached_property def metadata(self): pathname = os.path.join(self.dirname, self.filename) name_ver = '%s-%s' % (self.name, self.version) info_dir = '%s.dist-info' % name_ver wrapper = codecs.getreader('utf-8') with ZipFile(pathname, 'r') as zf: wheel_metadata = self.get_wheel_metadata(zf) wv = wheel_metadata['Wheel-Version'].split('.', 1) file_version = tuple([int(i) for i in wv]) if file_version < (1, 1): fn = 'METADATA' else: fn = METADATA_FILENAME try: metadata_filename = posixpath.join(info_dir, fn) with zf.open(metadata_filename) as bf: wf = wrapper(bf) result = Metadata(fileobj=wf) except KeyError: raise ValueError('Invalid wheel, because %s is ' 'missing' % fn) return result def get_wheel_metadata(self, zf): name_ver = '%s-%s' % (self.name, self.version) info_dir = '%s.dist-info' % name_ver metadata_filename = posixpath.join(info_dir, 'WHEEL') with zf.open(metadata_filename) as bf: wf = codecs.getreader('utf-8')(bf) message = message_from_file(wf) return dict(message) @cached_property def info(self): pathname = os.path.join(self.dirname, self.filename) with ZipFile(pathname, 'r') as zf: result = self.get_wheel_metadata(zf) return result def process_shebang(self, data): m = SHEBANG_RE.match(data) if m: end = m.end() shebang, data_after_shebang = data[:end], data[end:] # Preserve any arguments after the interpreter if b'pythonw' in shebang.lower(): shebang_python = SHEBANG_PYTHONW else: shebang_python = SHEBANG_PYTHON m = SHEBANG_DETAIL_RE.match(shebang) if m: args = b' ' + m.groups()[-1] else: args = b'' shebang = shebang_python + args data = shebang + data_after_shebang else: cr = data.find(b'\r') lf = data.find(b'\n') if cr < 0 or cr > lf: term = b'\n' else: if data[cr:cr + 2] == b'\r\n': term = b'\r\n' else: term = b'\r' data = SHEBANG_PYTHON + term + data return data def get_hash(self, data, hash_kind=None): if hash_kind is None: hash_kind = self.hash_kind try: hasher = getattr(hashlib, hash_kind) except AttributeError: raise DistlibException('Unsupported hash algorithm: %r' % hash_kind) result = hasher(data).digest() result = base64.urlsafe_b64encode(result).rstrip(b'=').decode('ascii') return hash_kind, result def write_record(self, records, record_path, base): records = list(records) # make a copy for sorting p = to_posix(os.path.relpath(record_path, base)) records.append((p, '', '')) records.sort() with CSVWriter(record_path) as writer: for row in records: writer.writerow(row) def write_records(self, info, libdir, archive_paths): records = [] distinfo, info_dir = info hasher = getattr(hashlib, self.hash_kind) for ap, p in archive_paths: with open(p, 'rb') as f: data = f.read() digest = '%s=%s' % self.get_hash(data) size = os.path.getsize(p) records.append((ap, digest, size)) p = os.path.join(distinfo, 'RECORD') self.write_record(records, p, libdir) ap = to_posix(os.path.join(info_dir, 'RECORD')) archive_paths.append((ap, p)) def build_zip(self, pathname, archive_paths): with ZipFile(pathname, 'w', zipfile.ZIP_DEFLATED) as zf: for ap, p in archive_paths: logger.debug('Wrote %s to %s in wheel', p, ap) zf.write(p, ap) def build(self, paths, tags=None, wheel_version=None): """ Build a wheel from files in specified paths, and use any specified tags when determining the name of the wheel. """ if tags is None: tags = {} libkey = list(filter(lambda o: o in paths, ('purelib', 'platlib')))[0] if libkey == 'platlib': is_pure = 'false' default_pyver = [IMPVER] default_abi = [ABI] default_arch = [ARCH] else: is_pure = 'true' default_pyver = [PYVER] default_abi = ['none'] default_arch = ['any'] self.pyver = tags.get('pyver', default_pyver) self.abi = tags.get('abi', default_abi) self.arch = tags.get('arch', default_arch) libdir = paths[libkey] name_ver = '%s-%s' % (self.name, self.version) data_dir = '%s.data' % name_ver info_dir = '%s.dist-info' % name_ver archive_paths = [] # First, stuff which is not in site-packages for key in ('data', 'headers', 'scripts'): if key not in paths: continue path = paths[key] if os.path.isdir(path): for root, dirs, files in os.walk(path): for fn in files: p = fsdecode(os.path.join(root, fn)) rp = os.path.relpath(p, path) ap = to_posix(os.path.join(data_dir, key, rp)) archive_paths.append((ap, p)) if key == 'scripts' and not p.endswith('.exe'): with open(p, 'rb') as f: data = f.read() data = self.process_shebang(data) with open(p, 'wb') as f: f.write(data) # Now, stuff which is in site-packages, other than the # distinfo stuff. path = libdir distinfo = None for root, dirs, files in os.walk(path): if root == path: # At the top level only, save distinfo for later # and skip it for now for i, dn in enumerate(dirs): dn = fsdecode(dn) if dn.endswith('.dist-info'): distinfo = os.path.join(root, dn) del dirs[i] break assert distinfo, '.dist-info directory expected, not found' for fn in files: # comment out next suite to leave .pyc files in if fsdecode(fn).endswith(('.pyc', '.pyo')): continue p = os.path.join(root, fn) rp = to_posix(os.path.relpath(p, path)) archive_paths.append((rp, p)) # Now distinfo. Assumed to be flat, i.e. os.listdir is enough. files = os.listdir(distinfo) for fn in files: if fn not in ('RECORD', 'INSTALLER', 'SHARED', 'WHEEL'): p = fsdecode(os.path.join(distinfo, fn)) ap = to_posix(os.path.join(info_dir, fn)) archive_paths.append((ap, p)) wheel_metadata = [ 'Wheel-Version: %d.%d' % (wheel_version or self.wheel_version), 'Generator: distlib %s' % __version__, 'Root-Is-Purelib: %s' % is_pure, ] for pyver, abi, arch in self.tags: wheel_metadata.append('Tag: %s-%s-%s' % (pyver, abi, arch)) p = os.path.join(distinfo, 'WHEEL') with open(p, 'w') as f: f.write('\n'.join(wheel_metadata)) ap = to_posix(os.path.join(info_dir, 'WHEEL')) archive_paths.append((ap, p)) # Now, at last, RECORD. # Paths in here are archive paths - nothing else makes sense. self.write_records((distinfo, info_dir), libdir, archive_paths) # Now, ready to build the zip file pathname = os.path.join(self.dirname, self.filename) self.build_zip(pathname, archive_paths) return pathname def install(self, paths, maker, **kwargs): """ Install a wheel to the specified paths. If kwarg ``warner`` is specified, it should be a callable, which will be called with two tuples indicating the wheel version of this software and the wheel version in the file, if there is a discrepancy in the versions. This can be used to issue any warnings to raise any exceptions. If kwarg ``lib_only`` is True, only the purelib/platlib files are installed, and the headers, scripts, data and dist-info metadata are not written. The return value is a :class_mod:`InstalledDistribution` instance unless ``options.lib_only`` is True, in which case the return value is ``None``. """ dry_run = maker.dry_run warner = kwargs.get('warner') lib_only = kwargs.get('lib_only', False) pathname = os.path.join(self.dirname, self.filename) name_ver = '%s-%s' % (self.name, self.version) data_dir = '%s.data' % name_ver info_dir = '%s.dist-info' % name_ver metadata_name = posixpath.join(info_dir, METADATA_FILENAME) wheel_metadata_name = posixpath.join(info_dir, 'WHEEL') record_name = posixpath.join(info_dir, 'RECORD') wrapper = codecs.getreader('utf-8') with ZipFile(pathname, 'r') as zf: with zf.open(wheel_metadata_name) as bwf: wf = wrapper(bwf) message = message_from_file(wf) wv = message['Wheel-Version'].split('.', 1) file_version = tuple([int(i) for i in wv]) if (file_version != self.wheel_version) and warner: warner(self.wheel_version, file_version) if message['Root-Is-Purelib'] == 'true': libdir = paths['purelib'] else: libdir = paths['platlib'] records = {} with zf.open(record_name) as bf: with CSVReader(stream=bf) as reader: for row in reader: p = row[0] records[p] = row data_pfx = posixpath.join(data_dir, '') info_pfx = posixpath.join(info_dir, '') script_pfx = posixpath.join(data_dir, 'scripts', '') # make a new instance rather than a copy of maker's, # as we mutate it fileop = FileOperator(dry_run=dry_run) fileop.record = True # so we can rollback if needed bc = not sys.dont_write_bytecode # Double negatives. Lovely! outfiles = [] # for RECORD writing # for script copying/shebang processing workdir = tempfile.mkdtemp() # set target dir later # we default add_launchers to False, as the # Python Launcher should be used instead maker.source_dir = workdir maker.target_dir = None try: for zinfo in zf.infolist(): arcname = zinfo.filename if isinstance(arcname, text_type): u_arcname = arcname else: u_arcname = arcname.decode('utf-8') # The signature file won't be in RECORD, # and we don't currently don't do anything with it if u_arcname.endswith('/RECORD.jws'): continue row = records[u_arcname] if row[2] and str(zinfo.file_size) != row[2]: raise DistlibException('size mismatch for ' '%s' % u_arcname) if row[1]: kind, value = row[1].split('=', 1) with zf.open(arcname) as bf: data = bf.read() _, digest = self.get_hash(data, kind) if digest != value: raise DistlibException('digest mismatch for ' '%s' % arcname) if lib_only and u_arcname.startswith((info_pfx, data_pfx)): logger.debug('lib_only: skipping %s', u_arcname) continue is_script = (u_arcname.startswith(script_pfx) and not u_arcname.endswith('.exe')) if u_arcname.startswith(data_pfx): _, where, rp = u_arcname.split('/', 2) outfile = os.path.join(paths[where], convert_path(rp)) else: # meant for site-packages. if u_arcname in (wheel_metadata_name, record_name): continue outfile = os.path.join(libdir, convert_path(u_arcname)) if not is_script: with zf.open(arcname) as bf: fileop.copy_stream(bf, outfile) outfiles.append(outfile) # Double check the digest of the written file if not dry_run and row[1]: with open(outfile, 'rb') as bf: data = bf.read() _, newdigest = self.get_hash(data, kind) if newdigest != digest: raise DistlibException('digest mismatch ' 'on write for ' '%s' % outfile) if bc and outfile.endswith('.py'): try: pyc = fileop.byte_compile(outfile) outfiles.append(pyc) except Exception: # Don't give up if byte-compilation fails, # but log it and perhaps warn the user logger.warning('Byte-compilation failed', exc_info=True) else: fn = os.path.basename(convert_path(arcname)) workname = os.path.join(workdir, fn) with zf.open(arcname) as bf: fileop.copy_stream(bf, workname) dn, fn = os.path.split(outfile) maker.target_dir = dn filenames = maker.make(fn) fileop.set_executable_mode(filenames) outfiles.extend(filenames) if lib_only: logger.debug('lib_only: returning None') dist = None else: # Generate scripts # Try to get pydist.json so we can see if there are # any commands to generate. If this fails (e.g. because # of a legacy wheel), log a warning but don't give up. commands = None file_version = self.info['Wheel-Version'] if file_version == '1.0': # Use legacy info ep = posixpath.join(info_dir, 'entry_points.txt') try: with zf.open(ep) as bwf: epdata = read_exports(bwf) commands = {} for key in ('console', 'gui'): k = '%s_scripts' % key if k in epdata: commands['wrap_%s' % key] = d = {} for v in epdata[k].values(): s = '%s:%s' % (v.prefix, v.suffix) if v.flags: s += ' %s' % v.flags d[v.name] = s except Exception: logger.warning('Unable to read legacy script ' 'metadata, so cannot generate ' 'scripts') else: try: with zf.open(metadata_name) as bwf: wf = wrapper(bwf) commands = json.load(wf).get('extensions') if commands: commands = commands.get('python.commands') except Exception: logger.warning('Unable to read JSON metadata, so ' 'cannot generate scripts') if commands: console_scripts = commands.get('wrap_console', {}) gui_scripts = commands.get('wrap_gui', {}) if console_scripts or gui_scripts: script_dir = paths.get('scripts', '') if not os.path.isdir(script_dir): raise ValueError('Valid script path not ' 'specified') maker.target_dir = script_dir for k, v in console_scripts.items(): script = '%s = %s' % (k, v) filenames = maker.make(script) fileop.set_executable_mode(filenames) if gui_scripts: options = {'gui': True } for k, v in gui_scripts.items(): script = '%s = %s' % (k, v) filenames = maker.make(script, options) fileop.set_executable_mode(filenames) p = os.path.join(libdir, info_dir) dist = InstalledDistribution(p) # Write SHARED paths = dict(paths) # don't change passed in dict del paths['purelib'] del paths['platlib'] paths['lib'] = libdir p = dist.write_shared_locations(paths, dry_run) if p: outfiles.append(p) # Write RECORD dist.write_installed_files(outfiles, paths['prefix'], dry_run) return dist except Exception: # pragma: no cover logger.exception('installation failed.') fileop.rollback() raise finally: shutil.rmtree(workdir) def _get_dylib_cache(self): global cache if cache is None: # Use native string to avoid issues on 2.x: see Python #20140. base = os.path.join(get_cache_base(), str('dylib-cache'), sys.version[:3]) cache = Cache(base) return cache def _get_extensions(self): pathname = os.path.join(self.dirname, self.filename) name_ver = '%s-%s' % (self.name, self.version) info_dir = '%s.dist-info' % name_ver arcname = posixpath.join(info_dir, 'EXTENSIONS') wrapper = codecs.getreader('utf-8') result = [] with ZipFile(pathname, 'r') as zf: try: with zf.open(arcname) as bf: wf = wrapper(bf) extensions = json.load(wf) cache = self._get_dylib_cache() prefix = cache.prefix_to_dir(pathname) cache_base = os.path.join(cache.base, prefix) if not os.path.isdir(cache_base): os.makedirs(cache_base) for name, relpath in extensions.items(): dest = os.path.join(cache_base, convert_path(relpath)) if not os.path.exists(dest): extract = True else: file_time = os.stat(dest).st_mtime file_time = datetime.datetime.fromtimestamp(file_time) info = zf.getinfo(relpath) wheel_time = datetime.datetime(*info.date_time) extract = wheel_time > file_time if extract: zf.extract(relpath, cache_base) result.append((name, dest)) except KeyError: pass return result def is_compatible(self): """ Determine if a wheel is compatible with the running system. """ return is_compatible(self) def is_mountable(self): """ Determine if a wheel is asserted as mountable by its metadata. """ return True # for now - metadata details TBD def mount(self, append=False): pathname = os.path.abspath(os.path.join(self.dirname, self.filename)) if not self.is_compatible(): msg = 'Wheel %s not compatible with this Python.' % pathname raise DistlibException(msg) if not self.is_mountable(): msg = 'Wheel %s is marked as not mountable.' % pathname raise DistlibException(msg) if pathname in sys.path: logger.debug('%s already in path', pathname) else: if append: sys.path.append(pathname) else: sys.path.insert(0, pathname) extensions = self._get_extensions() if extensions: if _hook not in sys.meta_path: sys.meta_path.append(_hook) _hook.add(pathname, extensions) def unmount(self): pathname = os.path.abspath(os.path.join(self.dirname, self.filename)) if pathname not in sys.path: logger.debug('%s not in path', pathname) else: sys.path.remove(pathname) if pathname in _hook.impure_wheels: _hook.remove(pathname) if not _hook.impure_wheels: if _hook in sys.meta_path: sys.meta_path.remove(_hook) def verify(self): pathname = os.path.join(self.dirname, self.filename) name_ver = '%s-%s' % (self.name, self.version) data_dir = '%s.data' % name_ver info_dir = '%s.dist-info' % name_ver metadata_name = posixpath.join(info_dir, METADATA_FILENAME) wheel_metadata_name = posixpath.join(info_dir, 'WHEEL') record_name = posixpath.join(info_dir, 'RECORD') wrapper = codecs.getreader('utf-8') with ZipFile(pathname, 'r') as zf: with zf.open(wheel_metadata_name) as bwf: wf = wrapper(bwf) message = message_from_file(wf) wv = message['Wheel-Version'].split('.', 1) file_version = tuple([int(i) for i in wv]) # TODO version verification records = {} with zf.open(record_name) as bf: with CSVReader(stream=bf) as reader: for row in reader: p = row[0] records[p] = row for zinfo in zf.infolist(): arcname = zinfo.filename if isinstance(arcname, text_type): u_arcname = arcname else: u_arcname = arcname.decode('utf-8') if '..' in u_arcname: raise DistlibException('invalid entry in ' 'wheel: %r' % u_arcname) # The signature file won't be in RECORD, # and we don't currently don't do anything with it if u_arcname.endswith('/RECORD.jws'): continue row = records[u_arcname] if row[2] and str(zinfo.file_size) != row[2]: raise DistlibException('size mismatch for ' '%s' % u_arcname) if row[1]: kind, value = row[1].split('=', 1) with zf.open(arcname) as bf: data = bf.read() _, digest = self.get_hash(data, kind) if digest != value: raise DistlibException('digest mismatch for ' '%s' % arcname) def update(self, modifier, dest_dir=None, **kwargs): """ Update the contents of a wheel in a generic way. The modifier should be a callable which expects a dictionary argument: its keys are archive-entry paths, and its values are absolute filesystem paths where the contents the corresponding archive entries can be found. The modifier is free to change the contents of the files pointed to, add new entries and remove entries, before returning. This method will extract the entire contents of the wheel to a temporary location, call the modifier, and then use the passed (and possibly updated) dictionary to write a new wheel. If ``dest_dir`` is specified, the new wheel is written there -- otherwise, the original wheel is overwritten. The modifier should return True if it updated the wheel, else False. This method returns the same value the modifier returns. """ def get_version(path_map, info_dir): version = path = None key = '%s/%s' % (info_dir, METADATA_FILENAME) if key not in path_map: key = '%s/PKG-INFO' % info_dir if key in path_map: path = path_map[key] version = Metadata(path=path).version return version, path def update_version(version, path): updated = None try: v = NormalizedVersion(version) i = version.find('-') if i < 0: updated = '%s+1' % version else: parts = [int(s) for s in version[i + 1:].split('.')] parts[-1] += 1 updated = '%s+%s' % (version[:i], '.'.join(str(i) for i in parts)) except UnsupportedVersionError: logger.debug('Cannot update non-compliant (PEP-440) ' 'version %r', version) if updated: md = Metadata(path=path) md.version = updated legacy = not path.endswith(METADATA_FILENAME) md.write(path=path, legacy=legacy) logger.debug('Version updated from %r to %r', version, updated) pathname = os.path.join(self.dirname, self.filename) name_ver = '%s-%s' % (self.name, self.version) info_dir = '%s.dist-info' % name_ver record_name = posixpath.join(info_dir, 'RECORD') with tempdir() as workdir: with ZipFile(pathname, 'r') as zf: path_map = {} for zinfo in zf.infolist(): arcname = zinfo.filename if isinstance(arcname, text_type): u_arcname = arcname else: u_arcname = arcname.decode('utf-8') if u_arcname == record_name: continue if '..' in u_arcname: raise DistlibException('invalid entry in ' 'wheel: %r' % u_arcname) zf.extract(zinfo, workdir) path = os.path.join(workdir, convert_path(u_arcname)) path_map[u_arcname] = path # Remember the version. original_version, _ = get_version(path_map, info_dir) # Files extracted. Call the modifier. modified = modifier(path_map, **kwargs) if modified: # Something changed - need to build a new wheel. current_version, path = get_version(path_map, info_dir) if current_version and (current_version == original_version): # Add or update local version to signify changes. update_version(current_version, path) # Decide where the new wheel goes. if dest_dir is None: fd, newpath = tempfile.mkstemp(suffix='.whl', prefix='wheel-update-', dir=workdir) os.close(fd) else: if not os.path.isdir(dest_dir): raise DistlibException('Not a directory: %r' % dest_dir) newpath = os.path.join(dest_dir, self.filename) archive_paths = list(path_map.items()) distinfo = os.path.join(workdir, info_dir) info = distinfo, info_dir self.write_records(info, workdir, archive_paths) self.build_zip(newpath, archive_paths) if dest_dir is None: shutil.copyfile(newpath, pathname) return modified def compatible_tags(): """ Return (pyver, abi, arch) tuples compatible with this Python. """ versions = [VER_SUFFIX] major = VER_SUFFIX[0] for minor in range(sys.version_info[1] - 1, - 1, -1): versions.append(''.join([major, str(minor)])) abis = [] for suffix, _, _ in imp.get_suffixes(): if suffix.startswith('.abi'): abis.append(suffix.split('.', 2)[1]) abis.sort() if ABI != 'none': abis.insert(0, ABI) abis.append('none') result = [] arches = [ARCH] if sys.platform == 'darwin': m = re.match('(\w+)_(\d+)_(\d+)_(\w+)$', ARCH) if m: name, major, minor, arch = m.groups() minor = int(minor) matches = [arch] if arch in ('i386', 'ppc'): matches.append('fat') if arch in ('i386', 'ppc', 'x86_64'): matches.append('fat3') if arch in ('ppc64', 'x86_64'): matches.append('fat64') if arch in ('i386', 'x86_64'): matches.append('intel') if arch in ('i386', 'x86_64', 'intel', 'ppc', 'ppc64'): matches.append('universal') while minor >= 0: for match in matches: s = '%s_%s_%s_%s' % (name, major, minor, match) if s != ARCH: # already there arches.append(s) minor -= 1 # Most specific - our Python version, ABI and arch for abi in abis: for arch in arches: result.append((''.join((IMP_PREFIX, versions[0])), abi, arch)) # where no ABI / arch dependency, but IMP_PREFIX dependency for i, version in enumerate(versions): result.append((''.join((IMP_PREFIX, version)), 'none', 'any')) if i == 0: result.append((''.join((IMP_PREFIX, version[0])), 'none', 'any')) # no IMP_PREFIX, ABI or arch dependency for i, version in enumerate(versions): result.append((''.join(('py', version)), 'none', 'any')) if i == 0: result.append((''.join(('py', version[0])), 'none', 'any')) return set(result) COMPATIBLE_TAGS = compatible_tags() del compatible_tags def is_compatible(wheel, tags=None): if not isinstance(wheel, Wheel): wheel = Wheel(wheel) # assume it's a filename result = False if tags is None: tags = COMPATIBLE_TAGS for ver, abi, arch in tags: if ver in wheel.pyver and abi in wheel.abi and arch in wheel.arch: result = True break return result
[ "cjoffrey@gmail.com" ]
cjoffrey@gmail.com
d355e03c44e9a7c2bf59b7e5d70f274fd921714e
e842b091e4d7236b2136ac906eb5108ef17bd8d9
/16_DynamicProgramming1/Step11/ys_11.py
737b22c9bb16ef5528ca6bf1c909b13559fda3f1
[ "MIT" ]
permissive
StudyForCoding/BEAKJOON
6eb4bfe9a302e29893b128eebe44e8db271546c8
84e1c5e463255e919ccf6b6a782978c205420dbf
refs/heads/master
2023-04-04T09:17:58.145493
2021-04-14T07:05:06
2021-04-14T07:05:06
266,000,040
0
0
MIT
2021-04-14T07:05:07
2020-05-22T02:30:13
Python
UTF-8
Python
false
false
563
py
N = int(input()) #수열개수 arr = input() arr = arr.split() arr= [int(i) for i in arr] #입력받은 수열배열 d = [0] * (N)#가장 긴 증가하는 부분수열 길이 max1 = 0 for i in range(0, N): min1 = 0 for j in range(0, i): if arr[i] > arr[j]: #앞수열보다 큰 수가 있으면 if min1 < d[j]: #현재 인덱스요소보다 작은 수 카운트 min1 = d[j] d[i] = min1 + 1 if max1 < d[i]: max1 = d[i] #d배열중 최대 값이 가장 긴 증가하는 부분수열이 print(max1)
[ "69625013+to06109@users.noreply.github.com" ]
69625013+to06109@users.noreply.github.com
523fb7f3aa1c456f2a0e4aa859cb776f8d3f2534
bcb47f8acc6530e48c7a013d459e453c2e5110c0
/urls.py
16f8f43a8d797fad08220b99def350e573c300ad
[]
no_license
gebisys/libreapp
09cef76e0635eb0fe1803769cba07b4320c29fc9
14dede79e47c7a0e0d9c492c5d87df9123ef59ec
refs/heads/master
2016-09-06T08:06:55.609876
2013-06-12T14:47:18
2013-06-12T14:47:18
null
0
0
null
null
null
null
UTF-8
Python
false
false
766
py
from django.conf.urls.defaults import * from django.views.generic.simple import direct_to_template from django.contrib import admin from django.conf import settings from news.feeds import ArchiveFeed from django.contrib import admin admin.autodiscover() urlpatterns = patterns('', url(r'^$', 'news.views.index_view'), url(r"^post/(?P<year>\d{4})/(?P<mounth>\d{1,2})/(?P<day>\d{2})/(?P<slug>[\w-]+)$", 'news.views.post'), url(r'^feed/$', ArchiveFeed()), # archivos estaticos url(r'^static/(?P<path>.*)$', 'django.views.static.serve', { 'document_root': settings.STATIC_ROOT }), url(r'^user/$', 'news.views.createuser'), # Uncomment the next line to enable the admin: url(r'^admin/', include(admin.site.urls)), )
[ "gebisys@gmail.com" ]
gebisys@gmail.com
c22cd593f5f83ae3732d104ca10c62e681b4363f
15f321878face2af9317363c5f6de1e5ddd9b749
/solutions_python/Problem_159/609.py
1d4cb7959af1112bc540d578dcf82f9dfd5fc3ae
[]
no_license
dr-dos-ok/Code_Jam_Webscraper
c06fd59870842664cd79c41eb460a09553e1c80a
26a35bf114a3aa30fc4c677ef069d95f41665cc0
refs/heads/master
2020-04-06T08:17:40.938460
2018-10-14T10:12:47
2018-10-14T10:12:47
null
0
0
null
null
null
null
UTF-8
Python
false
false
784
py
f = open('A-large.in') #f = open('test.in') count = int(f.readline()) output = '' for x in xrange(1, count + 1): platesCount = int(f.readline()) arr = f.readline().split() case1 = 0 case2 = 0 case2MaxGap = 0 for i in xrange(0, platesCount - 1): curPlate = int(arr[i]) nextPlate = int(arr[i+1]) gap = curPlate - nextPlate case2MaxGap = max(case2MaxGap, gap) if gap > 0: case1 += gap for j in xrange(0, platesCount - 1): curPlate = int(arr[j]) if curPlate < case2MaxGap: case2 += curPlate else: case2 += case2MaxGap output += 'Case #' + str(x) + ': ' + str(case1) + ' ' + str(case2) + '\n' print(output) newf = open('output.txt','w') newf.write(output)
[ "miliar1732@gmail.com" ]
miliar1732@gmail.com
081d7e5abd1e6534a8b35c34749bb42b37674a75
41e59ef6eed28c5c89fc5e5ad54605077819f812
/mnist_archs.py
e8337f2311dd87a1c7e47e738a2387ba43589f6c
[]
no_license
zueigung1419/IntrinsicDimDeep
498185a6bf27d61eb2073c5dfa1ba0514611e142
2c5140009eb3efccca1d774dc66cb043c04b26d8
refs/heads/master
2022-04-11T11:50:42.203540
2020-03-13T11:38:19
2020-03-13T11:38:19
null
0
0
null
null
null
null
UTF-8
Python
false
false
1,795
py
import torch import torch.nn as nn import torch.nn.functional as F # F.Chollet architecture class Net(nn.Module): def __init__(self): super(Net, self).__init__() self.conv1 = nn.Conv2d(1, 32, kernel_size=3) self.conv2 = nn.Conv2d(32, 64, kernel_size=3) # dropout 0.25 self.fc1 = nn.Linear(1600, 128) # dropout 0.5 self.fc2 = nn.Linear(128, 10) def forward(self, x): x = F.relu(F.max_pool2d( self.conv1(x), 2 ) ) x = F.relu(F.max_pool2d( self.conv2(x), 2 ) ) x = x.view(-1, 1600) x = F.relu(self.fc1(x)) x = self.fc2(x) return F.log_softmax(x, dim=1) def extract(self, x,verbose=False): out1 = F.relu(F.max_pool2d(self.conv1(x), 2 ) ) out2 = F.relu(F.max_pool2d(self.conv2(out1), 2 ) ) t = out2.view(-1, 1600) out3 = F.relu(self.fc1(t)) t = self.fc2(out3) out4 = F.log_softmax(t, dim=1) if verbose == True: print(out1.size()) print(out2.size()) print(out3.size()) print(out4.size()) return out1, out2, out3, out4 def extract_all(self, x,verbose=False): out1 = self.conv1(x) out2 = F.relu(F.max_pool2d(out1,2)) out3 = self.conv2(out2) out4 = F.relu(F.max_pool2d(out3,2)) t = out4.view(-1, 1600) out5 = F.relu(self.fc1(t)) t = self.fc2(out5) out6 = F.log_softmax(t, dim=1) if verbose == True: print(out1.size()) print(out2.size()) print(out3.size()) print(out4.size()) print(out5.size()) print(out6.size()) return out1, out2, out3, out4, out5, out6
[ "alessioansuini@gmail.com" ]
alessioansuini@gmail.com
05502558d8297623342bf39c101073a7049393f0
140003818ac3cf9a7d427ee5f0e87b1bc4f770ec
/metrics/mAP.py
cf8920732faacba76ba21e41f2c07bd950366558
[ "MIT" ]
permissive
chisyliu/RADDet
e495a80a7bdbd3ecd473130386c00809df5ff4f0
5c0020dc24024f5b87e864554769f5926b73593a
refs/heads/main
2023-04-18T22:21:57.054785
2021-05-02T14:16:51
2021-05-02T14:16:51
364,338,015
1
0
MIT
2021-05-04T17:40:57
2021-05-04T17:40:57
null
UTF-8
Python
false
false
4,450
py
import tensorflow as tf import tensorflow.keras as K import numpy as np import util.helper as helper def getTruePositive(pred, gt, input_size, iou_threshold=0.5, mode="3D"): """ output tp (true positive) with size [num_pred, ] """ assert mode in ["3D", "2D"] tp = np.zeros(len(pred)) detected_gt_boxes = [] for i in range(len(pred)): current_pred = pred[i] if mode == "3D": current_pred_box = current_pred[:6] current_pred_score = current_pred[6] current_pred_class = current_pred[7] gt_box = gt[..., :6] gt_class = gt[..., 6] else: current_pred_box = current_pred[:4] current_pred_score = current_pred[4] current_pred_class = current_pred[5] gt_box = gt[..., :4] gt_class = gt[..., 4] if len(detected_gt_boxes) == len(gt): break if mode == "3D": iou = helper.iou3d(current_pred_box[np.newaxis, ...], gt_box, input_size) else: iou = helper.iou2d(current_pred_box[np.newaxis, ...], gt_box) iou_max_idx = np.argmax(iou) iou_max = iou[iou_max_idx] if iou_max >= iou_threshold and iou_max_idx not in detected_gt_boxes: tp[i] = 1. detected_gt_boxes.append(iou_max_idx) fp = 1. - tp return tp, fp def computeAP(tp, fp, num_gt_class): """ Compute Average Precision """ tp_cumsum = np.cumsum(tp).astype(np.float32) fp_cumsum = np.cumsum(fp).astype(np.float32) recall = tp_cumsum / (num_gt_class + 1e-16) precision = tp_cumsum / (tp_cumsum + fp_cumsum) ########## NOTE: the following is under the reference of the repo ########### recall = np.insert(recall, 0, 0.0) recall = np.append(recall, 1.0) precision = np.insert(precision, 0, 0.0) precision = np.append(precision, 0.0) mrec = recall.copy() mpre = precision.copy() for i in range(len(mpre)-2, -1, -1): mpre[i] = max(mpre[i], mpre[i+1]) i_list = [] for i in range(1, len(mrec)): if mrec[i] != mrec[i-1]: i_list.append(i) # if it was matlab would be i + 1 ap = 0.0 for i in i_list: ap += ((mrec[i]-mrec[i-1])*mpre[i]) return ap, mrec, mpre def mAP(predictions, gts, input_size, ap_each_class, tp_iou_threshold=0.5, mode="3D"): """ Main function for calculating mAP Args: predictions -> [num_pred, 6 + score + class] gts -> [num_gt, 6 + class]""" gts = gts[gts[..., :6].any(axis=-1) > 0] all_gt_classes = np.unique(gts[:, 6]) ap_all = [] # ap_all_classes = np.zeros(num_all_classes).astype(np.float32) for class_i in all_gt_classes: ### NOTE: get the prediction per class and sort it ### pred_class = predictions[predictions[..., 7] == class_i] pred_class = pred_class[np.argsort(pred_class[..., 6])[::-1]] ### NOTE: get the ground truth per class ### gt_class = gts[gts[..., 6] == class_i] tp, fp = getTruePositive(pred_class, gt_class, input_size, \ iou_threshold=tp_iou_threshold, mode=mode) ap, mrecall, mprecision = computeAP(tp, fp, len(gt_class)) ap_all.append(ap) ap_each_class[int(class_i)].append(ap) mean_ap = np.mean(ap_all) return mean_ap, ap_each_class def mAP2D(predictions, gts, input_size, ap_each_class, tp_iou_threshold=0.5, mode="2D"): """ Main function for calculating mAP Args: predictions -> [num_pred, 4 + score + class] gts -> [num_gt, 4 + class]""" gts = gts[gts[..., :4].any(axis=-1) > 0] all_gt_classes = np.unique(gts[:, 4]) ap_all = [] for class_i in all_gt_classes: ### NOTE: get the prediction per class and sort it ### pred_class = predictions[predictions[..., 5] == class_i] pred_class = pred_class[np.argsort(pred_class[..., 4])[::-1]] ### NOTE: get the ground truth per class ### gt_class = gts[gts[..., 4] == class_i] tp, fp = getTruePositive(pred_class, gt_class, input_size, \ iou_threshold=tp_iou_threshold, mode=mode) ap, mrecall, mprecision = computeAP(tp, fp, len(gt_class)) ap_all.append(ap) ap_each_class[int(class_i)].append(ap) mean_ap = np.mean(ap_all) return mean_ap, ap_each_class
[ "zhangaocanada@gmail.com" ]
zhangaocanada@gmail.com
ea09bbbb1e03af80da8927d9c004f5b2689e5ce9
31abd09a08ebbec64951b7724fb94b178546bfd0
/Classifier/binary_classificationTree.py
287452ddfad30ccab4cb3b759930b3a3cb1bbde9
[]
no_license
juancho618/TAI_Project
9839ab014a14cc5c030217f12e72ab164eada297
63d55659cc1e4e7f36461762e99aaaa5f4b2b500
refs/heads/master
2021-01-23T02:18:58.628859
2017-05-31T07:27:33
2017-05-31T07:27:33
92,917,385
0
0
null
null
null
null
UTF-8
Python
false
false
2,551
py
import pandas as pd import matplotlib.pyplot as plt # plot library from sklearn import tree from dataProcessing import * # dummy encoding from sklearn.preprocessing import OneHotEncoder # best codification for categorical data df = pd.read_csv('../finalDS.csv', header=0) # print df # preprocess dataset with original values # Convert all the nominal values to integers (dummy version) df2 = encodeColumnDummy(df,0) # changing health insurance df2 = encodeColumnDummy(df2, 2) # changing Diagnosis df2 = encodeColumnDummy(df2, 3) # changing Speciality df2 = encodeColumnDummy(df2, 4) # changing Gender df2 = encodeColumnDummy(df2, 5) # changing Day # ------- functional way to get the data -------- x = df2.iloc[:,:6] # data with the attributes y = df2.iloc[:,6] # results # dataset spliter from sklearn.model_selection import train_test_split # import split and test functionality x_train, x_test, y_train, y_test = train_test_split(x,y, test_size = .8) # tree classifier algorithm clf = tree.DecisionTreeClassifier() # calling the decision tree clasifier # Naive Bayes classifier algorithm from sklearn.naive_bayes import MultinomialNB # import gaussian classi nb_clf = MultinomialNB() # --- Trying one hot encoder ------ enc = OneHotEncoder(categorical_features =[0, 2, 3, 4, 5]) # One Hot encoder Specifying the categorical attributes. enc.fit(x) #fit the encoder to the data clf.fit(enc.transform(x_train), y_train) # create the learninf instance nb_clf.fit(enc.transform(x_train), y_train) # Nive Bayes - Multinomial model # prediction predictions = clf.predict(enc.transform(x_test)) prediction_NB = nb_clf.predict(enc.transform(x_test)) # import the result metrics from sklearn import metrics print("Tree Classifier Precision", metrics.precision_score(y_test, predictions)) print("Tree Classifier Recall", metrics.recall_score(y_test, predictions)) print("Tree Classifier Beta Score 1", metrics.f1_score(y_test, predictions)) print("Tree Classifier Beta Score 0.5", metrics.fbeta_score(y_test, predictions, beta=0.5)) print("Tree Classifier Beta Score 2", metrics.fbeta_score(y_test, predictions, beta=2)) print("Naive Bayes Classifier Precision", metrics.precision_score(y_test,prediction_NB )) print("Naive Bayes Recall", metrics.recall_score(y_test, prediction_NB)) print("Naive Bayes Beta Score 1", metrics.f1_score(y_test, prediction_NB)) print("Naive Bayes Beta Score 0.5", metrics.fbeta_score(y_test, prediction_NB, beta=0.5)) print("Naive Bayes Beta Score 2", metrics.fbeta_score(y_test, prediction_NB, beta=2))
[ "jjsorianoe@gmail.com" ]
jjsorianoe@gmail.com
5532be931e0b45443ac49288433d6dbcf3e6db85
ad297efe52cb150997f6ee7828bb9fbd8512d60b
/scout/Miner/amazon_scraper.py
e2ef13a8379a6122cffd3765cc16b0aad7b9db73
[]
no_license
happyxiaoxu/zonmine
8b43f5bbb4dbadda324140bddd04aa603f7f51e8
e016cbc77c4be05ffdcd74bd98f8da898956a419
refs/heads/master
2022-12-16T09:53:04.427931
2019-04-03T02:02:38
2019-04-03T02:02:38
179,192,426
1
0
null
2022-12-08T01:21:41
2019-04-03T02:11:17
CSS
UTF-8
Python
false
false
13,415
py
# from .config import Config # from .product import Product from scout.Miner.config import Config from scout.Miner.product import Product from scout.Miner.grepwords_api import KeywordsApi import lxml.html import re # from fuzzywuzzy import fuzz from difflib import SequenceMatcher from urllib.parse import quote_plus from collections import OrderedDict import warnings warnings.filterwarnings("ignore") ##################################################################################### # Custom method to get first element or empty element by xpath def xpath_first(self, elm_xpath): elements = self.xpath(elm_xpath) empty_elm = lxml.html.HtmlElement() empty_elm.text = "-" return next(iter(elements), empty_elm) def xpath_get_text(self, elm_xpath): # gets the first element and trims the text element = self.xpath_first(elm_xpath) return " ".join(element.text_content().split()) lxml.html.HtmlElement.xpath_first = xpath_first lxml.html.HtmlElement.xpath_get_text = xpath_get_text ##################################################################################### def check_similarity(a, b): i = SequenceMatcher(None, a, b) return i.ratio() class AmazonScraper(object): def __init__(self, scraper_obj): # Scraper.__init__(self) self.scraper_obj = scraper_obj self.kw_api = KeywordsApi() return None # def asin_search(self, product_obj): asin = product_obj.asin print("ASIN : {}".format(asin)) # https://www.amazon.com/dp/B075QDGZX9 asin_url = Config.amazon_asin_url.format(asin) # asin_resp = self.session.get(asin_url) asin_resp = self.scraper_obj.make_request(asin_url) if asin_resp.status_code != 200: print("Skipping ASIN Search as status code is: {}".format(asin_resp.status_code)) return product_obj asin_xml = lxml.html.fromstring(asin_resp.content) print(asin_resp) print(asin_resp.url) product_obj.asin = asin product_obj.url = asin_url # product_obj.title = asin_xml.xpath_get_text('//span[@id="productTitle"]') product_obj.bread_crumbs = asin_xml.xpath_get_text('//div[contains(@class, "a-breadcrumb")]') # brand = asin_xml.xpath_first('//a[@id="bylineInfo"] | //a[@id="brand"]').attrib.get("href") # if brand: # product_obj.brand = brand.split("/")[1] product_obj.brand = asin_xml.xpath_get_text('//a[@id="bylineInfo"] | //a[@id="brand"]') product_obj.description = asin_xml.xpath_get_text('//div[@id="productDescription" and contains(@class, "a-section")]//p') # product_obj.weight = asin_xml.xpath_get_text(''' # //th[contains(text(), "Weight")]/following-sibling::td | //text()[contains(.,'Weight')]/ancestor::li # ''') product_obj.item_model_number = asin_xml.xpath_get_text(''' //*[contains(text(), "Item model number")]/following-sibling::td | //text()[contains(.,'Item model number')]/ancestor::li ''') # product_obj.item_dimensions = asin_xml.xpath_get_text(''' # //th[contains(text(), "Product Dimensions")]/following-sibling::td | //th[contains(text(), "Item Dimensions")]/following-sibling::td # ''') # BSR Data bsr_data = asin_xml.xpath_first(''' //*[contains(text(), "Best Sellers Rank")]/following-sibling::td ''').text.strip("(").strip().split("in") + ["-", "-"] product_obj.bsr = bsr_data[0] product_obj.bsr_category = bsr_data[1] # # Dimensions, in inches # //th[contains(text(), "Dimensions")]/following-sibling::td | //text()[contains(.,'Dimensions')]/ancestor::li dimensions_text = asin_xml.xpath_get_text(''' //*[contains(text(), "Dimensions")]/following-sibling::td | //text()[contains(.,'Dimensions')]/ancestor::li ''').split("inches")[0] dimensions = re.findall(r"[-+]?\d*\.\d+|\d+", dimensions_text) if dimensions: dimensions = [float(i) for i in sorted(dimensions, reverse=True)] dimensions = (dimensions + ["-", "-", "-"])[:3] product_obj.item_dimensions_length = dimensions[0] product_obj.item_dimensions_width = dimensions[1] product_obj.item_dimensions_thickness = dimensions[2] ######## # Item Weight, convert ounces to pounds by dividing with 16 # Alt. Xpaths: # //th[contains(text(), "Shipping Weight")]/following-sibling::td | //th[contains(text(), "Item Weight")]/following-sibling::td # //*[contains(text(), "Item Weight")]/following-sibling::td | //*[contains(text(), "Shipping Weight")]/following-sibling::td item_weight_text = asin_xml.xpath_get_text(''' //*[contains(text(), "Weight")]/following-sibling::td ''') item_weight = re.search(r"[-+]?\d*\.\d+|\d+", item_weight_text) if item_weight: item_weight = item_weight.group() if "ounce" in item_weight_text: print("Converting ounce to pound") item_weight = float(item_weight) / 16 product_obj.item_weight = item_weight # Fullfillment # print(result.xpath_first('.//span[contains(@class, "s-sponsored-info-icon")]').tag) if asin_xml.xpath('//text()[contains(., "sold by Amazon")]'): product_obj.is_amz = True elif asin_xml.xpath('//text()[contains(., "Fulfilled by Amazon")]'): product_obj.is_fba = True else: product_obj.is_fbm = True ############################ # /html/body//text()[matches(.,'test', 'i')] product_obj.manufacturer = asin_xml.xpath_get_text('//th[contains(text(), "Manufacturer")]/following-sibling::td') # product_obj.sold_by = asin_xml.xpath_get_text('//div[@id="merchant-info"]//a[1] | //div[@id="merchant-info"]') # product_obj.sold_by = asin_xml.xpath_get_text('//*[@id="merchant-info"]//a[1] | //*[@id="merchant-info"]') product_obj.sold_by = asin_xml.xpath_get_text('//*[@id="merchant-info"]//a') # product_obj.in_stock = asin_xml.xpath_get_text('//div[@id="availability"]') product_obj.in_stock = asin_xml.xpath_get_text('//*[@id="availability"]') # Extracting Features features_elm = asin_xml.xpath('//div[@id="feature-bullets"]//li//span[not(descendant:: a) and contains(@class, "a-list-item")]/text()') #not complete features = [" ".join(each_feature.split()) for each_feature in features_elm] + ["-", "-", "-", "-", "-"] features = features[:5] product_obj.feature1 = features[0] product_obj.feature2 = features[1] product_obj.feature3 = features[2] product_obj.feature4 = features[3] product_obj.feature5 = features[4] ##### sellers_count = re.search(r'\((.*?)\)', asin_xml.xpath_first('//div[@id="olp_feature_div"]').text_content()) if sellers_count: product_obj.sellers_count = sellers_count.group(1) print(product_obj.sellers_count) #Extracting Images # images_elm = asin_xml.xpath_first('//script[contains(text(), "ImageBlockATF") and not(contains(text(), "imageBlockATF"))]')[0] thumb_images = asin_xml.xpath('//li[contains(@class,"item")]//span[contains(@class,"a-button-thumbnail")]//img/@src') images = [re.sub(pattern = r'_.+_.', string=img, repl = "") for img in thumb_images] + ["-", "-", "-", "-", "-", "-", "-", "-"] images = images[:7] product_obj.image1 = images[0] product_obj.image2 = images[1] product_obj.image3 = images[2] product_obj.image4 = images[3] product_obj.image5 = images[4] product_obj.image6 = images[5] product_obj.image7 = images[6] # product_obj.image_8 = images[7] # Review Percentage product_obj.five_star_percentage = asin_xml.xpath_get_text('//a[contains(@class, "5star histogram-review-count")]') product_obj.four_star_percentage = asin_xml.xpath_get_text('//a[contains(@class, "4star histogram-review-count")]') product_obj.three_star_percentage = asin_xml.xpath_get_text('//a[contains(@class, "3star histogram-review-count")]') product_obj.two_star_percentage = asin_xml.xpath_get_text('//a[contains(@class, "2star histogram-review-count")]') product_obj.one_star_percentage = asin_xml.xpath_get_text('//a[contains(@class, "1star histogram-review-count")]') del([asin, asin_url, asin_resp, asin_xml, features_elm, thumb_images, images]) return product_obj # def keyword_search(self, keyword,job,db_handler): job.refresh_from_db() # print("Searching Amazon") products = [] print(keyword) print(quote_plus(keyword)) print("Current Keyword {}".format(keyword)) search_url = Config.amazon_search_url.format(quote_plus(keyword)) print(search_url) search_resp = self.scraper_obj.make_request(search_url) print(search_resp) print(search_resp.url) if search_resp.status_code != 200: print("Skipping keywords as status code is: {}".format(search_resp.status_code)) return products search_xml = lxml.html.fromstring(search_resp.content) print(search_xml) print('///////////////////') kw_stats = self.kw_api.get_stats([keyword]) cpc = kw_stats[keyword][0] monthly_search_volume = kw_stats[keyword][1] competition = kw_stats[keyword][2] # Edit xpath to include sponsored listings # search_results = search_xml.xpath('//div[@id="resultsCol"]//li[not(.//h5) and contains(@class, "s-result-item")]') # search_results = search_xml.xpath('//div[contains(@class, "s-result-list")]//div[@data-asin]')[:1] search_results = search_xml.xpath('//div[contains(@class, "s-result-list")]//div[@data-asin]') # search_results_1 = search_xml.xpath('//div[@id="resultsCol"]//li[@data-asin]//h2') # search_results = search_xml.xpath('//div[contains(@class, "s-result-list")]//div[@data-asin]') print(len(search_results)) # if not search_results: # with open("test.html", "wb") as oo: # oo.write(search_resp.content) if not search_results: print("No results found for keyword : {}".format(keyword)) return products for result in search_results: product_obj = Product() url = result.xpath('.//span[contains(@class, "a-text-normal")]/parent::a/@href') title = result.xpath_first('.//span[contains(@class, "a-text-normal")]//text() | .//img[contains(@class, "s-image")]//@alt') print(title.encode()) asin = result.attrib.get("data-asin") price = result.xpath_get_text('.//span[@class="a-offscreen"]') primary_image = result.xpath_first('.//img[@srcset and @data-image-load]/@src') review_count = result.xpath_first('.//a[contains(@href, "customerReviews")]').text_content().strip() review_score = result.xpath_first('.//i[contains(@class,"a-icon-star")]//span').text_content().split(" ")[0] is_prime = False is_addon = False is_sponsored = False # print(result.xpath_first('.//span[contains(@class, "s-sponsored-info-icon")]').tag) if result.xpath('.//i[contains(@class,"a-icon-prime")]'): is_prime = True if result.xpath('.//i[contains(@class,"a-icon-addon")]'): is_addon = True if result.xpath('.//div[@data-component-type="sp-sponsored-result"]'): print("Inside is sponsored") is_sponsored = True #Updating Counter to rotate proxies # Get details from ASIN # self.update_counter() product_obj.search_keyword = keyword product_obj.search_rank = search_results.index(result) + 1 product_obj.url = url product_obj.asin = asin product_obj.title = title product_obj.price = price product_obj.primary_image = primary_image product_obj.review_count = review_count product_obj.review_score = review_score product_obj.is_prime = is_prime product_obj.is_addon = is_addon product_obj.sponsored = is_sponsored product_obj.cpc = cpc product_obj.monthly_search_volume = monthly_search_volume product_obj.competition = competition print(is_sponsored) # products.append(product_obj) products.append(self.asin_search(product_obj)) db_handler.save_product(product_obj,job) job.scout_results_counter += 1 job.save() job.refresh_from_db() del([product_obj]) # break # with open("test.html", "wb") as oo: # oo.write(search_resp.content) return products
[ "happyxiaoxu@outlook.com" ]
happyxiaoxu@outlook.com
0121ed472138d492c9faefbd6f0c04308f5c2a4a
69cbb90a54ef4c312d7939c31d1da4060776306a
/airWritingPrediction.py
d0716f98214463884e6f81fe5825557eb6af1cd4
[]
no_license
priyankagohil/AirWritingRecognition
602d08131e7908b6fc13204dc4d0e9f27ca6e041
5228798157f7f21b59e6ef688e8f33e076657416
refs/heads/master
2023-08-24T10:48:51.588127
2021-01-06T17:52:56
2021-01-06T17:52:56
312,326,541
3
0
null
2021-10-18T05:43:02
2020-11-12T16:01:41
Python
UTF-8
Python
false
false
8,577
py
import cv2 import numpy as np import copy import torch import torchvision.transforms as transforms import torch.nn as nn import torch.nn.functional as F from PIL import Image rect1_tl = (320, 140) rect2_tl = (320, 240) rect3_tl = (320, 340) rect4_tl = (240, 270) rect5_tl = (400, 270) height = 30 width = 30 """CNN architecture of the model""" class Cnn(nn.Module): def __init__(self): super(Cnn, self).__init__() # convolutional layer self.conv1 = nn.Conv2d(1, 16, 3, padding=1) self.conv2 = nn.Conv2d(16, 32, 3, padding=1) self.conv3 = nn.Conv2d(32, 64, 3, padding=1) self.pool = nn.MaxPool2d(2, 2) self.fc1 = nn.Linear(64 * 3 * 3, 256) self.fc2 = nn.Linear(256, 26) self.dropout = nn.Dropout(0.25) def forward(self, x): # Adding sequence of convolutional and max pooling layers x = self.pool(F.relu(self.conv1(x))) x = self.pool(F.relu(self.conv2(x))) x = self.pool(F.relu(self.conv3(x))) x = x.view(-1, 64 * 3 * 3) x = self.dropout(x) x = F.relu(self.fc1(x)) x = self.dropout(x) x = self.fc2(x) return x def get_histogram(frame): roi1 = frame[rect1_tl[1]:rect1_tl[1] + width, rect1_tl[0]:rect1_tl[0] + height] roi2 = frame[rect2_tl[1]:rect2_tl[1] + width, rect2_tl[0]:rect2_tl[0] + height] roi3 = frame[rect3_tl[1]:rect3_tl[1] + width, rect3_tl[0]:rect3_tl[0] + height] roi4 = frame[rect4_tl[1]:rect4_tl[1] + width, rect4_tl[0]:rect4_tl[0] + height] roi5 = frame[rect5_tl[1]:rect5_tl[1] + width, rect5_tl[0]:rect5_tl[0] + height] roi = np.concatenate((roi1, roi2, roi3, roi4, roi5), axis=0) roi_hsv = cv2.cvtColor(roi, cv2.COLOR_BGR2HSV) return cv2.calcHist([roi_hsv], [0, 1], None, [180, 256], [0, 180, 0, 256]) def draw_rectangles(frame=0): frame_with_rect = frame cv2.rectangle(frame_with_rect, rect1_tl, tuple(np.array(rect1_tl) + np.array((height, width))), (0, 0, 255), 1) cv2.rectangle(frame_with_rect, rect2_tl, tuple(np.array(rect2_tl) + np.array((height, width))), (0, 0, 255), 1) cv2.rectangle(frame_with_rect, rect3_tl, tuple(np.array(rect3_tl) + np.array((height, width))), (0, 0, 255), 1) cv2.rectangle(frame_with_rect, rect4_tl, tuple(np.array(rect4_tl) + np.array((height, width))), (0, 0, 255), 1) cv2.rectangle(frame_with_rect, rect5_tl, tuple(np.array(rect5_tl) + np.array((height, width))), (0, 0, 255), 1) return frame_with_rect def get_mask(frame, histogram): frame_hsv = cv2.cvtColor(frame, cv2.COLOR_BGR2HSV) mask = cv2.calcBackProject([frame_hsv], [0, 1], histogram, [0, 180, 0, 256], 1) _, mask = cv2.threshold(mask, 10, 255, cv2.THRESH_BINARY) kernel = cv2.getStructuringElement(cv2.MORPH_ELLIPSE, (10, 10)) mask = cv2.filter2D(mask, -1, kernel) kernel1 = np.ones((7, 7), np.uint8) mask = cv2.morphologyEx(mask, cv2.MORPH_OPEN, kernel1) mask = cv2.morphologyEx(mask, cv2.MORPH_CLOSE, kernel) mask = cv2.bilateralFilter(mask, 5, 75, 75) return mask def get_max_contour(mask): contours = cv2.findContours(mask, cv2.RETR_TREE, cv2.CHAIN_APPROX_SIMPLE)[0] max = 0 mi = 0 for i in range(len(contours)): area = cv2.contourArea(contours[i]) if area > 1500: max = area mi = i return contours[mi] def draw_defects(frame_with_rect, max_contour, hull): defects = cv2.convexityDefects(max_contour, hull) for i in range(defects.shape[0]): s, e, f, d = defects[i, 0] start = tuple(max_contour[s][0]) end = tuple(max_contour[e][0]) far = tuple(max_contour[f][0]) cv2.line(frame_with_rect, start, far, [255, 0, 0], 2) cv2.line(frame_with_rect, far, end, [0, 255, 0], 2) cv2.circle(frame_with_rect, far, 5, [0, 0, 255], -1) def get_centroid(contour): m = cv2.moments(contour) cx = int(m['m10'] / m['m00']) cy = int(m['m01'] / m['m00']) return cx, cy def get_farthest_point(defects, contour, centroid): if defects is not None and centroid is not None: s = defects[:, 0][:, 0] cx, cy = centroid x = np.array(contour[s][:, 0][:, 0], dtype=np.float) y = np.array(contour[s][:, 0][:, 1], dtype=np.float) xp = cv2.pow(cv2.subtract(x, cx), 2) yp = cv2.pow(cv2.subtract(y, cy), 2) dist = cv2.sqrt(cv2.add(xp, yp)) dist_max_i = np.argmax(dist) if dist_max_i < len(s): farthest_defect = s[dist_max_i] farthest_point = tuple(contour[farthest_defect][0]) return farthest_point else: return None def get_ROI(canvas): gray = cv2.bitwise_not(canvas) ret, thresh = cv2.threshold(gray, 5, 255, cv2.THRESH_BINARY_INV) _, ctrs, _ = cv2.findContours(thresh, cv2.RETR_TREE, cv2.CHAIN_APPROX_SIMPLE) areas = [] for i in range(len(ctrs)): x, y, w, h = cv2.boundingRect(ctrs[i]) areas.append((w * h, i)) def sort_second(val): return val[0] areas.sort(key=sort_second, reverse=True) x, y, w, h = cv2.boundingRect(ctrs[areas[1][1]]) cv2.rectangle(canvas, (x, y), (x + w, y + h), (255, 255, 0), 1) roi = gray[y:y + h, x:x + w] return roi def character_prediction(roi, model): """Predicts character written with image processing""" img = cv2.resize(roi, (28, 28)) img = cv2.GaussianBlur(img, (3, 3), 0) img = Image.fromarray(img) normalize = transforms.Normalize( mean=[0.5, 0.5, 0.5], std=[0.5, 0.5, 0.5] ) preprocess = transforms.Compose([ transforms.Resize((28, 28)), transforms.ToTensor(), normalize ]) p_img = preprocess(img) model.eval() p_img = p_img.reshape([1, 1, 28, 28]).float() output = model(torch.transpose(p_img, 2, 3)) _, preds_tensor = torch.max(output, 1) preds = np.squeeze(preds_tensor.numpy()) return preds def main(): video = cv2.VideoCapture(0) canvas = np.zeros((720, 1280), np.uint8) far_points = [] pressed = False is_drawing = False made_prediction = False # Creating the model model = Cnn() model.load_state_dict(torch.load('model_emnist.pt', map_location='cpu')) # Actions to perform with each key while True: _, frame = video.read() frame = cv2.flip(frame, flipCode=1) original_frame = copy.deepcopy(frame) original_frame = draw_rectangles(original_frame) canvas[:, :] = 255 key = cv2.waitKey(1) # ready to draw if key & 0xFF == ord('s'): pressed = True histogram = get_histogram(frame) # To start drawing if key & 0xFF == ord('d'): is_drawing = True # To clear drawing if key & 0xFF == ord('c'): canvas[:, :] = 255 is_drawing = False far_points.clear() made_prediction = False if is_drawing: if len(far_points) > 100: far_points.pop(0) far_points.append(far) for i in range(len(far_points) - 1): cv2.line(original_frame, far_points[i], far_points[i + 1], (255, 5, 255), 20) cv2.line(canvas, far_points[i], far_points[i + 1], (0, 0, 0), 20) # To predict the character drawn if key & 0xFF == ord('p'): is_drawing = False roi = get_ROI(canvas) prediction = character_prediction(roi, model) print(prediction) made_prediction = True name = str(prediction) + '.jpg' cv2.imwrite(name, roi) if pressed: mask = get_mask(frame, histogram) max_contour = get_max_contour(mask) hull = cv2.convexHull(max_contour, returnPoints=False) draw_defects(original_frame, max_contour, hull) defects = cv2.convexityDefects(max_contour, hull) far = get_farthest_point(defects, max_contour, get_centroid(max_contour)) cv2.circle(original_frame, far, 10, [0, 200, 255], -1) if made_prediction: font = cv2.FONT_HERSHEY_SIMPLEX cv2.putText(original_frame, 'Character written : ' + chr(prediction + 65), (10, 500), font, 4, (255, 255, 255), 2, cv2.LINE_AA) # To quit the drawing if key & 0xFF == ord('q'): break cv2.imshow('frame', original_frame) video.release() cv2.destroyAllWindows() if __name__ == '__main__': main()
[ "priyanka@onlinedegree@iitm.ac.in" ]
priyanka@onlinedegree@iitm.ac.in
afbade6cf7187497d35c8ebf245bc2caf3732775
c0e41a88ff45bdb7bbdc8e4276427b276eec5d13
/kyo/python/1_function/1_args.py
a93d7b955d358025108b2595479cdd21473a90fb
[]
no_license
sunzhqiiang888/KingSun
7440748825ed72c411f8ff63c478ad412efa51ac
7473cc30d398f1ae96cbc7175198052d518fef94
refs/heads/master
2020-03-08T08:17:17.284210
2018-04-15T03:15:25
2018-04-15T03:15:25
null
0
0
null
null
null
null
UTF-8
Python
false
false
858
py
#!/usr/bin/env python3 def test(name='未知', age=17, sex='男', phone="13602578723"): print("学生信息: %s(%d) %s %s" % (name, age, sex, phone)) def testV(name, *args, phone="13877777777", **kwargs): # print(args, type(args), kwargs, type(kwargs)) print(name, args, phone, kwargs) if __name__ == "__main__": def testArgs(): test() test("李四") test("张三", 23) test("王二", 17, '女') test("马六", 19, '未知', '13877669900') test("李小四", phone='110') test(phone='119') def testVargs(): # testV() testV("李四") testV("张三", 23, en=34, cn=56) testV("王二", 17, '女') testV("马六", 19, '未知', '13877669900') testV("李小四", phone='110') testV("王五", phone='119') testVargs()
[ "354340684@qq.com" ]
354340684@qq.com
79397345c631e26cb88ef3385cd94d8a3e22ec70
678b5e4378d809f5048a9284de1f845a3c4540f6
/src/scenes/cards/Card.py
266f093ce35443c5e4d2b72124060ef11cd7b63b
[ "MIT" ]
permissive
codingblocks/get-outta-here
2fccc4c97104078fdd70c2a9b5ea8cfbaf6a7706
8d37047876a93a050f0116086a470df123421cb9
refs/heads/main
2023-02-23T14:52:04.986791
2021-01-24T22:32:08
2021-01-24T22:32:08
332,569,558
0
1
MIT
2021-03-25T06:33:31
2021-01-24T22:24:59
Python
UTF-8
Python
false
false
6,676
py
import pygame from src.config import (CARD_FRONT_BACKGROUND_IMAGE, CHARACTER_SPRITE_PATH, MAP_TILE_HEIGHT, MAP_TILE_WIDTH, MAIN_FONT_FILE, MAP_TILE_SCALE, CARD_TEXT_SIZE, CARD_TEXT_COLOR, CARD_IMAGE_SIZE, CARD_TITLE_SIZE, CARD_TITLE_IMAGE, CARD_TEXT_IMAGE, CARD_TYPE_IMAGE, CARD_TYPE_SIZE, CARD_TYPE_COLOR, PERSONNEL_CARD_DIR, EFFECT_CARD_DIR, SPECIAL_CARD_DIR) from src.scenes.text_writer import draw_text import src.scenes.globals as g class Card(pygame.sprite.Sprite): character_sheet = None def __init__(self, config: dict): if self.character_sheet is None: self.character_sheet = pygame.image.load(CHARACTER_SPRITE_PATH).convert_alpha() pygame.sprite.Sprite.__init__(self) self.config = config if self.config['type'].lower() == "special": self.card_dir = f"{SPECIAL_CARD_DIR}" elif self.config['type'].lower() == "effect": self.card_dir = f"{EFFECT_CARD_DIR}" else: self.card_dir = f"{PERSONNEL_CARD_DIR}" card_background_path = f"{self.card_dir}\\{CARD_FRONT_BACKGROUND_IMAGE}" print(card_background_path) card_background = pygame.image.load(card_background_path).convert_alpha() self._draw_portrait(card_background) self._draw_title(card_background) self._draw_text(card_background) self._draw_type(card_background) self.image = card_background self.rect = self.image.get_rect() self.rect.x = -1000 self.rect.y = -1000 def update(self, position: tuple): self.rect = self.image.get_rect() self.rect.x = position[0] self.rect.y = position[1] def can_be_played(self): can_be_played = True modifiers = self.config.get("modifiers", {}) if modifiers.get("energy", 0) and modifiers["energy"] + g.resources.energy <= 0: can_be_played = False if modifiers.get("fuel", 0) and modifiers["fuel"] + g.resources.fuel <= 0: can_be_played = False if modifiers.get("shields", 0) and modifiers["shields"] + g.resources.shields <= 0: can_be_played = False return can_be_played def play(self): modifiers = self.config.get("modifiers", {}) for k,v in modifiers.items(): g.resources.modify(k, v) def is_single_use(self): return self.config['type'].lower() == "personnel" def _draw_portrait(self, card_background): if "sheet" in self.config['sprite']: self._draw_item_portrait(card_background) else: self._draw_character_portrait(card_background) def _draw_item_portrait(self, card_background): config = self.config sheet = pygame.image.load(config['sprite']['sheet']).convert_alpha() tile_size = self.config['sprite']['sheet_tile_size'] sprite_position_on_sheet = (config['sprite']['x'], config['sprite']['y']) sprite_rect = pygame.Rect(sprite_position_on_sheet[0] * tile_size, sprite_position_on_sheet[1] * tile_size, tile_size, tile_size) portrait_image = pygame.Surface((tile_size, tile_size), pygame.SRCALPHA) portrait_image.blit(sheet, (0, 0), sprite_rect) scaled_portrait = pygame.transform.scale(portrait_image, (MAP_TILE_WIDTH * 3, MAP_TILE_HEIGHT * 3)) card_center = card_background.get_rect().center card_center = (card_center[0] - scaled_portrait.get_rect().w // 2, card_center[1] - 190) card_background.blit(scaled_portrait, card_center) pass def _draw_character_portrait(self, card_background): config = self.config portrait_position = (config['sprite']['x'], config['sprite']['y']) blank_image = pygame.Surface((MAP_TILE_WIDTH // MAP_TILE_SCALE, MAP_TILE_HEIGHT // MAP_TILE_SCALE), pygame.SRCALPHA) sprite_rect = pygame.Rect(portrait_position[0] * MAP_TILE_WIDTH // MAP_TILE_SCALE, portrait_position[1] * MAP_TILE_HEIGHT // MAP_TILE_SCALE, MAP_TILE_WIDTH // MAP_TILE_SCALE, MAP_TILE_HEIGHT // MAP_TILE_SCALE) blank_image.blit(self.character_sheet, (0, 0), sprite_rect) scaled_portrait = pygame.transform.scale(blank_image, (MAP_TILE_WIDTH * 3, MAP_TILE_HEIGHT * 3)) portrait_rect = scaled_portrait.get_rect() card_center = card_background.get_rect().center card_center = (card_center[0] - scaled_portrait.get_rect().w // 2, card_center[1] - 190) card_background.blit(scaled_portrait, card_center, portrait_rect) def _draw_title(self, card_background): plate_image = pygame.image.load(f"{self.card_dir}\\{CARD_TITLE_IMAGE}").convert_alpha() font = pygame.font.Font(MAIN_FONT_FILE, CARD_TITLE_SIZE) font_render = font.render(self.config['title'], True, CARD_TEXT_COLOR) font_render.get_rect().center = card_background.get_rect().center font_left = plate_image.get_rect().center[0] - font_render.get_rect().w // 2 font_top = plate_image.get_rect().center[1] - font_render.get_rect().h // 2 font_pos = (font_left, font_top - 8) plate_image.blit(font_render, font_pos) card_background.blit(plate_image, (31, 214)) def _draw_type(self, card_background): plate_image = pygame.image.load(f"{self.card_dir}\\{CARD_TYPE_IMAGE}").convert_alpha() font = pygame.font.Font(MAIN_FONT_FILE, CARD_TYPE_SIZE) font_render = font.render(self.config['type'], True, CARD_TYPE_COLOR) font_render.get_rect().center = card_background.get_rect().center font_left = plate_image.get_rect().center[0] - font_render.get_rect().w // 2 font_top = plate_image.get_rect().center[1] - font_render.get_rect().h // 2 font_pos = (font_left, font_top) plate_image.blit(font_render, font_pos) card_background.blit(plate_image, (73, 259)) def _draw_text(self, card_background): plate_image = pygame.image.load(f"{self.card_dir}\\{CARD_TEXT_IMAGE}").convert_alpha() font = pygame.font.Font(MAIN_FONT_FILE, CARD_TEXT_SIZE) card_center = card_background.get_rect().center cr = card_background.get_rect() left_buffer = 70 text_area = pygame.Rect(cr.x + left_buffer, card_center[1] + 80, CARD_IMAGE_SIZE[0] - left_buffer * 2, 200) card_background.blit(plate_image, (25, 276)) draw_text(card_background, self.config['text'], CARD_TEXT_COLOR, text_area, font) # Card title, cost
[ "me@joezack.com" ]
me@joezack.com
72e036decfd9852b8b0e9d8170a5cb7d4a854ab6
efd5d310c1a43335a70fbcb89a2480aa7aa1f423
/goto.py
f3a86a748de3a5238c0a2cc5d452c397e6c070cc
[ "Unlicense" ]
permissive
tianjinghai1978/python-goto
aee690c59ef4e5fecedd742a7db8a44e76de5f52
acbe736221d2238df3d09beab457d0bb19d05812
refs/heads/master
2023-03-19T08:27:22.300630
2019-12-15T12:48:22
2019-12-15T13:21:51
null
0
0
null
null
null
null
UTF-8
Python
false
false
7,676
py
import dis import struct import array import types import functools try: _array_to_bytes = array.array.tobytes except AttributeError: _array_to_bytes = array.array.tostring class _Bytecode: def __init__(self): code = (lambda: x if x else y).__code__.co_code opcode, oparg = struct.unpack_from('BB', code, 2) # Starting with Python 3.6, the bytecode format has changed, using # 16-bit words (8-bit opcode + 8-bit argument) for each instruction, # as opposed to previously 24 bit (8-bit opcode + 16-bit argument) # for instructions that expect an argument and otherwise 8 bit. # https://bugs.python.org/issue26647 if dis.opname[opcode] == 'POP_JUMP_IF_FALSE': self.argument = struct.Struct('B') self.have_argument = 0 # As of Python 3.6, jump targets are still addressed by their # byte unit. This is matter to change, so that jump targets, # in the future might refer to code units (address in bytes / 2). # https://bugs.python.org/issue26647 self.jump_unit = 8 // oparg else: self.argument = struct.Struct('<H') self.have_argument = dis.HAVE_ARGUMENT self.jump_unit = 1 @property def argument_bits(self): return self.argument.size * 8 _BYTECODE = _Bytecode() def _make_code(code, codestring): args = [ code.co_argcount, code.co_nlocals, code.co_stacksize, code.co_flags, codestring, code.co_consts, code.co_names, code.co_varnames, code.co_filename, code.co_name, code.co_firstlineno, code.co_lnotab, code.co_freevars, code.co_cellvars ] try: args.insert(1, code.co_kwonlyargcount) # PY3 except AttributeError: pass return types.CodeType(*args) def _parse_instructions(code): extended_arg = 0 extended_arg_offset = None pos = 0 while pos < len(code): offset = pos if extended_arg_offset is not None: offset = extended_arg_offset opcode = struct.unpack_from('B', code, pos)[0] pos += 1 oparg = None if opcode >= _BYTECODE.have_argument: oparg = extended_arg | _BYTECODE.argument.unpack_from(code, pos)[0] pos += _BYTECODE.argument.size if opcode == dis.EXTENDED_ARG: extended_arg = oparg << _BYTECODE.argument_bits extended_arg_offset = offset continue extended_arg = 0 extended_arg_offset = None yield (dis.opname[opcode], oparg, offset) def _get_instruction_size(opname, oparg=0): size = 1 extended_arg = oparg >> _BYTECODE.argument_bits if extended_arg != 0: size += _get_instruction_size('EXTENDED_ARG', extended_arg) oparg &= (1 << _BYTECODE.argument_bits) - 1 opcode = dis.opmap[opname] if opcode >= _BYTECODE.have_argument: size += _BYTECODE.argument.size return size def _get_instructions_size(ops): size = 0 for op in ops: if isinstance(op, str): size += _get_instruction_size(op) else: size += _get_instruction_size(*op) return size def _write_instruction(buf, pos, opname, oparg=0): extended_arg = oparg >> _BYTECODE.argument_bits if extended_arg != 0: pos = _write_instruction(buf, pos, 'EXTENDED_ARG', extended_arg) oparg &= (1 << _BYTECODE.argument_bits) - 1 opcode = dis.opmap[opname] buf[pos] = opcode pos += 1 if opcode >= _BYTECODE.have_argument: _BYTECODE.argument.pack_into(buf, pos, oparg) pos += _BYTECODE.argument.size return pos def _write_instructions(buf, pos, ops): for op in ops: if isinstance(op, str): pos = _write_instruction(buf, pos, op) else: pos = _write_instruction(buf, pos, *op) return pos def _find_labels_and_gotos(code): labels = {} gotos = [] block_stack = [] block_counter = 0 opname1 = oparg1 = offset1 = None opname2 = oparg2 = offset2 = None opname3 = oparg3 = offset3 = None for opname4, oparg4, offset4 in _parse_instructions(code.co_code): if opname1 in ('LOAD_GLOBAL', 'LOAD_NAME'): if opname2 == 'LOAD_ATTR' and opname3 == 'POP_TOP': name = code.co_names[oparg1] if name == 'label': if oparg2 in labels: raise SyntaxError('Ambiguous label {0!r}'.format( code.co_names[oparg2] )) labels[oparg2] = (offset1, offset4, tuple(block_stack)) elif name == 'goto': gotos.append((offset1, offset4, oparg2, tuple(block_stack))) elif opname1 in ('SETUP_LOOP', 'SETUP_EXCEPT', 'SETUP_FINALLY', 'SETUP_WITH', 'SETUP_ASYNC_WITH'): block_counter += 1 block_stack.append(block_counter) elif opname1 == 'POP_BLOCK' and block_stack: block_stack.pop() opname1, oparg1, offset1 = opname2, oparg2, offset2 opname2, oparg2, offset2 = opname3, oparg3, offset3 opname3, oparg3, offset3 = opname4, oparg4, offset4 return labels, gotos def _inject_nop_sled(buf, pos, end): while pos < end: pos = _write_instruction(buf, pos, 'NOP') def _patch_code(code): labels, gotos = _find_labels_and_gotos(code) buf = array.array('B', code.co_code) for pos, end, _ in labels.values(): _inject_nop_sled(buf, pos, end) for pos, end, label, origin_stack in gotos: try: _, target, target_stack = labels[label] except KeyError: raise SyntaxError('Unknown label {0!r}'.format( code.co_names[label] )) target_depth = len(target_stack) if origin_stack[:target_depth] != target_stack: raise SyntaxError('Jump into different block') ops = [] for i in range(len(origin_stack) - target_depth): ops.append('POP_BLOCK') ops.append(('JUMP_ABSOLUTE', target // _BYTECODE.jump_unit)) if pos + _get_instructions_size(ops) > end: # not enough space, add code at buffer end and jump there buf_end = len(buf) go_to_end_ops = [('JUMP_ABSOLUTE', buf_end // _BYTECODE.jump_unit)] if pos + _get_instructions_size(go_to_end_ops) > end: # not sure if reachable raise SyntaxError('Goto in an incredibly huge function') pos = _write_instructions(buf, pos, go_to_end_ops) _inject_nop_sled(buf, pos, end) buf.extend([0] * _get_instructions_size(ops)) _write_instructions(buf, buf_end, ops) else: pos = _write_instructions(buf, pos, ops) _inject_nop_sled(buf, pos, end) return _make_code(code, _array_to_bytes(buf)) def with_goto(func_or_code): if isinstance(func_or_code, types.CodeType): return _patch_code(func_or_code) return functools.update_wrapper( types.FunctionType( _patch_code(func_or_code.__code__), func_or_code.__globals__, func_or_code.__name__, func_or_code.__defaults__, func_or_code.__closure__, ), func_or_code )
[ "sebastian.noack@gmail.com" ]
sebastian.noack@gmail.com
4d71975bc09e3c0a6e6ee256fd6840bf15111f68
cf0ab8503d4d704045070deea1e2125375711e86
/apps/sockets/tests/test_importer.py
d4c7afc7997e84b3c918c6e4dc2634f31084008f
[]
no_license
faierbol/syncano-platform
c3c6468600115752fd9fa5e46a0ad59f75f6bc9c
879111874d1ef70418b4890cf970720b0a2be4d8
refs/heads/master
2023-07-20T10:13:40.066127
2021-02-08T15:01:13
2021-02-08T15:01:13
null
0
0
null
null
null
null
UTF-8
Python
false
false
8,405
py
# coding=UTF8 from unittest import mock from django.test import TestCase from django.utils import timezone from apps.sockets.exceptions import ObjectProcessingError, SocketConfigValidationError, SocketMissingFile from apps.sockets.importer import INTERVAL_REGEX, SocketImporter from apps.sockets.models import Socket from apps.sockets.validators import CustomSocketConfigValidator @mock.patch('apps.sockets.signal_handlers.SocketProcessorTask', mock.MagicMock()) @mock.patch('apps.sockets.download_utils.ZipDownloadFileHandler.get_socket_spec') class TestSocketImporter(TestCase): importer_class = SocketImporter @mock.patch('apps.sockets.download_utils.ZipDownloadFileHandler.read_file', mock.Mock(side_effect=SocketMissingFile('error'))) def process_socket(self, download_mock, socket_source, **kwargs): socket = Socket(created_at=timezone.now(), **kwargs) download_mock.return_value = socket_source return socket, self.importer_class(socket).process() def assert_validation(self, download_mock, error_msg, socket_source, line=None): with self.assertRaisesMessage(ObjectProcessingError, error_msg) as cm: self.process_socket(download_mock, socket_source) if line is not None: self.assertEqual(cm.exception.lineno, line, 'Lines not equal for: "{}"; Expected: {}, got: {}.'.format(str(cm.exception), line, cm.exception.lineno)) def assert_validation_with_config(self, download_mock, error_msg, socket_source, config=None): with self.assertRaisesMessage(SocketConfigValidationError, error_msg): socket, _ = self.process_socket(download_mock, socket_source, config=config or {}) CustomSocketConfigValidator().validate(socket_config=socket.config, meta_config=socket.metadata.get('config') or {}) def test_serializer_validation(self, download_mock): self.assert_validation(download_mock, 'No calls defined', """ endpoints: my_endpoint_#1: script: script_endpoint_1 """, line=3) def test_basic_validation(self, download_mock): self.assert_validation(download_mock, 'Too many properties', '\n'.join(['name{}: name'.format(i) for i in range(self.importer_class.max_number_of_keys + 1)])) self.assert_validation(download_mock, 'Wrong format', '- wrong format') def test_endpoints_validation(self, download_mock): self.assert_validation(download_mock, 'No calls defined', """ endpoints: endpoint1: {} """, line=3) def test_cache_validation(self, download_mock): self.assert_validation(download_mock, 'Invalid cache value', """ endpoints: endpoint1: cache: 100000 source: | print 1 """, line=3) def test_timeout_validation(self, download_mock): self.assert_validation(download_mock, 'Invalid timeout value', """ endpoints: endpoint1: timeout: 100000 source: | print 1 """, line=3) def test_script_endpoints_format_validation(self, download_mock): self.assert_validation(download_mock, 'Wrong format', """ endpoints: - endpoint1 """, line=3) self.assert_validation(download_mock, 'Wrong format', """ endpoints: endpoint1: - script """, line=4) self.assert_validation(download_mock, 'Wrong format', """ endpoints: endpoint1: file: - script.py """, line=5) self.assert_validation(download_mock, 'Source file path contains invalid characters', """ endpoints: endpoint1: file: <script.py """, line=3) self.assert_validation(download_mock, 'Source file path is too long', """ endpoints: endpoint1: file: {} """.format('a' * 500), line=3) self.assert_validation(download_mock, 'Wrong format', """ endpoints: endpoint1: POST: - script """, line=5) def test_channel_endpoints_format_validation(self, download_mock): self.assert_validation(download_mock, 'Wrong format', """ endpoints: endpoint1: channel: - script """, line=5) self.assert_validation(download_mock, 'Wrong format', """ endpoints: endpoint1: channel: something.{a!bc}.{user} """, line=4) self.process_socket(download_mock, """ endpoints: endpoint1: channel: something.{ABC}.{user} """) self.process_socket(download_mock, """ endpoints: endpoint1: | channels.publish("a") """) def test_config_validation(self, download_mock): self.assert_validation_with_config( download_mock, 'Error validating socket config. "user_key" is required.', """ config: secret_key: value: some value user_key: required: true value: some value """) for socket_yml in ( """ config: key: null """, """ config: - value """): self.assert_validation_with_config( download_mock, 'Error validating socket config. Wrong format.', socket_yml) def test_event_handlers_validation(self, download_mock): self.assert_validation(download_mock, 'Wrong format', """ event_handlers: - eh """, line=3) self.assert_validation(download_mock, 'Wrong format', """ event_handlers: data.user.create: - src """, line=4) self.assert_validation(download_mock, 'Unsupported event handler type', """ event_handlers: something.bla.bla: | print 1 """, line=3) def test_data_event_handlers_validation(self, download_mock): self.assert_validation(download_mock, 'Wrong format for data event handler', """ event_handlers: data.usercreate: | print 1 """, line=3) def test_schedule_event_handlers_validation(self, download_mock): self.assert_validation(download_mock, 'Wrong format for schedule event handler', """ event_handlers: schedule.interval#5_minutes: | print 1 """, line=3) self.assert_validation(download_mock, 'Wrong format for schedule interval', """ event_handlers: schedule.interval.5_zonks: | print 1 """, line=3) self.assert_validation(download_mock, 'Wrong type of schedule event handler', """ event_handlers: schedule.intercal.5_minutes: | print 1 """, line=3) def test_custom_event_handlers_validation(self, download_mock): self.assert_validation(download_mock, 'Wrong format for event handler', """ event_handlers: events: | print 1 """, line=3) self.assert_validation(download_mock, 'Wrong format for event handler', """ event_handlers: events.socket1.event2.suffix: | print 1 """, line=3) class TestSocketEventHandler(TestCase): def calculate_interval(self, interval_str): match = INTERVAL_REGEX.match(interval_str) if not match: return None interval_dict = match.groupdict(0) return int(interval_dict['hours']) * 60 * 60 + int(interval_dict['minutes']) * 60 + \ int(interval_dict['seconds']) def test_schedule_interval_regex(self): for interval_str, value in ( ('5h', 5 * 60 * 60), ('5m', 5 * 60), ('5s', 5), ('5_hours_10_minutes_30_seconds', 5 * 60 * 60 + 10 * 60 + 30), ('1_hour_1_minute_1_second', 1 * 60 * 60 + 1 * 60 + 1), ('1h_2m_3s', 1 * 60 * 60 + 2 * 60 + 3), ('1h_2m_3s', 1 * 60 * 60 + 2 * 60 + 3), ('3s_2m', None), ('2m_1h', None), ('1_hor', None), ): self.assertEqual(self.calculate_interval(interval_str), value)
[ "rk@23doors.com" ]
rk@23doors.com
ec7a051e8f0312346a58fd5aba93972c0e87435a
3a517a9b62e24eccfa44baf8a93857f14e1042bc
/Infy Assigns/Virus.py
d38330ed11c02ce27a4527cce963040597b38385
[]
no_license
milindaj/codingground
d20831323603e040f3c1285f222a11eb9e653662
ee9035d539d33e673df54c681102307a41e93585
refs/heads/master
2016-09-06T00:46:05.624018
2015-01-22T02:29:19
2015-01-22T02:29:19
29,096,387
0
0
null
null
null
null
UTF-8
Python
false
false
2,737
py
# Hello World program in Python print "Hello World!\n" #input = "90 120 Infected,90 150 NotInfected,100 140 Infected,80 130 NotInfected#95 125,95 145,75 160" input = "80 120 Infected,70 145 Infected,90 100 Infected,80 150 NotInfected,80 80 NotInfected,100 120 NotInfected#120 148,75 148,60 90" split = input.split('#') catVals = split[0].split(",") #catVals = [int(cat) for cat in catVals] trvVals = split[1].split(",") #trvVals = [int(trv) for trv in trvVals] #inpArray = dict() inTempList = [] #infected temp inPrsList = [] # infected pressure unTempList = [] # uninfected temp unPrsList = [] # uninfected pressure for cat in catVals: catTmp = cat.split(" ") if catTmp[2] == "Infected": inTempList.append(int(catTmp[0])) inPrsList.append(int(catTmp[1])) else: unTempList.append(int(catTmp[0])) unPrsList.append(int(catTmp[1])) inTempList.sort() inPrsList.sort() unTempList.sort() unPrsList.sort() tmpRangeType = "" prsRangeType = "" rangeType = "" if inTempList[0] < unTempList[0]: tmpRangeType = "Infected" elif inTempList[0] > unTempList[0]: tmpRangeType = "NotInfected" if inPrsList[len(inPrsList)-1] > unPrsList[len(unPrsList)-1]: prsRangeType = "Infected" elif inPrsList[len(inPrsList)-1] < unPrsList[len(unPrsList)-1]: prsRangeType = "NotInfected" if tmpRangeType == prsRangeType: rangeType = prsRangeType else: rangeType = "Unknown" print( "range type is :" + rangeType) ansList = [] for trv in trvVals: trvTmp = trv.split(" ") tmp = int(trvTmp[0]) prs = int(trvTmp[1]) if (tmp >= inTempList[0] and tmp <= inTempList[len(inTempList)-1]) and (prs >= inPrsList[0] and prs <= inPrsList[len(inPrsList)-1]): ansList.append("Infected") elif (tmp >= unTempList[0] and tmp <= unTempList[len(unTempList)-1]) and (prs >= unPrsList[0] and prs <= unPrsList[len(unPrsList)-1]): ansList.append("Notinfected") else: if rangeType != "Unknown": if rangeType == "Infected": if tmp < inTempList[0] and prs > inPrsList[len(inTempList)-1]: ansList.append("Infected") else: ansList.append("Unknown") else: if tmp < unTempList[0] and prs > unPrsList[len(unTempList)-1]: ansList.append("Notinfected") else: ansList.append("Unknown") else: ansList.append("Unknown") print(catVals) print(trvVals) print("\n") print("infected List :") print(inTempList) print(inPrsList) print("\n") print("Not infected List :") print(unTempList) print(unPrsList) print("\n") print(ansList)
[ "milindaj@gmail.com" ]
milindaj@gmail.com
6a4054dd06199db6b9309cb200c08c31c28b270c
4d861accd86c2c4fbffeb75e963aa4c099d6dd0a
/Practica5/EjercicioB.py
6803cb510d11ca5901c54ad48f76ddf90b878768
[]
no_license
adwolfart/ModeladoyP
98e7118b600815587ae0840134f84eaab2475a14
7bab5fb3a837c14d80077641dc9be91f325b5d9a
refs/heads/master
2020-07-12T08:08:34.563441
2019-11-15T20:42:17
2019-11-15T20:42:17
204,761,474
0
0
null
null
null
null
UTF-8
Python
false
false
10,988
py
import threading from threading import Thread,Semaphore import time #semaforo para manipular los problemas de concurrencia semaforo = Semaphore(1); #Crear variable semáforo #listas que son utilizadas como bandas transportadoras listaPiezas = [] listaPiezasEnsamblador = [] listaPiezasEnsamblador1 = [] listaPaqueteFinalizados = [] mataHilos = True #se usa para verificar si se esta ejecutando un hilo en algun momento entro = False #cuenta las piezas que se van creando contadorPiezas = 0 #clase que modela los paquetes class Paquete(object): def __init__(self, tipo, listaPiezasPaquete): self.tipo = tipo self.listaPiezasPaquete = listaPiezasPaquete milisegundos = int(round(time.time() * 1000)) self.identificador = tipo+str(milisegundos) def getTipo(self): return self.tipo def getListaPiezasPaquete(self): return self.listaPiezasPaquete def getIdentificador(self): return self.identificador #clase que modela las piezas class Pieza(object): def __init__(self, tipo): self.tipo = tipo milisegundos = int(round(time.time() * 1000)) self.identificador = tipo+str(milisegundos) def getTipoPieza(self): return self.tipo def getIdentificador(self): return self.identificador #clase que modela al productor de piezas. se puede crear varias piezas y dependiendo de la pieza la guarda en el #lugar correspondiente class Productor(threading.Thread): def __init__(self, tipoPieza, tiempo): self.tiempo = tiempo self.tipoPieza = tipoPieza threading.Thread.__init__(self) def crearPieza(self): nuevaPieza = Pieza(self.tipoPieza) if revisaPiezasCreadas(): global mataHilos mataHilos = False time.sleep(self.tiempo) return nuevaPieza #funcion que guarda la pieza en la banda correspondiente depende de la pieza que es creada def run (self): global contadorPiezas while mataHilos: pieza = self.crearPieza() contadorPiezas = contadorPiezas + 1 if pieza.getTipoPieza() == "C": listaPiezasEnsamblador.append(pieza) elif pieza.getTipoPieza() == "D": if entro: semaforo.acquire(); listaPiezasEnsamblador1.append(pieza) else: listaPiezas.append(pieza) print("Se crea la pieza: "+str(contadorPiezas)+ " con id: "+pieza.getIdentificador()) def getTiempo(self): return self.tiempo #clase que modela a Ensamblador class Ensamblador(threading.Thread): def __init__(self, tipoPieza, tiempo, listaCantTipo): self.tipoPieza = tipoPieza self.tiempo = tiempo self.listaCantTipo = listaCantTipo self.listaPiezasRecolectadas = [] threading.Thread.__init__(self) #funcion que crea piezas cada cierto tiempo. crear las piezas si en la banda hay las piezas correspondientes def crearPiezaEnsamblador(self, i): global contadorPiezas cantidadTotalPiezas = 0 for np in self.listaCantTipo: cantidadTotalPiezas = cantidadTotalPiezas + np[1] tupla = self.listaCantTipo[i] tipoPiezaS = tupla[0] cantidadPiezasS = tupla[1] j = 0 while j < len(listaPiezas): piezaP = listaPiezas[j] if tipoPiezaS == piezaP.getTipoPieza(): cuantasPiezas = 0 for k in self.listaPiezasRecolectadas: if k.getTipoPieza() == piezaP.getTipoPieza(): cuantasPiezas = cuantasPiezas + 1 if cuantasPiezas < cantidadPiezasS: self.listaPiezasRecolectadas.append(piezaP) del listaPiezas[j] break j = j + 1 if len(self.listaPiezasRecolectadas)==cantidadTotalPiezas: nuevaPieza = Pieza(self.tipoPieza) listaPiezasEnsamblador.append(nuevaPieza) contadorPiezas = contadorPiezas + 1 print("Se crea la pieza: "+str(contadorPiezas)+" con id: "+nuevaPieza.getIdentificador()) del self.listaPiezasRecolectadas[0:len(self.listaPiezasRecolectadas)] if revisaPiezasCreadas(): global mataHilos mataHilos = False time.sleep(self.tiempo) #funcion que se ejecuta hasta que se cunple la condicion def run (self): i = 0 while mataHilos: if i == len(self.listaCantTipo): i = 0 self.crearPiezaEnsamblador(i) i = i + 1 def getTiempo(self): return self.tiempo #clase que modela el segundo Ensamblador que hay en la mecanismo class Ensamblador1(threading.Thread): def __init__(self, tipoPieza, tiempo, listaCantTipo): self.tipoPieza = tipoPieza self.tiempo = tiempo self.listaCantTipo = listaCantTipo self.listaPiezasRecolectadas = [] threading.Thread.__init__(self) #funcion que crea piezas cada cierto tiempo. crear las piezas si en la banda hay las piezas correspondientes def crearPiezaEnsamblador(self, i): global contadorPiezas cantidadTotalPiezas = 0 for np in self.listaCantTipo: cantidadTotalPiezas = cantidadTotalPiezas + np[1] tupla = self.listaCantTipo[i] tipoPiezaS = tupla[0] cantidadPiezasS = tupla[1] j = 0 while j < len(listaPiezasEnsamblador): piezaP = listaPiezasEnsamblador[j] if tipoPiezaS == piezaP.getTipoPieza(): cuantasPiezas = 0 for k in self.listaPiezasRecolectadas: if k.getTipoPieza() == piezaP.getTipoPieza(): cuantasPiezas = cuantasPiezas + 1 if cuantasPiezas < cantidadPiezasS: self.listaPiezasRecolectadas.append(piezaP) del listaPiezasEnsamblador[j] break j = j + 1 if len(self.listaPiezasRecolectadas)==cantidadTotalPiezas: nuevaPieza = Pieza(self.tipoPieza) if entro: semaforo.acquire() listaPiezasEnsamblador1.append(nuevaPieza) contadorPiezas = contadorPiezas + 1 print("Se crea la pieza: "+str(contadorPiezas)+ " con id: "+nuevaPieza.getIdentificador()) del self.listaPiezasRecolectadas[0:len(self.listaPiezasRecolectadas)] if revisaPiezasCreadas(): global mataHilos mataHilos = False time.sleep(self.tiempo) #funcion que se ejecuta hasta que se cunple la condicion def run (self): i = 0 while mataHilos: if i == len(self.listaCantTipo): i = 0 self.crearPiezaEnsamblador(i) i = i + 1 def getTiempo(self): return self.tiempo #clase que modela a un empaquetador. class Empaquetador(threading.Thread): def __init__(self, tipoPieza, tipoPiezaEntrada, tiempo, tamPaquete): self.tiempo = tiempo self.tipoPieza = tipoPieza self.tamPaquete = tamPaquete self.listaPiezasPaquete = [] self.tipoPiezaEntrada = tipoPiezaEntrada threading.Thread.__init__(self) #crea paquetes dependiendo de las especificaciones def crearPaquete(self): global entro i = 0 entro = True while i < len(listaPiezasEnsamblador1): piezaP = listaPiezasEnsamblador1[i] if piezaP.getTipoPieza() == self.tipoPiezaEntrada: self.listaPiezasPaquete.append(piezaP) del listaPiezasEnsamblador1[i] if len(self.listaPiezasPaquete) == self.tamPaquete: nuevoPaquete = Paquete(self.tipoPieza, self.listaPiezasPaquete) print("Se creo el paquete:"+str(nuevoPaquete.getIdentificador())) listaPaqueteFinalizados.append(nuevoPaquete) del self.listaPiezasPaquete[0:len(self.listaPiezasPaquete)] if revisaPiezasCreadas(): global mataHilos mataHilos = False i = i + 1 entro = False semaforo.release() #funcion que se ejecuta hasta que se cumple la condicion de finalizar def run (self): while mataHilos: self.crearPaquete() #clase que modela a un empaquetador. class Empaquetador1(threading.Thread): def __init__(self, tipoPieza, tipoPiezaEntrada, tiempo, tamPaquete): self.tiempo = tiempo self.tipoPieza = tipoPieza self.tamPaquete = tamPaquete self.listaPiezasPaquete = [] self.tipoPiezaEntrada = tipoPiezaEntrada threading.Thread.__init__(self) #crea paquetes dependiendo de las especificaciones def crearPaquete(self): global entro entro = True i = 0 while i < len(listaPiezasEnsamblador1): piezaP = listaPiezasEnsamblador1[i] if piezaP.getTipoPieza() == self.tipoPiezaEntrada: self.listaPiezasPaquete.append(piezaP) del listaPiezasEnsamblador1[i] if len(self.listaPiezasPaquete) == self.tamPaquete: nuevoPaquete = Paquete(self.tipoPieza, self.listaPiezasPaquete) print("Se creo el paquete:"+str(nuevoPaquete.getIdentificador())) listaPaqueteFinalizados.append(nuevoPaquete) del self.listaPiezasPaquete[0:len(self.listaPiezasPaquete)] if revisaPiezasCreadas(): global mataHilos mataHilos = False i = i + 1 semaforo.release() entro = False def run (self): while mataHilos: self.crearPaquete() #funcion que se ejecuta hasta que se cumple la condicion de finalizar def revisaPiezasCreadas(): if contadorPiezas >= 200: return True else: return False #funcion que crea el ejercicio B def ejercicioA(): #productor de tipos A,B,C y D. se crean cada 2 segundos p = Productor("A", 2) p1 = Productor("B", 2) p2 = Productor("C", 2) p3 = Productor("D", 2) #Ensambladores de tipo E y F, se crean paquetes cada 2 segundos, las listas representan los tipos de piezas #y la cantidad que necesita ensamblador1 = Ensamblador("E", 2, [("A", 2), ("B", 2)]) ensamblador2 = Ensamblador1("F", 2, [("C", 2), ("E", 1)]) #Empaquetador que crea paquetes de tipo P y Q, necesita piezas de tipo F y D, los crea cada 0 sengundos y #para crear un paquete necesita 5 y 10 piezas empaquetador1 = Empaquetador("P", "F", 0, 5) empaquetador2 = Empaquetador1("Q", "D", 0, 10) #inicia los hilos p.start() p1.start() p2.start() p3.start() ensamblador1.start() ensamblador2.start() empaquetador1.start() empaquetador2.start() ejercicioA()
[ "lopo@lopo-pc" ]
lopo@lopo-pc
0f3ce92a2ff9742a1df0452ef3c71ce7e361bd2b
f8ad6963bfc851657ea50c6a036cfad29cdd7f60
/Books/LearningTensorFlow/Chapter5_Text_Sequence_Tensorboard/scan_example.py
4cf7e1f4fa42316220ed1621d22dc6ddfdcbd77a
[]
no_license
foru120/PythonRepository
e1ab0265c0f50ef2e9acdf7447237c913560692b
db6b6be0f9fb91b0a81a3b6a2ec5631daab10f98
refs/heads/master
2021-01-01T06:53:11.728109
2019-04-25T13:52:50
2019-04-25T13:52:50
97,541,222
4
0
null
null
null
null
UTF-8
Python
false
false
234
py
import numpy as np import tensorflow as tf elems = np.array(['T', 'e', 'n', 's', 'o', 'r', ' ', 'F', 'l', 'o', 'w']) scan_sum = tf.scan(lambda a, x: a + x, elems) sess = tf.InteractiveSession() print(sess.run(scan_sum)) sess.close()
[ "broodsky1122@hanmail.net" ]
broodsky1122@hanmail.net
d667c4b00b34a3dea6dc5a59897ad12547204d3a
62f723af2777f987ed25ff4026aa2ad1ef9e5ce7
/model/modeling.py
8ca4d7e742ceae0d99665ca8af9b33e4998a5447
[]
no_license
renjie-liang/Flatten_Net
0c311b7625323c848d3aac72a14d97647b8ca6fe
fe8ff7932cf05fb36a8bfb674d8061e52a6027ab
refs/heads/master
2022-12-21T08:02:40.841405
2019-10-06T19:09:42
2019-10-06T19:09:42
211,687,528
1
0
null
null
null
null
UTF-8
Python
false
false
383
py
import torch import torch.nn as nn from model.net import * from lib.config import cfg class get_Model(nn.Module): def __init__(self): super(get_Model, self).__init__() self.net = eval(cfg.MODEL.NET_NAME)() def forward(self, x): x = self.net(x) return x def to_stirng(self): return '{}'.format(str(self.net)) # BASE_Decoder
[ "allen_rj@163.com" ]
allen_rj@163.com
5639c8a9787426db742cd1e8e2e6518dd45c03b3
6c83083e1ddff6028a004b6beb40921d3d723a29
/gofish2.py
0fb1af0b9d48b1d623a3735d78c095dd193e02cc
[]
no_license
pfigz/python-files-projects
62255d38ace41150df8c26831be3f631f713a802
6ff153dece345ed5d93de4c46df8f0158e634861
refs/heads/main
2023-07-17T08:27:47.491472
2021-09-09T14:19:37
2021-09-09T14:19:37
404,751,407
0
0
null
null
null
null
UTF-8
Python
false
false
3,251
py
#Go fish example 2 import random def MakeDeck(): deck = [] c = 0 values = ["Ace", "2", "3", "4", "5", "6", "7", "8", "9", "10", "Jack", "Queen", "King"] suits = ["Hearts", "Spades", "Diamonds", "Clubs"] for v in values: for s in suits: deck.append([v,s]) random.shuffle(deck) return deck import random, time fish_deck = deck.MakeDeck() #for i in fish_deck: # print(i[0]+" of "+i[1]) class fisherman(): name = "" hand = [] sets = [] def ask(player1, player2): pause = random.randint(2,5) has = [] choose = randint(0,len(player1.hand)-1) value = player1.hand[choose][0] for card in player2.hand: if card[0] == value: has.append(card) for card in has: player2.hand.remove(card) for card in has: player1.hand.append(card) return_string = player1.name+" asked "+player2.name+" for "+value+"s. " print(return_string) return_string = player2.name+" had "+str(len(has))+". " print(return_string) if len(has) == 0: draw(player1) return_string = player1.name+" had to go fish." print(return_string) def draw(player): card = fish_deck.pop() player.hand.append(card) def set_check(player): amount = {} for card in player.hand: if card[0] not in amount.keys(): amount[card[0]] = 1 if card[0] in amount.keys(): amount[card[0]] += 1 for count in amount.keys(): if amount[count] == 4: print(player.name+" got a set of "+count+"s.") player.sets.append(count) player.hand[:] = [card for card in player.hand if card[0] == count] john = fisherman() john.name = "John" tim = fisherman() tim.name = "Tim" sara = fisherman() sara.name = "Sara" kris = fisherman() kris.name = "Kris" def play(player1, player2, player3, player4, deck): turn = 0 size = 7 dealt = 0 order = [player1, player2, player3, player4] random.shuffle(order) while dealt < size: draw(order[0]) draw(order[1]) draw(order[2]) draw(order[3]) dealt += 1 while len(deck) != 0: for player in order: count = 0 hand = player.name+"'s hand: " for card in player.hand: if count < len(player.hand)-1: hand += card[0]+" of "+card[1]+", " count += 1 elif count == len(player.hand)-1: hand += card[0]+" of "+card[1]+"." print(hand) count = 0 sets = player.name+"'s sets: " for set in player.sets: if count < len(player.sets)-1: sets += set+"s, " elif count == len(player.sets)-1: sets += set+"s." print(sets) other_player = turn while other_player == turn: other_player = random.randint(0,3) ask(order[turn], order[other_player]) set_check(order[turn]) if turn >= 3: turn = 0 else: turn += 1 time.sleep(10) print("=========================================") play(john, tim, sara, kris, fish_deck)
[ "pfigdev@gmail.com" ]
pfigdev@gmail.com
3bdc87a55f1a198db75b03987f4bf990b86d2541
363245bca6658463411aba57e89f96bc6cb9434e
/tests/test_gamesession.py
3733e371c58ac5a2dadaa0606f0521c2b6e4e054
[ "MIT" ]
permissive
daleecoding/HQ-Trivia-Homework
75b5de2e955e08eb51b82c64cd9b56e9dbe85c79
c41e014458539bf9b3a0cc9fb360ececb8a029a3
refs/heads/master
2022-12-22T05:49:15.754811
2020-08-04T12:49:32
2020-08-04T12:49:32
302,655,081
0
0
null
null
null
null
UTF-8
Python
false
false
12,432
py
import asyncio import unittest from unittest.mock import AsyncMock from unittest.mock import MagicMock from unittest.mock import call from unittest.mock import patch from websockets.exceptions import ConnectionClosed import hqtrivia.config as config from hqtrivia.gamesession import GameSession from hqtrivia.messages import * from hqtrivia.question import Question from hqtrivia.player import Player class GameSessionTest(unittest.TestCase): def setUp(self): # Set to a shorter time for faster test self.original_round_duration = config.CONFIG_ROUND_DURATION config.CONFIG_ROUND_DURATION = min(2, config.CONFIG_ROUND_DURATION) def tearDown(self): config.CONFIG_ROUND_DURATION = self.original_round_duration def get_mocked_player(self, answer): """Helper method to get a mocked up player for use """ player = MagicMock() send_announcement = AsyncMock() player.attach_mock(send_announcement, 'send_announcement') send_question = AsyncMock() player.attach_mock(send_question, 'send_question') send_answers = AsyncMock() player.attach_mock(send_answers, 'send_answers') recv_answer = AsyncMock() player.attach_mock(recv_answer, 'recv_answer') recv_answer.return_value = answer return (player, send_announcement, send_question, send_answers, recv_answer) def test_run(self): """Tests that run() method calls execute_next_round() only once if it returns False. """ game = GameSession(0, []) game.abort_game = AsyncMock() game.execute_next_round = AsyncMock() # First time returns True. On second time returns False. game.execute_next_round.side_effect = [True, False] asyncio.run(game.run()) game.abort_game.assert_not_called() game.execute_next_round.assert_has_calls([call(), call()]) def test_run_when_exception_thrown(self): """Tests that run() method calls execute_next_round() only once if it returns False. """ self.assertRaises(Exception, GameSession.run) game = GameSession(0, []) game.abort_game = AsyncMock() game.execute_next_round = AsyncMock( return_value=True, side_effect=Exception) with self.assertRaises(Exception): asyncio.run(game.run()) game.abort_game.assert_called_once() game.execute_next_round.assert_called_once() def test_abort_game(self): """Tests that abort_game() method calls handle_eliminated_players() once """ player = MagicMock() send_announcement = AsyncMock() player.attach_mock(send_announcement, 'send_announcement') game = GameSession(0, [player]) game.handle_eliminated_players = MagicMock() asyncio.run(game.abort_game()) game.handle_eliminated_players.assert_called_once() send_announcement.assert_called_once_with( MESSAGE_NETWORK_ERROR_OCCURRED) @patch('hqtrivia.question.Question.generate', new_callable=AsyncMock) def test_execute_next_round_identifies_winner(self, generate): """Tests that execute_next_round() method correctly identifies winner from the game. """ question = Question('Question 1', ['A', 'B', 'C', 'D'], 'C') generate.return_value = question # Create player 1 with wrong answer, and player 2 with the right answer (player1, send_announcement1, send_question1, send_answers1, recv_answer1) = self.get_mocked_player('D') (player2, send_announcement2, send_question2, send_answers2, recv_answer2) = self.get_mocked_player('C') game = GameSession(0, [player1, player2]) result = asyncio.run(game.execute_next_round()) self.assertFalse( result, "execute_next_round() should return False since game ended with a winner.") self.assertEqual(len(game.players), 0, "No player should remain in the game") # Check that the question was sent to the players send_question1.assert_any_call(question) send_question2.assert_any_call(question) # Check that the count of the answers from each participant was sent send_answers1.assert_any_call(question, [0, 0, 1, 1]) send_answers2.assert_any_call(question, [0, 0, 1, 1]) # Check that these message were sent last send_announcement1.assert_called_with(MESSAGE_YOU_ARE_ELIMINATED) send_announcement2.assert_called_with(MESSAGE_YOU_ARE_THE_WINNER) @patch('hqtrivia.question.Question.generate', new_callable=AsyncMock) def test_execute_next_round_player_times_out(self, generate): """Tests that execute_next_round() eliminates player who did not respond within timeout. """ question = Question('Question 1', ['A', 'B', 'C', 'D'], 'C') generate.return_value = question # Create player 1 with the wrong answer, and player 2 will timeout (player1, send_announcement1, send_question1, send_answers1, recv_answer1) = self.get_mocked_player('D') async def over_sleep(): await asyncio.sleep(config.CONFIG_ROUND_DURATION+1) (player2, send_announcement2, send_question2, send_answers2, recv_answer2) = self.get_mocked_player(None) player2.attach_mock(over_sleep, 'recv_answer') game = GameSession(0, [player1, player2]) result = asyncio.run(game.execute_next_round()) self.assertFalse( result, "execute_next_round() should return False since game ended.") self.assertEqual(len(game.players), 0, "No player should remain in the game") # Check that the question was sent to the players # Check that the question was sent to the players send_question1.assert_any_call(question) send_question2.assert_any_call(question) # Check that the count of the answers from each participant was sent send_answers1.assert_any_call(question, [0, 0, 0, 1]) send_answers2.assert_any_call(question, [0, 0, 0, 1]) # Check that these message were sent last send_announcement1.assert_called_with(MESSAGE_YOU_ARE_ELIMINATED) send_announcement1.assert_called_with(MESSAGE_YOU_ARE_ELIMINATED) @patch('hqtrivia.question.Question.generate', new_callable=AsyncMock) def test_execute_next_round_player_recvMessage_exception(self, generate): """Tests that execute_next_round() eliminates player who did not respond within timeout. """ question = Question('Question 1', ['A', 'B', 'C', 'D'], 'C') generate.return_value = question # Create player 1 with wrong answer, and player 2 with the right answer (player1, send_announcement1, send_question1, send_answers1, recv_answer1) = self.get_mocked_player('D') (player2, send_announcement2, send_question2, send_answers2, recv_answer2) = self.get_mocked_player('C') recv_answer2 = AsyncMock(side_effect=Exception( 'General Error during recv_answer()')) player2.attach_mock(recv_answer2, 'recv_answer') game = GameSession(0, [player1, player2]) result = asyncio.run(game.execute_next_round()) self.assertFalse( result, "execute_next_round() should return False since game ended.") self.assertEqual(len(game.players), 0, "No player should remain in the game") # Check that the question was sent to the players send_question1.assert_any_call(question) send_question2.assert_any_call(question) # Check that the count of the answers from each participant was sent send_answers1.assert_any_call(question, [0, 0, 0, 1]) send_answers2.assert_any_call(question, [0, 0, 0, 1]) # Check that these message were sent last send_announcement1.assert_called_with(MESSAGE_YOU_ARE_ELIMINATED) send_announcement1.assert_called_with(MESSAGE_YOU_ARE_ELIMINATED) @patch('hqtrivia.question.Question.generate', new_callable=AsyncMock) def test_execute_next_round_move_to_next_round(self, generate): """Tests that execute_next_round() method has to continue to the next round. """ question = Question('Question 1', ['A', 'B', 'C', 'D'], 'C') generate.return_value = question # Both players gives the correct answer (player1, send_announcement1, send_question1, send_answers1, recv_answer1) = self.get_mocked_player('C') (player2, send_announcement2, send_question2, send_answers2, recv_answer2) = self.get_mocked_player('C') game = GameSession(0, [player1, player2]) result = asyncio.run(game.execute_next_round()) self.assertTrue( result, "execute_next_round() should return True since no single winner yet.") self.assertEqual(len(game.players), 2, "Two players should still be in the game") # Check that the question was sent to the players send_question1.assert_any_call(question) send_question2.assert_any_call(question) # Check that the count of the answers from each participant was sent send_answers1.assert_any_call(question, [0, 0, 2, 0]) send_answers2.assert_any_call(question, [0, 0, 2, 0]) # Check that these message were sent last send_announcement1.assert_called_with(MESSAGE_CORRECT_ANSWER) send_announcement1.assert_called_with(MESSAGE_CORRECT_ANSWER) @patch('hqtrivia.question.Question.generate', new_callable=AsyncMock) def test_execute_next_round_all_eliminated(self, generate): """Tests that execute_next_round() method eliminates everyone without winner """ question = Question('Question 1', ['A', 'B', 'C', 'D'], 'C') generate.return_value = question # Both players gives wrong answer (player1, send_announcement1, send_question1, send_answers1, recv_answer1) = self.get_mocked_player('A') (player2, send_announcement2, send_question2, send_answers2, recv_answer2) = self.get_mocked_player('B') game = GameSession(0, [player1, player2]) result = asyncio.run(game.execute_next_round()) self.assertFalse( result, "execute_next_round() should return False since everyone eliminated.") self.assertEqual(len(game.players), 0, "No one should remain in the game") # Check that the question was sent to the players send_question1.assert_any_call(question) send_question2.assert_any_call(question) # Check that the count of the answers from each participant was sent send_answers1.assert_any_call(question, [1, 1, 0, 0]) send_answers2.assert_any_call(question, [1, 1, 0, 0]) # Check that these message were sent last send_announcement1.assert_called_with(MESSAGE_YOU_ARE_ELIMINATED) send_announcement1.assert_called_with(MESSAGE_YOU_ARE_ELIMINATED) @patch('hqtrivia.question.Question.generate', new_callable=AsyncMock) def test_run_exception_question_generator(self, generate): """Tests that run() method eliminates everyone without winner """ question = Question('Question 1', ['A', 'B', 'C', 'D'], 'C') generate.return_value = question generate.side_effect = Exception("Network Error!!!") # Both players give wrong answer, but not important since Question.generate will throw exception (player1, send_announcement1, send_question1, send_answers1, recv_answer1) = self.get_mocked_player('A') (player2, send_announcement2, send_question2, send_answers2, recv_answer2) = self.get_mocked_player('B') game = GameSession(0, [player1, player2]) with self.assertRaises(Exception): asyncio.run(game.run()) self.assertEqual(len(game.players), 0, "No one should remain in the game") # Check that these message were sent last send_announcement1.assert_called_once_with( MESSAGE_NETWORK_ERROR_OCCURRED) send_announcement2.assert_called_once_with( MESSAGE_NETWORK_ERROR_OCCURRED)
[ "68974415+dongkyutoronto@users.noreply.github.com" ]
68974415+dongkyutoronto@users.noreply.github.com
8ffe18e059c8aa73c2d15b8bc652cbaba00d8265
2236b86215a6c6895fbb9bc10451497121f70c0a
/rate_prediction/apv25buffering/config.py
054bb4e5bd3c5ac10362fdc27da32e6ed7a72133
[]
no_license
perhansson/burst-trigger-toy
b754a5462e175e8098ccb08f1187b32697cf6679
6e49bee473e761c36abd8f73c7c9474a40b9e109
refs/heads/master
2021-01-19T22:02:02.686562
2015-10-28T20:42:10
2015-10-28T20:42:10
35,178,918
0
0
null
null
null
null
UTF-8
Python
false
false
47
py
debug = False batch = True Nperiods = 10000000
[ "per.hansson80@gmail.com" ]
per.hansson80@gmail.com
20635d8de0d79ef721fcfb9b23358878187fd7dd
e1032ba0bff93df82f6418f67edcf896dd134c68
/Hackerrank/sorted_sums.py
a4a2accb0fc5099a0a231c30429d05c63257d35f
[]
no_license
webdastur/ProgrammingProblems
2956e1700866eeb957ccc027e4e14bb2a9ec37a2
8677ecb4d3e1377e5b73351eee19fe68b6febd9c
refs/heads/master
2022-11-09T16:43:04.469734
2022-10-29T13:51:22
2022-10-29T13:51:22
183,794,167
0
0
null
null
null
null
UTF-8
Python
false
false
948
py
class SortedSums: def __init__(self, size): self.size = size self.arr = [0] * self.size def add(self, x, val): if x == 0: self.arr[0] += val return while self.size > x: self.arr[x] += val x += x & (-x) def rank(self, x): if 0 > x: return 0 res = self.arr[0] while 0 < x: res += self.arr[x] x &= x - 1 return res def sortedSum(a): pre = SortedSums(10 ** 6 + 1) post = SortedSums(10 ** 6 + 1) temp = total = ans = 0 n = len(a) for i in range(n): pos = pre.rank(a[i]) + 1 g = total - post.rank(a[i]) temp = (temp + pos * a[i] + g) % (10 ** 9 + 7) ans = (ans + temp) % (10 ** 9 + 7) total += a[i] pre.add(a[i], 1) post.add(a[i], a[i]) return ans if __name__ == '__main__': print(sortedSum([9, 5, 8]))
[ "webdasturuz@gmail.com" ]
webdasturuz@gmail.com
26f5feec50c17f59e09b58ec23e082f0893c8b39
98c62fa21899dc823cfa6ccdfb89ba5bf5836b1a
/fileupload/admin.py
25bf6c18faed3ecad700fb19bc7fea68a4c48939
[]
no_license
sagar-spkt/DjangoAWS
1ae3bab19303d0344973eb026972c8a7853a5df4
a7d032370f1697e18054740dc439d5f08087fd88
refs/heads/master
2021-08-16T21:46:39.690026
2018-12-16T17:29:35
2018-12-16T18:17:34
162,022,099
0
0
null
2021-06-10T21:03:34
2018-12-16T16:56:20
Python
UTF-8
Python
false
false
198
py
from django.contrib import admin from fileupload.models import PublicDocument, PrivateDocument # Register your models here. admin.site.register(PublicDocument) admin.site.register(PrivateDocument)
[ "spkt.sagar@gmailcom" ]
spkt.sagar@gmailcom
e9880252c3b8871fcba2e2f278da14e2c40131dc
9ae6ce54bf9a2a86201961fdbd5e7b0ec913ff56
/google/ads/googleads/v10/services/services/custom_conversion_goal_service/client.py
6a590a9ac28792be1d45d85fcb1db11c168b6b0e
[ "Apache-2.0" ]
permissive
GerhardusM/google-ads-python
73b275a06e5401e6b951a6cd99af98c247e34aa3
676ac5fcb5bec0d9b5897f4c950049dac5647555
refs/heads/master
2022-07-06T19:05:50.932553
2022-06-17T20:41:17
2022-06-17T20:41:17
207,535,443
0
0
Apache-2.0
2019-09-10T10:58:55
2019-09-10T10:58:55
null
UTF-8
Python
false
false
21,247
py
# -*- coding: utf-8 -*- # Copyright 2022 Google LLC # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # from collections import OrderedDict import os import re from typing import Dict, Optional, Sequence, Tuple, Type, Union import pkg_resources from google.api_core import client_options as client_options_lib from google.api_core import gapic_v1 from google.api_core import retry as retries from google.auth import credentials as ga_credentials # type: ignore from google.auth.transport import mtls # type: ignore from google.auth.transport.grpc import SslCredentials # type: ignore from google.auth.exceptions import MutualTLSChannelError # type: ignore from google.oauth2 import service_account # type: ignore try: OptionalRetry = Union[retries.Retry, gapic_v1.method._MethodDefault] except AttributeError: # pragma: NO COVER OptionalRetry = Union[retries.Retry, object] # type: ignore from google.ads.googleads.v10.services.types import ( custom_conversion_goal_service, ) from .transports.base import ( CustomConversionGoalServiceTransport, DEFAULT_CLIENT_INFO, ) from .transports.grpc import CustomConversionGoalServiceGrpcTransport class CustomConversionGoalServiceClientMeta(type): """Metaclass for the CustomConversionGoalService client. This provides class-level methods for building and retrieving support objects (e.g. transport) without polluting the client instance objects. """ _transport_registry = ( OrderedDict() ) # type: Dict[str, Type[CustomConversionGoalServiceTransport]] _transport_registry["grpc"] = CustomConversionGoalServiceGrpcTransport def get_transport_class( cls, label: str = None, ) -> Type[CustomConversionGoalServiceTransport]: """Returns an appropriate transport class. Args: label: The name of the desired transport. If none is provided, then the first transport in the registry is used. Returns: The transport class to use. """ # If a specific transport is requested, return that one. if label: return cls._transport_registry[label] # No transport is requested; return the default (that is, the first one # in the dictionary). return next(iter(cls._transport_registry.values())) class CustomConversionGoalServiceClient( metaclass=CustomConversionGoalServiceClientMeta ): """Service to manage custom conversion goal.""" @staticmethod def _get_default_mtls_endpoint(api_endpoint): """Converts api endpoint to mTLS endpoint. Convert "*.sandbox.googleapis.com" and "*.googleapis.com" to "*.mtls.sandbox.googleapis.com" and "*.mtls.googleapis.com" respectively. Args: api_endpoint (Optional[str]): the api endpoint to convert. Returns: str: converted mTLS api endpoint. """ if not api_endpoint: return api_endpoint mtls_endpoint_re = re.compile( r"(?P<name>[^.]+)(?P<mtls>\.mtls)?(?P<sandbox>\.sandbox)?(?P<googledomain>\.googleapis\.com)?" ) m = mtls_endpoint_re.match(api_endpoint) name, mtls, sandbox, googledomain = m.groups() if mtls or not googledomain: return api_endpoint if sandbox: return api_endpoint.replace( "sandbox.googleapis.com", "mtls.sandbox.googleapis.com" ) return api_endpoint.replace(".googleapis.com", ".mtls.googleapis.com") DEFAULT_ENDPOINT = "googleads.googleapis.com" DEFAULT_MTLS_ENDPOINT = _get_default_mtls_endpoint.__func__( # type: ignore DEFAULT_ENDPOINT ) @classmethod def from_service_account_info(cls, info: dict, *args, **kwargs): """Creates an instance of this client using the provided credentials info. Args: info (dict): The service account private key info. args: Additional arguments to pass to the constructor. kwargs: Additional arguments to pass to the constructor. Returns: CustomConversionGoalServiceClient: The constructed client. """ credentials = service_account.Credentials.from_service_account_info( info ) kwargs["credentials"] = credentials return cls(*args, **kwargs) @classmethod def from_service_account_file(cls, filename: str, *args, **kwargs): """Creates an instance of this client using the provided credentials file. Args: filename (str): The path to the service account private key json file. args: Additional arguments to pass to the constructor. kwargs: Additional arguments to pass to the constructor. Returns: CustomConversionGoalServiceClient: The constructed client. """ credentials = service_account.Credentials.from_service_account_file( filename ) kwargs["credentials"] = credentials return cls(*args, **kwargs) from_service_account_json = from_service_account_file @property def transport(self) -> CustomConversionGoalServiceTransport: """Returns the transport used by the client instance. Returns: CustomConversionGoalServiceTransport: The transport used by the client instance. """ return self._transport def __enter__(self): return self def __exit__(self, type, value, traceback): """Releases underlying transport's resources. .. warning:: ONLY use as a context manager if the transport is NOT shared with other clients! Exiting the with block will CLOSE the transport and may cause errors in other clients! """ self.transport.close() @staticmethod def conversion_action_path( customer_id: str, conversion_action_id: str, ) -> str: """Returns a fully-qualified conversion_action string.""" return "customers/{customer_id}/conversionActions/{conversion_action_id}".format( customer_id=customer_id, conversion_action_id=conversion_action_id, ) @staticmethod def parse_conversion_action_path(path: str) -> Dict[str, str]: """Parses a conversion_action path into its component segments.""" m = re.match( r"^customers/(?P<customer_id>.+?)/conversionActions/(?P<conversion_action_id>.+?)$", path, ) return m.groupdict() if m else {} @staticmethod def custom_conversion_goal_path( customer_id: str, goal_id: str, ) -> str: """Returns a fully-qualified custom_conversion_goal string.""" return "customers/{customer_id}/customConversionGoals/{goal_id}".format( customer_id=customer_id, goal_id=goal_id, ) @staticmethod def parse_custom_conversion_goal_path(path: str) -> Dict[str, str]: """Parses a custom_conversion_goal path into its component segments.""" m = re.match( r"^customers/(?P<customer_id>.+?)/customConversionGoals/(?P<goal_id>.+?)$", path, ) return m.groupdict() if m else {} @staticmethod def common_billing_account_path( billing_account: str, ) -> str: """Returns a fully-qualified billing_account string.""" return "billingAccounts/{billing_account}".format( billing_account=billing_account, ) @staticmethod def parse_common_billing_account_path(path: str) -> Dict[str, str]: """Parse a billing_account path into its component segments.""" m = re.match(r"^billingAccounts/(?P<billing_account>.+?)$", path) return m.groupdict() if m else {} @staticmethod def common_folder_path( folder: str, ) -> str: """Returns a fully-qualified folder string.""" return "folders/{folder}".format( folder=folder, ) @staticmethod def parse_common_folder_path(path: str) -> Dict[str, str]: """Parse a folder path into its component segments.""" m = re.match(r"^folders/(?P<folder>.+?)$", path) return m.groupdict() if m else {} @staticmethod def common_organization_path( organization: str, ) -> str: """Returns a fully-qualified organization string.""" return "organizations/{organization}".format( organization=organization, ) @staticmethod def parse_common_organization_path(path: str) -> Dict[str, str]: """Parse a organization path into its component segments.""" m = re.match(r"^organizations/(?P<organization>.+?)$", path) return m.groupdict() if m else {} @staticmethod def common_project_path( project: str, ) -> str: """Returns a fully-qualified project string.""" return "projects/{project}".format( project=project, ) @staticmethod def parse_common_project_path(path: str) -> Dict[str, str]: """Parse a project path into its component segments.""" m = re.match(r"^projects/(?P<project>.+?)$", path) return m.groupdict() if m else {} @staticmethod def common_location_path( project: str, location: str, ) -> str: """Returns a fully-qualified location string.""" return "projects/{project}/locations/{location}".format( project=project, location=location, ) @staticmethod def parse_common_location_path(path: str) -> Dict[str, str]: """Parse a location path into its component segments.""" m = re.match( r"^projects/(?P<project>.+?)/locations/(?P<location>.+?)$", path ) return m.groupdict() if m else {} def __init__( self, *, credentials: Optional[ga_credentials.Credentials] = None, transport: Union[ str, CustomConversionGoalServiceTransport, None ] = None, client_options: Optional[client_options_lib.ClientOptions] = None, client_info: gapic_v1.client_info.ClientInfo = DEFAULT_CLIENT_INFO, ) -> None: """Instantiates the custom conversion goal service client. Args: credentials (Optional[google.auth.credentials.Credentials]): The authorization credentials to attach to requests. These credentials identify the application to the service; if none are specified, the client will attempt to ascertain the credentials from the environment. transport (Union[str, CustomConversionGoalServiceTransport]): The transport to use. If set to None, a transport is chosen automatically. client_options (google.api_core.client_options.ClientOptions): Custom options for the client. It won't take effect if a ``transport`` instance is provided. (1) The ``api_endpoint`` property can be used to override the default endpoint provided by the client. GOOGLE_API_USE_MTLS_ENDPOINT environment variable can also be used to override the endpoint: "always" (always use the default mTLS endpoint), "never" (always use the default regular endpoint) and "auto" (auto switch to the default mTLS endpoint if client certificate is present, this is the default value). However, the ``api_endpoint`` property takes precedence if provided. (2) If GOOGLE_API_USE_CLIENT_CERTIFICATE environment variable is "true", then the ``client_cert_source`` property can be used to provide client certificate for mutual TLS transport. If not provided, the default SSL client certificate will be used if present. If GOOGLE_API_USE_CLIENT_CERTIFICATE is "false" or not set, no client certificate will be used. client_info (google.api_core.gapic_v1.client_info.ClientInfo): The client info used to send a user-agent string along with API requests. If ``None``, then default info will be used. Generally, you only need to set this if you're developing your own client library. Raises: google.auth.exceptions.MutualTLSChannelError: If mutual TLS transport creation failed for any reason. """ if isinstance(client_options, dict): client_options = client_options_lib.from_dict(client_options) if client_options is None: client_options = client_options_lib.ClientOptions() # Create SSL credentials for mutual TLS if needed. if os.getenv("GOOGLE_API_USE_CLIENT_CERTIFICATE", "false") not in ( "true", "false", ): raise ValueError( "Environment variable `GOOGLE_API_USE_CLIENT_CERTIFICATE` must be either `true` or `false`" ) use_client_cert = ( os.getenv("GOOGLE_API_USE_CLIENT_CERTIFICATE", "false") == "true" ) client_cert_source_func = None is_mtls = False if use_client_cert: if client_options.client_cert_source: is_mtls = True client_cert_source_func = client_options.client_cert_source else: is_mtls = mtls.has_default_client_cert_source() if is_mtls: client_cert_source_func = mtls.default_client_cert_source() else: client_cert_source_func = None # Figure out which api endpoint to use. if client_options.api_endpoint is not None: api_endpoint = client_options.api_endpoint else: use_mtls_env = os.getenv("GOOGLE_API_USE_MTLS_ENDPOINT", "auto") if use_mtls_env == "never": api_endpoint = self.DEFAULT_ENDPOINT elif use_mtls_env == "always": api_endpoint = self.DEFAULT_MTLS_ENDPOINT elif use_mtls_env == "auto": api_endpoint = ( self.DEFAULT_MTLS_ENDPOINT if is_mtls else self.DEFAULT_ENDPOINT ) else: raise MutualTLSChannelError( "Unsupported GOOGLE_API_USE_MTLS_ENDPOINT value. Accepted " "values: never, auto, always" ) # Save or instantiate the transport. # Ordinarily, we provide the transport, but allowing a custom transport # instance provides an extensibility point for unusual situations. if isinstance(transport, CustomConversionGoalServiceTransport): # transport is a CustomConversionGoalServiceTransport instance. if credentials or client_options.credentials_file: raise ValueError( "When providing a transport instance, " "provide its credentials directly." ) if client_options.scopes: raise ValueError( "When providing a transport instance, provide its scopes " "directly." ) self._transport = transport else: Transport = type(self).get_transport_class(transport) self._transport = Transport( credentials=credentials, credentials_file=client_options.credentials_file, host=api_endpoint, scopes=client_options.scopes, client_cert_source_for_mtls=client_cert_source_func, quota_project_id=client_options.quota_project_id, client_info=client_info, always_use_jwt_access=True, ) def mutate_custom_conversion_goals( self, request: Union[ custom_conversion_goal_service.MutateCustomConversionGoalsRequest, dict, ] = None, *, customer_id: str = None, operations: Sequence[ custom_conversion_goal_service.CustomConversionGoalOperation ] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, timeout: float = None, metadata: Sequence[Tuple[str, str]] = (), ) -> custom_conversion_goal_service.MutateCustomConversionGoalsResponse: r"""Creates, updates or removes custom conversion goals. Operation statuses are returned. Args: request (Union[google.ads.googleads.v10.services.types.MutateCustomConversionGoalsRequest, dict]): The request object. Request message for [CustomConversionGoalService.MutateCustomConversionGoals][google.ads.googleads.v10.services.CustomConversionGoalService.MutateCustomConversionGoals]. customer_id (str): Required. The ID of the customer whose custom conversion goals are being modified. This corresponds to the ``customer_id`` field on the ``request`` instance; if ``request`` is provided, this should not be set. operations (Sequence[google.ads.googleads.v10.services.types.CustomConversionGoalOperation]): Required. The list of operations to perform on individual custom conversion goal. This corresponds to the ``operations`` field on the ``request`` instance; if ``request`` is provided, this should not be set. retry (google.api_core.retry.Retry): Designation of what errors, if any, should be retried. timeout (float): The timeout for this request. metadata (Sequence[Tuple[str, str]]): Strings which should be sent along with the request as metadata. Returns: google.ads.googleads.v10.services.types.MutateCustomConversionGoalsResponse: Response message for a custom conversion goal mutate. """ # Create or coerce a protobuf request object. # Quick check: If we got a request object, we should *not* have # gotten any keyword arguments that map to the request. has_flattened_params = any([customer_id, operations]) if request is not None and has_flattened_params: raise ValueError( "If the `request` argument is set, then none of " "the individual field arguments should be set." ) # Minor optimization to avoid making a copy if the user passes # in a custom_conversion_goal_service.MutateCustomConversionGoalsRequest. # There's no risk of modifying the input as we've already verified # there are no flattened fields. if not isinstance( request, custom_conversion_goal_service.MutateCustomConversionGoalsRequest, ): request = custom_conversion_goal_service.MutateCustomConversionGoalsRequest( request ) # If we have keyword arguments corresponding to fields on the # request, apply these. if customer_id is not None: request.customer_id = customer_id if operations is not None: request.operations = operations # Wrap the RPC method; this adds retry and timeout information, # and friendly error handling. rpc = self._transport._wrapped_methods[ self._transport.mutate_custom_conversion_goals ] # Certain fields should be provided within the metadata header; # add these here. metadata = tuple(metadata) + ( gapic_v1.routing_header.to_grpc_metadata( (("customer_id", request.customer_id),) ), ) # Send the request. response = rpc( request, retry=retry, timeout=timeout, metadata=metadata, ) # Done; return the response. return response try: DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo( gapic_version=pkg_resources.get_distribution( "google-ads", ).version, ) except pkg_resources.DistributionNotFound: DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo() __all__ = ("CustomConversionGoalServiceClient",)
[ "noreply@github.com" ]
GerhardusM.noreply@github.com
5184f3b5a7ed92581d910e979f58f31d1a589646
cad83699bb72bc1d560815d5b1c34fc9b371f163
/cartridge/project_template/manage.py
4330dc2e1a9c13927f25c69c19985527dbbe4267
[ "BSD-3-Clause" ]
permissive
BeUnique/cartridge
a1903ecc78029a576e57c3832b16357c7661b1b8
e5b887fed96d01ab93237f345fc420bcbe56d027
refs/heads/master
2021-01-19T06:00:01.090292
2011-08-31T01:04:33
2011-08-31T01:04:33
null
0
0
null
null
null
null
UTF-8
Python
false
false
1,284
py
#!/usr/bin/env python # When project_template is used as the actual project during Mezzanine # development, insert the development path into sys.path so that the # development version of Mezzanine is used rather than the installed version. import os import sys project_path = os.path.dirname(os.path.abspath(__file__)) project_dir = project_path.split(os.sep)[-1] if project_dir == "project_template": dev_path = os.path.abspath(os.path.join(project_path, "..", "..")) if dev_path not in sys.path: sys.path.insert(0, dev_path) import cartridge cartridge_path = os.path.dirname(os.path.abspath(cartridge.__file__)) assert os.path.abspath(os.path.join(cartridge_path, "..")) == dev_path from django.core.management import execute_manager try: import settings # Assumed to be in the same directory. except ImportError: import sys sys.stderr.write("Error: Can't find the file 'settings.py' in the " "directory containing %r. It appears you've customized things.\n" "You'll have to run django-admin.py, passing it your settings module.\n" "(If the file settings.py does indeed exist, it's causing an " "ImportError somehow.)\n" % __file__) sys.exit(1) if __name__ == "__main__": execute_manager(settings)
[ "steve@jupo.org" ]
steve@jupo.org
6e7f711485de28b5120b7854383a32ff11f98c7d
cb48956cb48a94da0a25e7742b87b623224664ed
/network/relighting.py
bef13c75f6a6c179329e32a4459a1d8c993cdd27
[ "Apache-2.0" ]
permissive
Siddharth-Shrivastava7/DANNet
e5f2fb03bd8cb4e19392863566a9b97efa59f955
8db10056a4e445d24fc899505923615457cae5b7
refs/heads/main
2023-06-20T20:33:32.602314
2021-07-30T17:52:00
2021-07-30T17:52:00
362,430,200
0
0
Apache-2.0
2021-04-28T10:38:22
2021-04-28T10:38:21
null
UTF-8
Python
false
false
7,379
py
import torch import torch.nn as nn import functools affine_par = True class ResnetBlock(nn.Module): """Define a Resnet block""" def __init__(self, dim, padding_type, norm_layer, use_dropout, use_bias): """Initialize the Resnet block A resnet block is a conv block with skip connections We construct a conv block with build_conv_block function, and implement skip connections in <forward> function. Original Resnet paper: https://arxiv.org/pdf/1512.03385.pdf """ super(ResnetBlock, self).__init__() self.conv_block = self.build_conv_block(dim, padding_type, norm_layer, use_dropout, use_bias) def build_conv_block(self, dim, padding_type, norm_layer, use_dropout, use_bias): """Construct a convolutional block. Parameters: dim (int) -- the number of channels in the conv layer. padding_type (str) -- the name of padding layer: reflect | replicate | zero norm_layer -- normalization layer use_dropout (bool) -- if use dropout layers. use_bias (bool) -- if the conv layer uses bias or not Returns a conv block (with a conv layer, a normalization layer, and a non-linearity layer (ReLU)) """ conv_block = [] p = 0 if padding_type == 'reflect': conv_block += [nn.ReflectionPad2d(1)] elif padding_type == 'replicate': conv_block += [nn.ReplicationPad2d(1)] elif padding_type == 'zero': p = 1 else: raise NotImplementedError('padding [%s] is not implemented' % padding_type) conv_block += [nn.Conv2d(dim, dim, kernel_size=3, padding=p, bias=use_bias), norm_layer(dim), nn.ReLU(True)] if use_dropout: conv_block += [nn.Dropout(0.5)] p = 0 if padding_type == 'reflect': conv_block += [nn.ReflectionPad2d(1)] elif padding_type == 'replicate': conv_block += [nn.ReplicationPad2d(1)] elif padding_type == 'zero': p = 1 else: raise NotImplementedError('padding [%s] is not implemented' % padding_type) conv_block += [nn.Conv2d(dim, dim, kernel_size=3, padding=p, bias=use_bias), norm_layer(dim)] return nn.Sequential(*conv_block) def forward(self, x): """Forward function (with skip connections)""" out = x + self.conv_block(x) # add skip connections return out class ResnetGenerator(nn.Module): """Resnet-based generator that consists of Resnet blocks between a few downsampling/upsampling operations. We adapt Torch code and idea from Justin Johnson's neural style transfer project(https://github.com/jcjohnson/fast-neural-style) """ def __init__(self, input_nc, output_nc, ngf=64, norm_layer=nn.BatchNorm2d, use_dropout=False, n_blocks=6, padding_type='reflect'): """Construct a Resnet-based generator Parameters: input_nc (int) -- the number of channels in input images output_nc (int) -- the number of channels in output images ngf (int) -- the number of filters in the last conv layer norm_layer -- normalization layer use_dropout (bool) -- if use dropout layers n_blocks (int) -- the number of ResNet blocks padding_type (str) -- the name of padding layer in conv layers: reflect | replicate | zero """ assert(n_blocks >= 0) super(ResnetGenerator, self).__init__() if type(norm_layer) == functools.partial: use_bias = norm_layer.func == nn.InstanceNorm2d else: use_bias = norm_layer == nn.InstanceNorm2d model = [nn.ReflectionPad2d(3), nn.Conv2d(input_nc, ngf, kernel_size=7, padding=0, bias=use_bias), norm_layer(ngf), nn.ReLU(True)] n_downsampling = 2 for i in range(n_downsampling): # add downsampling layers mult = 2 ** i model += [nn.Conv2d(ngf * mult, ngf * mult * 2, kernel_size=3, stride=2, padding=1, bias=use_bias), norm_layer(ngf * mult * 2), nn.ReLU(True)] mult = 2 ** n_downsampling for i in range(n_blocks): # add ResNet blocks model += [ResnetBlock(ngf * mult, padding_type=padding_type, norm_layer=norm_layer, use_dropout=use_dropout, use_bias=use_bias)] for i in range(n_downsampling): # add upsampling layers mult = 2 ** (n_downsampling - i) model += [nn.ConvTranspose2d(ngf * mult, int(ngf * mult / 2), kernel_size=3, stride=2, padding=1, output_padding=1, bias=use_bias), norm_layer(int(ngf * mult / 2)), nn.ReLU(True)] model += [nn.ReflectionPad2d(3)] model += [nn.Conv2d(ngf, output_nc, kernel_size=7, padding=0)] model += [nn.Tanh()] self.model = nn.Sequential(*model) def forward(self, input): """Standard forward""" return self.model(input) def LightNet(): model = ResnetGenerator(3, 3, 64, norm_layer=nn.BatchNorm2d, use_dropout=False, n_blocks=3) return model class L_exp_z(nn.Module): def __init__(self, patch_size): super(L_exp_z, self).__init__() self.pool = nn.AvgPool2d(patch_size) def forward(self, x, mean_val): x = torch.mean(x, 1, keepdim=True) mean = self.pool(x) d = torch.mean(torch.pow(mean - torch.FloatTensor([mean_val]).cuda(), 2)) return d class L_TV(nn.Module): def __init__(self, TVLoss_weight=1): super(L_TV, self).__init__() self.TVLoss_weight = TVLoss_weight def forward(self, x): batch_size = x.size()[0] h_x = x.size()[2] w_x = x.size()[3] count_h = (x.size()[2] - 1) * x.size()[3] count_w = x.size()[2] * (x.size()[3] - 1) h_tv = torch.pow((x[:, :, 1:, :] - x[:, :, :h_x - 1, :]), 2).sum() w_tv = torch.pow((x[:, :, :, 1:] - x[:, :, :, :w_x - 1]), 2).sum() return self.TVLoss_weight * 2 * (h_tv / count_h + w_tv / count_w) / batch_size class SSIM(nn.Module): """Layer to compute the SSIM loss between a pair of images """ def __init__(self): super(SSIM, self).__init__() self.mu_x_pool = nn.AvgPool2d(3, 1) self.mu_y_pool = nn.AvgPool2d(3, 1) self.sig_x_pool = nn.AvgPool2d(3, 1) self.sig_y_pool = nn.AvgPool2d(3, 1) self.sig_xy_pool = nn.AvgPool2d(3, 1) self.refl = nn.ReflectionPad2d(1) self.C1 = 0.01 ** 2 self.C2 = 0.03 ** 2 def forward(self, x, y): x = self.refl(x) y = self.refl(y) mu_x = self.mu_x_pool(x) mu_y = self.mu_y_pool(y) sigma_x = self.sig_x_pool(x ** 2) - mu_x ** 2 sigma_y = self.sig_y_pool(y ** 2) - mu_y ** 2 sigma_xy = self.sig_xy_pool(x * y) - mu_x * mu_y SSIM_n = (2 * mu_x * mu_y + self.C1) * (2 * sigma_xy + self.C2) SSIM_d = (mu_x ** 2 + mu_y ** 2 + self.C1) * (sigma_x + sigma_y + self.C2) return torch.clamp((1 - SSIM_n / SSIM_d) / 2, 0, 1)
[ "noreply@github.com" ]
Siddharth-Shrivastava7.noreply@github.com
963a9147e8343c232ca5f41b05658557f4a21e1b
9ae0fe3620c5be4af7801462acc7564e8d4a56f0
/OneDrive_4_8-26-2019/TSqlVisitor.py
41ad406c9430d77fc567e1940c78148efda05311
[]
no_license
MacJei/antlr
66ea15b3e8b063354664031dac9427db4c02ad5c
7e945a6424caad00e58f7a4b41b0297556f00294
refs/heads/master
2022-03-20T17:49:06.241851
2019-12-18T12:27:47
2019-12-18T12:27:47
null
0
0
null
null
null
null
UTF-8
Python
false
false
44,697
py
# Generated from TSql.g4 by ANTLR 4.7.2 from antlr4 import * if __name__ is not None and "." in __name__: from .TSqlParser import TSqlParser else: from TSqlParser import TSqlParser # This class defines a complete generic visitor for a parse tree produced by TSqlParser. class TSqlVisitor(ParseTreeVisitor): # Visit a parse tree produced by TSqlParser#tsql_file. def visitTsql_file(self, ctx:TSqlParser.Tsql_fileContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#batch. def visitBatch(self, ctx:TSqlParser.BatchContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#sql_clauses. def visitSql_clauses(self, ctx:TSqlParser.Sql_clausesContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#sql_clause. def visitSql_clause(self, ctx:TSqlParser.Sql_clauseContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#dml_clause. def visitDml_clause(self, ctx:TSqlParser.Dml_clauseContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#cfl_statement. def visitCfl_statement(self, ctx:TSqlParser.Cfl_statementContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#block_statement. def visitBlock_statement(self, ctx:TSqlParser.Block_statementContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#break_statement. def visitBreak_statement(self, ctx:TSqlParser.Break_statementContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#continue_statement. def visitContinue_statement(self, ctx:TSqlParser.Continue_statementContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#goto_statement. def visitGoto_statement(self, ctx:TSqlParser.Goto_statementContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#return_statement. def visitReturn_statement(self, ctx:TSqlParser.Return_statementContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#if_statement. def visitIf_statement(self, ctx:TSqlParser.If_statementContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#throw_statement. def visitThrow_statement(self, ctx:TSqlParser.Throw_statementContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#throw_error_number. def visitThrow_error_number(self, ctx:TSqlParser.Throw_error_numberContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#throw_message. def visitThrow_message(self, ctx:TSqlParser.Throw_messageContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#throw_state. def visitThrow_state(self, ctx:TSqlParser.Throw_stateContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#try_catch_statement. def visitTry_catch_statement(self, ctx:TSqlParser.Try_catch_statementContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#waitfor_statement. def visitWaitfor_statement(self, ctx:TSqlParser.Waitfor_statementContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#while_statement. def visitWhile_statement(self, ctx:TSqlParser.While_statementContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#print_statement. def visitPrint_statement(self, ctx:TSqlParser.Print_statementContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#raiseerror_statement. def visitRaiseerror_statement(self, ctx:TSqlParser.Raiseerror_statementContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#empty_statement. def visitEmpty_statement(self, ctx:TSqlParser.Empty_statementContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#another_statement. def visitAnother_statement(self, ctx:TSqlParser.Another_statementContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#file_path. def visitFile_path(self, ctx:TSqlParser.File_pathContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#file_directory_path_separator. def visitFile_directory_path_separator(self, ctx:TSqlParser.File_directory_path_separatorContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#event_session_predicate_expression. def visitEvent_session_predicate_expression(self, ctx:TSqlParser.Event_session_predicate_expressionContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#event_session_predicate_factor. def visitEvent_session_predicate_factor(self, ctx:TSqlParser.Event_session_predicate_factorContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#event_session_predicate_leaf. def visitEvent_session_predicate_leaf(self, ctx:TSqlParser.Event_session_predicate_leafContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#create_queue. def visitCreate_queue(self, ctx:TSqlParser.Create_queueContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#queue_settings. def visitQueue_settings(self, ctx:TSqlParser.Queue_settingsContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#alter_queue. def visitAlter_queue(self, ctx:TSqlParser.Alter_queueContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#queue_action. def visitQueue_action(self, ctx:TSqlParser.Queue_actionContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#queue_rebuild_options. def visitQueue_rebuild_options(self, ctx:TSqlParser.Queue_rebuild_optionsContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#create_contract. def visitCreate_contract(self, ctx:TSqlParser.Create_contractContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#conversation_statement. def visitConversation_statement(self, ctx:TSqlParser.Conversation_statementContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#message_statement. def visitMessage_statement(self, ctx:TSqlParser.Message_statementContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#merge_statement. def visitMerge_statement(self, ctx:TSqlParser.Merge_statementContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#merge_matched. def visitMerge_matched(self, ctx:TSqlParser.Merge_matchedContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#merge_not_matched. def visitMerge_not_matched(self, ctx:TSqlParser.Merge_not_matchedContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#delete_statement. def visitDelete_statement(self, ctx:TSqlParser.Delete_statementContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#delete_statement_from. def visitDelete_statement_from(self, ctx:TSqlParser.Delete_statement_fromContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#insert_statement. def visitInsert_statement(self, ctx:TSqlParser.Insert_statementContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#insert_statement_value. def visitInsert_statement_value(self, ctx:TSqlParser.Insert_statement_valueContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#receive_statement. def visitReceive_statement(self, ctx:TSqlParser.Receive_statementContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#select_statement. def visitSelect_statement(self, ctx:TSqlParser.Select_statementContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#time. def visitTime(self, ctx:TSqlParser.TimeContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#update_statement. def visitUpdate_statement(self, ctx:TSqlParser.Update_statementContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#output_clause. def visitOutput_clause(self, ctx:TSqlParser.Output_clauseContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#output_dml_list_elem. def visitOutput_dml_list_elem(self, ctx:TSqlParser.Output_dml_list_elemContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#output_column_name. def visitOutput_column_name(self, ctx:TSqlParser.Output_column_nameContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#rowset_function_limited. def visitRowset_function_limited(self, ctx:TSqlParser.Rowset_function_limitedContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#openquery. def visitOpenquery(self, ctx:TSqlParser.OpenqueryContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#opendatasource. def visitOpendatasource(self, ctx:TSqlParser.OpendatasourceContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#declare_statement. def visitDeclare_statement(self, ctx:TSqlParser.Declare_statementContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#cursor_statement. def visitCursor_statement(self, ctx:TSqlParser.Cursor_statementContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#execute_statement. def visitExecute_statement(self, ctx:TSqlParser.Execute_statementContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#execute_body. def visitExecute_body(self, ctx:TSqlParser.Execute_bodyContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#execute_statement_arg. def visitExecute_statement_arg(self, ctx:TSqlParser.Execute_statement_argContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#execute_var_string. def visitExecute_var_string(self, ctx:TSqlParser.Execute_var_stringContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#security_statement. def visitSecurity_statement(self, ctx:TSqlParser.Security_statementContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#create_certificate. def visitCreate_certificate(self, ctx:TSqlParser.Create_certificateContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#existing_keys. def visitExisting_keys(self, ctx:TSqlParser.Existing_keysContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#private_key_options. def visitPrivate_key_options(self, ctx:TSqlParser.Private_key_optionsContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#generate_new_keys. def visitGenerate_new_keys(self, ctx:TSqlParser.Generate_new_keysContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#date_options. def visitDate_options(self, ctx:TSqlParser.Date_optionsContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#open_key. def visitOpen_key(self, ctx:TSqlParser.Open_keyContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#close_key. def visitClose_key(self, ctx:TSqlParser.Close_keyContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#create_key. def visitCreate_key(self, ctx:TSqlParser.Create_keyContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#key_options. def visitKey_options(self, ctx:TSqlParser.Key_optionsContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#algorithm. def visitAlgorithm(self, ctx:TSqlParser.AlgorithmContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#encryption_mechanism. def visitEncryption_mechanism(self, ctx:TSqlParser.Encryption_mechanismContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#decryption_mechanism. def visitDecryption_mechanism(self, ctx:TSqlParser.Decryption_mechanismContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#grant_permission. def visitGrant_permission(self, ctx:TSqlParser.Grant_permissionContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#set_statement. def visitSet_statement(self, ctx:TSqlParser.Set_statementContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#transaction_statement. def visitTransaction_statement(self, ctx:TSqlParser.Transaction_statementContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#go_statement. def visitGo_statement(self, ctx:TSqlParser.Go_statementContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#use_statement. def visitUse_statement(self, ctx:TSqlParser.Use_statementContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#setuser_statement. def visitSetuser_statement(self, ctx:TSqlParser.Setuser_statementContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#execute_clause. def visitExecute_clause(self, ctx:TSqlParser.Execute_clauseContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#declare_local. def visitDeclare_local(self, ctx:TSqlParser.Declare_localContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#table_type_definition. def visitTable_type_definition(self, ctx:TSqlParser.Table_type_definitionContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#xml_type_definition. def visitXml_type_definition(self, ctx:TSqlParser.Xml_type_definitionContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#xml_schema_collection. def visitXml_schema_collection(self, ctx:TSqlParser.Xml_schema_collectionContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#column_def_table_constraints. def visitColumn_def_table_constraints(self, ctx:TSqlParser.Column_def_table_constraintsContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#column_def_table_constraint. def visitColumn_def_table_constraint(self, ctx:TSqlParser.Column_def_table_constraintContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#column_definition. def visitColumn_definition(self, ctx:TSqlParser.Column_definitionContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#materialized_column_definition. def visitMaterialized_column_definition(self, ctx:TSqlParser.Materialized_column_definitionContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#column_constraint. def visitColumn_constraint(self, ctx:TSqlParser.Column_constraintContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#table_constraint. def visitTable_constraint(self, ctx:TSqlParser.Table_constraintContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#on_delete. def visitOn_delete(self, ctx:TSqlParser.On_deleteContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#on_update. def visitOn_update(self, ctx:TSqlParser.On_updateContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#index_options. def visitIndex_options(self, ctx:TSqlParser.Index_optionsContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#index_option. def visitIndex_option(self, ctx:TSqlParser.Index_optionContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#declare_cursor. def visitDeclare_cursor(self, ctx:TSqlParser.Declare_cursorContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#declare_set_cursor_common. def visitDeclare_set_cursor_common(self, ctx:TSqlParser.Declare_set_cursor_commonContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#declare_set_cursor_common_partial. def visitDeclare_set_cursor_common_partial(self, ctx:TSqlParser.Declare_set_cursor_common_partialContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#fetch_cursor. def visitFetch_cursor(self, ctx:TSqlParser.Fetch_cursorContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#set_special. def visitSet_special(self, ctx:TSqlParser.Set_specialContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#constant_LOCAL_ID. def visitConstant_LOCAL_ID(self, ctx:TSqlParser.Constant_LOCAL_IDContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#expression. def visitExpression(self, ctx:TSqlParser.ExpressionContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#primitive_expression. def visitPrimitive_expression(self, ctx:TSqlParser.Primitive_expressionContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#case_expression. def visitCase_expression(self, ctx:TSqlParser.Case_expressionContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#unary_operator_expression. def visitUnary_operator_expression(self, ctx:TSqlParser.Unary_operator_expressionContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#bracket_expression. def visitBracket_expression(self, ctx:TSqlParser.Bracket_expressionContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#constant_expression. def visitConstant_expression(self, ctx:TSqlParser.Constant_expressionContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#subquery. def visitSubquery(self, ctx:TSqlParser.SubqueryContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#with_expression. def visitWith_expression(self, ctx:TSqlParser.With_expressionContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#common_table_expression. def visitCommon_table_expression(self, ctx:TSqlParser.Common_table_expressionContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#update_elem. def visitUpdate_elem(self, ctx:TSqlParser.Update_elemContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#search_condition_list. def visitSearch_condition_list(self, ctx:TSqlParser.Search_condition_listContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#search_condition. def visitSearch_condition(self, ctx:TSqlParser.Search_conditionContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#search_condition_and. def visitSearch_condition_and(self, ctx:TSqlParser.Search_condition_andContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#search_condition_not. def visitSearch_condition_not(self, ctx:TSqlParser.Search_condition_notContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#predicate. def visitPredicate(self, ctx:TSqlParser.PredicateContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#query_expression. def visitQuery_expression(self, ctx:TSqlParser.Query_expressionContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#sql_union. def visitSql_union(self, ctx:TSqlParser.Sql_unionContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#query_specification. def visitQuery_specification(self, ctx:TSqlParser.Query_specificationContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#top_clause. def visitTop_clause(self, ctx:TSqlParser.Top_clauseContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#top_percent. def visitTop_percent(self, ctx:TSqlParser.Top_percentContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#top_count. def visitTop_count(self, ctx:TSqlParser.Top_countContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#order_by_clause. def visitOrder_by_clause(self, ctx:TSqlParser.Order_by_clauseContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#for_clause. def visitFor_clause(self, ctx:TSqlParser.For_clauseContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#xml_common_directives. def visitXml_common_directives(self, ctx:TSqlParser.Xml_common_directivesContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#order_by_expression. def visitOrder_by_expression(self, ctx:TSqlParser.Order_by_expressionContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#group_by_item. def visitGroup_by_item(self, ctx:TSqlParser.Group_by_itemContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#option_clause. def visitOption_clause(self, ctx:TSqlParser.Option_clauseContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#option. def visitOption(self, ctx:TSqlParser.OptionContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#optimize_for_arg. def visitOptimize_for_arg(self, ctx:TSqlParser.Optimize_for_argContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#select_list. def visitSelect_list(self, ctx:TSqlParser.Select_listContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#udt_method_arguments. def visitUdt_method_arguments(self, ctx:TSqlParser.Udt_method_argumentsContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#asterisk. def visitAsterisk(self, ctx:TSqlParser.AsteriskContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#column_elem. def visitColumn_elem(self, ctx:TSqlParser.Column_elemContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#udt_elem. def visitUdt_elem(self, ctx:TSqlParser.Udt_elemContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#expression_elem. def visitExpression_elem(self, ctx:TSqlParser.Expression_elemContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#select_list_elem. def visitSelect_list_elem(self, ctx:TSqlParser.Select_list_elemContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#table_sources. def visitTable_sources(self, ctx:TSqlParser.Table_sourcesContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#table_source. def visitTable_source(self, ctx:TSqlParser.Table_sourceContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#table_source_item_joined. def visitTable_source_item_joined(self, ctx:TSqlParser.Table_source_item_joinedContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#table_source_item. def visitTable_source_item(self, ctx:TSqlParser.Table_source_itemContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#open_xml. def visitOpen_xml(self, ctx:TSqlParser.Open_xmlContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#schema_declaration. def visitSchema_declaration(self, ctx:TSqlParser.Schema_declarationContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#column_declaration. def visitColumn_declaration(self, ctx:TSqlParser.Column_declarationContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#change_table. def visitChange_table(self, ctx:TSqlParser.Change_tableContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#join_part. def visitJoin_part(self, ctx:TSqlParser.Join_partContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#pivot_clause. def visitPivot_clause(self, ctx:TSqlParser.Pivot_clauseContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#unpivot_clause. def visitUnpivot_clause(self, ctx:TSqlParser.Unpivot_clauseContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#full_column_name_list. def visitFull_column_name_list(self, ctx:TSqlParser.Full_column_name_listContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#table_name_with_hint. def visitTable_name_with_hint(self, ctx:TSqlParser.Table_name_with_hintContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#rowset_function. def visitRowset_function(self, ctx:TSqlParser.Rowset_functionContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#bulk_option. def visitBulk_option(self, ctx:TSqlParser.Bulk_optionContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#derived_table. def visitDerived_table(self, ctx:TSqlParser.Derived_tableContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#RANKING_WINDOWED_FUNC. def visitRANKING_WINDOWED_FUNC(self, ctx:TSqlParser.RANKING_WINDOWED_FUNCContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#AGGREGATE_WINDOWED_FUNC. def visitAGGREGATE_WINDOWED_FUNC(self, ctx:TSqlParser.AGGREGATE_WINDOWED_FUNCContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#ANALYTIC_WINDOWED_FUNC. def visitANALYTIC_WINDOWED_FUNC(self, ctx:TSqlParser.ANALYTIC_WINDOWED_FUNCContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#SCALAR_FUNCTION. def visitSCALAR_FUNCTION(self, ctx:TSqlParser.SCALAR_FUNCTIONContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#BINARY_CHECKSUM. def visitBINARY_CHECKSUM(self, ctx:TSqlParser.BINARY_CHECKSUMContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#CAST. def visitCAST(self, ctx:TSqlParser.CASTContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#CONVERT. def visitCONVERT(self, ctx:TSqlParser.CONVERTContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#CHECKSUM. def visitCHECKSUM(self, ctx:TSqlParser.CHECKSUMContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#COALESCE. def visitCOALESCE(self, ctx:TSqlParser.COALESCEContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#CURRENT_TIMESTAMP. def visitCURRENT_TIMESTAMP(self, ctx:TSqlParser.CURRENT_TIMESTAMPContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#CURRENT_USER. def visitCURRENT_USER(self, ctx:TSqlParser.CURRENT_USERContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#DATEADD. def visitDATEADD(self, ctx:TSqlParser.DATEADDContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#DATEDIFF. def visitDATEDIFF(self, ctx:TSqlParser.DATEDIFFContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#DATENAME. def visitDATENAME(self, ctx:TSqlParser.DATENAMEContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#DATEPART. def visitDATEPART(self, ctx:TSqlParser.DATEPARTContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#GETDATE. def visitGETDATE(self, ctx:TSqlParser.GETDATEContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#GETUTCDATE. def visitGETUTCDATE(self, ctx:TSqlParser.GETUTCDATEContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#IDENTITY. def visitIDENTITY(self, ctx:TSqlParser.IDENTITYContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#MIN_ACTIVE_ROWVERSION. def visitMIN_ACTIVE_ROWVERSION(self, ctx:TSqlParser.MIN_ACTIVE_ROWVERSIONContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#NULLIF. def visitNULLIF(self, ctx:TSqlParser.NULLIFContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#STUFF. def visitSTUFF(self, ctx:TSqlParser.STUFFContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#SESSION_USER. def visitSESSION_USER(self, ctx:TSqlParser.SESSION_USERContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#SYSTEM_USER. def visitSYSTEM_USER(self, ctx:TSqlParser.SYSTEM_USERContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#ISNULL. def visitISNULL(self, ctx:TSqlParser.ISNULLContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#XML_DATA_TYPE_FUNC. def visitXML_DATA_TYPE_FUNC(self, ctx:TSqlParser.XML_DATA_TYPE_FUNCContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#IFF. def visitIFF(self, ctx:TSqlParser.IFFContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#xml_data_type_methods. def visitXml_data_type_methods(self, ctx:TSqlParser.Xml_data_type_methodsContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#value_method. def visitValue_method(self, ctx:TSqlParser.Value_methodContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#query_method. def visitQuery_method(self, ctx:TSqlParser.Query_methodContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#exist_method. def visitExist_method(self, ctx:TSqlParser.Exist_methodContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#modify_method. def visitModify_method(self, ctx:TSqlParser.Modify_methodContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#nodes_method. def visitNodes_method(self, ctx:TSqlParser.Nodes_methodContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#switch_section. def visitSwitch_section(self, ctx:TSqlParser.Switch_sectionContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#switch_search_condition_section. def visitSwitch_search_condition_section(self, ctx:TSqlParser.Switch_search_condition_sectionContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#as_column_alias. def visitAs_column_alias(self, ctx:TSqlParser.As_column_aliasContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#as_table_alias. def visitAs_table_alias(self, ctx:TSqlParser.As_table_aliasContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#table_alias. def visitTable_alias(self, ctx:TSqlParser.Table_aliasContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#with_table_hints. def visitWith_table_hints(self, ctx:TSqlParser.With_table_hintsContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#insert_with_table_hints. def visitInsert_with_table_hints(self, ctx:TSqlParser.Insert_with_table_hintsContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#table_hint. def visitTable_hint(self, ctx:TSqlParser.Table_hintContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#index_value. def visitIndex_value(self, ctx:TSqlParser.Index_valueContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#column_alias_list. def visitColumn_alias_list(self, ctx:TSqlParser.Column_alias_listContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#column_alias. def visitColumn_alias(self, ctx:TSqlParser.Column_aliasContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#table_value_constructor. def visitTable_value_constructor(self, ctx:TSqlParser.Table_value_constructorContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#expression_list. def visitExpression_list(self, ctx:TSqlParser.Expression_listContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#ranking_windowed_function. def visitRanking_windowed_function(self, ctx:TSqlParser.Ranking_windowed_functionContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#aggregate_windowed_function. def visitAggregate_windowed_function(self, ctx:TSqlParser.Aggregate_windowed_functionContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#analytic_windowed_function. def visitAnalytic_windowed_function(self, ctx:TSqlParser.Analytic_windowed_functionContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#all_distinct_expression. def visitAll_distinct_expression(self, ctx:TSqlParser.All_distinct_expressionContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#over_clause. def visitOver_clause(self, ctx:TSqlParser.Over_clauseContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#row_or_range_clause. def visitRow_or_range_clause(self, ctx:TSqlParser.Row_or_range_clauseContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#window_frame_extent. def visitWindow_frame_extent(self, ctx:TSqlParser.Window_frame_extentContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#window_frame_bound. def visitWindow_frame_bound(self, ctx:TSqlParser.Window_frame_boundContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#window_frame_preceding. def visitWindow_frame_preceding(self, ctx:TSqlParser.Window_frame_precedingContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#window_frame_following. def visitWindow_frame_following(self, ctx:TSqlParser.Window_frame_followingContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#full_table_name. def visitFull_table_name(self, ctx:TSqlParser.Full_table_nameContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#table_name. def visitTable_name(self, ctx:TSqlParser.Table_nameContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#func_proc_name_schema. def visitFunc_proc_name_schema(self, ctx:TSqlParser.Func_proc_name_schemaContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#func_proc_name_database_schema. def visitFunc_proc_name_database_schema(self, ctx:TSqlParser.Func_proc_name_database_schemaContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#func_proc_name_server_database_schema. def visitFunc_proc_name_server_database_schema(self, ctx:TSqlParser.Func_proc_name_server_database_schemaContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#ddl_object. def visitDdl_object(self, ctx:TSqlParser.Ddl_objectContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#full_column_name. def visitFull_column_name(self, ctx:TSqlParser.Full_column_nameContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#column_name_list_with_order. def visitColumn_name_list_with_order(self, ctx:TSqlParser.Column_name_list_with_orderContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#column_name_list. def visitColumn_name_list(self, ctx:TSqlParser.Column_name_listContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#cursor_name. def visitCursor_name(self, ctx:TSqlParser.Cursor_nameContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#on_off. def visitOn_off(self, ctx:TSqlParser.On_offContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#clustered. def visitClustered(self, ctx:TSqlParser.ClusteredContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#null_notnull. def visitNull_notnull(self, ctx:TSqlParser.Null_notnullContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#null_or_default. def visitNull_or_default(self, ctx:TSqlParser.Null_or_defaultContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#scalar_function_name. def visitScalar_function_name(self, ctx:TSqlParser.Scalar_function_nameContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#begin_conversation_timer. def visitBegin_conversation_timer(self, ctx:TSqlParser.Begin_conversation_timerContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#begin_conversation_dialog. def visitBegin_conversation_dialog(self, ctx:TSqlParser.Begin_conversation_dialogContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#contract_name. def visitContract_name(self, ctx:TSqlParser.Contract_nameContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#service_name. def visitService_name(self, ctx:TSqlParser.Service_nameContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#end_conversation. def visitEnd_conversation(self, ctx:TSqlParser.End_conversationContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#waitfor_conversation. def visitWaitfor_conversation(self, ctx:TSqlParser.Waitfor_conversationContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#get_conversation. def visitGet_conversation(self, ctx:TSqlParser.Get_conversationContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#queue_id. def visitQueue_id(self, ctx:TSqlParser.Queue_idContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#send_conversation. def visitSend_conversation(self, ctx:TSqlParser.Send_conversationContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#data_type. def visitData_type(self, ctx:TSqlParser.Data_typeContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#constant. def visitConstant(self, ctx:TSqlParser.ConstantContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#sign. def visitSign(self, ctx:TSqlParser.SignContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#identifier. def visitIdentifier(self, ctx:TSqlParser.IdentifierContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#simple_id. def visitSimple_id(self, ctx:TSqlParser.Simple_idContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#comparison_operator. def visitComparison_operator(self, ctx:TSqlParser.Comparison_operatorContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#assignment_operator. def visitAssignment_operator(self, ctx:TSqlParser.Assignment_operatorContext): return self.visitChildren(ctx) del TSqlParser
[ "noreply@github.com" ]
MacJei.noreply@github.com
1215d8edd938231f4d95f34270b4973a1bf63dbd
4db6ffdad89c6a3d1cf9731b815dbb7177b10043
/code/landcover_model/config.py
39913c695156dfd96525beffad1fcfbf51bf2308
[]
no_license
hugisveitti/thesis
55002c38ce58ad5b99dc07900e0a2121cbe096c0
a99a034fa684eae4d9c7f7cad391a521ff0525db
refs/heads/master
2023-07-27T05:05:16.535638
2021-08-15T19:26:43
2021-08-15T19:26:43
364,648,515
0
0
null
null
null
null
UTF-8
Python
false
false
411
py
import torch IMAGE_WIDTH = 256 IMAGE_HEIGHT = 256 num_classes = 9 tensor_type = torch.HalfTensor device = "cuda" cross_entropy_weights = torch.tensor([1.75827705e-05, 2.63985891e-02, 4.93802954e-01, 0.00000000e+00, 9.56997597e-02, 6.52813402e-02, 2.43301976e-01, 2.19168076e-02, 0.00000000e+00, 2.51651604e-02, 2.09771106e-02, 3.17152767e-03, 0.00000000e+00, 4.26719143e-03]).to(device)
[ "hugiholm1@gmail.com" ]
hugiholm1@gmail.com
788ffc9f63b276da6cc4ebcab3f6f95eb32e6321
81384c792f7c8d85e08318abb0b79b1ae8768718
/controllers/draft.py
29ddfaecee941213b208aec42a7aac05f804e5a8
[ "LicenseRef-scancode-public-domain" ]
permissive
drewbeller/team_drafter
b2037afed29268b253a4a03a6dd6122396ad89ae
152f421d947c6c33442261f229dac4535b1b7bc2
refs/heads/master
2021-01-01T17:17:20.900147
2015-08-16T20:52:59
2015-08-16T20:52:59
23,207,245
1
0
null
null
null
null
UTF-8
Python
false
false
2,589
py
# coding: utf8 import ast def generate_picks(): # load draft parameters draft_order = db(db.draft_parameter.draft_parameter == 'Draft Order').select().first()['draft_parameter_value'] rounds = int(db(db.draft_parameter.draft_parameter == 'Rounds').select().first()['draft_parameter_value']) # pick_times = db(db.draft_parameter.draft_parameter == 'Pick Time').select().first()['draft_parameter_value'] # pick_times = ast.literal_eval(pick_times) # select all teams and add to dict with id and order teams = db(db.team).select(orderby=db.team.team_order) db(db.pick).delete() pick_num = 0 for round in range(rounds): round_num = round + 1 # round starts at zero if (round_num % 2 == 0) and (draft_order == 'snake'): teams = teams.sort(lambda team: team.team_order, reverse=True) else: teams = teams.sort(lambda team: team.team_order) print round_num for team in teams: pick_num = pick_num + 1 pick_id = db.pick.insert(pick_num=pick_num, pick_round=round_num, pick_team=team) session.flash = "Draft Board is ready for %s teams and %s rounds!" % (len(teams), rounds) redirect(URL('appadmin', 'manage/league')) ##table-pick def manage_draft(): #picks = db(db.pick).select(orderby=db.pick.id) fields = [db.pick.id, db.pick.pick_num, db.pick.pick_round, db.pick.pick_team, db.pick.pick_owner, db.pick.pick_player, db.player.player_first_name, db.player.player_last_name, db.player.player_position, db.player.player_pro_team] picks = db(db.pick).select(*fields, left=[db.player.on(db.pick.pick_player == db.player.id)]) teams = db(db.team).select(orderby=db.team.team_order) rounds = int(db(db.draft_parameter.draft_parameter == 'Rounds').select().first()['draft_parameter_value']) if request.get_vars['pick'] is None: request.get_vars['pick'] = 1 pick_id = db(db.pick.pick_num == request.get_vars['pick']).select().first().id form=crud.update(db.pick, pick_id, deletable=False) if form.accepts(request.vars, session): redirect(URL('manage_draft', vars=dict(pick=int(request.get_vars['pick'])+1))) if (int(request.get_vars['pick'])<11): pick_time = 180 elif (int(request.get_vars['pick'])<131): pick_time = 90 else: pick_time = 30 return dict(picks=picks, teams=teams, rounds=rounds, form=form, pick_time=pick_time) def trade_pick(): return dict(message=T('Hello World')) def trade_player(): return dict(message=T('Hello World'))
[ "Drew@drews-mbp.bellerhome" ]
Drew@drews-mbp.bellerhome