blob_id
stringlengths
40
40
directory_id
stringlengths
40
40
path
stringlengths
3
281
content_id
stringlengths
40
40
detected_licenses
listlengths
0
57
license_type
stringclasses
2 values
repo_name
stringlengths
6
116
snapshot_id
stringlengths
40
40
revision_id
stringlengths
40
40
branch_name
stringclasses
313 values
visit_date
timestamp[us]
revision_date
timestamp[us]
committer_date
timestamp[us]
github_id
int64
18.2k
668M
star_events_count
int64
0
102k
fork_events_count
int64
0
38.2k
gha_license_id
stringclasses
17 values
gha_event_created_at
timestamp[us]
gha_created_at
timestamp[us]
gha_language
stringclasses
107 values
src_encoding
stringclasses
20 values
language
stringclasses
1 value
is_vendor
bool
2 classes
is_generated
bool
2 classes
length_bytes
int64
4
6.02M
extension
stringclasses
78 values
content
stringlengths
2
6.02M
authors
listlengths
1
1
author
stringlengths
0
175
6a4054dd06199db6b9309cb200c08c31c28b270c
4d861accd86c2c4fbffeb75e963aa4c099d6dd0a
/Practica5/EjercicioB.py
6803cb510d11ca5901c54ad48f76ddf90b878768
[]
no_license
adwolfart/ModeladoyP
98e7118b600815587ae0840134f84eaab2475a14
7bab5fb3a837c14d80077641dc9be91f325b5d9a
refs/heads/master
2020-07-12T08:08:34.563441
2019-11-15T20:42:17
2019-11-15T20:42:17
204,761,474
0
0
null
null
null
null
UTF-8
Python
false
false
10,988
py
import threading from threading import Thread,Semaphore import time #semaforo para manipular los problemas de concurrencia semaforo = Semaphore(1); #Crear variable semáforo #listas que son utilizadas como bandas transportadoras listaPiezas = [] listaPiezasEnsamblador = [] listaPiezasEnsamblador1 = [] listaPaqueteFinalizados = [] mataHilos = True #se usa para verificar si se esta ejecutando un hilo en algun momento entro = False #cuenta las piezas que se van creando contadorPiezas = 0 #clase que modela los paquetes class Paquete(object): def __init__(self, tipo, listaPiezasPaquete): self.tipo = tipo self.listaPiezasPaquete = listaPiezasPaquete milisegundos = int(round(time.time() * 1000)) self.identificador = tipo+str(milisegundos) def getTipo(self): return self.tipo def getListaPiezasPaquete(self): return self.listaPiezasPaquete def getIdentificador(self): return self.identificador #clase que modela las piezas class Pieza(object): def __init__(self, tipo): self.tipo = tipo milisegundos = int(round(time.time() * 1000)) self.identificador = tipo+str(milisegundos) def getTipoPieza(self): return self.tipo def getIdentificador(self): return self.identificador #clase que modela al productor de piezas. se puede crear varias piezas y dependiendo de la pieza la guarda en el #lugar correspondiente class Productor(threading.Thread): def __init__(self, tipoPieza, tiempo): self.tiempo = tiempo self.tipoPieza = tipoPieza threading.Thread.__init__(self) def crearPieza(self): nuevaPieza = Pieza(self.tipoPieza) if revisaPiezasCreadas(): global mataHilos mataHilos = False time.sleep(self.tiempo) return nuevaPieza #funcion que guarda la pieza en la banda correspondiente depende de la pieza que es creada def run (self): global contadorPiezas while mataHilos: pieza = self.crearPieza() contadorPiezas = contadorPiezas + 1 if pieza.getTipoPieza() == "C": listaPiezasEnsamblador.append(pieza) elif pieza.getTipoPieza() == "D": if entro: semaforo.acquire(); listaPiezasEnsamblador1.append(pieza) else: listaPiezas.append(pieza) print("Se crea la pieza: "+str(contadorPiezas)+ " con id: "+pieza.getIdentificador()) def getTiempo(self): return self.tiempo #clase que modela a Ensamblador class Ensamblador(threading.Thread): def __init__(self, tipoPieza, tiempo, listaCantTipo): self.tipoPieza = tipoPieza self.tiempo = tiempo self.listaCantTipo = listaCantTipo self.listaPiezasRecolectadas = [] threading.Thread.__init__(self) #funcion que crea piezas cada cierto tiempo. crear las piezas si en la banda hay las piezas correspondientes def crearPiezaEnsamblador(self, i): global contadorPiezas cantidadTotalPiezas = 0 for np in self.listaCantTipo: cantidadTotalPiezas = cantidadTotalPiezas + np[1] tupla = self.listaCantTipo[i] tipoPiezaS = tupla[0] cantidadPiezasS = tupla[1] j = 0 while j < len(listaPiezas): piezaP = listaPiezas[j] if tipoPiezaS == piezaP.getTipoPieza(): cuantasPiezas = 0 for k in self.listaPiezasRecolectadas: if k.getTipoPieza() == piezaP.getTipoPieza(): cuantasPiezas = cuantasPiezas + 1 if cuantasPiezas < cantidadPiezasS: self.listaPiezasRecolectadas.append(piezaP) del listaPiezas[j] break j = j + 1 if len(self.listaPiezasRecolectadas)==cantidadTotalPiezas: nuevaPieza = Pieza(self.tipoPieza) listaPiezasEnsamblador.append(nuevaPieza) contadorPiezas = contadorPiezas + 1 print("Se crea la pieza: "+str(contadorPiezas)+" con id: "+nuevaPieza.getIdentificador()) del self.listaPiezasRecolectadas[0:len(self.listaPiezasRecolectadas)] if revisaPiezasCreadas(): global mataHilos mataHilos = False time.sleep(self.tiempo) #funcion que se ejecuta hasta que se cunple la condicion def run (self): i = 0 while mataHilos: if i == len(self.listaCantTipo): i = 0 self.crearPiezaEnsamblador(i) i = i + 1 def getTiempo(self): return self.tiempo #clase que modela el segundo Ensamblador que hay en la mecanismo class Ensamblador1(threading.Thread): def __init__(self, tipoPieza, tiempo, listaCantTipo): self.tipoPieza = tipoPieza self.tiempo = tiempo self.listaCantTipo = listaCantTipo self.listaPiezasRecolectadas = [] threading.Thread.__init__(self) #funcion que crea piezas cada cierto tiempo. crear las piezas si en la banda hay las piezas correspondientes def crearPiezaEnsamblador(self, i): global contadorPiezas cantidadTotalPiezas = 0 for np in self.listaCantTipo: cantidadTotalPiezas = cantidadTotalPiezas + np[1] tupla = self.listaCantTipo[i] tipoPiezaS = tupla[0] cantidadPiezasS = tupla[1] j = 0 while j < len(listaPiezasEnsamblador): piezaP = listaPiezasEnsamblador[j] if tipoPiezaS == piezaP.getTipoPieza(): cuantasPiezas = 0 for k in self.listaPiezasRecolectadas: if k.getTipoPieza() == piezaP.getTipoPieza(): cuantasPiezas = cuantasPiezas + 1 if cuantasPiezas < cantidadPiezasS: self.listaPiezasRecolectadas.append(piezaP) del listaPiezasEnsamblador[j] break j = j + 1 if len(self.listaPiezasRecolectadas)==cantidadTotalPiezas: nuevaPieza = Pieza(self.tipoPieza) if entro: semaforo.acquire() listaPiezasEnsamblador1.append(nuevaPieza) contadorPiezas = contadorPiezas + 1 print("Se crea la pieza: "+str(contadorPiezas)+ " con id: "+nuevaPieza.getIdentificador()) del self.listaPiezasRecolectadas[0:len(self.listaPiezasRecolectadas)] if revisaPiezasCreadas(): global mataHilos mataHilos = False time.sleep(self.tiempo) #funcion que se ejecuta hasta que se cunple la condicion def run (self): i = 0 while mataHilos: if i == len(self.listaCantTipo): i = 0 self.crearPiezaEnsamblador(i) i = i + 1 def getTiempo(self): return self.tiempo #clase que modela a un empaquetador. class Empaquetador(threading.Thread): def __init__(self, tipoPieza, tipoPiezaEntrada, tiempo, tamPaquete): self.tiempo = tiempo self.tipoPieza = tipoPieza self.tamPaquete = tamPaquete self.listaPiezasPaquete = [] self.tipoPiezaEntrada = tipoPiezaEntrada threading.Thread.__init__(self) #crea paquetes dependiendo de las especificaciones def crearPaquete(self): global entro i = 0 entro = True while i < len(listaPiezasEnsamblador1): piezaP = listaPiezasEnsamblador1[i] if piezaP.getTipoPieza() == self.tipoPiezaEntrada: self.listaPiezasPaquete.append(piezaP) del listaPiezasEnsamblador1[i] if len(self.listaPiezasPaquete) == self.tamPaquete: nuevoPaquete = Paquete(self.tipoPieza, self.listaPiezasPaquete) print("Se creo el paquete:"+str(nuevoPaquete.getIdentificador())) listaPaqueteFinalizados.append(nuevoPaquete) del self.listaPiezasPaquete[0:len(self.listaPiezasPaquete)] if revisaPiezasCreadas(): global mataHilos mataHilos = False i = i + 1 entro = False semaforo.release() #funcion que se ejecuta hasta que se cumple la condicion de finalizar def run (self): while mataHilos: self.crearPaquete() #clase que modela a un empaquetador. class Empaquetador1(threading.Thread): def __init__(self, tipoPieza, tipoPiezaEntrada, tiempo, tamPaquete): self.tiempo = tiempo self.tipoPieza = tipoPieza self.tamPaquete = tamPaquete self.listaPiezasPaquete = [] self.tipoPiezaEntrada = tipoPiezaEntrada threading.Thread.__init__(self) #crea paquetes dependiendo de las especificaciones def crearPaquete(self): global entro entro = True i = 0 while i < len(listaPiezasEnsamblador1): piezaP = listaPiezasEnsamblador1[i] if piezaP.getTipoPieza() == self.tipoPiezaEntrada: self.listaPiezasPaquete.append(piezaP) del listaPiezasEnsamblador1[i] if len(self.listaPiezasPaquete) == self.tamPaquete: nuevoPaquete = Paquete(self.tipoPieza, self.listaPiezasPaquete) print("Se creo el paquete:"+str(nuevoPaquete.getIdentificador())) listaPaqueteFinalizados.append(nuevoPaquete) del self.listaPiezasPaquete[0:len(self.listaPiezasPaquete)] if revisaPiezasCreadas(): global mataHilos mataHilos = False i = i + 1 semaforo.release() entro = False def run (self): while mataHilos: self.crearPaquete() #funcion que se ejecuta hasta que se cumple la condicion de finalizar def revisaPiezasCreadas(): if contadorPiezas >= 200: return True else: return False #funcion que crea el ejercicio B def ejercicioA(): #productor de tipos A,B,C y D. se crean cada 2 segundos p = Productor("A", 2) p1 = Productor("B", 2) p2 = Productor("C", 2) p3 = Productor("D", 2) #Ensambladores de tipo E y F, se crean paquetes cada 2 segundos, las listas representan los tipos de piezas #y la cantidad que necesita ensamblador1 = Ensamblador("E", 2, [("A", 2), ("B", 2)]) ensamblador2 = Ensamblador1("F", 2, [("C", 2), ("E", 1)]) #Empaquetador que crea paquetes de tipo P y Q, necesita piezas de tipo F y D, los crea cada 0 sengundos y #para crear un paquete necesita 5 y 10 piezas empaquetador1 = Empaquetador("P", "F", 0, 5) empaquetador2 = Empaquetador1("Q", "D", 0, 10) #inicia los hilos p.start() p1.start() p2.start() p3.start() ensamblador1.start() ensamblador2.start() empaquetador1.start() empaquetador2.start() ejercicioA()
[ "lopo@lopo-pc" ]
lopo@lopo-pc
d355e03c44e9a7c2bf59b7e5d70f274fd921714e
e842b091e4d7236b2136ac906eb5108ef17bd8d9
/16_DynamicProgramming1/Step11/ys_11.py
737b22c9bb16ef5528ca6bf1c909b13559fda3f1
[ "MIT" ]
permissive
StudyForCoding/BEAKJOON
6eb4bfe9a302e29893b128eebe44e8db271546c8
84e1c5e463255e919ccf6b6a782978c205420dbf
refs/heads/master
2023-04-04T09:17:58.145493
2021-04-14T07:05:06
2021-04-14T07:05:06
266,000,040
0
0
MIT
2021-04-14T07:05:07
2020-05-22T02:30:13
Python
UTF-8
Python
false
false
563
py
N = int(input()) #수열개수 arr = input() arr = arr.split() arr= [int(i) for i in arr] #입력받은 수열배열 d = [0] * (N)#가장 긴 증가하는 부분수열 길이 max1 = 0 for i in range(0, N): min1 = 0 for j in range(0, i): if arr[i] > arr[j]: #앞수열보다 큰 수가 있으면 if min1 < d[j]: #현재 인덱스요소보다 작은 수 카운트 min1 = d[j] d[i] = min1 + 1 if max1 < d[i]: max1 = d[i] #d배열중 최대 값이 가장 긴 증가하는 부분수열이 print(max1)
[ "69625013+to06109@users.noreply.github.com" ]
69625013+to06109@users.noreply.github.com
5184f3b5a7ed92581d910e979f58f31d1a589646
cad83699bb72bc1d560815d5b1c34fc9b371f163
/cartridge/project_template/manage.py
4330dc2e1a9c13927f25c69c19985527dbbe4267
[ "BSD-3-Clause" ]
permissive
BeUnique/cartridge
a1903ecc78029a576e57c3832b16357c7661b1b8
e5b887fed96d01ab93237f345fc420bcbe56d027
refs/heads/master
2021-01-19T06:00:01.090292
2011-08-31T01:04:33
2011-08-31T01:04:33
null
0
0
null
null
null
null
UTF-8
Python
false
false
1,284
py
#!/usr/bin/env python # When project_template is used as the actual project during Mezzanine # development, insert the development path into sys.path so that the # development version of Mezzanine is used rather than the installed version. import os import sys project_path = os.path.dirname(os.path.abspath(__file__)) project_dir = project_path.split(os.sep)[-1] if project_dir == "project_template": dev_path = os.path.abspath(os.path.join(project_path, "..", "..")) if dev_path not in sys.path: sys.path.insert(0, dev_path) import cartridge cartridge_path = os.path.dirname(os.path.abspath(cartridge.__file__)) assert os.path.abspath(os.path.join(cartridge_path, "..")) == dev_path from django.core.management import execute_manager try: import settings # Assumed to be in the same directory. except ImportError: import sys sys.stderr.write("Error: Can't find the file 'settings.py' in the " "directory containing %r. It appears you've customized things.\n" "You'll have to run django-admin.py, passing it your settings module.\n" "(If the file settings.py does indeed exist, it's causing an " "ImportError somehow.)\n" % __file__) sys.exit(1) if __name__ == "__main__": execute_manager(settings)
[ "steve@jupo.org" ]
steve@jupo.org
523fb7f3aa1c456f2a0e4aa859cb776f8d3f2534
bcb47f8acc6530e48c7a013d459e453c2e5110c0
/urls.py
16f8f43a8d797fad08220b99def350e573c300ad
[]
no_license
gebisys/libreapp
09cef76e0635eb0fe1803769cba07b4320c29fc9
14dede79e47c7a0e0d9c492c5d87df9123ef59ec
refs/heads/master
2016-09-06T08:06:55.609876
2013-06-12T14:47:18
2013-06-12T14:47:18
null
0
0
null
null
null
null
UTF-8
Python
false
false
766
py
from django.conf.urls.defaults import * from django.views.generic.simple import direct_to_template from django.contrib import admin from django.conf import settings from news.feeds import ArchiveFeed from django.contrib import admin admin.autodiscover() urlpatterns = patterns('', url(r'^$', 'news.views.index_view'), url(r"^post/(?P<year>\d{4})/(?P<mounth>\d{1,2})/(?P<day>\d{2})/(?P<slug>[\w-]+)$", 'news.views.post'), url(r'^feed/$', ArchiveFeed()), # archivos estaticos url(r'^static/(?P<path>.*)$', 'django.views.static.serve', { 'document_root': settings.STATIC_ROOT }), url(r'^user/$', 'news.views.createuser'), # Uncomment the next line to enable the admin: url(r'^admin/', include(admin.site.urls)), )
[ "gebisys@gmail.com" ]
gebisys@gmail.com
8ffe18e059c8aa73c2d15b8bc652cbaba00d8265
2236b86215a6c6895fbb9bc10451497121f70c0a
/rate_prediction/apv25buffering/config.py
054bb4e5bd3c5ac10362fdc27da32e6ed7a72133
[]
no_license
perhansson/burst-trigger-toy
b754a5462e175e8098ccb08f1187b32697cf6679
6e49bee473e761c36abd8f73c7c9474a40b9e109
refs/heads/master
2021-01-19T22:02:02.686562
2015-10-28T20:42:10
2015-10-28T20:42:10
35,178,918
0
0
null
null
null
null
UTF-8
Python
false
false
47
py
debug = False batch = True Nperiods = 10000000
[ "per.hansson80@gmail.com" ]
per.hansson80@gmail.com
64764f6b4a24d45418ef96845068613a38b12ef6
921e479e9fe4f72c0621207b262f7464b6812d9b
/ants.py
c2ba7b96cbf5bdf1c537ec32ec1f333bdb5b6ce0
[]
no_license
ajai-sharma-backup/ants
af60fc671ffe7c9eede2002189906e6095defc96
9df7ebc8907132415d088302e9a9fe2f0bca2116
refs/heads/master
2021-06-13T17:52:18.431694
2015-06-02T21:17:37
2015-06-02T21:17:37
null
0
0
null
null
null
null
UTF-8
Python
false
false
21,763
py
"""The ants module implements game logic for Ants Vs. SomeBees.""" # Name: Ajai Sharma # Email: ajai.sharma@gmail.com import random import sys from ucb import main, interact, trace from collections import OrderedDict ################ # Core Classes # ################ class Place: """A Place holds insects and has an exit to another Place.""" def __init__(self, name, exit=None): """Create a Place with the given exit. name -- A string; the name of this Place. exit -- The Place reached by exiting this Place (may be None). """ self.name = name self.exit = exit self.bees = [] # A list of Bees self.ant = None # An Ant self.entrance = None # A Place # Phase 1: Add an entrance to the exit if self.exit is not None: self.exit.entrance = self def add_insect(self, insect): """Add an Insect to this Place. There can be at most one Ant in a Place, unless exactly one of them is a BodyguardAnt (Phase 2), in which case there can be two. If add_insect tries to add more Ants than is allowed, an assertion error is raised. There can be any number of Bees in a Place. """ if insect.is_ant(): # Phase 2: Special handling for BodyguardAnt if self.ant: if self.ant.can_contain(insect): self.ant.contain_ant(insect) elif insect.can_contain(self.ant): insect.contain_ant(self.ant) self.ant = insect else: assert self.ant is None, 'Two ants in {0}'.format(self) else: self.ant = insect else: self.bees.append(insect) insect.place = self def remove_insect(self, insect): """Remove an Insect from this Place.""" if not insect.is_ant(): self.bees.remove(insect) elif type(insect) == QueenAnt and insect.armor > 0: return else: assert self.ant == insect, '{0} is not in {1}'.format(insect, self) if self.ant.container and self.ant.ant: self.ant = self.ant.ant else: self.ant = None insect.place = None def __str__(self): return self.name class Insect: """An Insect, the base class of Ant and Bee, has armor and a Place.""" watersafe = False def __init__(self, armor, place=None): """Create an Insect with an armor amount and a starting Place.""" self.armor = armor self.place = place # set by Place.add_insect and Place.remove_insect def reduce_armor(self, amount): """Reduce armor by amount, and remove the insect from its place if it has no armor remaining. >>> test_insect = Insect(5) >>> test_insect.reduce_armor(2) >>> test_insect.armor 3 """ self.armor -= amount if self.armor <= 0: game_print('{0} ran out of armor and expired'.format(self)) self.place.remove_insect(self) def action(self, colony): """Perform the default action that this Insect takes each turn. colony -- The AntColony, used to access game state information. """ def is_ant(self): """Return whether this Insect is an Ant.""" return False def __repr__(self): cname = type(self).__name__ return '{0}({1}, {2})'.format(cname, self.armor, self.place) class Bee(Insect): """A Bee moves from place to place, following exits and stinging ants.""" name = 'Bee' watersafe = True def sting(self, ant): """Attack an Ant, reducing the Ant's armor by 1.""" ant.reduce_armor(1) def move_to(self, place): """Move from the Bee's current Place to a new Place.""" self.place.remove_insect(self) place.add_insect(self) def blocked(self): """Return True if this Bee cannot advance to the next Place.""" # Phase 2: Special handling for NinjaAnt return self.place.ant and self.place.ant.blocks_path def action(self, colony): """A Bee's action stings the Ant that blocks its exit if it is blocked, or moves to the exit of its current place otherwise. colony -- The AntColony, used to access game state information. """ if self.blocked(): self.sting(self.place.ant) elif self.place is not colony.hive and self.armor > 0: self.move_to(self.place.exit) class Ant(Insect): """An Ant occupies a place and does work for the colony.""" implemented = False # Only implemented Ant classes should be instantiated damage = 0 food_cost = 0 container = False blocks_path = True def __init__(self, armor=1): """Create an Ant with an armor quantity.""" Insect.__init__(self, armor) def is_ant(self): return True def can_contain(self, other): return self.container and self.ant is None and not other.container class HarvesterAnt(Ant): """HarvesterAnt produces 1 additional food per turn for the colony.""" name = 'Harvester' food_cost = 2 implemented = True def action(self, colony): """Produce 1 additional food for the colony. colony -- The AntColony, used to access game state information. """ colony.food += 1 def random_or_none(l): """Return a random element of list l, or return None if l is empty.""" return random.choice(l) if l else None class ThrowerAnt(Ant): """ThrowerAnt throws a leaf each turn at the nearest Bee in its range.""" name = 'Thrower' implemented = True damage = 1 food_cost = 4 min_range = 0 max_range = 10 def nearest_bee(self, hive): """Return the nearest Bee in a Place that is not the Hive, connected to the ThrowerAnt's Place by following entrances. This method returns None if there is no such Bee. Problem B5: This method returns None if there is no Bee in range. """ current_place = self.place distance = 0 while current_place != hive and distance <= self.max_range: if current_place.bees and distance >= self.min_range: return random_or_none(current_place.bees) current_place = current_place.entrance distance += 1 def throw_at(self, target): """Throw a leaf at the target Bee, reducing its armor.""" if target is not None: target.reduce_armor(self.damage) def action(self, colony): """Throw a leaf at the nearest Bee in range.""" self.throw_at(self.nearest_bee(colony.hive)) class Hive(Place): """The Place from which the Bees launch their assault. assault_plan -- An AssaultPlan; when & where bees enter the colony. """ def __init__(self, assault_plan): self.name = 'Hive' self.assault_plan = assault_plan self.bees = [] for bee in assault_plan.all_bees: self.add_insect(bee) # The following attributes are always None for a Hive self.entrance = None self.ant = None self.exit = None def strategy(self, colony): exits = [p for p in colony.places.values() if p.entrance is self] for bee in self.assault_plan.get(colony.time, []): bee.move_to(random.choice(exits)) class AntColony: """An ant collective that manages global game state and simulates time. Attributes: time -- elapsed time food -- the colony's available food total queen -- the place where the queen resides places -- A list of all places in the colony (including a Hive) bee_entrances -- A list of places that bees can enter """ def __init__(self, strategy, hive, ant_types, create_places, food=2): """Create an AntColony for simulating a game. Arguments: strategy -- a function to deploy ants to places hive -- a Hive full of bees ant_types -- a list of ant constructors create_places -- a function that creates the set of places """ self.time = 0 self.food = food self.strategy = strategy self.hive = hive self.ant_types = OrderedDict((a.name, a) for a in ant_types) self.configure(hive, create_places) def configure(self, hive, create_places): """Configure the places in the colony.""" self.queen = Place('AntQueen') self.places = OrderedDict() self.bee_entrances = [] def register_place(place, is_bee_entrance): self.places[place.name] = place if is_bee_entrance: place.entrance = hive self.bee_entrances.append(place) register_place(self.hive, False) create_places(self.queen, register_place) def simulate(self): """Simulate an attack on the ant colony (i.e., play the game).""" while len(self.queen.bees) == 0 and len(self.bees) > 0: self.hive.strategy(self) # Bees invade self.strategy(self) # Ants deploy for ant in self.ants: # Ants take actions if ant.armor > 0: ant.action(self) for bee in self.bees: # Bees take actions if bee.armor > 0: bee.action(self) self.time += 1 if len(self.queen.bees) > 0: game_print('The ant queen has perished. Please try again.') else: game_print('All bees are vanquished. You win!') def deploy_ant(self, place_name, ant_type_name): """Place an ant if enough food is available. This method is called by the current strategy to deploy ants. """ constructor = self.ant_types[ant_type_name] if self.food < constructor.food_cost: game_print('Not enough food remains to place ' + ant_type_name) else: self.places[place_name].add_insect(constructor()) self.food -= constructor.food_cost def remove_ant(self, place_name): """Remove an Ant from the Colony.""" place = self.places[place_name] if place.ant is not None: place.remove_insect(place.ant) @property def ants(self): return [p.ant for p in self.places.values() if p.ant is not None] @property def bees(self): return [b for p in self.places.values() for b in p.bees] @property def insects(self): return self.ants + self.bees def __str__(self): status = ' (Food: {0}, Time: {1})'.format(self.food, self.time) return str([str(i) for i in self.ants + self.bees]) + status def ant_types(): """Return a list of all implemented Ant classes.""" all_ant_types = [] new_types = [Ant] while new_types: new_types = [t for c in new_types for t in c.__subclasses__()] all_ant_types.extend(new_types) return [t for t in all_ant_types if t.implemented] def interactive_strategy(colony): """A strategy that starts an interactive session and lets the user make changes to the colony. For example, one might deploy a ThrowerAnt to the first tunnel by invoking: colony.deploy_ant('tunnel_0_0', 'Thrower') """ game_print('colony: ' + str(colony)) msg = '<Control>-D (<Control>-Z <Enter> on Windows) completes a turn.\n' interact(msg) def start_with_strategy(args, strategy): """Reads command-line arguments and starts Ants vs. SomeBees with those options.""" import argparse parser = argparse.ArgumentParser(description="Play Ants vs. SomeBees") parser.add_argument('-t', '--ten', action='store_true', help='start with ten food') parser.add_argument('-f', '--full', action='store_true', help='loads a full layout and assault plan') parser.add_argument('-w', '--water', action='store_true', help='loads a full layout with water') parser.add_argument('-i', '--insane', action='store_true', help='loads a difficult assault plan') args = parser.parse_args() assault_plan = make_test_assault_plan() layout = test_layout food = 2 if args.ten: food = 10 if args.full: assault_plan = make_full_assault_plan() layout = dry_layout if args.water: layout = mixed_layout if args.insane: assault_plan = make_insane_assault_plan() hive = Hive(assault_plan) AntColony(strategy, hive, ant_types(), layout, food).simulate() ########### # Layouts # ########### def mixed_layout(queen, register_place, length=8, tunnels=3, moat_frequency=3): """Register Places with the colony.""" for tunnel in range(tunnels): exit = queen for step in range(length): if moat_frequency != 0 and (step + 1) % moat_frequency == 0: exit = Water('water_{0}_{1}'.format(tunnel, step), exit) else: exit = Place('tunnel_{0}_{1}'.format(tunnel, step), exit) register_place(exit, step == length - 1) def test_layout(queen, register_place, length=8, tunnels=1): mixed_layout(queen, register_place, length, tunnels, 0) def dry_layout(queen, register_place, length=8, tunnels=3): mixed_layout(queen, register_place, length, tunnels, 0) ################# # Assault Plans # ################# class AssaultPlan(dict): """The Bees' plan of attack for the Colony. Attacks come in timed waves. An AssaultPlan is a dictionary from times (int) to waves (list of Bees). >>> AssaultPlan().add_wave(4, 2) {4: [Bee(3, None), Bee(3, None)]} """ def __init__(self, bee_armor=3): self.bee_armor = bee_armor def add_wave(self, time, count): """Add a wave at time with count Bees that have the specified armor.""" bees = [Bee(self.bee_armor) for _ in range(count)] self.setdefault(time, []).extend(bees) return self @property def all_bees(self): """Place all Bees in the hive and return the list of Bees.""" return [bee for wave in self.values() for bee in wave] def make_test_assault_plan(): return AssaultPlan().add_wave(2, 1).add_wave(3, 1) def make_full_assault_plan(): plan = AssaultPlan().add_wave(2, 1) for time in range(3, 15, 2): plan.add_wave(time, 1) return plan.add_wave(15, 8) def make_insane_assault_plan(): plan = AssaultPlan(4).add_wave(1, 2) for time in range(3, 15): plan.add_wave(time, 1) return plan.add_wave(15, 20) ############## # Extensions # ############## class Water(Place): """Water is a place that can only hold 'watersafe' insects.""" def add_insect(self, insect): """Add insect if it is watersafe, otherwise reduce its armor to 0.""" game_print('added', insect, insect.watersafe) super().add_insect(insect) if not insect.watersafe: insect.reduce_armor(insect.armor) class FireAnt(Ant): """FireAnt cooks any Bee in its Place when it expires.""" name = 'Fire' damage = 3 food_cost = 4 implemented = True def reduce_armor(self, amount): if amount >= self.armor: for bee in self.place.bees[:]: bee.reduce_armor(self.damage) super().reduce_armor(amount) class LongThrower(ThrowerAnt): """A ThrowerAnt that only throws leaves at Bees at least 4 places away.""" name = 'Long' food_cost = 3 min_range = 4 implemented = True class ShortThrower(ThrowerAnt): """A ThrowerAnt that only throws leaves at Bees within 3 places.""" name = 'Short' food_cost = 3 max_range = 2 implemented = True class WallAnt(Ant): """WallAnt is an Ant which has a large amount of armor.""" name = 'Wall' food_cost = 4 implemented = True def __init__(self): Ant.__init__(self) self.armor = 4 class NinjaAnt(Ant): """NinjaAnt is an Ant which does not block the path and does 1 damage to all Bees in the exact same Place.""" name = 'Ninja' food_cost = 6 blocks_path = False damage = 1 implemented = True def action(self, colony): #list(map(lambda bee: getattr(bee, 'reduce_armor')(self.damage), self.place.bees)) print(self.place.bees) for bee in self.place.bees[:]: bee.reduce_armor(self.damage) class ScubaThrower(ThrowerAnt): """ScubaThrower is a ThrowerAnt which is watersafe.""" name = 'Scuba' food_cost = 5 watersafe = True implemented = True class HungryAnt(Ant): """HungryAnt will take three "turns" to eat a Bee in the same space as it. While eating, the HungryAnt can't eat another Bee. """ name = 'Hungry' food_cost = 4 time_to_digest = 3 implemented = True def __init__(self): Ant.__init__(self) self.digesting = 0 def eat_bee(self, bee): bee.reduce_armor(bee.armor) self.digesting = self.time_to_digest def action(self, colony): if self.digesting > 0: self.digesting -= 1 elif self.place.bees: self.eat_bee(random_or_none(self.place.bees)) class BodyguardAnt(Ant): """BodyguardAnt provides protection to other Ants.""" name = 'Bodyguard' food_cost = 4 armor = 2 container = True implemented = True def __init__(self): Ant.__init__(self, 2) self.ant = None # The Ant hidden in this bodyguard def contain_ant(self, ant): self.ant = ant def action(self, colony): if self.ant: self.ant.action(colony) class QueenPlace: """A place that represents both places in which the bees find the queen. (1) The original colony queen location at the end of all tunnels, and (2) The place in which the QueenAnt resides. """ def __init__(self, colony_queen, ant_queen): self.colony_queen = colony_queen self.ant_queen = ant_queen @property def bees(self): return self.colony_queen.bees + self.ant_queen.bees class QueenAnt(ScubaThrower): """The Queen of the colony. The game is over if a bee enters her place.""" name = 'Queen' implemented = True food_cost = 6 imposter = False def __init__(self): ScubaThrower.__init__(self) self.doubled_ants = [self] if QueenAnt.imposter: self.armor = 0 QueenAnt.imposter = True def buff_ant(self, place): ant_list = [place.ant, place.ant.ant if place.ant and place.ant.container else None] for ant in ant_list: if ant and ant not in self.doubled_ants: self.doubled_ants.append(ant) ant.damage *= 2 def action(self, colony): """A queen ant throws a leaf, but also doubles the damage of ants in her tunnel. Impostor queens do only one thing: die.""" def iterate_over_tunnel(start, action, direction='both'): '''Performs an action at every place in the tunnel of the starting place''' action(start) if direction == 'both': for direction in ('entrance', 'exit'): next_place = getattr(start, direction) if next_place: iterate_over_tunnel(next_place, action, direction) else: next_place = getattr(start, direction) if next_place: iterate_over_tunnel(next_place, action, direction) if type(colony.queen) != QueenPlace: colony.queen = QueenPlace(colony.queen, self.place) if self.armor > 0: iterate_over_tunnel(self.place, self.buff_ant) super().action(colony) class AntRemover(Ant): """Allows the player to remove ants from the board in the GUI.""" name = 'Remover' implemented = True def __init__(self): Ant.__init__(self, 0) ################## # Status Effects # ################## def make_slow(action): """Return a new action method that calls action every other turn. action -- An action method of some Bee """ def slowed(colony): if colony.time % 2 == 0: action(colony) return slowed def make_stun(action): """Return a new action method that does nothing. action -- An action method of some Bee """ def stunned(colony): pass return stunned def apply_effect(effect, bee, duration): """Apply a status effect to a Bee that lasts for duration turns.""" timer = duration original = bee.action altered = effect(original) def timed_effect(colony): nonlocal timer if timer == 0: bee.action = original bee.action(colony) else: altered(colony) timer -= 1 bee.action = timed_effect class SlowThrower(ThrowerAnt): """ThrowerAnt that causes Slow on Bees.""" name = 'Slow' food_cost = 4 implemented = True def throw_at(self, target): if target: apply_effect(make_slow, target, 3) class StunThrower(ThrowerAnt): """ThrowerAnt that causes Stun on Bees.""" name = 'Stun' food_cost = 6 implemented = True def throw_at(self, target): if target: apply_effect(make_stun, target, 1) def game_print(*args, **kargs): print(*args, **kargs) @main def run(*args): start_with_strategy(args, interactive_strategy)
[ "ajai.sharma@gmail.com" ]
ajai.sharma@gmail.com
5278d0ebc39489eb80a4b0a82ecaf609f72027a7
ca7aa979e7059467e158830b76673f5b77a0f5a3
/Python_codes/p03813/s431513228.py
c4298e11e8a394fafce0121831f9fbfa51e6a6ab
[]
no_license
Aasthaengg/IBMdataset
7abb6cbcc4fb03ef5ca68ac64ba460c4a64f8901
f33f1c5c3b16d0ea8d1f5a7d479ad288bb3f48d8
refs/heads/main
2023-04-22T10:22:44.763102
2021-05-13T17:27:22
2021-05-13T17:27:22
367,112,348
0
0
null
null
null
null
UTF-8
Python
false
false
260
py
def read_int(): return int(input().strip()) def read_ints(): return list(map(int, input().strip().split(' '))) def solve(): x = read_int() if x < 1200: return 'ABC' return 'ARC' if __name__ == '__main__': print(solve())
[ "66529651+Aastha2104@users.noreply.github.com" ]
66529651+Aastha2104@users.noreply.github.com
5639c8a9787426db742cd1e8e2e6518dd45c03b3
6c83083e1ddff6028a004b6beb40921d3d723a29
/gofish2.py
0fb1af0b9d48b1d623a3735d78c095dd193e02cc
[]
no_license
pfigz/python-files-projects
62255d38ace41150df8c26831be3f631f713a802
6ff153dece345ed5d93de4c46df8f0158e634861
refs/heads/main
2023-07-17T08:27:47.491472
2021-09-09T14:19:37
2021-09-09T14:19:37
404,751,407
0
0
null
null
null
null
UTF-8
Python
false
false
3,251
py
#Go fish example 2 import random def MakeDeck(): deck = [] c = 0 values = ["Ace", "2", "3", "4", "5", "6", "7", "8", "9", "10", "Jack", "Queen", "King"] suits = ["Hearts", "Spades", "Diamonds", "Clubs"] for v in values: for s in suits: deck.append([v,s]) random.shuffle(deck) return deck import random, time fish_deck = deck.MakeDeck() #for i in fish_deck: # print(i[0]+" of "+i[1]) class fisherman(): name = "" hand = [] sets = [] def ask(player1, player2): pause = random.randint(2,5) has = [] choose = randint(0,len(player1.hand)-1) value = player1.hand[choose][0] for card in player2.hand: if card[0] == value: has.append(card) for card in has: player2.hand.remove(card) for card in has: player1.hand.append(card) return_string = player1.name+" asked "+player2.name+" for "+value+"s. " print(return_string) return_string = player2.name+" had "+str(len(has))+". " print(return_string) if len(has) == 0: draw(player1) return_string = player1.name+" had to go fish." print(return_string) def draw(player): card = fish_deck.pop() player.hand.append(card) def set_check(player): amount = {} for card in player.hand: if card[0] not in amount.keys(): amount[card[0]] = 1 if card[0] in amount.keys(): amount[card[0]] += 1 for count in amount.keys(): if amount[count] == 4: print(player.name+" got a set of "+count+"s.") player.sets.append(count) player.hand[:] = [card for card in player.hand if card[0] == count] john = fisherman() john.name = "John" tim = fisherman() tim.name = "Tim" sara = fisherman() sara.name = "Sara" kris = fisherman() kris.name = "Kris" def play(player1, player2, player3, player4, deck): turn = 0 size = 7 dealt = 0 order = [player1, player2, player3, player4] random.shuffle(order) while dealt < size: draw(order[0]) draw(order[1]) draw(order[2]) draw(order[3]) dealt += 1 while len(deck) != 0: for player in order: count = 0 hand = player.name+"'s hand: " for card in player.hand: if count < len(player.hand)-1: hand += card[0]+" of "+card[1]+", " count += 1 elif count == len(player.hand)-1: hand += card[0]+" of "+card[1]+"." print(hand) count = 0 sets = player.name+"'s sets: " for set in player.sets: if count < len(player.sets)-1: sets += set+"s, " elif count == len(player.sets)-1: sets += set+"s." print(sets) other_player = turn while other_player == turn: other_player = random.randint(0,3) ask(order[turn], order[other_player]) set_check(order[turn]) if turn >= 3: turn = 0 else: turn += 1 time.sleep(10) print("=========================================") play(john, tim, sara, kris, fish_deck)
[ "pfigdev@gmail.com" ]
pfigdev@gmail.com
1215d8edd938231f4d95f34270b4973a1bf63dbd
4db6ffdad89c6a3d1cf9731b815dbb7177b10043
/code/landcover_model/config.py
39913c695156dfd96525beffad1fcfbf51bf2308
[]
no_license
hugisveitti/thesis
55002c38ce58ad5b99dc07900e0a2121cbe096c0
a99a034fa684eae4d9c7f7cad391a521ff0525db
refs/heads/master
2023-07-27T05:05:16.535638
2021-08-15T19:26:43
2021-08-15T19:26:43
364,648,515
0
0
null
null
null
null
UTF-8
Python
false
false
411
py
import torch IMAGE_WIDTH = 256 IMAGE_HEIGHT = 256 num_classes = 9 tensor_type = torch.HalfTensor device = "cuda" cross_entropy_weights = torch.tensor([1.75827705e-05, 2.63985891e-02, 4.93802954e-01, 0.00000000e+00, 9.56997597e-02, 6.52813402e-02, 2.43301976e-01, 2.19168076e-02, 0.00000000e+00, 2.51651604e-02, 2.09771106e-02, 3.17152767e-03, 0.00000000e+00, 4.26719143e-03]).to(device)
[ "hugiholm1@gmail.com" ]
hugiholm1@gmail.com
9e2010ab225e7382d1bcaaae37d358d2bdd39500
042962d67c8742d01f66f254375b1ceb9136d182
/sensibleServer/__init__.py
b4047c4be8d5508123a5789445b9f26a212307ed
[]
no_license
Sensibility/sensibleServer
d8b0afefd846e31d454912f579067fc90526d046
a4125fa36e9a47e3ea05287ab6b54ca22110eb0d
refs/heads/master
2021-05-26T07:46:00.499527
2019-11-07T17:00:31
2019-11-07T17:00:31
127,950,558
1
0
null
null
null
null
UTF-8
Python
false
false
1,868
py
""" This package provides a simple http(s) server that serves content with optional cgi scripting available. Usage: sensibleServer DOCUMENT_ROOT [ -c --enable-cgi ] """ __author__ = "ocket8888" __version__ = "0.0.1" import argparse ROOT, CGI = None, None def main() -> int: """ Runs the program. """ global ROOT, CGI, __version__ parser = argparse.ArgumentParser(description="A simple http(s) server that serves content with optional cgi scripting available.") parser.add_argument("DOCUMENT_ROOT", type=str, help="The root of the web content to serve.") parser.add_argument("-c", "--enable-cgi", action="store_const", const=True, default=False, help="Enable the running of python scripts as Common Gateway Interface scripts.") parser.add_argument("-v", "--version", action="store_const", const=True, default=False, help="Print out the version number and exit.") args = parser.parse_args() if args.version: print(__version__) return 0 import os from sys import stderr ROOT = os.path.abspath(args.DOCUMENT_ROOT) if not os.path.isdir(ROOT): print("No such directory: '%s'" % ROOT, file=stderr) return 1 os.chdir(ROOT) CGI = args.enable_cgi if CGI: from . import server os.environ['DOCUMENT_ROOT'] = ROOT handler = server.CGIServer with Server.HTTPServer(("", 8000), handler) as server: print("Serving at port 8000") try: server.serve_forever() except KeyboardInterrupt: pass else: from .server import Server, RequestHandler with Server(("", 8080), RequestHandler) as server: print("Serving at port 8080", file=stderr) if __debug__: from time import time print("II:", time(), "Directory Contents:", ' '.join(i for i in os.listdir(ROOT)), sep='\t', file=stderr) try: server.serve_forever() except KeyboardInterrupt: pass return 0 if __name__ == '__main__': exit(main())
[ "ocket8888@gmail.com" ]
ocket8888@gmail.com
5a7644e93929e110f2439ee13de2a41444e4af81
bd749eca9a6aa414745bf13106c948fb2794d72b
/blog_app/migrations/0002_auto_20190910_1519.py
eb89b3692ac1da5756d047aa055b459d93bc40d4
[]
no_license
DreamInit/my_blog
115b07b40a32fc1d50fb642281a501dcce1ef46c
1309664c9504a93f36cbf7bbf30cc87cb096608d
refs/heads/master
2022-11-03T09:28:26.098193
2019-09-11T11:54:56
2019-09-11T11:54:56
207,740,133
0
1
null
2022-10-09T12:20:34
2019-09-11T06:33:17
Python
UTF-8
Python
false
false
340
py
# Generated by Django 2.0 on 2019-09-10 07:19 from django.db import migrations class Migration(migrations.Migration): dependencies = [ ('blog_app', '0001_initial'), ] operations = [ migrations.AlterModelOptions( name='blog', options={'ordering': ['-created_time']}, ), ]
[ "chuan.huang@scaleflux.com" ]
chuan.huang@scaleflux.com
3bdc87a55f1a198db75b03987f4bf990b86d2541
363245bca6658463411aba57e89f96bc6cb9434e
/tests/test_gamesession.py
3733e371c58ac5a2dadaa0606f0521c2b6e4e054
[ "MIT" ]
permissive
daleecoding/HQ-Trivia-Homework
75b5de2e955e08eb51b82c64cd9b56e9dbe85c79
c41e014458539bf9b3a0cc9fb360ececb8a029a3
refs/heads/master
2022-12-22T05:49:15.754811
2020-08-04T12:49:32
2020-08-04T12:49:32
302,655,081
0
0
null
null
null
null
UTF-8
Python
false
false
12,432
py
import asyncio import unittest from unittest.mock import AsyncMock from unittest.mock import MagicMock from unittest.mock import call from unittest.mock import patch from websockets.exceptions import ConnectionClosed import hqtrivia.config as config from hqtrivia.gamesession import GameSession from hqtrivia.messages import * from hqtrivia.question import Question from hqtrivia.player import Player class GameSessionTest(unittest.TestCase): def setUp(self): # Set to a shorter time for faster test self.original_round_duration = config.CONFIG_ROUND_DURATION config.CONFIG_ROUND_DURATION = min(2, config.CONFIG_ROUND_DURATION) def tearDown(self): config.CONFIG_ROUND_DURATION = self.original_round_duration def get_mocked_player(self, answer): """Helper method to get a mocked up player for use """ player = MagicMock() send_announcement = AsyncMock() player.attach_mock(send_announcement, 'send_announcement') send_question = AsyncMock() player.attach_mock(send_question, 'send_question') send_answers = AsyncMock() player.attach_mock(send_answers, 'send_answers') recv_answer = AsyncMock() player.attach_mock(recv_answer, 'recv_answer') recv_answer.return_value = answer return (player, send_announcement, send_question, send_answers, recv_answer) def test_run(self): """Tests that run() method calls execute_next_round() only once if it returns False. """ game = GameSession(0, []) game.abort_game = AsyncMock() game.execute_next_round = AsyncMock() # First time returns True. On second time returns False. game.execute_next_round.side_effect = [True, False] asyncio.run(game.run()) game.abort_game.assert_not_called() game.execute_next_round.assert_has_calls([call(), call()]) def test_run_when_exception_thrown(self): """Tests that run() method calls execute_next_round() only once if it returns False. """ self.assertRaises(Exception, GameSession.run) game = GameSession(0, []) game.abort_game = AsyncMock() game.execute_next_round = AsyncMock( return_value=True, side_effect=Exception) with self.assertRaises(Exception): asyncio.run(game.run()) game.abort_game.assert_called_once() game.execute_next_round.assert_called_once() def test_abort_game(self): """Tests that abort_game() method calls handle_eliminated_players() once """ player = MagicMock() send_announcement = AsyncMock() player.attach_mock(send_announcement, 'send_announcement') game = GameSession(0, [player]) game.handle_eliminated_players = MagicMock() asyncio.run(game.abort_game()) game.handle_eliminated_players.assert_called_once() send_announcement.assert_called_once_with( MESSAGE_NETWORK_ERROR_OCCURRED) @patch('hqtrivia.question.Question.generate', new_callable=AsyncMock) def test_execute_next_round_identifies_winner(self, generate): """Tests that execute_next_round() method correctly identifies winner from the game. """ question = Question('Question 1', ['A', 'B', 'C', 'D'], 'C') generate.return_value = question # Create player 1 with wrong answer, and player 2 with the right answer (player1, send_announcement1, send_question1, send_answers1, recv_answer1) = self.get_mocked_player('D') (player2, send_announcement2, send_question2, send_answers2, recv_answer2) = self.get_mocked_player('C') game = GameSession(0, [player1, player2]) result = asyncio.run(game.execute_next_round()) self.assertFalse( result, "execute_next_round() should return False since game ended with a winner.") self.assertEqual(len(game.players), 0, "No player should remain in the game") # Check that the question was sent to the players send_question1.assert_any_call(question) send_question2.assert_any_call(question) # Check that the count of the answers from each participant was sent send_answers1.assert_any_call(question, [0, 0, 1, 1]) send_answers2.assert_any_call(question, [0, 0, 1, 1]) # Check that these message were sent last send_announcement1.assert_called_with(MESSAGE_YOU_ARE_ELIMINATED) send_announcement2.assert_called_with(MESSAGE_YOU_ARE_THE_WINNER) @patch('hqtrivia.question.Question.generate', new_callable=AsyncMock) def test_execute_next_round_player_times_out(self, generate): """Tests that execute_next_round() eliminates player who did not respond within timeout. """ question = Question('Question 1', ['A', 'B', 'C', 'D'], 'C') generate.return_value = question # Create player 1 with the wrong answer, and player 2 will timeout (player1, send_announcement1, send_question1, send_answers1, recv_answer1) = self.get_mocked_player('D') async def over_sleep(): await asyncio.sleep(config.CONFIG_ROUND_DURATION+1) (player2, send_announcement2, send_question2, send_answers2, recv_answer2) = self.get_mocked_player(None) player2.attach_mock(over_sleep, 'recv_answer') game = GameSession(0, [player1, player2]) result = asyncio.run(game.execute_next_round()) self.assertFalse( result, "execute_next_round() should return False since game ended.") self.assertEqual(len(game.players), 0, "No player should remain in the game") # Check that the question was sent to the players # Check that the question was sent to the players send_question1.assert_any_call(question) send_question2.assert_any_call(question) # Check that the count of the answers from each participant was sent send_answers1.assert_any_call(question, [0, 0, 0, 1]) send_answers2.assert_any_call(question, [0, 0, 0, 1]) # Check that these message were sent last send_announcement1.assert_called_with(MESSAGE_YOU_ARE_ELIMINATED) send_announcement1.assert_called_with(MESSAGE_YOU_ARE_ELIMINATED) @patch('hqtrivia.question.Question.generate', new_callable=AsyncMock) def test_execute_next_round_player_recvMessage_exception(self, generate): """Tests that execute_next_round() eliminates player who did not respond within timeout. """ question = Question('Question 1', ['A', 'B', 'C', 'D'], 'C') generate.return_value = question # Create player 1 with wrong answer, and player 2 with the right answer (player1, send_announcement1, send_question1, send_answers1, recv_answer1) = self.get_mocked_player('D') (player2, send_announcement2, send_question2, send_answers2, recv_answer2) = self.get_mocked_player('C') recv_answer2 = AsyncMock(side_effect=Exception( 'General Error during recv_answer()')) player2.attach_mock(recv_answer2, 'recv_answer') game = GameSession(0, [player1, player2]) result = asyncio.run(game.execute_next_round()) self.assertFalse( result, "execute_next_round() should return False since game ended.") self.assertEqual(len(game.players), 0, "No player should remain in the game") # Check that the question was sent to the players send_question1.assert_any_call(question) send_question2.assert_any_call(question) # Check that the count of the answers from each participant was sent send_answers1.assert_any_call(question, [0, 0, 0, 1]) send_answers2.assert_any_call(question, [0, 0, 0, 1]) # Check that these message were sent last send_announcement1.assert_called_with(MESSAGE_YOU_ARE_ELIMINATED) send_announcement1.assert_called_with(MESSAGE_YOU_ARE_ELIMINATED) @patch('hqtrivia.question.Question.generate', new_callable=AsyncMock) def test_execute_next_round_move_to_next_round(self, generate): """Tests that execute_next_round() method has to continue to the next round. """ question = Question('Question 1', ['A', 'B', 'C', 'D'], 'C') generate.return_value = question # Both players gives the correct answer (player1, send_announcement1, send_question1, send_answers1, recv_answer1) = self.get_mocked_player('C') (player2, send_announcement2, send_question2, send_answers2, recv_answer2) = self.get_mocked_player('C') game = GameSession(0, [player1, player2]) result = asyncio.run(game.execute_next_round()) self.assertTrue( result, "execute_next_round() should return True since no single winner yet.") self.assertEqual(len(game.players), 2, "Two players should still be in the game") # Check that the question was sent to the players send_question1.assert_any_call(question) send_question2.assert_any_call(question) # Check that the count of the answers from each participant was sent send_answers1.assert_any_call(question, [0, 0, 2, 0]) send_answers2.assert_any_call(question, [0, 0, 2, 0]) # Check that these message were sent last send_announcement1.assert_called_with(MESSAGE_CORRECT_ANSWER) send_announcement1.assert_called_with(MESSAGE_CORRECT_ANSWER) @patch('hqtrivia.question.Question.generate', new_callable=AsyncMock) def test_execute_next_round_all_eliminated(self, generate): """Tests that execute_next_round() method eliminates everyone without winner """ question = Question('Question 1', ['A', 'B', 'C', 'D'], 'C') generate.return_value = question # Both players gives wrong answer (player1, send_announcement1, send_question1, send_answers1, recv_answer1) = self.get_mocked_player('A') (player2, send_announcement2, send_question2, send_answers2, recv_answer2) = self.get_mocked_player('B') game = GameSession(0, [player1, player2]) result = asyncio.run(game.execute_next_round()) self.assertFalse( result, "execute_next_round() should return False since everyone eliminated.") self.assertEqual(len(game.players), 0, "No one should remain in the game") # Check that the question was sent to the players send_question1.assert_any_call(question) send_question2.assert_any_call(question) # Check that the count of the answers from each participant was sent send_answers1.assert_any_call(question, [1, 1, 0, 0]) send_answers2.assert_any_call(question, [1, 1, 0, 0]) # Check that these message were sent last send_announcement1.assert_called_with(MESSAGE_YOU_ARE_ELIMINATED) send_announcement1.assert_called_with(MESSAGE_YOU_ARE_ELIMINATED) @patch('hqtrivia.question.Question.generate', new_callable=AsyncMock) def test_run_exception_question_generator(self, generate): """Tests that run() method eliminates everyone without winner """ question = Question('Question 1', ['A', 'B', 'C', 'D'], 'C') generate.return_value = question generate.side_effect = Exception("Network Error!!!") # Both players give wrong answer, but not important since Question.generate will throw exception (player1, send_announcement1, send_question1, send_answers1, recv_answer1) = self.get_mocked_player('A') (player2, send_announcement2, send_question2, send_answers2, recv_answer2) = self.get_mocked_player('B') game = GameSession(0, [player1, player2]) with self.assertRaises(Exception): asyncio.run(game.run()) self.assertEqual(len(game.players), 0, "No one should remain in the game") # Check that these message were sent last send_announcement1.assert_called_once_with( MESSAGE_NETWORK_ERROR_OCCURRED) send_announcement2.assert_called_once_with( MESSAGE_NETWORK_ERROR_OCCURRED)
[ "68974415+dongkyutoronto@users.noreply.github.com" ]
68974415+dongkyutoronto@users.noreply.github.com
26f5feec50c17f59e09b58ec23e082f0893c8b39
98c62fa21899dc823cfa6ccdfb89ba5bf5836b1a
/fileupload/admin.py
25bf6c18faed3ecad700fb19bc7fea68a4c48939
[]
no_license
sagar-spkt/DjangoAWS
1ae3bab19303d0344973eb026972c8a7853a5df4
a7d032370f1697e18054740dc439d5f08087fd88
refs/heads/master
2021-08-16T21:46:39.690026
2018-12-16T17:29:35
2018-12-16T18:17:34
162,022,099
0
0
null
2021-06-10T21:03:34
2018-12-16T16:56:20
Python
UTF-8
Python
false
false
198
py
from django.contrib import admin from fileupload.models import PublicDocument, PrivateDocument # Register your models here. admin.site.register(PublicDocument) admin.site.register(PrivateDocument)
[ "spkt.sagar@gmailcom" ]
spkt.sagar@gmailcom
10ba62e4e2abefa1a7435eb6853c7606548f54aa
92d5555325b328549b25d3c2af9b8c577d1511cd
/log/config.py
e612e8fe669bfae1ce5e9f6fbdc77cedebcd275c
[]
no_license
zhang29845987/tpshop_login_api_1
a5d213ead57376fb1e2b91d0554b2bc6541f06f7
58a78a60e6378d960ad03c9de4a4f3a006078f89
refs/heads/master
2020-09-29T18:12:52.191624
2019-12-10T10:32:26
2019-12-10T10:32:26
227,091,400
0
0
null
null
null
null
UTF-8
Python
false
false
1,129
py
import logging import os # 获取项目根目录 from logging.handlers import TimedRotatingFileHandler BASE_DIR = os.path.dirname(os.path.abspath(__file__)) print(os.path.dirname(os.path.abspath(__file__))) # 初始化日志配置 def init_log_config(): #添加日志器 默认日志root logger=logging.getLogger() logger.setLevel(logging.INFO) #处理器 输出到控制台 shl=logging.StreamHandler() #处理器输出到文件到指定路径,when以小时为单位存文件,interval一个小时存一个文件,backcount默认保存全部文件,encoding 使用utf8编码格式 trfhl=TimedRotatingFileHandler(filename=BASE_DIR+"/log/mylog.log",when="h",interval=1,backupCount=0,encoding="utf8") #格式化器 fmter=logging.Formatter(fmt="%(asctime)s %(levelname)s [%(name)s] [%(filename)s(%(funcName)s:%(lineno)d)] - %(message)s") #处理器添加格式化器 shl.setFormatter(fmter) # 处理器添加格式化器 trfhl.setFormatter(fmter) #日志器添加处理器 logger.addHandler(shl) #日志器添加处理器 logger.addHandler(trfhl)
[ "zpb0990@163.com" ]
zpb0990@163.com
963a9147e8343c232ca5f41b05658557f4a21e1b
9ae0fe3620c5be4af7801462acc7564e8d4a56f0
/OneDrive_4_8-26-2019/TSqlVisitor.py
41ad406c9430d77fc567e1940c78148efda05311
[]
no_license
MacJei/antlr
66ea15b3e8b063354664031dac9427db4c02ad5c
7e945a6424caad00e58f7a4b41b0297556f00294
refs/heads/master
2022-03-20T17:49:06.241851
2019-12-18T12:27:47
2019-12-18T12:27:47
null
0
0
null
null
null
null
UTF-8
Python
false
false
44,697
py
# Generated from TSql.g4 by ANTLR 4.7.2 from antlr4 import * if __name__ is not None and "." in __name__: from .TSqlParser import TSqlParser else: from TSqlParser import TSqlParser # This class defines a complete generic visitor for a parse tree produced by TSqlParser. class TSqlVisitor(ParseTreeVisitor): # Visit a parse tree produced by TSqlParser#tsql_file. def visitTsql_file(self, ctx:TSqlParser.Tsql_fileContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#batch. def visitBatch(self, ctx:TSqlParser.BatchContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#sql_clauses. def visitSql_clauses(self, ctx:TSqlParser.Sql_clausesContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#sql_clause. def visitSql_clause(self, ctx:TSqlParser.Sql_clauseContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#dml_clause. def visitDml_clause(self, ctx:TSqlParser.Dml_clauseContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#cfl_statement. def visitCfl_statement(self, ctx:TSqlParser.Cfl_statementContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#block_statement. def visitBlock_statement(self, ctx:TSqlParser.Block_statementContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#break_statement. def visitBreak_statement(self, ctx:TSqlParser.Break_statementContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#continue_statement. def visitContinue_statement(self, ctx:TSqlParser.Continue_statementContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#goto_statement. def visitGoto_statement(self, ctx:TSqlParser.Goto_statementContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#return_statement. def visitReturn_statement(self, ctx:TSqlParser.Return_statementContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#if_statement. def visitIf_statement(self, ctx:TSqlParser.If_statementContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#throw_statement. def visitThrow_statement(self, ctx:TSqlParser.Throw_statementContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#throw_error_number. def visitThrow_error_number(self, ctx:TSqlParser.Throw_error_numberContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#throw_message. def visitThrow_message(self, ctx:TSqlParser.Throw_messageContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#throw_state. def visitThrow_state(self, ctx:TSqlParser.Throw_stateContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#try_catch_statement. def visitTry_catch_statement(self, ctx:TSqlParser.Try_catch_statementContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#waitfor_statement. def visitWaitfor_statement(self, ctx:TSqlParser.Waitfor_statementContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#while_statement. def visitWhile_statement(self, ctx:TSqlParser.While_statementContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#print_statement. def visitPrint_statement(self, ctx:TSqlParser.Print_statementContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#raiseerror_statement. def visitRaiseerror_statement(self, ctx:TSqlParser.Raiseerror_statementContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#empty_statement. def visitEmpty_statement(self, ctx:TSqlParser.Empty_statementContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#another_statement. def visitAnother_statement(self, ctx:TSqlParser.Another_statementContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#file_path. def visitFile_path(self, ctx:TSqlParser.File_pathContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#file_directory_path_separator. def visitFile_directory_path_separator(self, ctx:TSqlParser.File_directory_path_separatorContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#event_session_predicate_expression. def visitEvent_session_predicate_expression(self, ctx:TSqlParser.Event_session_predicate_expressionContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#event_session_predicate_factor. def visitEvent_session_predicate_factor(self, ctx:TSqlParser.Event_session_predicate_factorContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#event_session_predicate_leaf. def visitEvent_session_predicate_leaf(self, ctx:TSqlParser.Event_session_predicate_leafContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#create_queue. def visitCreate_queue(self, ctx:TSqlParser.Create_queueContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#queue_settings. def visitQueue_settings(self, ctx:TSqlParser.Queue_settingsContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#alter_queue. def visitAlter_queue(self, ctx:TSqlParser.Alter_queueContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#queue_action. def visitQueue_action(self, ctx:TSqlParser.Queue_actionContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#queue_rebuild_options. def visitQueue_rebuild_options(self, ctx:TSqlParser.Queue_rebuild_optionsContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#create_contract. def visitCreate_contract(self, ctx:TSqlParser.Create_contractContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#conversation_statement. def visitConversation_statement(self, ctx:TSqlParser.Conversation_statementContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#message_statement. def visitMessage_statement(self, ctx:TSqlParser.Message_statementContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#merge_statement. def visitMerge_statement(self, ctx:TSqlParser.Merge_statementContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#merge_matched. def visitMerge_matched(self, ctx:TSqlParser.Merge_matchedContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#merge_not_matched. def visitMerge_not_matched(self, ctx:TSqlParser.Merge_not_matchedContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#delete_statement. def visitDelete_statement(self, ctx:TSqlParser.Delete_statementContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#delete_statement_from. def visitDelete_statement_from(self, ctx:TSqlParser.Delete_statement_fromContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#insert_statement. def visitInsert_statement(self, ctx:TSqlParser.Insert_statementContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#insert_statement_value. def visitInsert_statement_value(self, ctx:TSqlParser.Insert_statement_valueContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#receive_statement. def visitReceive_statement(self, ctx:TSqlParser.Receive_statementContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#select_statement. def visitSelect_statement(self, ctx:TSqlParser.Select_statementContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#time. def visitTime(self, ctx:TSqlParser.TimeContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#update_statement. def visitUpdate_statement(self, ctx:TSqlParser.Update_statementContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#output_clause. def visitOutput_clause(self, ctx:TSqlParser.Output_clauseContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#output_dml_list_elem. def visitOutput_dml_list_elem(self, ctx:TSqlParser.Output_dml_list_elemContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#output_column_name. def visitOutput_column_name(self, ctx:TSqlParser.Output_column_nameContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#rowset_function_limited. def visitRowset_function_limited(self, ctx:TSqlParser.Rowset_function_limitedContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#openquery. def visitOpenquery(self, ctx:TSqlParser.OpenqueryContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#opendatasource. def visitOpendatasource(self, ctx:TSqlParser.OpendatasourceContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#declare_statement. def visitDeclare_statement(self, ctx:TSqlParser.Declare_statementContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#cursor_statement. def visitCursor_statement(self, ctx:TSqlParser.Cursor_statementContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#execute_statement. def visitExecute_statement(self, ctx:TSqlParser.Execute_statementContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#execute_body. def visitExecute_body(self, ctx:TSqlParser.Execute_bodyContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#execute_statement_arg. def visitExecute_statement_arg(self, ctx:TSqlParser.Execute_statement_argContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#execute_var_string. def visitExecute_var_string(self, ctx:TSqlParser.Execute_var_stringContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#security_statement. def visitSecurity_statement(self, ctx:TSqlParser.Security_statementContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#create_certificate. def visitCreate_certificate(self, ctx:TSqlParser.Create_certificateContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#existing_keys. def visitExisting_keys(self, ctx:TSqlParser.Existing_keysContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#private_key_options. def visitPrivate_key_options(self, ctx:TSqlParser.Private_key_optionsContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#generate_new_keys. def visitGenerate_new_keys(self, ctx:TSqlParser.Generate_new_keysContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#date_options. def visitDate_options(self, ctx:TSqlParser.Date_optionsContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#open_key. def visitOpen_key(self, ctx:TSqlParser.Open_keyContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#close_key. def visitClose_key(self, ctx:TSqlParser.Close_keyContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#create_key. def visitCreate_key(self, ctx:TSqlParser.Create_keyContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#key_options. def visitKey_options(self, ctx:TSqlParser.Key_optionsContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#algorithm. def visitAlgorithm(self, ctx:TSqlParser.AlgorithmContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#encryption_mechanism. def visitEncryption_mechanism(self, ctx:TSqlParser.Encryption_mechanismContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#decryption_mechanism. def visitDecryption_mechanism(self, ctx:TSqlParser.Decryption_mechanismContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#grant_permission. def visitGrant_permission(self, ctx:TSqlParser.Grant_permissionContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#set_statement. def visitSet_statement(self, ctx:TSqlParser.Set_statementContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#transaction_statement. def visitTransaction_statement(self, ctx:TSqlParser.Transaction_statementContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#go_statement. def visitGo_statement(self, ctx:TSqlParser.Go_statementContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#use_statement. def visitUse_statement(self, ctx:TSqlParser.Use_statementContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#setuser_statement. def visitSetuser_statement(self, ctx:TSqlParser.Setuser_statementContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#execute_clause. def visitExecute_clause(self, ctx:TSqlParser.Execute_clauseContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#declare_local. def visitDeclare_local(self, ctx:TSqlParser.Declare_localContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#table_type_definition. def visitTable_type_definition(self, ctx:TSqlParser.Table_type_definitionContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#xml_type_definition. def visitXml_type_definition(self, ctx:TSqlParser.Xml_type_definitionContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#xml_schema_collection. def visitXml_schema_collection(self, ctx:TSqlParser.Xml_schema_collectionContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#column_def_table_constraints. def visitColumn_def_table_constraints(self, ctx:TSqlParser.Column_def_table_constraintsContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#column_def_table_constraint. def visitColumn_def_table_constraint(self, ctx:TSqlParser.Column_def_table_constraintContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#column_definition. def visitColumn_definition(self, ctx:TSqlParser.Column_definitionContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#materialized_column_definition. def visitMaterialized_column_definition(self, ctx:TSqlParser.Materialized_column_definitionContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#column_constraint. def visitColumn_constraint(self, ctx:TSqlParser.Column_constraintContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#table_constraint. def visitTable_constraint(self, ctx:TSqlParser.Table_constraintContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#on_delete. def visitOn_delete(self, ctx:TSqlParser.On_deleteContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#on_update. def visitOn_update(self, ctx:TSqlParser.On_updateContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#index_options. def visitIndex_options(self, ctx:TSqlParser.Index_optionsContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#index_option. def visitIndex_option(self, ctx:TSqlParser.Index_optionContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#declare_cursor. def visitDeclare_cursor(self, ctx:TSqlParser.Declare_cursorContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#declare_set_cursor_common. def visitDeclare_set_cursor_common(self, ctx:TSqlParser.Declare_set_cursor_commonContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#declare_set_cursor_common_partial. def visitDeclare_set_cursor_common_partial(self, ctx:TSqlParser.Declare_set_cursor_common_partialContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#fetch_cursor. def visitFetch_cursor(self, ctx:TSqlParser.Fetch_cursorContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#set_special. def visitSet_special(self, ctx:TSqlParser.Set_specialContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#constant_LOCAL_ID. def visitConstant_LOCAL_ID(self, ctx:TSqlParser.Constant_LOCAL_IDContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#expression. def visitExpression(self, ctx:TSqlParser.ExpressionContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#primitive_expression. def visitPrimitive_expression(self, ctx:TSqlParser.Primitive_expressionContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#case_expression. def visitCase_expression(self, ctx:TSqlParser.Case_expressionContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#unary_operator_expression. def visitUnary_operator_expression(self, ctx:TSqlParser.Unary_operator_expressionContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#bracket_expression. def visitBracket_expression(self, ctx:TSqlParser.Bracket_expressionContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#constant_expression. def visitConstant_expression(self, ctx:TSqlParser.Constant_expressionContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#subquery. def visitSubquery(self, ctx:TSqlParser.SubqueryContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#with_expression. def visitWith_expression(self, ctx:TSqlParser.With_expressionContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#common_table_expression. def visitCommon_table_expression(self, ctx:TSqlParser.Common_table_expressionContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#update_elem. def visitUpdate_elem(self, ctx:TSqlParser.Update_elemContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#search_condition_list. def visitSearch_condition_list(self, ctx:TSqlParser.Search_condition_listContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#search_condition. def visitSearch_condition(self, ctx:TSqlParser.Search_conditionContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#search_condition_and. def visitSearch_condition_and(self, ctx:TSqlParser.Search_condition_andContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#search_condition_not. def visitSearch_condition_not(self, ctx:TSqlParser.Search_condition_notContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#predicate. def visitPredicate(self, ctx:TSqlParser.PredicateContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#query_expression. def visitQuery_expression(self, ctx:TSqlParser.Query_expressionContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#sql_union. def visitSql_union(self, ctx:TSqlParser.Sql_unionContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#query_specification. def visitQuery_specification(self, ctx:TSqlParser.Query_specificationContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#top_clause. def visitTop_clause(self, ctx:TSqlParser.Top_clauseContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#top_percent. def visitTop_percent(self, ctx:TSqlParser.Top_percentContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#top_count. def visitTop_count(self, ctx:TSqlParser.Top_countContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#order_by_clause. def visitOrder_by_clause(self, ctx:TSqlParser.Order_by_clauseContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#for_clause. def visitFor_clause(self, ctx:TSqlParser.For_clauseContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#xml_common_directives. def visitXml_common_directives(self, ctx:TSqlParser.Xml_common_directivesContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#order_by_expression. def visitOrder_by_expression(self, ctx:TSqlParser.Order_by_expressionContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#group_by_item. def visitGroup_by_item(self, ctx:TSqlParser.Group_by_itemContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#option_clause. def visitOption_clause(self, ctx:TSqlParser.Option_clauseContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#option. def visitOption(self, ctx:TSqlParser.OptionContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#optimize_for_arg. def visitOptimize_for_arg(self, ctx:TSqlParser.Optimize_for_argContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#select_list. def visitSelect_list(self, ctx:TSqlParser.Select_listContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#udt_method_arguments. def visitUdt_method_arguments(self, ctx:TSqlParser.Udt_method_argumentsContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#asterisk. def visitAsterisk(self, ctx:TSqlParser.AsteriskContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#column_elem. def visitColumn_elem(self, ctx:TSqlParser.Column_elemContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#udt_elem. def visitUdt_elem(self, ctx:TSqlParser.Udt_elemContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#expression_elem. def visitExpression_elem(self, ctx:TSqlParser.Expression_elemContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#select_list_elem. def visitSelect_list_elem(self, ctx:TSqlParser.Select_list_elemContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#table_sources. def visitTable_sources(self, ctx:TSqlParser.Table_sourcesContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#table_source. def visitTable_source(self, ctx:TSqlParser.Table_sourceContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#table_source_item_joined. def visitTable_source_item_joined(self, ctx:TSqlParser.Table_source_item_joinedContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#table_source_item. def visitTable_source_item(self, ctx:TSqlParser.Table_source_itemContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#open_xml. def visitOpen_xml(self, ctx:TSqlParser.Open_xmlContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#schema_declaration. def visitSchema_declaration(self, ctx:TSqlParser.Schema_declarationContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#column_declaration. def visitColumn_declaration(self, ctx:TSqlParser.Column_declarationContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#change_table. def visitChange_table(self, ctx:TSqlParser.Change_tableContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#join_part. def visitJoin_part(self, ctx:TSqlParser.Join_partContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#pivot_clause. def visitPivot_clause(self, ctx:TSqlParser.Pivot_clauseContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#unpivot_clause. def visitUnpivot_clause(self, ctx:TSqlParser.Unpivot_clauseContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#full_column_name_list. def visitFull_column_name_list(self, ctx:TSqlParser.Full_column_name_listContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#table_name_with_hint. def visitTable_name_with_hint(self, ctx:TSqlParser.Table_name_with_hintContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#rowset_function. def visitRowset_function(self, ctx:TSqlParser.Rowset_functionContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#bulk_option. def visitBulk_option(self, ctx:TSqlParser.Bulk_optionContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#derived_table. def visitDerived_table(self, ctx:TSqlParser.Derived_tableContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#RANKING_WINDOWED_FUNC. def visitRANKING_WINDOWED_FUNC(self, ctx:TSqlParser.RANKING_WINDOWED_FUNCContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#AGGREGATE_WINDOWED_FUNC. def visitAGGREGATE_WINDOWED_FUNC(self, ctx:TSqlParser.AGGREGATE_WINDOWED_FUNCContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#ANALYTIC_WINDOWED_FUNC. def visitANALYTIC_WINDOWED_FUNC(self, ctx:TSqlParser.ANALYTIC_WINDOWED_FUNCContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#SCALAR_FUNCTION. def visitSCALAR_FUNCTION(self, ctx:TSqlParser.SCALAR_FUNCTIONContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#BINARY_CHECKSUM. def visitBINARY_CHECKSUM(self, ctx:TSqlParser.BINARY_CHECKSUMContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#CAST. def visitCAST(self, ctx:TSqlParser.CASTContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#CONVERT. def visitCONVERT(self, ctx:TSqlParser.CONVERTContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#CHECKSUM. def visitCHECKSUM(self, ctx:TSqlParser.CHECKSUMContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#COALESCE. def visitCOALESCE(self, ctx:TSqlParser.COALESCEContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#CURRENT_TIMESTAMP. def visitCURRENT_TIMESTAMP(self, ctx:TSqlParser.CURRENT_TIMESTAMPContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#CURRENT_USER. def visitCURRENT_USER(self, ctx:TSqlParser.CURRENT_USERContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#DATEADD. def visitDATEADD(self, ctx:TSqlParser.DATEADDContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#DATEDIFF. def visitDATEDIFF(self, ctx:TSqlParser.DATEDIFFContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#DATENAME. def visitDATENAME(self, ctx:TSqlParser.DATENAMEContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#DATEPART. def visitDATEPART(self, ctx:TSqlParser.DATEPARTContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#GETDATE. def visitGETDATE(self, ctx:TSqlParser.GETDATEContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#GETUTCDATE. def visitGETUTCDATE(self, ctx:TSqlParser.GETUTCDATEContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#IDENTITY. def visitIDENTITY(self, ctx:TSqlParser.IDENTITYContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#MIN_ACTIVE_ROWVERSION. def visitMIN_ACTIVE_ROWVERSION(self, ctx:TSqlParser.MIN_ACTIVE_ROWVERSIONContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#NULLIF. def visitNULLIF(self, ctx:TSqlParser.NULLIFContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#STUFF. def visitSTUFF(self, ctx:TSqlParser.STUFFContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#SESSION_USER. def visitSESSION_USER(self, ctx:TSqlParser.SESSION_USERContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#SYSTEM_USER. def visitSYSTEM_USER(self, ctx:TSqlParser.SYSTEM_USERContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#ISNULL. def visitISNULL(self, ctx:TSqlParser.ISNULLContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#XML_DATA_TYPE_FUNC. def visitXML_DATA_TYPE_FUNC(self, ctx:TSqlParser.XML_DATA_TYPE_FUNCContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#IFF. def visitIFF(self, ctx:TSqlParser.IFFContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#xml_data_type_methods. def visitXml_data_type_methods(self, ctx:TSqlParser.Xml_data_type_methodsContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#value_method. def visitValue_method(self, ctx:TSqlParser.Value_methodContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#query_method. def visitQuery_method(self, ctx:TSqlParser.Query_methodContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#exist_method. def visitExist_method(self, ctx:TSqlParser.Exist_methodContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#modify_method. def visitModify_method(self, ctx:TSqlParser.Modify_methodContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#nodes_method. def visitNodes_method(self, ctx:TSqlParser.Nodes_methodContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#switch_section. def visitSwitch_section(self, ctx:TSqlParser.Switch_sectionContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#switch_search_condition_section. def visitSwitch_search_condition_section(self, ctx:TSqlParser.Switch_search_condition_sectionContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#as_column_alias. def visitAs_column_alias(self, ctx:TSqlParser.As_column_aliasContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#as_table_alias. def visitAs_table_alias(self, ctx:TSqlParser.As_table_aliasContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#table_alias. def visitTable_alias(self, ctx:TSqlParser.Table_aliasContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#with_table_hints. def visitWith_table_hints(self, ctx:TSqlParser.With_table_hintsContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#insert_with_table_hints. def visitInsert_with_table_hints(self, ctx:TSqlParser.Insert_with_table_hintsContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#table_hint. def visitTable_hint(self, ctx:TSqlParser.Table_hintContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#index_value. def visitIndex_value(self, ctx:TSqlParser.Index_valueContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#column_alias_list. def visitColumn_alias_list(self, ctx:TSqlParser.Column_alias_listContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#column_alias. def visitColumn_alias(self, ctx:TSqlParser.Column_aliasContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#table_value_constructor. def visitTable_value_constructor(self, ctx:TSqlParser.Table_value_constructorContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#expression_list. def visitExpression_list(self, ctx:TSqlParser.Expression_listContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#ranking_windowed_function. def visitRanking_windowed_function(self, ctx:TSqlParser.Ranking_windowed_functionContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#aggregate_windowed_function. def visitAggregate_windowed_function(self, ctx:TSqlParser.Aggregate_windowed_functionContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#analytic_windowed_function. def visitAnalytic_windowed_function(self, ctx:TSqlParser.Analytic_windowed_functionContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#all_distinct_expression. def visitAll_distinct_expression(self, ctx:TSqlParser.All_distinct_expressionContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#over_clause. def visitOver_clause(self, ctx:TSqlParser.Over_clauseContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#row_or_range_clause. def visitRow_or_range_clause(self, ctx:TSqlParser.Row_or_range_clauseContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#window_frame_extent. def visitWindow_frame_extent(self, ctx:TSqlParser.Window_frame_extentContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#window_frame_bound. def visitWindow_frame_bound(self, ctx:TSqlParser.Window_frame_boundContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#window_frame_preceding. def visitWindow_frame_preceding(self, ctx:TSqlParser.Window_frame_precedingContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#window_frame_following. def visitWindow_frame_following(self, ctx:TSqlParser.Window_frame_followingContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#full_table_name. def visitFull_table_name(self, ctx:TSqlParser.Full_table_nameContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#table_name. def visitTable_name(self, ctx:TSqlParser.Table_nameContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#func_proc_name_schema. def visitFunc_proc_name_schema(self, ctx:TSqlParser.Func_proc_name_schemaContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#func_proc_name_database_schema. def visitFunc_proc_name_database_schema(self, ctx:TSqlParser.Func_proc_name_database_schemaContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#func_proc_name_server_database_schema. def visitFunc_proc_name_server_database_schema(self, ctx:TSqlParser.Func_proc_name_server_database_schemaContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#ddl_object. def visitDdl_object(self, ctx:TSqlParser.Ddl_objectContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#full_column_name. def visitFull_column_name(self, ctx:TSqlParser.Full_column_nameContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#column_name_list_with_order. def visitColumn_name_list_with_order(self, ctx:TSqlParser.Column_name_list_with_orderContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#column_name_list. def visitColumn_name_list(self, ctx:TSqlParser.Column_name_listContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#cursor_name. def visitCursor_name(self, ctx:TSqlParser.Cursor_nameContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#on_off. def visitOn_off(self, ctx:TSqlParser.On_offContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#clustered. def visitClustered(self, ctx:TSqlParser.ClusteredContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#null_notnull. def visitNull_notnull(self, ctx:TSqlParser.Null_notnullContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#null_or_default. def visitNull_or_default(self, ctx:TSqlParser.Null_or_defaultContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#scalar_function_name. def visitScalar_function_name(self, ctx:TSqlParser.Scalar_function_nameContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#begin_conversation_timer. def visitBegin_conversation_timer(self, ctx:TSqlParser.Begin_conversation_timerContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#begin_conversation_dialog. def visitBegin_conversation_dialog(self, ctx:TSqlParser.Begin_conversation_dialogContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#contract_name. def visitContract_name(self, ctx:TSqlParser.Contract_nameContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#service_name. def visitService_name(self, ctx:TSqlParser.Service_nameContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#end_conversation. def visitEnd_conversation(self, ctx:TSqlParser.End_conversationContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#waitfor_conversation. def visitWaitfor_conversation(self, ctx:TSqlParser.Waitfor_conversationContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#get_conversation. def visitGet_conversation(self, ctx:TSqlParser.Get_conversationContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#queue_id. def visitQueue_id(self, ctx:TSqlParser.Queue_idContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#send_conversation. def visitSend_conversation(self, ctx:TSqlParser.Send_conversationContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#data_type. def visitData_type(self, ctx:TSqlParser.Data_typeContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#constant. def visitConstant(self, ctx:TSqlParser.ConstantContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#sign. def visitSign(self, ctx:TSqlParser.SignContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#identifier. def visitIdentifier(self, ctx:TSqlParser.IdentifierContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#simple_id. def visitSimple_id(self, ctx:TSqlParser.Simple_idContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#comparison_operator. def visitComparison_operator(self, ctx:TSqlParser.Comparison_operatorContext): return self.visitChildren(ctx) # Visit a parse tree produced by TSqlParser#assignment_operator. def visitAssignment_operator(self, ctx:TSqlParser.Assignment_operatorContext): return self.visitChildren(ctx) del TSqlParser
[ "noreply@github.com" ]
noreply@github.com
ff1fd7cc8123d0a5423c973f1719f3e816799522
2fe08f181c53b7d94c5e97f08016b86705488594
/app.py
c2170f933f0edeab3758e05f01956f0539cd387b
[]
no_license
KidoThunder/ExportXiamiList
521871639e67bd2cc235de46ce4078610bb44e65
b769740e4613f58ff29381f1e8865adbf6368ac2
refs/heads/master
2021-06-16T20:00:49.713299
2017-04-24T06:40:18
2017-04-24T06:40:18
null
0
0
null
null
null
null
UTF-8
Python
false
false
4,606
py
# python3 # author: fyl00 # source: https://github.com/fyl00/ExportXiamiList import logging import re import sys from PyQt5.QtCore import QObject, QThread, pyqtSignal from PyQt5.QtGui import QTextCursor, QIcon from PyQt5.QtWidgets import QMainWindow, QApplication, QMessageBox, QFileDialog from lxml import etree from XiamiList.tips import * from XiamiList.xiami import XiamiHandle, XiamiLink from ui import Ui_MainWindow import images_qr # 打印输出到 logTextEdit class QtLogHandler(logging.Handler): def __init__(self): logging.Handler.__init__(self) def emit(self, record): record = self.format(record) if record: EmittingStream.stdout().write('%s\n' % record) class EmittingStream(QObject): _stdout = None _stderr = None textWritten = pyqtSignal(str) def write(self, text): if not self.signalsBlocked(): self.textWritten.emit(str(text)) def flush(self): pass def fileno(self): return -1 @staticmethod def stdout(): if not EmittingStream._stdout: EmittingStream._stdout = EmittingStream() sys.stdout = EmittingStream._stdout return EmittingStream._stdout @staticmethod def stderr(): if not EmittingStream._stderr: EmittingStream._stderr = EmittingStream() sys.stderr = EmittingStream._stderr return EmittingStream._stderr # 后台抓取,防止界面未响应 class XiamiThread(QThread): finished = pyqtSignal(str) def __init__(self, url): QThread.__init__(self) self.url = url def run(self): xmlstr = XiamiHandle().get_list(self.url) self.finished.emit(xmlstr) # 界面窗口 class AppWindow(QMainWindow): def __init__(self): QMainWindow.__init__(self) self.ui = Ui_MainWindow() self.ui.setupUi(self) self._enbale_source_link() self.ui.startButton.clicked.connect(self.click_start_button) EmittingStream.stdout().textWritten.connect(self._logout) EmittingStream.stderr().textWritten.connect(self._logout) self.ui.linkLineEdit.setPlaceholderText("请输入歌单链接") self.ui.linkLineEdit.setFocus() self._logout(GET_LINK) # Storing a reference to the thread after it's been created # http://stackoverflow.com/questions/15702782/qthread-destroyed-while-thread-is-still-running self.threads = [] def click_start_button(self): url = self.ui.linkLineEdit.text() if not self._check_url(url): return thread = XiamiThread(url) self.threads.append(thread) thread.finished.connect(self._task_finished) thread.start() self.ui.startButton.setDisabled(True) def _check_url(self, url): link = XiamiLink(url) if link.is_collect is None: title = "链接格式错误" QMessageBox.critical(self, title, LINK_ERROR_TIPS) return False return True def _task_finished(self, value): self._save_xml(value) self.ui.startButton.setDisabled(False) def _enbale_source_link(self): link_text = "源码:<a href='https://github.com/fyl00/ExportXiamiList'>GitHub</a>" self.ui.sourceLabel.setText(link_text) self.ui.sourceLabel.setOpenExternalLinks(True) def _logout(self, outstr): cursor = self.ui.logTextEdit.textCursor() cursor.insertText(outstr) self.ui.logTextEdit.moveCursor(QTextCursor.End) def _save_xml(self, xmlstr): options = QFileDialog.Options() options |= QFileDialog.DontUseNativeDialog filename, _ = QFileDialog.getSaveFileName(self, "QFileDialog.getSaveFileName()", "songs.kgl", "Kugou/Netease Files (*.kgl)", options=options) if filename: r = re.search("\.kgl$", filename) if not r: filename = "%s.kgl" % filename print("** 导出文件位置:%s" % filename) root = etree.fromstring(xmlstr) etree.ElementTree(root).write(filename, xml_declaration=True, encoding="utf8", pretty_print=True) if __name__ == "__main__": app = QApplication(sys.argv) window = AppWindow() app.setWindowIcon(QIcon(':/static/favicon.ico')) window.show() sys.exit(app.exec_())
[ "fyl00lee@gmail.com" ]
fyl00lee@gmail.com
d035e9fd898748e129cda959b3d579e13d600d7b
aae29f1c7c8eabda658a00e39f471377b6da0052
/pygame_project/main.py
dec238616052d072703a724685d1660c382f7741
[]
no_license
dreamerror/practice_projects
52041645ba3c7421b9c7fb1a1f6b376915bd9fd9
9d4010e92cf82f49e7938995ee74a8365ef28c80
refs/heads/master
2023-06-27T20:33:13.936427
2021-07-31T03:48:11
2021-07-31T03:48:11
389,886,395
0
0
null
null
null
null
UTF-8
Python
false
false
7,446
py
import pygame import random import sys from pygame.locals import DOUBLEBUF, KEYUP, K_RIGHT, K_LEFT, K_UP, K_DOWN, K_SPACE, QUIT pygame.init() NUM_SHAPES = 5 PUZZLE_COLUMNS = 6 PUZZLE_ROWS = 12 SHAPE_WIDTH = 50 SHAPE_HEIGHT = 50 FPS = 15 WINDOW_WIDTH = PUZZLE_COLUMNS * SHAPE_WIDTH WINDOW_HEIGHT = PUZZLE_ROWS * SHAPE_HEIGHT + 75 BACKGROUND = pygame.image.load("images/background.png") JEWEL_1 = pygame.image.load("images/jewel_1.png") JEWEL_2 = pygame.image.load("images/jewel_2.png") JEWEL_3 = pygame.image.load("images/jewel_3.png") JEWEL_4 = pygame.image.load("images/jewel_4.png") JEWEL_5 = pygame.image.load("images/jewel_5.png") SHAPES_LIST = [JEWEL_1, JEWEL_2, JEWEL_3, JEWEL_4, JEWEL_5] BLANK = pygame.image.load("images/background.png") WHITE = (255, 255, 255) BLACK = (0, 0, 0) FONT_SIZE = 36 TEXT_OFFSET = 5 MINIMUM_MATCH = 3 SINGLE_POINTS = 1 DOUBLE_POINTS = 3 TRIPLE_POINTS = 9 EXTRA_LENGTH_POINTS = 0 RANDOM_POINTS = 1 FPS_CLOCK = pygame.time.Clock() DISPLAY_SURFACE = pygame.display.set_mode((WINDOW_WIDTH, WINDOW_HEIGHT), DOUBLEBUF) pygame.display.set_caption("Puzzle") def main(): global score global selector jewels_board = generate_random_board() selector = (0, 0) score = 0 last_move_time = pygame.time.get_ticks() blit_board(jewels_board) draw_selector(selector) remove_matches(jewels_board, selector) blit_board(jewels_board) blit_score(score) draw_selector(selector) while True: for event in pygame.event.get(): if event.type == KEYUP: if event.key == K_RIGHT and selector[0] < (PUZZLE_COLUMNS - 2): selector = (selector[0] + 1, selector[1]) if event.key == K_LEFT and selector[0] > 0: selector = (selector[0] - 1, selector[1]) if event.key == K_DOWN and selector[1] < (PUZZLE_ROWS - 1): selector = (selector[0], selector[1] + 1) if event.key == K_UP and selector[1] > 0: selector = (selector[0], selector[1] - 1) if event.key == K_SPACE: last_move_time = pygame.time.get_ticks() swap_pieces(selector, jewels_board) if find_matches(jewels_board): remove_matches(jewels_board, selector) else: swap_pieces(selector, jewels_board) if event.type == QUIT: pygame.quit() sys.exit() blit_board(jewels_board) blit_score(score) draw_selector(selector) pygame.display.update() FPS_CLOCK.tick(FPS) def generate_random_board(): return [[random.choice(SHAPES_LIST) for _ in range(PUZZLE_COLUMNS)] for _ in range(PUZZLE_ROWS)] def blit_score(score): font = pygame.font.Font(None, FONT_SIZE) text = font.render("Score: " + str(score), True, BLACK) DISPLAY_SURFACE.blit(text, (TEXT_OFFSET, WINDOW_HEIGHT - FONT_SIZE)) def blit_board(board): DISPLAY_SURFACE.blit(BACKGROUND, (0, 0)) row_num = 0 for row in board: column_num = 0 for shape in row: DISPLAY_SURFACE.blit(shape, (SHAPE_WIDTH * column_num, SHAPE_HEIGHT * row_num)) column_num += 1 row_num += 1 def draw_selector(position): top_left = (position[0] * SHAPE_WIDTH, position[1] * SHAPE_HEIGHT) top_right = (top_left[0] + SHAPE_WIDTH * 2, top_left[1]) bottom_left = (top_left[0], top_left[1] + SHAPE_HEIGHT) bottom_right = (top_right[0], top_right[1] + SHAPE_HEIGHT) pygame.draw.lines(DISPLAY_SURFACE, WHITE, True, [top_left, top_right, bottom_right, bottom_left], 3) def swap_pieces(position, board): x, y = position board[y][x + 1], board[y][x] = board[y][x], board[y][x + 1] def remove_matches(board, selector): matches = find_matches(board) while matches: score_matches(board, selector, matches) clear_matches(board, matches) refill_columns(board) matches = find_matches(board) selector = (0, 0) def score_matches(board, selector, matches): global score player_matches = [] selector = (selector[1], selector[0]) for match in matches: for position in match: if (position == selector or position == (selector[0], selector[1] + 1)) and (not match in player_matches): player_matches.append(match) if len(player_matches) == 1: score += SINGLE_POINTS elif len(player_matches) == 2: score += DOUBLE_POINTS elif len(player_matches) == 3: score += TRIPLE_POINTS for match in player_matches: score += int((len(match) - MINIMUM_MATCH) * EXTRA_LENGTH_POINTS) for match in matches: if match not in player_matches: score += RANDOM_POINTS def find_matches(board): clear_list = [] for column in range(PUZZLE_COLUMNS): length = 1 for row in range(1, PUZZLE_ROWS): if board[row][column] == board[row - 1][column]: length += 1 if not board[row][column] == board[row - 1][column]: if length >= MINIMUM_MATCH: match = [] for clearRow in range(row - length, row): match.append((clearRow, column)) clear_list.append(match) length = 1 if row == PUZZLE_ROWS - 1: if length >= MINIMUM_MATCH: match = [] for clearRow in range(row - (length - 1), row + 1): match.append((clearRow, column)) clear_list.append(match) for row in range(PUZZLE_ROWS): length = 1 for column in range(1, PUZZLE_COLUMNS): if board[row][column] == board[row][column - 1]: length += 1 if not board[row][column] == board[row][column - 1]: if length >= MINIMUM_MATCH: match = [] for clear_column in range(column - length, column): match.append((row, clear_column)) clear_list.append(match) length = 1 if column == PUZZLE_COLUMNS - 1: if length >= MINIMUM_MATCH: match = [] for clear_column in range(column - (length - 1), column + 1): match.append((row, clear_column)) clear_list.append(match) return clear_list def clear_matches(board, matches): for match in matches: for position in match: row, column = position board[row][column] = BLANK def refill_columns(board): for column in range(PUZZLE_COLUMNS): for row in range(PUZZLE_ROWS): if board[row][column] == BLANK: test = 0 length = 0 while row + test < PUZZLE_ROWS and board[row + test][column] == BLANK: length += 1 test += 1 for blank_row in range(row, PUZZLE_ROWS): try: board[blank_row][column] = board[blank_row + length][column] except IndexError: board[blank_row][column] = SHAPES_LIST[random.randrange(0, len(SHAPES_LIST))] if __name__ == '__main__': main()
[ "66074257+dreamerror@users.noreply.github.com" ]
66074257+dreamerror@users.noreply.github.com
23eee071d0d6157fda786b1c2e638cb649ec482e
10261f7a8864a8db3970d9e8ecc26645cf7f901c
/custom_bandits/custom_bandits/envs/custom_twoArm_env.py
f2b8aeadf05c39267ec261d52a8c36356451f023
[]
no_license
abhikmr778/CS698R-ABHINAV-KUMAR-16907018-ASSIGN-1
6219bf94e86b37a55082e761ec291a788ad50351
ff722b7d9d27fbe6571063a26e04e48efb8f10f3
refs/heads/main
2023-07-29T06:47:26.097700
2021-09-11T11:13:32
2021-09-11T11:13:32
405,250,244
0
0
null
null
null
null
UTF-8
Python
false
false
2,723
py
########################################################################### # # # Environment Class template followed from # # https://stable-baselines.readthedocs.io/en/master/guide/custom_env.html # # # ########################################################################### import gym from gym import spaces import numpy as np class TwoArmBandit(gym.Env): # metadata = {'render.modes':['human']} def __init__(self, alpha, beta, seed): super(TwoArmBandit, self).__init__() N_DISCRETE_ACTIONS = 2 LEFT = 0 RIGHT = 1 N_DISCRETE_STATES = 3 self.alpha = alpha self.beta = beta self.action_space = spaces.Discrete(N_DISCRETE_ACTIONS) self.observation_space = spaces.Discrete(N_DISCRETE_STATES) # Data structure to store MDP of 2-arm Bernoulli Bandit self.P = {} self.P[0] = { LEFT: [[self.alpha, 1, 0, True], [1-self.alpha, 2, 1, True]], RIGHT: [[self.beta, 2, 1, True], [1-self.beta, 1, 0, True]] } self.P[1] = { LEFT: [[1,1,0,True]], RIGHT: [[1,1,0,True]] } self.P[2] = { LEFT: [[1,2,0,True]], RIGHT: [[1,2,0,True]] } self.q_value = np.array([1-self.alpha,self.beta]) self.agent_position = self.reset() self.seed(seed) def step(self, action): # get transition from MDP dynamics probabilities = [] next_states = [] rewards = [] dones = [] # sample transition according to the probabilities of transition function for dynamic in self.P[self.agent_position][action]: probabilities.append(dynamic[0]) idx = [i for i in range(len(self.P[self.agent_position][action]))] j = int(np.random.choice(a=idx,size=1,p=probabilities)) _, observation, reward, done = self.P[self.agent_position][action][j] # update agent's position self.agent_position = observation info = {} return observation, reward, done, info def reset(self): self.agent_position = 0 return self.agent_position # reward, done, info can't be included def seed(self, seed=0): np.random.seed(seed) def render(self, mode='human'): raise NotImplementedError def close (self): raise NotImplementedError
[ "abhi.kumar778@gmail.com" ]
abhi.kumar778@gmail.com
a9280167358863b7781ffed296d8c21e28ee3189
f0a1b45309785a89bf74eccb75e229fd38a0aaa8
/math_utils.py
48245f597c0819e491d64268d76f8b6fe85245c1
[]
no_license
mikaelmello/python-rsa-implementation
f8f8ecb7f79d8808b4d9ed520c14666560315a2b
7b501253e41fc4c9f07db02d73896c9c189c8ebf
refs/heads/master
2020-09-21T13:06:19.271833
2020-04-07T18:51:35
2020-04-07T18:53:23
224,797,536
0
0
null
null
null
null
UTF-8
Python
false
false
2,051
py
import secrets import math def extended_euclid(a, b): """Extended euclidean algorithm""" s = 0 t = 1 r = b old_s = 1 old_t = 0 old_r = a while r != 0: quotient = old_r // r old_r, r = (r, old_r - quotient * r) old_s, s = (s, old_s - quotient * s) old_t, t = (t, old_t - quotient * t) return old_r, old_s, old_t def mrand(l, r): """Generates a random number between l and r, inclusive""" sz = r - l + 1 return l + secrets.randbelow(sz) def bits(a): count = 0 while a > 0: a = a // 2 count += 1 return count def mulmod(a, b, c): """Returns (a*b)%c""" return (a*b) % c def fexp(num, exponent, mod): """Fast exponentiation, returns (num^exponent)%mod.""" res = 1 while exponent > 0: if exponent % 2 == 1: res = (res * num) % mod exponent = exponent // 2 num = (num * num) % mod return res def is_prime(n): """Checks whether a number is prime""" if n <= 1: return False if n <= 3: return True s = 0 d = n - 1 while(d % 2 == 0): d = d // 2 s = s + 1 for k in range(64): a = mrand(2, n) x = fexp(a, d, n) if x != 1 and x != n-1: for r in range(1, s): x = mulmod(x, x, n) if x == 1: return False if x == n-1: break if x != n-1: return False return True def gen_prime(bit_length): prime = secrets.randbits(bit_length) prime = prime | 1 prime = prime | (1 << (bit_length-1)) prime = prime | (1 << (bit_length-2)) while not is_prime(prime): prime += 2 return prime def get_large_primes(): """Generates two large primes""" n = mrand(246, 256) base_num = 2**n x = base_num while(not is_prime(x)): x = x + 1 while(not is_prime(base_num)): base_num = base_num - 1 return (x, base_num)
[ "git@mikaelmello.com" ]
git@mikaelmello.com
0e10cbdfa5b1cd030ad2bdd01ff695d9fdb60938
d88868b88864e4d10009c58b27323034715b0277
/projects/barter/deployments/docker/barter/test.py
1ff68bc01c1f1c3d50a8992e741a015cb27fa14d
[ "Apache-2.0" ]
permissive
shamal112mn/packer-1
795ebd9e0fee444f0cbb01897a50e199b73f1307
be2720e9cb61bf1110a880e94e32a6767a341588
refs/heads/master
2023-07-10T09:33:10.516559
2021-08-17T02:25:50
2021-08-17T02:25:50
327,763,823
0
0
null
2021-01-08T01:08:49
2021-01-08T01:08:48
null
UTF-8
Python
false
false
388
py
import requests import json url = "https://investors-exchange-iex-trading.p.rapidapi.com/stock/tsla/effective-spread" headers = { 'x-rapidapi-key': "158cd4f9cdmsh0d92f8b92b1d427p1947b6jsn857aa1252e0b", 'x-rapidapi-host': "investors-exchange-iex-trading.p.rapidapi.com" } response = requests.request("GET", url, headers=headers) print(json.dumps(response.json(), indent=2))
[ "you@example.com" ]
you@example.com
e9880252c3b8871fcba2e2f278da14e2c40131dc
9ae6ce54bf9a2a86201961fdbd5e7b0ec913ff56
/google/ads/googleads/v10/services/services/custom_conversion_goal_service/client.py
6a590a9ac28792be1d45d85fcb1db11c168b6b0e
[ "Apache-2.0" ]
permissive
GerhardusM/google-ads-python
73b275a06e5401e6b951a6cd99af98c247e34aa3
676ac5fcb5bec0d9b5897f4c950049dac5647555
refs/heads/master
2022-07-06T19:05:50.932553
2022-06-17T20:41:17
2022-06-17T20:41:17
207,535,443
0
0
Apache-2.0
2019-09-10T10:58:55
2019-09-10T10:58:55
null
UTF-8
Python
false
false
21,247
py
# -*- coding: utf-8 -*- # Copyright 2022 Google LLC # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # from collections import OrderedDict import os import re from typing import Dict, Optional, Sequence, Tuple, Type, Union import pkg_resources from google.api_core import client_options as client_options_lib from google.api_core import gapic_v1 from google.api_core import retry as retries from google.auth import credentials as ga_credentials # type: ignore from google.auth.transport import mtls # type: ignore from google.auth.transport.grpc import SslCredentials # type: ignore from google.auth.exceptions import MutualTLSChannelError # type: ignore from google.oauth2 import service_account # type: ignore try: OptionalRetry = Union[retries.Retry, gapic_v1.method._MethodDefault] except AttributeError: # pragma: NO COVER OptionalRetry = Union[retries.Retry, object] # type: ignore from google.ads.googleads.v10.services.types import ( custom_conversion_goal_service, ) from .transports.base import ( CustomConversionGoalServiceTransport, DEFAULT_CLIENT_INFO, ) from .transports.grpc import CustomConversionGoalServiceGrpcTransport class CustomConversionGoalServiceClientMeta(type): """Metaclass for the CustomConversionGoalService client. This provides class-level methods for building and retrieving support objects (e.g. transport) without polluting the client instance objects. """ _transport_registry = ( OrderedDict() ) # type: Dict[str, Type[CustomConversionGoalServiceTransport]] _transport_registry["grpc"] = CustomConversionGoalServiceGrpcTransport def get_transport_class( cls, label: str = None, ) -> Type[CustomConversionGoalServiceTransport]: """Returns an appropriate transport class. Args: label: The name of the desired transport. If none is provided, then the first transport in the registry is used. Returns: The transport class to use. """ # If a specific transport is requested, return that one. if label: return cls._transport_registry[label] # No transport is requested; return the default (that is, the first one # in the dictionary). return next(iter(cls._transport_registry.values())) class CustomConversionGoalServiceClient( metaclass=CustomConversionGoalServiceClientMeta ): """Service to manage custom conversion goal.""" @staticmethod def _get_default_mtls_endpoint(api_endpoint): """Converts api endpoint to mTLS endpoint. Convert "*.sandbox.googleapis.com" and "*.googleapis.com" to "*.mtls.sandbox.googleapis.com" and "*.mtls.googleapis.com" respectively. Args: api_endpoint (Optional[str]): the api endpoint to convert. Returns: str: converted mTLS api endpoint. """ if not api_endpoint: return api_endpoint mtls_endpoint_re = re.compile( r"(?P<name>[^.]+)(?P<mtls>\.mtls)?(?P<sandbox>\.sandbox)?(?P<googledomain>\.googleapis\.com)?" ) m = mtls_endpoint_re.match(api_endpoint) name, mtls, sandbox, googledomain = m.groups() if mtls or not googledomain: return api_endpoint if sandbox: return api_endpoint.replace( "sandbox.googleapis.com", "mtls.sandbox.googleapis.com" ) return api_endpoint.replace(".googleapis.com", ".mtls.googleapis.com") DEFAULT_ENDPOINT = "googleads.googleapis.com" DEFAULT_MTLS_ENDPOINT = _get_default_mtls_endpoint.__func__( # type: ignore DEFAULT_ENDPOINT ) @classmethod def from_service_account_info(cls, info: dict, *args, **kwargs): """Creates an instance of this client using the provided credentials info. Args: info (dict): The service account private key info. args: Additional arguments to pass to the constructor. kwargs: Additional arguments to pass to the constructor. Returns: CustomConversionGoalServiceClient: The constructed client. """ credentials = service_account.Credentials.from_service_account_info( info ) kwargs["credentials"] = credentials return cls(*args, **kwargs) @classmethod def from_service_account_file(cls, filename: str, *args, **kwargs): """Creates an instance of this client using the provided credentials file. Args: filename (str): The path to the service account private key json file. args: Additional arguments to pass to the constructor. kwargs: Additional arguments to pass to the constructor. Returns: CustomConversionGoalServiceClient: The constructed client. """ credentials = service_account.Credentials.from_service_account_file( filename ) kwargs["credentials"] = credentials return cls(*args, **kwargs) from_service_account_json = from_service_account_file @property def transport(self) -> CustomConversionGoalServiceTransport: """Returns the transport used by the client instance. Returns: CustomConversionGoalServiceTransport: The transport used by the client instance. """ return self._transport def __enter__(self): return self def __exit__(self, type, value, traceback): """Releases underlying transport's resources. .. warning:: ONLY use as a context manager if the transport is NOT shared with other clients! Exiting the with block will CLOSE the transport and may cause errors in other clients! """ self.transport.close() @staticmethod def conversion_action_path( customer_id: str, conversion_action_id: str, ) -> str: """Returns a fully-qualified conversion_action string.""" return "customers/{customer_id}/conversionActions/{conversion_action_id}".format( customer_id=customer_id, conversion_action_id=conversion_action_id, ) @staticmethod def parse_conversion_action_path(path: str) -> Dict[str, str]: """Parses a conversion_action path into its component segments.""" m = re.match( r"^customers/(?P<customer_id>.+?)/conversionActions/(?P<conversion_action_id>.+?)$", path, ) return m.groupdict() if m else {} @staticmethod def custom_conversion_goal_path( customer_id: str, goal_id: str, ) -> str: """Returns a fully-qualified custom_conversion_goal string.""" return "customers/{customer_id}/customConversionGoals/{goal_id}".format( customer_id=customer_id, goal_id=goal_id, ) @staticmethod def parse_custom_conversion_goal_path(path: str) -> Dict[str, str]: """Parses a custom_conversion_goal path into its component segments.""" m = re.match( r"^customers/(?P<customer_id>.+?)/customConversionGoals/(?P<goal_id>.+?)$", path, ) return m.groupdict() if m else {} @staticmethod def common_billing_account_path( billing_account: str, ) -> str: """Returns a fully-qualified billing_account string.""" return "billingAccounts/{billing_account}".format( billing_account=billing_account, ) @staticmethod def parse_common_billing_account_path(path: str) -> Dict[str, str]: """Parse a billing_account path into its component segments.""" m = re.match(r"^billingAccounts/(?P<billing_account>.+?)$", path) return m.groupdict() if m else {} @staticmethod def common_folder_path( folder: str, ) -> str: """Returns a fully-qualified folder string.""" return "folders/{folder}".format( folder=folder, ) @staticmethod def parse_common_folder_path(path: str) -> Dict[str, str]: """Parse a folder path into its component segments.""" m = re.match(r"^folders/(?P<folder>.+?)$", path) return m.groupdict() if m else {} @staticmethod def common_organization_path( organization: str, ) -> str: """Returns a fully-qualified organization string.""" return "organizations/{organization}".format( organization=organization, ) @staticmethod def parse_common_organization_path(path: str) -> Dict[str, str]: """Parse a organization path into its component segments.""" m = re.match(r"^organizations/(?P<organization>.+?)$", path) return m.groupdict() if m else {} @staticmethod def common_project_path( project: str, ) -> str: """Returns a fully-qualified project string.""" return "projects/{project}".format( project=project, ) @staticmethod def parse_common_project_path(path: str) -> Dict[str, str]: """Parse a project path into its component segments.""" m = re.match(r"^projects/(?P<project>.+?)$", path) return m.groupdict() if m else {} @staticmethod def common_location_path( project: str, location: str, ) -> str: """Returns a fully-qualified location string.""" return "projects/{project}/locations/{location}".format( project=project, location=location, ) @staticmethod def parse_common_location_path(path: str) -> Dict[str, str]: """Parse a location path into its component segments.""" m = re.match( r"^projects/(?P<project>.+?)/locations/(?P<location>.+?)$", path ) return m.groupdict() if m else {} def __init__( self, *, credentials: Optional[ga_credentials.Credentials] = None, transport: Union[ str, CustomConversionGoalServiceTransport, None ] = None, client_options: Optional[client_options_lib.ClientOptions] = None, client_info: gapic_v1.client_info.ClientInfo = DEFAULT_CLIENT_INFO, ) -> None: """Instantiates the custom conversion goal service client. Args: credentials (Optional[google.auth.credentials.Credentials]): The authorization credentials to attach to requests. These credentials identify the application to the service; if none are specified, the client will attempt to ascertain the credentials from the environment. transport (Union[str, CustomConversionGoalServiceTransport]): The transport to use. If set to None, a transport is chosen automatically. client_options (google.api_core.client_options.ClientOptions): Custom options for the client. It won't take effect if a ``transport`` instance is provided. (1) The ``api_endpoint`` property can be used to override the default endpoint provided by the client. GOOGLE_API_USE_MTLS_ENDPOINT environment variable can also be used to override the endpoint: "always" (always use the default mTLS endpoint), "never" (always use the default regular endpoint) and "auto" (auto switch to the default mTLS endpoint if client certificate is present, this is the default value). However, the ``api_endpoint`` property takes precedence if provided. (2) If GOOGLE_API_USE_CLIENT_CERTIFICATE environment variable is "true", then the ``client_cert_source`` property can be used to provide client certificate for mutual TLS transport. If not provided, the default SSL client certificate will be used if present. If GOOGLE_API_USE_CLIENT_CERTIFICATE is "false" or not set, no client certificate will be used. client_info (google.api_core.gapic_v1.client_info.ClientInfo): The client info used to send a user-agent string along with API requests. If ``None``, then default info will be used. Generally, you only need to set this if you're developing your own client library. Raises: google.auth.exceptions.MutualTLSChannelError: If mutual TLS transport creation failed for any reason. """ if isinstance(client_options, dict): client_options = client_options_lib.from_dict(client_options) if client_options is None: client_options = client_options_lib.ClientOptions() # Create SSL credentials for mutual TLS if needed. if os.getenv("GOOGLE_API_USE_CLIENT_CERTIFICATE", "false") not in ( "true", "false", ): raise ValueError( "Environment variable `GOOGLE_API_USE_CLIENT_CERTIFICATE` must be either `true` or `false`" ) use_client_cert = ( os.getenv("GOOGLE_API_USE_CLIENT_CERTIFICATE", "false") == "true" ) client_cert_source_func = None is_mtls = False if use_client_cert: if client_options.client_cert_source: is_mtls = True client_cert_source_func = client_options.client_cert_source else: is_mtls = mtls.has_default_client_cert_source() if is_mtls: client_cert_source_func = mtls.default_client_cert_source() else: client_cert_source_func = None # Figure out which api endpoint to use. if client_options.api_endpoint is not None: api_endpoint = client_options.api_endpoint else: use_mtls_env = os.getenv("GOOGLE_API_USE_MTLS_ENDPOINT", "auto") if use_mtls_env == "never": api_endpoint = self.DEFAULT_ENDPOINT elif use_mtls_env == "always": api_endpoint = self.DEFAULT_MTLS_ENDPOINT elif use_mtls_env == "auto": api_endpoint = ( self.DEFAULT_MTLS_ENDPOINT if is_mtls else self.DEFAULT_ENDPOINT ) else: raise MutualTLSChannelError( "Unsupported GOOGLE_API_USE_MTLS_ENDPOINT value. Accepted " "values: never, auto, always" ) # Save or instantiate the transport. # Ordinarily, we provide the transport, but allowing a custom transport # instance provides an extensibility point for unusual situations. if isinstance(transport, CustomConversionGoalServiceTransport): # transport is a CustomConversionGoalServiceTransport instance. if credentials or client_options.credentials_file: raise ValueError( "When providing a transport instance, " "provide its credentials directly." ) if client_options.scopes: raise ValueError( "When providing a transport instance, provide its scopes " "directly." ) self._transport = transport else: Transport = type(self).get_transport_class(transport) self._transport = Transport( credentials=credentials, credentials_file=client_options.credentials_file, host=api_endpoint, scopes=client_options.scopes, client_cert_source_for_mtls=client_cert_source_func, quota_project_id=client_options.quota_project_id, client_info=client_info, always_use_jwt_access=True, ) def mutate_custom_conversion_goals( self, request: Union[ custom_conversion_goal_service.MutateCustomConversionGoalsRequest, dict, ] = None, *, customer_id: str = None, operations: Sequence[ custom_conversion_goal_service.CustomConversionGoalOperation ] = None, retry: OptionalRetry = gapic_v1.method.DEFAULT, timeout: float = None, metadata: Sequence[Tuple[str, str]] = (), ) -> custom_conversion_goal_service.MutateCustomConversionGoalsResponse: r"""Creates, updates or removes custom conversion goals. Operation statuses are returned. Args: request (Union[google.ads.googleads.v10.services.types.MutateCustomConversionGoalsRequest, dict]): The request object. Request message for [CustomConversionGoalService.MutateCustomConversionGoals][google.ads.googleads.v10.services.CustomConversionGoalService.MutateCustomConversionGoals]. customer_id (str): Required. The ID of the customer whose custom conversion goals are being modified. This corresponds to the ``customer_id`` field on the ``request`` instance; if ``request`` is provided, this should not be set. operations (Sequence[google.ads.googleads.v10.services.types.CustomConversionGoalOperation]): Required. The list of operations to perform on individual custom conversion goal. This corresponds to the ``operations`` field on the ``request`` instance; if ``request`` is provided, this should not be set. retry (google.api_core.retry.Retry): Designation of what errors, if any, should be retried. timeout (float): The timeout for this request. metadata (Sequence[Tuple[str, str]]): Strings which should be sent along with the request as metadata. Returns: google.ads.googleads.v10.services.types.MutateCustomConversionGoalsResponse: Response message for a custom conversion goal mutate. """ # Create or coerce a protobuf request object. # Quick check: If we got a request object, we should *not* have # gotten any keyword arguments that map to the request. has_flattened_params = any([customer_id, operations]) if request is not None and has_flattened_params: raise ValueError( "If the `request` argument is set, then none of " "the individual field arguments should be set." ) # Minor optimization to avoid making a copy if the user passes # in a custom_conversion_goal_service.MutateCustomConversionGoalsRequest. # There's no risk of modifying the input as we've already verified # there are no flattened fields. if not isinstance( request, custom_conversion_goal_service.MutateCustomConversionGoalsRequest, ): request = custom_conversion_goal_service.MutateCustomConversionGoalsRequest( request ) # If we have keyword arguments corresponding to fields on the # request, apply these. if customer_id is not None: request.customer_id = customer_id if operations is not None: request.operations = operations # Wrap the RPC method; this adds retry and timeout information, # and friendly error handling. rpc = self._transport._wrapped_methods[ self._transport.mutate_custom_conversion_goals ] # Certain fields should be provided within the metadata header; # add these here. metadata = tuple(metadata) + ( gapic_v1.routing_header.to_grpc_metadata( (("customer_id", request.customer_id),) ), ) # Send the request. response = rpc( request, retry=retry, timeout=timeout, metadata=metadata, ) # Done; return the response. return response try: DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo( gapic_version=pkg_resources.get_distribution( "google-ads", ).version, ) except pkg_resources.DistributionNotFound: DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo() __all__ = ("CustomConversionGoalServiceClient",)
[ "noreply@github.com" ]
noreply@github.com
b9aae4c3e88a7792acd442dd2d9d158dd3d47ae4
fb8cbebdf034b2f478943752d5443afc82c6eef5
/tuirer/venv/lib/python3.6/site-packages/IPython/core/inputtransformer.py
44ec5a1aae19b8708a617982258ffa6ec5f11bd0
[]
no_license
fariasjr/CitiTuirer
f64e0ec93ef088f8140bb0961d2ad4ed3b59448a
deb3f7a9c2d45b8a7f54639037f097b99abdac11
refs/heads/master
2020-03-24T05:10:36.261050
2018-08-01T20:24:30
2018-08-01T20:24:30
142,477,521
0
0
null
null
null
null
UTF-8
Python
false
false
18,242
py
"""Input transformer classes to support IPython special syntax. This includes the machinery to recognise and transform ``%magic`` commands, ``!system`` commands, ``help?`` querying, prompt stripping, and so forth. """ import abc import functools import re from io import StringIO from IPython.core.splitinput import LineInfo from IPython.utils import tokenize2 from IPython.utils.tokenize2 import TokenError, generate_tokens, untokenize #----------------------------------------------------------------------------- # Globals #----------------------------------------------------------------------------- # The escape sequences that define the syntax transformations IPython will # apply to user input. These can NOT be just changed here: many regular # expressions and other parts of the code may use their hardcoded values, and # for all intents and purposes they constitute the 'IPython syntax', so they # should be considered fixed. ESC_SHELL = '!' # Send line to underlying system shell ESC_SH_CAP = '!!' # Send line to system shell and capture output ESC_HELP = '?' # Find information about object ESC_HELP2 = '??' # Find extra-detailed information about object ESC_MAGIC = '%' # Call magic function ESC_MAGIC2 = '%%' # Call cell-magic function ESC_QUOTE = ',' # Split args on whitespace, quote each as string and call ESC_QUOTE2 = ';' # Quote all args as a single string, call ESC_PAREN = '/' # Call first argument with rest of line as arguments ESC_SEQUENCES = [ESC_SHELL, ESC_SH_CAP, ESC_HELP ,\ ESC_HELP2, ESC_MAGIC, ESC_MAGIC2,\ ESC_QUOTE, ESC_QUOTE2, ESC_PAREN ] class InputTransformer(metaclass=abc.ABCMeta): """Abstract base class for line-based input transformers.""" @abc.abstractmethod def push(self, line): """Send a line of input to the transformer, returning the transformed input or None if the transformer is waiting for more input. Must be overridden by subclasses. Implementations may raise ``SyntaxError`` if the input is invalid. No other exceptions may be raised. """ pass @abc.abstractmethod def reset(self): """Return, transformed any lines that the transformer has accumulated, and reset its internal state. Must be overridden by subclasses. """ pass @classmethod def wrap(cls, func): """Can be used by subclasses as a decorator, to return a factory that will allow instantiation with the decorated object. """ @functools.wraps(func) def transformer_factory(**kwargs): return cls(func, **kwargs) return transformer_factory class StatelessInputTransformer(InputTransformer): """Wrapper for a stateless input transformer implemented as a function.""" def __init__(self, func): self.func = func def __repr__(self): return "StatelessInputTransformer(func={0!r})".format(self.func) def push(self, line): """Send a line of input to the transformer, returning the transformed input.""" return self.func(line) def reset(self): """No-op - exists for compatibility.""" pass class CoroutineInputTransformer(InputTransformer): """Wrapper for an input transformer implemented as a coroutine.""" def __init__(self, coro, **kwargs): # Prime it self.coro = coro(**kwargs) next(self.coro) def __repr__(self): return "CoroutineInputTransformer(coro={0!r})".format(self.coro) def push(self, line): """Send a line of input to the transformer, returning the transformed input or None if the transformer is waiting for more input. """ return self.coro.send(line) def reset(self): """Return, transformed any lines that the transformer has accumulated, and reset its internal state. """ return self.coro.send(None) class TokenInputTransformer(InputTransformer): """Wrapper for a token-based input transformer. func should accept a list of tokens (5-tuples, see tokenize docs), and return an iterable which can be passed to tokenize.untokenize(). """ def __init__(self, func): self.func = func self.buf = [] self.reset_tokenizer() def reset_tokenizer(self): it = iter(self.buf) self.tokenizer = generate_tokens(it.__next__) def push(self, line): self.buf.append(line + '\n') if all(l.isspace() for l in self.buf): return self.reset() tokens = [] stop_at_NL = False try: for intok in self.tokenizer: tokens.append(intok) t = intok[0] if t == tokenize2.NEWLINE or (stop_at_NL and t == tokenize2.NL): # Stop before we try to pull a line we don't have yet break elif t == tokenize2.ERRORTOKEN: stop_at_NL = True except TokenError: # Multi-line statement - stop and try again with the next line self.reset_tokenizer() return None return self.output(tokens) def output(self, tokens): self.buf.clear() self.reset_tokenizer() return untokenize(self.func(tokens)).rstrip('\n') def reset(self): l = ''.join(self.buf) self.buf.clear() self.reset_tokenizer() if l: return l.rstrip('\n') class assemble_python_lines(TokenInputTransformer): def __init__(self): super(assemble_python_lines, self).__init__(None) def output(self, tokens): return self.reset() @CoroutineInputTransformer.wrap def assemble_logical_lines(): """Join lines following explicit line continuations (\)""" line = '' while True: line = (yield line) if not line or line.isspace(): continue parts = [] while line is not None: if line.endswith('\\') and (not has_comment(line)): parts.append(line[:-1]) line = (yield None) # Get another line else: parts.append(line) break # Output line = ''.join(parts) # Utilities def _make_help_call(target, esc, lspace, next_input=None): """Prepares a pinfo(2)/psearch call from a target name and the escape (i.e. ? or ??)""" method = 'pinfo2' if esc == '??' \ else 'psearch' if '*' in target \ else 'pinfo' arg = " ".join([method, target]) #Prepare arguments for get_ipython().run_line_magic(magic_name, magic_args) t_magic_name, _, t_magic_arg_s = arg.partition(' ') t_magic_name = t_magic_name.lstrip(ESC_MAGIC) if next_input is None: return '%sget_ipython().run_line_magic(%r, %r)' % (lspace, t_magic_name, t_magic_arg_s) else: return '%sget_ipython().set_next_input(%r);get_ipython().run_line_magic(%r, %r)' % \ (lspace, next_input, t_magic_name, t_magic_arg_s) # These define the transformations for the different escape characters. def _tr_system(line_info): "Translate lines escaped with: !" cmd = line_info.line.lstrip().lstrip(ESC_SHELL) return '%sget_ipython().system(%r)' % (line_info.pre, cmd) def _tr_system2(line_info): "Translate lines escaped with: !!" cmd = line_info.line.lstrip()[2:] return '%sget_ipython().getoutput(%r)' % (line_info.pre, cmd) def _tr_help(line_info): "Translate lines escaped with: ?/??" # A naked help line should just fire the intro help screen if not line_info.line[1:]: return 'get_ipython().show_usage()' return _make_help_call(line_info.ifun, line_info.esc, line_info.pre) def _tr_magic(line_info): "Translate lines escaped with: %" tpl = '%sget_ipython().run_line_magic(%r, %r)' if line_info.line.startswith(ESC_MAGIC2): return line_info.line cmd = ' '.join([line_info.ifun, line_info.the_rest]).strip() #Prepare arguments for get_ipython().run_line_magic(magic_name, magic_args) t_magic_name, _, t_magic_arg_s = cmd.partition(' ') t_magic_name = t_magic_name.lstrip(ESC_MAGIC) return tpl % (line_info.pre, t_magic_name, t_magic_arg_s) def _tr_quote(line_info): "Translate lines escaped with: ," return '%s%s("%s")' % (line_info.pre, line_info.ifun, '", "'.join(line_info.the_rest.split()) ) def _tr_quote2(line_info): "Translate lines escaped with: ;" return '%s%s("%s")' % (line_info.pre, line_info.ifun, line_info.the_rest) def _tr_paren(line_info): "Translate lines escaped with: /" return '%s%s(%s)' % (line_info.pre, line_info.ifun, ", ".join(line_info.the_rest.split())) tr = { ESC_SHELL : _tr_system, ESC_SH_CAP : _tr_system2, ESC_HELP : _tr_help, ESC_HELP2 : _tr_help, ESC_MAGIC : _tr_magic, ESC_QUOTE : _tr_quote, ESC_QUOTE2 : _tr_quote2, ESC_PAREN : _tr_paren } @StatelessInputTransformer.wrap def escaped_commands(line): """Transform escaped commands - %magic, !system, ?help + various autocalls. """ if not line or line.isspace(): return line lineinf = LineInfo(line) if lineinf.esc not in tr: return line return tr[lineinf.esc](lineinf) _initial_space_re = re.compile(r'\s*') _help_end_re = re.compile(r"""(%{0,2} [a-zA-Z_*][\w*]* # Variable name (\.[a-zA-Z_*][\w*]*)* # .etc.etc ) (\?\??)$ # ? or ?? """, re.VERBOSE) # Extra pseudotokens for multiline strings and data structures _MULTILINE_STRING = object() _MULTILINE_STRUCTURE = object() def _line_tokens(line): """Helper for has_comment and ends_in_comment_or_string.""" readline = StringIO(line).readline toktypes = set() try: for t in generate_tokens(readline): toktypes.add(t[0]) except TokenError as e: # There are only two cases where a TokenError is raised. if 'multi-line string' in e.args[0]: toktypes.add(_MULTILINE_STRING) else: toktypes.add(_MULTILINE_STRUCTURE) return toktypes def has_comment(src): """Indicate whether an input line has (i.e. ends in, or is) a comment. This uses tokenize, so it can distinguish comments from # inside strings. Parameters ---------- src : string A single line input string. Returns ------- comment : bool True if source has a comment. """ return (tokenize2.COMMENT in _line_tokens(src)) def ends_in_comment_or_string(src): """Indicates whether or not an input line ends in a comment or within a multiline string. Parameters ---------- src : string A single line input string. Returns ------- comment : bool True if source ends in a comment or multiline string. """ toktypes = _line_tokens(src) return (tokenize2.COMMENT in toktypes) or (_MULTILINE_STRING in toktypes) @StatelessInputTransformer.wrap def help_end(line): """Translate lines with ?/?? at the end""" m = _help_end_re.search(line) if m is None or ends_in_comment_or_string(line): return line target = m.group(1) esc = m.group(3) lspace = _initial_space_re.match(line).group(0) # If we're mid-command, put it back on the next prompt for the user. next_input = line.rstrip('?') if line.strip() != m.group(0) else None return _make_help_call(target, esc, lspace, next_input) @CoroutineInputTransformer.wrap def cellmagic(end_on_blank_line=False): """Captures & transforms cell magics. After a cell magic is started, this stores up any lines it gets until it is reset (sent None). """ tpl = 'get_ipython().run_cell_magic(%r, %r, %r)' cellmagic_help_re = re.compile('%%\w+\?') line = '' while True: line = (yield line) # consume leading empty lines while not line: line = (yield line) if not line.startswith(ESC_MAGIC2): # This isn't a cell magic, idle waiting for reset then start over while line is not None: line = (yield line) continue if cellmagic_help_re.match(line): # This case will be handled by help_end continue first = line body = [] line = (yield None) while (line is not None) and \ ((line.strip() != '') or not end_on_blank_line): body.append(line) line = (yield None) # Output magic_name, _, first = first.partition(' ') magic_name = magic_name.lstrip(ESC_MAGIC2) line = tpl % (magic_name, first, u'\n'.join(body)) def _strip_prompts(prompt_re, initial_re=None, turnoff_re=None): """Remove matching input prompts from a block of input. Parameters ---------- prompt_re : regular expression A regular expression matching any input prompt (including continuation) initial_re : regular expression, optional A regular expression matching only the initial prompt, but not continuation. If no initial expression is given, prompt_re will be used everywhere. Used mainly for plain Python prompts, where the continuation prompt ``...`` is a valid Python expression in Python 3, so shouldn't be stripped. If initial_re and prompt_re differ, only initial_re will be tested against the first line. If any prompt is found on the first two lines, prompts will be stripped from the rest of the block. """ if initial_re is None: initial_re = prompt_re line = '' while True: line = (yield line) # First line of cell if line is None: continue out, n1 = initial_re.subn('', line, count=1) if turnoff_re and not n1: if turnoff_re.match(line): # We're in e.g. a cell magic; disable this transformer for # the rest of the cell. while line is not None: line = (yield line) continue line = (yield out) if line is None: continue # check for any prompt on the second line of the cell, # because people often copy from just after the first prompt, # so we might not see it in the first line. out, n2 = prompt_re.subn('', line, count=1) line = (yield out) if n1 or n2: # Found a prompt in the first two lines - check for it in # the rest of the cell as well. while line is not None: line = (yield prompt_re.sub('', line, count=1)) else: # Prompts not in input - wait for reset while line is not None: line = (yield line) @CoroutineInputTransformer.wrap def classic_prompt(): """Strip the >>>/... prompts of the Python interactive shell.""" # FIXME: non-capturing version (?:...) usable? prompt_re = re.compile(r'^(>>>|\.\.\.)( |$)') initial_re = re.compile(r'^>>>( |$)') # Any %magic/!system is IPython syntax, so we needn't look for >>> prompts turnoff_re = re.compile(r'^[%!]') return _strip_prompts(prompt_re, initial_re, turnoff_re) @CoroutineInputTransformer.wrap def ipy_prompt(): """Strip IPython's In [1]:/...: prompts.""" # FIXME: non-capturing version (?:...) usable? prompt_re = re.compile(r'^(In \[\d+\]: |\s*\.{3,}: ?)') # Disable prompt stripping inside cell magics turnoff_re = re.compile(r'^%%') return _strip_prompts(prompt_re, turnoff_re=turnoff_re) @CoroutineInputTransformer.wrap def leading_indent(): """Remove leading indentation. If the first line starts with a spaces or tabs, the same whitespace will be removed from each following line until it is reset. """ space_re = re.compile(r'^[ \t]+') line = '' while True: line = (yield line) if line is None: continue m = space_re.match(line) if m: space = m.group(0) while line is not None: if line.startswith(space): line = line[len(space):] line = (yield line) else: # No leading spaces - wait for reset while line is not None: line = (yield line) _assign_pat = \ r'''(?P<lhs>(\s*) ([\w\.]+) # Initial identifier (\s*,\s* \*?[\w\.]+)* # Further identifiers for unpacking \s*?,? # Trailing comma ) \s*=\s* ''' assign_system_re = re.compile(r'{}!\s*(?P<cmd>.*)'.format(_assign_pat), re.VERBOSE) assign_system_template = '%s = get_ipython().getoutput(%r)' @StatelessInputTransformer.wrap def assign_from_system(line): """Transform assignment from system commands (e.g. files = !ls)""" m = assign_system_re.match(line) if m is None: return line return assign_system_template % m.group('lhs', 'cmd') assign_magic_re = re.compile(r'{}%\s*(?P<cmd>.*)'.format(_assign_pat), re.VERBOSE) assign_magic_template = '%s = get_ipython().run_line_magic(%r, %r)' @StatelessInputTransformer.wrap def assign_from_magic(line): """Transform assignment from magic commands (e.g. a = %who_ls)""" m = assign_magic_re.match(line) if m is None: return line #Prepare arguments for get_ipython().run_line_magic(magic_name, magic_args) m_lhs, m_cmd = m.group('lhs', 'cmd') t_magic_name, _, t_magic_arg_s = m_cmd.partition(' ') t_magic_name = t_magic_name.lstrip(ESC_MAGIC) return assign_magic_template % (m_lhs, t_magic_name, t_magic_arg_s)
[ "jornadaciti@ug4c08.windows.cin.ufpe.br" ]
jornadaciti@ug4c08.windows.cin.ufpe.br
633d089ce731df7f5e583fa5d1a666bf329b9e9a
446d500d0c7ca5b46c21c70db820d1b2dac11a38
/comp11 - intro/proj1/helper.py
0859f8f24e5f4e69c0e1b3f1565de9c0b3ee4771
[]
no_license
vladhugec/Course-Work
3397d52bb4a649196e336c48d0cab340126b2d8b
27356f32d33aae497ede109fa9e66a19516f79fb
refs/heads/master
2020-12-18T12:59:25.542999
2020-01-21T21:58:34
2020-01-21T21:58:34
235,391,035
3
0
null
null
null
null
UTF-8
Python
false
false
361
py
# helper.py # Written by: Kalina Allen, kallen07 # 02/23/2018 # # Purpose: provide a helpful function for comp11 project 1 from string import ascii_letters # arguments: # mystr: string to clean # returns: s but with all non-alphabet characters removed def clean_word(mystr): return (''.join(char for char in mystr if char in ascii_letters)).lower()
[ "vladhugec@me.com" ]
vladhugec@me.com
cbf504b8cd17c0a67f5038d1dce21a3000e0629c
6366498f7fcaf6c73c34fb0428bbbb899c8a8703
/Math/ProjectEuler005.py
1378b169df4cb2a308f4608ae22d4592a5586cc2
[]
no_license
bibhuty-did-this/MySolutions
2a3da0369cec5ced00c167d80008b2c11e8edbd5
4c6cad6853ac5626ada4302dd4770ec0cbcc99f3
refs/heads/master
2023-06-14T08:38:29.220996
2023-06-01T04:49:05
2023-06-01T04:49:05
102,737,994
1
0
null
null
null
null
UTF-8
Python
false
false
304
py
# Algorithm: # You have to find the lcm of all the numbers combinedly # Generate a lookup table from fractions import gcd lcm=[] lcm.append(1) for i in range(1,41): lcm.append(((lcm[i-1]*i)/gcd(lcm[i-1],i))) # Display the result for _ in range(int(raw_input())): print lcm[int(raw_input())]
[ "emailofpanda@yahoo.com" ]
emailofpanda@yahoo.com
bd0f712c9aea151cb5ebee92c726e449ed9bc413
5cefde592dcc5c40b73734d9488199f084622912
/MNIST/attacks_base.py
63dba65da2e82fb07601154582ce167dbff86b16
[]
no_license
RemiBERNHARD/Membership_Inference_Attacks
d41f2e1f1966c76cc94f0f6f65f858c30ecb7f5a
a4b68a5bd1d0993ca60a2bacb2d5e445b51cdb10
refs/heads/main
2023-03-01T13:14:04.257446
2021-02-12T13:00:39
2021-02-12T13:00:39
338,311,636
2
0
null
null
null
null
UTF-8
Python
false
false
2,851
py
import numpy as np from keras.utils import np_utils from keras.models import Model from keras.layers import Dense, Input from keras.optimizers import Adam ######################### def attack_label(model, X_vec, y_vec): y_pred = np.argmax(model.predict(X_vec), axis=1) memb_pred = np.equal(y_pred, y_vec) return(memb_pred) ######################### def c_ent(vec): vec = np.clip(vec, 1e-12, 1. - 1e-12) ce = -np.sum(vec*np.log(vec+1e-9)) return(ce) def attack_ce(model, X_vec, threshold): conf_pred = model.predict(X_vec) ce_pred = np.apply_along_axis(c_ent, 1, conf_pred) memb_pred = np.less_equal(ce_pred, threshold) return(memb_pred) ######################### def c_ent_loss(vec_pred, vec_true): vec_pred = np.clip(vec_pred, 1e-12, 1. - 1e-12) ce = -np.sum(vec_true*np.log(vec_pred+1e-9)) return(ce) def attack_lossce(model, X_vec, y_vec, threshold): Y_vec = np_utils.to_categorical(y_vec,10) conf_pred = model.predict(X_vec) lossce_pred = np.zeros(conf_pred.shape[0]) for i in range(0, conf_pred.shape[0]): lossce_pred[i] = c_ent_loss(conf_pred[i], Y_vec[i]) memb_pred = np.less_equal(lossce_pred, threshold) return(memb_pred) ######################### def attack_conf(model, X_vec, y_vec, threshold): Y_vec = np_utils.to_categorical(y_vec,10) conf_pred = model.predict(X_vec) conf_label_pred = conf_pred[np.where(Y_vec==1)] memb_pred = np.greater_equal(conf_label_pred, threshold) return(memb_pred) ######################### def mentr(conf_vec, Y_vec): conf_y = conf_vec[np.where(Y_vec==1)][0] conf_noty = conf_vec[np.where(Y_vec!=1)] mentr_val = -(1 - conf_y)*np.log(conf_y) - c_ent(conf_noty) return(mentr_val) def attack_mentr(model, X_vec, y_vec, threshold): Y_vec = np_utils.to_categorical(y_vec,10) conf_pred = model.predict(X_vec) mentr_pred = np.array([mentr(conf_pred[i], Y_vec[i]) for i in range(0, conf_pred.shape[0])]) memb_pred = np.less_equal(mentr_pred, threshold) return(memb_pred) ######################### def attack_shokri(model, X_vec, memb_true): conf_pred = model.predict(X_vec) Y_member = np_utils.to_categorical(memb_true[memb_true==1],2) Y_nonmember = np_utils.to_categorical(memb_true[memb_true==0],2) X_data = conf_pred Y_data = np.concatenate((Y_member, Y_nonmember)) input_shape = (10,) inputs = Input(shape=input_shape) l = Dense(64, activation="tanh")(inputs) outputs = Dense(2, activation="softmax")(l) model_f = Model(inputs, outputs) model_f.compile(loss='categorical_crossentropy', optimizer=Adam(0.001), metrics=['accuracy']) model_f.fit(X_data, Y_data, epochs=50, batch_size=10, verbose=0) memb_pred = np.argmax(model_f.predict(X_data), axis=1) return(memb_pred, model_f)
[ "noreply@github.com" ]
noreply@github.com
7824b29dc495b887680fcb309cd946bcb8051fb5
34229ca69dda0acf53daba3a9696f93ae341d226
/lab/lab01/lab01.py
82ec65a20af4baf650c6642bf906bf0d4aaea5b9
[]
no_license
Unc1eWu/CS61A
60687f0ac3722997af9d3aef4c1190611942197c
18a379a30b2d16629f649e091623892dc2c15df7
refs/heads/master
2020-04-29T16:59:55.469735
2019-09-22T09:08:29
2019-09-22T09:08:29
176,282,456
0
0
null
null
null
null
UTF-8
Python
false
false
362
py
"""Lab 1: Expressions and Control Structures""" # Q3 def sum_digits(n): """Sum all the digits of n. >>> sum_digits(10) # 1 + 0 = 1 1 >>> sum_digits(4224) # 4 + 2 + 2 + 4 = 12 12 >>> sum_digits(1234567890) 45 """ "*** YOUR CODE HERE ***" sum = 0 while n > 0: sum += n % 10 n = n // 10 return sum
[ "Weijie.Wu17@student.xjtlu.edu.cn" ]
Weijie.Wu17@student.xjtlu.edu.cn
46276e90877480976304b67321686d3487c1d5fe
a4fd0e2674264327b50cc913312d959043610ef9
/My_Social_Project/My_Social_Project/urls.py
867f51eecae02f6db439256148b143bc629e1ddb
[]
no_license
ztarikul/InstaShohor-Social-Media-Site
7dedc88ecdffd26501063935ac448482613e3dfd
08d56d1c185435997930a4a7d8b768442acd4be9
refs/heads/master
2023-02-13T12:28:25.359635
2021-01-09T23:32:03
2021-01-09T23:32:03
328,261,703
0
0
null
null
null
null
UTF-8
Python
false
false
517
py
from django.contrib import admin from django.urls import path, include from django.conf import settings from django.contrib.staticfiles.urls import static, staticfiles_urlpatterns from App_Posts import views urlpatterns = [ path('admin/', admin.site.urls), path('accounts/', include('App_Login.urls')), path('post/', include('App_Posts.urls')), path('', views.home, name='home'), ] urlpatterns += staticfiles_urlpatterns() urlpatterns += static(settings.MEDIA_URL, document_root=settings.MEDIA_ROOT)
[ "ztarikulislam@gmail.com" ]
ztarikulislam@gmail.com
2cdcc3f42550aff81547194ab8337407a6b72240
0d6c0df740fad947cd489bcd81c067c2a50e6508
/projects/hog/tests/05.py
29eebe503fa8fa044c6c7cefc608ea2cdc7710da
[ "MIT" ]
permissive
louieyan/CS61A
39f317db01c25219ef813dd8b0af788441ba700f
efa7e7d1becbea776ccfa545197a50c20b85d4fb
refs/heads/master
2022-03-08T03:41:57.052682
2022-02-28T02:03:30
2022-02-28T02:03:30
148,614,155
2
3
null
null
null
null
UTF-8
Python
false
false
6,687
py
test = { 'name': 'Question 5', 'points': 2, 'suites': [ { 'cases': [ { 'answer': 'While score0 and score1 are both less than goal', 'choices': [ 'While score0 and score1 are both less than goal', 'While at least one of score0 or score1 is less than goal', 'While score0 is less than goal', 'While score1 is less than goal' ], 'hidden': False, 'locked': False, 'question': r""" The variables score0 and score1 are the scores for Player 0 and Player 1, respectively. Under what conditions should the game continue? """ }, { 'answer': 'A function that returns the number of dice a player will roll', 'choices': [ 'The number of dice a player will roll', 'A function that returns the number of dice a player will roll', "A player's desired turn outcome" ], 'hidden': False, 'locked': False, 'question': 'What is a strategy in the context of this game?' }, { 'answer': 'strategy1(score1, score0)', 'choices': [ 'strategy1(score1, score0)', 'strategy1(score0, score1)', 'strategy1(score1)', 'strategy1(score0)' ], 'hidden': False, 'locked': False, 'question': r""" If strategy1 is Player 1's strategy function, score0 is Player 0's current score, and score1 is Player 1's current score, then which of the following demonstrates correct usage of strategy1? """ } ], 'scored': False, 'type': 'concept' }, { 'cases': [ { 'code': r""" >>> # >>> # Play function stops at goal >>> s0, s1 = hog.play(always(5), always(3), score0=91, score1=10, dice=always_three) >>> s0 106 >>> s1 10 """, 'hidden': False, 'locked': False }, { 'code': r""" >>> # >>> # Goal score is not hardwired >>> s0, s1 = hog.play(always(5), always(5), goal=10, dice=always_three) >>> s0 15 >>> s1 0 """, 'hidden': False, 'locked': False } ], 'scored': True, 'setup': r""" >>> import hog >>> always_three = hog.make_test_dice(3) >>> always = hog.always_roll """, 'teardown': '', 'type': 'doctest' }, { 'cases': [ { 'code': r""" >>> # >>> # Use strategies >>> # We recommend working this out turn-by-turn on a piece of paper. >>> strat0 = lambda score, opponent: opponent % 10 >>> strat1 = lambda score, opponent: score // 10 >>> s0, s1 = hog.play(strat0, strat1, score0=41, score1=80, dice=always_three) >>> s0 51 >>> s1 104 """, 'hidden': False, 'locked': False } ], 'scored': True, 'setup': r""" >>> import hog >>> always_three = hog.make_test_dice(3) """, 'teardown': '', 'type': 'doctest' }, { 'cases': [ { 'code': r""" >>> # >>> # Goal edge case >>> s0, s1 = hog.play(always(4), always(3), score0=88, score1=20, dice=always_three) >>> s0 20 >>> s1 100 """, 'hidden': False, 'locked': False }, { 'code': r""" >>> # >>> # Player 1 win >>> s0, s1 = hog.play(always(4), always(4), score0=87, score1=88, dice=always_three) >>> s0 99 >>> s1 100 """, 'hidden': False, 'locked': False }, { 'code': r""" >>> # Check strategies are actually used correctly >>> strat0 = lambda score, opponent: opponent % 10 >>> strat1 = lambda score, opponent: opponent // 10 >>> s0, s1 = hog.play(strat0, strat1, score0=40, score1=92, dice=always_three) >>> s0 101 >>> s1 73 """, 'hidden': False, 'locked': False }, { 'code': r""" >>> # >>> # Swine swap applies during Player 1 turn >>> s0, s1 = hog.play(always(4), always(4), score0=42, score1=96, dice=always_three) >>> s0 108 >>> s1 54 """, 'hidden': False, 'locked': False }, { 'code': r""" >>> # >>> # Free bacon refers to correct opponent score >>> s0, s1 = hog.play(always(0), always(0), score0=11, score1=99, dice=always_three) >>> s0 13 >>> s1 103 """, 'hidden': False, 'locked': False }, { 'code': r""" >>> # >>> # Handle multiple turns with many swaps >>> s0, s1 = hog.play(always(0), always(2), goal=15, dice=always_three) >>> s0 16 >>> s1 2 """, 'hidden': False, 'locked': False } ], 'scored': True, 'setup': r""" >>> import hog >>> always_three = hog.make_test_dice(3) >>> always = hog.always_roll """, 'teardown': '', 'type': 'doctest' }, { 'cases': [ { 'code': r""" >>> tests.play_utils.check_play_function(hog) """, 'hidden': False, 'locked': False } ], 'scored': True, 'setup': r""" >>> # Fuzz Testing >>> # Plays a lot of random games, and calculates a secret value. >>> # Failing this test means something is wrong, but you should >>> # look at other tests to see where the problem might be. >>> # Hint: make sure you're only calling take_turn once per turn! >>> # >>> import hog, importlib >>> importlib.reload(hog) >>> import tests.play_utils """, 'teardown': r""" """, 'type': 'doctest' } ] }
[ "noreply@github.com" ]
noreply@github.com
d0e79b2550921ec82c01575be4ab9c242dc8b522
0313d27c61d83f4cc783611d23cd11dd9af20d12
/flyback.py
4ec898bebf9c00c54e9346439ae546d11eceda39
[ "MIT" ]
permissive
di2mot/Flyback_registrator
fe45e3fa5701e8550f21a274276a157e2d38c709
1cd887b16bf31b60f0cbb8166fc813dd96c812da
refs/heads/main
2023-03-01T20:46:15.724438
2021-02-08T19:20:32
2021-02-08T19:20:32
337,182,177
0
0
null
null
null
null
UTF-8
Python
false
false
5,310
py
from selenium import webdriver from selenium.webdriver.support.ui import Select from selenium.webdriver.chrome.options import Options from selenium.common.exceptions import NoSuchElementException import time ''' СУПЕРТ ТУПАЯ ПРОГА для регистрации на форуме flyback.org.ru Да, решить просто, но на тот момент я не знал методов расчёта, но умел в Python. Ниже решение за 1 минуту. Для условия типа: Даны резисторыR1=40ом R2=310ом R3=300ом. все элементы схемы идеальны, падение напряжение на диоде равно нулю Входное напряжение: синусоидальное 50гц, с действующим значением 54В. Найти до какого значения зарядится конденсатор. Решение: 54В * (R2 / (R1 + R2)) = 47.83 ДА, всё максимально просто. Хотите сложнее? Запустите скрипт))) ''' text = { 'username': 'you_login', 'email': 'you_email', 'bday_day': '1', 'bday_year': '1970', 'new_password': 'strongPassword', 'password_confirm': 'strongPassword', } # если нужно найти вольтаж, то начните с близкого к этому значению, если энергию, то с 0.01 volt = 40.00 start = time.time() link = 'http://flyback.org.ru/user.php?mode=register&agreed=true' # Мне было лень думать над первым селектором selector_0 : str = 'body > table:nth-child(6) > tbody > tr > td > form > table.forumline > tbody > tr:nth-child(7) > td:nth-child(2)' selector_1 : str = 'username' selector_2 : str = 'email' selector_3 : str = 'bday_day' selector_4 : str = 'bday_month' selector_5 : str = '6' selector_6 : str = 'bday_year' selector_7 : str = 'new_password' selector_8 : str = 'password_confirm' selector_9 : str = 'answer' selector_10 : str = 'apply' selector_11: str = 'submit' # Ну так калечно run = True try: # Это с отключённым графическим режимом chrome_options = Options() chrome_options.add_argument("--headless") #стандартная инициализация browser = webdriver.Chrome(options=chrome_options) # Это для графичес кого режима # browser = webdriver.Chrome() browser.get(link) # тут получаю текст задания что бы было task = browser.find_element_by_css_selector(selector_0) print(task.text) # вставляю текст в форму username = browser.find_element_by_name(selector_1) username.send_keys(text['username']) # вставляю текст в форму email = browser.find_element_by_name(selector_2) email.send_keys(text['email']) # вставляю текст в форму bday_day = browser.find_element_by_name(selector_3) bday_day.clear() bday_day.send_keys(text['bday_day']) # выбираем месяц bday_day = Select(browser.find_element_by_name(selector_4)) bday_day.select_by_value(selector_5) # вставляю текст в форму bday_year = browser.find_element_by_name(selector_6) bday_year.clear() bday_year.send_keys(text['bday_year']) # выбираю параметр apply = browser.find_element_by_name(selector_10) apply.click() while run: # вставляю текст в форму new_password = browser.find_element_by_name(selector_7) new_password.send_keys(text['new_password']) # вставляю текст в форму password_confirm = browser.find_element_by_name(selector_8) password_confirm.send_keys(text['password_confirm']) # вставляю текст в форму answer = browser.find_element_by_name(selector_9) answer.clear() # что бы адекватно округляло imp_answer = "{0:0.2f}".format(volt) answer.send_keys(imp_answer) # клик по кнопке button = browser.find_element_by_name(selector_11) button.click() try: findText = browser.find_element_by_tag_name('h1') if findText.text == 'Неверный ответ': volt += 0.01 browser.back() else: print(volt, imp_answer) run = False except NoSuchElementException: print(volt, imp_answer) except NoSuchElementException: print('Finish!') finally: # А тут чисто что бы глянуть результат # получаем исходный код html = browser.page_source # Пишим в файл. Не надо? Закомитить with open('page.html', 'w') as file: file.write(html) # закрываем браузер после всех манипуляций browser.quit() fin = time.time() rez_time = fin - start # Просто оценить потраченное время print(f'Finish!\n It took:\n second: {rez_time}\n minutes: {rez_time/60}\n hour: {rez_time/3600}')
[ "motsar.dm@gmail.com" ]
motsar.dm@gmail.com
9e2fd5f4161aa84c62285366896ec09d1372f331
522e5c63bd53dccf824c17e882f564e8423331ea
/gbce_trading_unittests.py
497ea315ef6066e57c1da334e7e5a3c796870a50
[]
no_license
pankaj-purohit/Assignment
df20a65bb2460cbb28e19b6156003a2611a36f82
9ce95b020673ef621e3696b16997c7ff2f989daa
refs/heads/master
2020-05-31T16:23:37.125624
2019-06-05T11:17:14
2019-06-05T11:17:14
190,380,093
0
0
null
null
null
null
UTF-8
Python
false
false
3,925
py
import datetime import unittest import gbce_trading_config from gbce_trading import GBCETrading, Stock, Trade class GBCEUnitTests(unittest.TestCase): def test_getDividendYield(self): stockComm = {'POP': {'Type': 'Common', 'Last Dividend': 8, 'Fixed Dividend': None, 'Par Value': 100, 'Trades': []}} testObj = GBCETrading(stockComm) self.assertEqual(testObj.getDividendYield(symbol='POP', price=160.0), 0.05) self.assertEqual(testObj.getDividendYield(symbol='POP', price=0), None) self.assertRaises(ValueError, testObj.getDividendYield, symbol='POP', price=-10.7) self.assertRaises(ValueError, testObj.getDividendYield, symbol = 'ABC', price=24.7) stockPref = {'JOE': {'Type': 'Preferred', 'Last Dividend': 18, 'Fixed Dividend': 4, 'Par Value': 200, 'Trades': []}} testObj2 = GBCETrading(stockPref) self.assertAlmostEqual(testObj2.getDividendYield(symbol='JOE', price=254.5), 3.143, 2) def test_getPERatio(self): stockComm = {'ALE': {'Type': 'Common', 'Last Dividend': 23, 'Fixed Dividend': None, 'Par Value': 60, 'Trades': []}} testObj = GBCETrading(stockComm) self.assertAlmostEqual(testObj.getPERatio(symbol='ALE', price=88.0), 336.695, 2) self.assertEqual(testObj.getPERatio(symbol='ALE', price=0), None) self.assertRaises(ValueError, testObj.getPERatio, symbol='POP', price=-75.7) self.assertRaises(ValueError, testObj.getPERatio, symbol='XYZ', price=190.2) stockPref = {'GIN': {'Type': 'Preferred', 'Last Dividend': 24, 'Fixed Dividend': 14, 'Par Value': 250, 'Trades': []}} testObj2 = GBCETrading(stockPref) self.assertEqual(testObj2.getPERatio(symbol='GIN', price=350), 35) def test_getVolumeWeightedStockPrice(self): trade1 = Trade(symbol='TEA', quantity=100, tradeType=gbce_trading_config.TradeType.BUY, price=120.5) trade2 = Trade(symbol='TEA', quantity=50, tradeType=gbce_trading_config.TradeType.SELL, price=125.4) trade3 = Trade(symbol='TEA', quantity=200, tradeType=gbce_trading_config.TradeType.BUY, price=110.5, timestamp=datetime.datetime.now() - datetime.timedelta(minutes=7)) stockGin = Stock(symbol='GIN') stockGin.addStockDetails(type=gbce_trading_config.StockType.COMMON, lastDividend=20, parValue=200) stockTea = Stock(symbol='TEA') stockTea.addStockDetails(type=gbce_trading_config.StockType.COMMON, lastDividend=0, parValue=100) stockTea.recordTrade(trade1) stockTea.recordTrade(trade2) stockTea.recordTrade(trade3) testObj = GBCETrading() self.assertAlmostEqual(testObj.getVolumeWeightedStockPrice(symbol='TEA'), 122.133, 2) self.assertRaises(ValueError, testObj.getVolumeWeightedStockPrice, symbol='PQR') self.assertEqual(testObj.getVolumeWeightedStockPrice(symbol='GIN'), None) def test_getGBCEAllShareIndex(self): trade1 = Trade(symbol='TEA', quantity=100, tradeType=gbce_trading_config.TradeType.BUY, price=120.5) trade2 = Trade(symbol='TEA', quantity=50, tradeType=gbce_trading_config.TradeType.SELL, price=125.4) stockGin = Stock(symbol='GIN') stockGin.addStockDetails(type=gbce_trading_config.StockType.COMMON, lastDividend=20, parValue=200) stockTea = Stock(symbol='TEA') stockTea.addStockDetails(type=gbce_trading_config.StockType.COMMON, lastDividend=0, parValue=100) stockTea.recordTrade(trade1) stockTea.recordTrade(trade2) testObj = GBCETrading() self.assertEqual(testObj.getGBCEAllShareIndex(), None) trade3 = Trade(symbol='GIN', quantity=150, tradeType=gbce_trading_config.TradeType.BUY, price=225.4) stockGin.recordTrade(trade3) self.assertAlmostEqual(testObj.getGBCEAllShareIndex(), 165.918, 2) if __name__ == '__main__': unittest.main()
[ "noreply@github.com" ]
noreply@github.com
61a68ab16ef1c92be2bcf040f1cde0f7934ad563
db26b08ad35594a8a2aba36a68bfe37f85728727
/DjangoAngularAssignment/conversation/views.py
4f355cbd6325591ee242efd26e5640a770b7f29c
[]
no_license
rafikkamal/django
83f3217f3e31fd13a5e1c4e81ed583ccfd4971fe
7d593707f0f33b074e22f75f9ed0a6e0ef0b39d1
refs/heads/master
2021-01-20T06:25:30.872254
2017-04-30T21:02:12
2017-04-30T21:02:12
89,877,099
1
0
null
null
null
null
UTF-8
Python
false
false
6,687
py
from django.db.models import Q from django.shortcuts import render,redirect from django.http import HttpResponse from django.views import generic from django.views.generic.edit import CreateView, UpdateView, DeleteView from django.contrib.auth import authenticate, login from .models import Conversation from django.views.generic import View from .forms import UserForm from django.contrib.auth.models import User from .models import Conversation,Conversations #from django.core.urlsolvers import reverse_lazy # Create your views here. def index(request): if request.user.is_authenticated(): users = User.objects.all().exclude(pk=request.user.id) #return HttpResponse(users) return render(request, 'conversation/index.html', {'users': users}) else: return render(request, 'conversation/login.html', {'message': 'HEllo !'}) def login_user(request): if request.method == "POST": username = request.POST['username'] password = request.POST['password'] user = authenticate(username=username, password=password) if user is not None: if user.is_active: login(request, user) return render(request, 'conversation/index.html', {'message': 'Logged in Successfully !'}) else: return render(request, 'conversation/login.html', {'error_message': 'Your account has been disabled'}) else: return render(request, 'conversation/login.html', {'error_message': 'Invalid login'}) return render(request, 'conversation/login.html') class DetailView(generic.DetailView): model = Conversation template_name = 'conversation/detail.html' class UserFormView(View): form_class = UserForm template_name = 'conversation/registration_form.html' def get(self, request): form = self.form_class(None) return render(request, self.template_name, {'form': form}) def post(self, request): form = self.form_class(request.POST) if form.is_valid(): user = form.save(commit=False) username = form.cleaned_data['username'] password = form.cleaned_data['password'] user.set_password(password) user.save() user = authenticate(username=username, password=password) if user is not None: if user.is_active: login(request, user) return redirect('conversation:index') return render(request, self.template_name, {'form': form}) def VarifyUser(pk): try: user = User.objects.get(pk=pk) except User.DoesNotExist: users = User.objects.all().exclude(pk=request.user.id) return render(request, 'conversation/index.html', {'users': users, 'message': 'No Such User Found'}) return user class ConversationListView(View): def get(self, request, pk, *args, **kwargs): user = VarifyUser(pk) conversations = Conversations.objects.get(Q(created_by=request.user.id, created_for=pk) | Q(created_by=pk, created_for=request.user.id)) #return HttpResponse(conversations) if conversations: conversation = Conversation.objects.filter(conversation=conversations) return render(request, 'conversation/conversation_template.html', {'user': user, 'conversation': conversation, 'id': conversations.id}) return render(request, 'conversation/connect.html', {'user': user, 'conversations': conversations}) class ConversationProcessor(View): def post(self, request, pk, *args, **kwargs): user = VarifyUser(pk) if request.method == "POST": message = request.POST['message'] conversations = Conversations.objects.get(Q(created_by=request.user.id, created_for=pk) | Q(created_by=pk, created_for=request.user.id)) if str(conversations.id) == request.POST['conversation']: conversation = Conversation() conversation.message = message conversation.user = request.user conversation.conversation = conversations conversation.save() return redirect('conversation:conversation_connect', pk=int(pk)) else: users = User.objects.all().exclude(pk=request.user.id) return render(request, 'conversation/index.html', {'users': users, 'message': 'You Dont have access!!!'}) class ConversationProcessorUpdate(View): def post(self, request, pk, *args, **kwargs): user = VarifyUser(pk) if request.method == "POST": message = request.POST['message'] conversations = Conversations.objects.get(Q(created_by=request.user.id, created_for=pk) | Q(created_by=pk, created_for=request.user.id)) if str(conversations.id) == request.POST['conversation']: conversation = Conversation.objects.get(pk=request.POST['conversation_id']) conversation.message = message conversation.save() return redirect('conversation:conversation_connect', pk=int(pk)) class ConversationProcessorDelete(View): def post(self, request, pk, *args, **kwargs): user = VarifyUser(pk) if request.method == "POST": conversations = Conversations.objects.get(Q(created_by=request.user.id, created_for=pk) | Q(created_by=pk, created_for=request.user.id)) if str(conversations.id) == request.POST['conversation']: conversation = Conversation.objects.get(pk=request.POST['conversation_id']) conversation.delete() return redirect('conversation:conversation_connect', pk=int(pk)) class ConversationProcessorArchive(View): def post(self, request, pk, *args, **kwargs): user = VarifyUser(pk) if request.method == "POST": conversations = Conversations.objects.get(Q(created_by=request.user.id, created_for=pk) | Q(created_by=pk, created_for=request.user.id)) if str(conversations.id) == request.POST['conversation']: conversation = Conversation.objects.get(pk=request.POST['conversation_id']) conversation.archive = True conversation.save() return redirect('conversation:conversation_connect', pk=int(pk))
[ "rafik.rkn@gmail.com" ]
rafik.rkn@gmail.com
6e7f711485de28b5120b7854383a32ff11f98c7d
cb48956cb48a94da0a25e7742b87b623224664ed
/network/relighting.py
bef13c75f6a6c179329e32a4459a1d8c993cdd27
[ "Apache-2.0" ]
permissive
Siddharth-Shrivastava7/DANNet
e5f2fb03bd8cb4e19392863566a9b97efa59f955
8db10056a4e445d24fc899505923615457cae5b7
refs/heads/main
2023-06-20T20:33:32.602314
2021-07-30T17:52:00
2021-07-30T17:52:00
362,430,200
0
0
Apache-2.0
2021-04-28T10:38:22
2021-04-28T10:38:21
null
UTF-8
Python
false
false
7,379
py
import torch import torch.nn as nn import functools affine_par = True class ResnetBlock(nn.Module): """Define a Resnet block""" def __init__(self, dim, padding_type, norm_layer, use_dropout, use_bias): """Initialize the Resnet block A resnet block is a conv block with skip connections We construct a conv block with build_conv_block function, and implement skip connections in <forward> function. Original Resnet paper: https://arxiv.org/pdf/1512.03385.pdf """ super(ResnetBlock, self).__init__() self.conv_block = self.build_conv_block(dim, padding_type, norm_layer, use_dropout, use_bias) def build_conv_block(self, dim, padding_type, norm_layer, use_dropout, use_bias): """Construct a convolutional block. Parameters: dim (int) -- the number of channels in the conv layer. padding_type (str) -- the name of padding layer: reflect | replicate | zero norm_layer -- normalization layer use_dropout (bool) -- if use dropout layers. use_bias (bool) -- if the conv layer uses bias or not Returns a conv block (with a conv layer, a normalization layer, and a non-linearity layer (ReLU)) """ conv_block = [] p = 0 if padding_type == 'reflect': conv_block += [nn.ReflectionPad2d(1)] elif padding_type == 'replicate': conv_block += [nn.ReplicationPad2d(1)] elif padding_type == 'zero': p = 1 else: raise NotImplementedError('padding [%s] is not implemented' % padding_type) conv_block += [nn.Conv2d(dim, dim, kernel_size=3, padding=p, bias=use_bias), norm_layer(dim), nn.ReLU(True)] if use_dropout: conv_block += [nn.Dropout(0.5)] p = 0 if padding_type == 'reflect': conv_block += [nn.ReflectionPad2d(1)] elif padding_type == 'replicate': conv_block += [nn.ReplicationPad2d(1)] elif padding_type == 'zero': p = 1 else: raise NotImplementedError('padding [%s] is not implemented' % padding_type) conv_block += [nn.Conv2d(dim, dim, kernel_size=3, padding=p, bias=use_bias), norm_layer(dim)] return nn.Sequential(*conv_block) def forward(self, x): """Forward function (with skip connections)""" out = x + self.conv_block(x) # add skip connections return out class ResnetGenerator(nn.Module): """Resnet-based generator that consists of Resnet blocks between a few downsampling/upsampling operations. We adapt Torch code and idea from Justin Johnson's neural style transfer project(https://github.com/jcjohnson/fast-neural-style) """ def __init__(self, input_nc, output_nc, ngf=64, norm_layer=nn.BatchNorm2d, use_dropout=False, n_blocks=6, padding_type='reflect'): """Construct a Resnet-based generator Parameters: input_nc (int) -- the number of channels in input images output_nc (int) -- the number of channels in output images ngf (int) -- the number of filters in the last conv layer norm_layer -- normalization layer use_dropout (bool) -- if use dropout layers n_blocks (int) -- the number of ResNet blocks padding_type (str) -- the name of padding layer in conv layers: reflect | replicate | zero """ assert(n_blocks >= 0) super(ResnetGenerator, self).__init__() if type(norm_layer) == functools.partial: use_bias = norm_layer.func == nn.InstanceNorm2d else: use_bias = norm_layer == nn.InstanceNorm2d model = [nn.ReflectionPad2d(3), nn.Conv2d(input_nc, ngf, kernel_size=7, padding=0, bias=use_bias), norm_layer(ngf), nn.ReLU(True)] n_downsampling = 2 for i in range(n_downsampling): # add downsampling layers mult = 2 ** i model += [nn.Conv2d(ngf * mult, ngf * mult * 2, kernel_size=3, stride=2, padding=1, bias=use_bias), norm_layer(ngf * mult * 2), nn.ReLU(True)] mult = 2 ** n_downsampling for i in range(n_blocks): # add ResNet blocks model += [ResnetBlock(ngf * mult, padding_type=padding_type, norm_layer=norm_layer, use_dropout=use_dropout, use_bias=use_bias)] for i in range(n_downsampling): # add upsampling layers mult = 2 ** (n_downsampling - i) model += [nn.ConvTranspose2d(ngf * mult, int(ngf * mult / 2), kernel_size=3, stride=2, padding=1, output_padding=1, bias=use_bias), norm_layer(int(ngf * mult / 2)), nn.ReLU(True)] model += [nn.ReflectionPad2d(3)] model += [nn.Conv2d(ngf, output_nc, kernel_size=7, padding=0)] model += [nn.Tanh()] self.model = nn.Sequential(*model) def forward(self, input): """Standard forward""" return self.model(input) def LightNet(): model = ResnetGenerator(3, 3, 64, norm_layer=nn.BatchNorm2d, use_dropout=False, n_blocks=3) return model class L_exp_z(nn.Module): def __init__(self, patch_size): super(L_exp_z, self).__init__() self.pool = nn.AvgPool2d(patch_size) def forward(self, x, mean_val): x = torch.mean(x, 1, keepdim=True) mean = self.pool(x) d = torch.mean(torch.pow(mean - torch.FloatTensor([mean_val]).cuda(), 2)) return d class L_TV(nn.Module): def __init__(self, TVLoss_weight=1): super(L_TV, self).__init__() self.TVLoss_weight = TVLoss_weight def forward(self, x): batch_size = x.size()[0] h_x = x.size()[2] w_x = x.size()[3] count_h = (x.size()[2] - 1) * x.size()[3] count_w = x.size()[2] * (x.size()[3] - 1) h_tv = torch.pow((x[:, :, 1:, :] - x[:, :, :h_x - 1, :]), 2).sum() w_tv = torch.pow((x[:, :, :, 1:] - x[:, :, :, :w_x - 1]), 2).sum() return self.TVLoss_weight * 2 * (h_tv / count_h + w_tv / count_w) / batch_size class SSIM(nn.Module): """Layer to compute the SSIM loss between a pair of images """ def __init__(self): super(SSIM, self).__init__() self.mu_x_pool = nn.AvgPool2d(3, 1) self.mu_y_pool = nn.AvgPool2d(3, 1) self.sig_x_pool = nn.AvgPool2d(3, 1) self.sig_y_pool = nn.AvgPool2d(3, 1) self.sig_xy_pool = nn.AvgPool2d(3, 1) self.refl = nn.ReflectionPad2d(1) self.C1 = 0.01 ** 2 self.C2 = 0.03 ** 2 def forward(self, x, y): x = self.refl(x) y = self.refl(y) mu_x = self.mu_x_pool(x) mu_y = self.mu_y_pool(y) sigma_x = self.sig_x_pool(x ** 2) - mu_x ** 2 sigma_y = self.sig_y_pool(y ** 2) - mu_y ** 2 sigma_xy = self.sig_xy_pool(x * y) - mu_x * mu_y SSIM_n = (2 * mu_x * mu_y + self.C1) * (2 * sigma_xy + self.C2) SSIM_d = (mu_x ** 2 + mu_y ** 2 + self.C1) * (sigma_x + sigma_y + self.C2) return torch.clamp((1 - SSIM_n / SSIM_d) / 2, 0, 1)
[ "noreply@github.com" ]
noreply@github.com
75f74b2bca9e9f3323f53b9941cda59e24b9064e
7f7898939786f7dae5693d66d35149b177b06a34
/mgw7510/forms.py
4390a9bff113b0c16f4699d6b0a82354e60ef680
[ "Apache-2.0" ]
permissive
gter1216/mgw7510-web
f8a6ef2cd3793fdafadf28ba6ec4d4493db69a7c
758e9053b1071655b5e23f714a82f8698cb562e6
refs/heads/master
2021-01-12T07:45:40.679559
2017-02-03T13:45:04
2017-02-03T13:45:04
77,008,390
2
5
null
2017-03-01T04:55:55
2016-12-21T02:25:14
HTML
UTF-8
Python
false
false
2,143
py
from django.forms import ModelForm from mgw7510.models import WebUser from django import forms # inherit from class ModelForm class WebUserForm(ModelForm): # username = forms.EmailField(error_messages="user name should not be empty") # password = forms.CharField(error_messages="password should not be empty") # password is an optional field password = forms.CharField(required=False) confirmPassword = forms.CharField(required=False) newPassword = forms.CharField(required=False) confirmNewPassword = forms.CharField(required=False) userWorkDir = forms.CharField(required=False) pakServerIp = forms.GenericIPAddressField(required=False) pakServerUsername = forms.CharField(required=False) pakServerPasswd = forms.CharField(required=False) pakServerFp = forms.CharField(required=False) seedVMIp = forms.GenericIPAddressField(required=False) seedVMUsername = forms.CharField(required=False) seedVMPasswd = forms.CharField(required=False) seedVMOpenrcAbsPath = forms.CharField(required=False) seedVMKeypairAbsPath = forms.CharField(required=False) yactServerIp = forms.GenericIPAddressField(required=False) yactServerUsername = forms.CharField(required=False) yactServerPasswd = forms.CharField(required=False) yactServerDIFAbsPath = forms.CharField(required=False) yactServerYactAbsPath = forms.CharField(required=False) userInputFile = forms.FileField(label='Select a file', required=False) tmpPath = forms.CharField(required=False) userInputFileName = forms.CharField(required=False) progressBarData = forms.CharField(required=False) userInputUploadedFlag = forms.CharField(required=False) ceDeployState = forms.CharField(required=False) ceSelectRel = forms.CharField(required=False) ceSelectPak = forms.CharField(required=False) ceDeployProcess = forms.CharField(required=False) swImageName = forms.CharField(required=False) class Meta: model = WebUser # inherit form WebUser fields = '__all__'
[ "xuxiao1216@126.com" ]
xuxiao1216@126.com
c8611799a1e24a658b75c368489cc107c804783e
d8ac9944ff402ec9f0bea82213a8418982c16f39
/python/Tushare_API_Scripts/DataFrame写入PG数据库.py
97b330f3c46b13c7ee8b6c6d9117eb63e282fef7
[]
no_license
zzgbird/PyAndShell_Learning
cdc50d23ac0d05dcb9293171bb3a9946485766d3
9f0f7b3a39d2ad8a6a5211dd612d805dac5de64c
refs/heads/master
2023-03-27T11:30:35.245342
2021-03-21T07:02:59
2021-03-21T07:02:59
300,139,221
0
0
null
null
null
null
UTF-8
Python
false
false
2,244
py
# DataFrame写入PG数据库 # 提前创建df数据集并安装psycopg2模块 pip install psycopg2 # 1. 使用pandas的to_sql导入 import pandas as pd from sqlalchemy import create_engine engine = create_engine('postgresql://postgres:ican00@127.0.0.1:5432/stockam') start = datetime.datetime.now() pd.io.sql.to_sql(df, 'stock_basic', engine,index= False, schema='ods',if_exists='append') # fail 如果表存在,啥也不做 # replace 如果表存在,删了表,再建立一个新表,把数据插入 # append 如果表存在,把数据插入,如果表不存在创建一个表!! end = datetime.datetime.now() print('time cost:',(end - start)) # 2.使用copy_from导入 这种方式直接用了psycopg2模块中的copy_from方法,写入速度最快。 # dataframe类型转换为IO缓冲区中的str类型 # 需要先建表并指定插入列名 import psycopg2 from io import StringIO df = pro.query('namechange',) output = StringIO() df.to_csv(output, sep='\t', index=False, header=False) output1 = output.getvalue() #conn = pgconnection() conn = psycopg2.connect(database="stockam", user="postgres", password="ican00", host="127.0.0.1", port="5432") cur = conn.cursor() # columns=['label1', 'label2'] cur.copy_from(StringIO(output1),'ods.namechange',) conn.commit() cur.close() conn.close() # 3. executemany()方法批量输入数据到数据库 import psycopg2 conn = psycopg2.connect(host=***, port=***, database=***, user=***, password=***) cur = conn.cursor() sql = "insert into " + table_name + " values(%s, %s, %s)" cur.executemany(sql, data) conn.commit() conn.close() # 强大的copy_from(),是postgresSQ的内置函数 import psycopg2 from sqlalchemy import create_engine import pandas as pd import StringIO from io import BytesIO data1 = pd.DataFrame(data) # dataframe类型转换为IO缓冲区中的str类型 output = BytesIO() data1.to_csv(output, sep='\t', index=False, header=False) output1 = output.getvalue() conn = psycopg2.connect(host=***, user=***, password=***, database=***) cur = conn.cursor() cur.copy_from(StringIO.StringIO(output1), table_name1, null='',columns=columns) conn.commit() cur.close() conn.close() print('done') 这儿需要特别记录下,copy_from
[ "zzg_30th@outlook.com" ]
zzg_30th@outlook.com
061a39f3108aeed656ad2dd3e0779406a17ee0f6
414daef846cbb984cb97c1183e7ea557ce466575
/alpha_sigma-master/record_node_backup/new_MCTS.py
45046ddcd18f5acf49821148d967979c791aa016
[]
no_license
yyh1999070901/GoAI
c0e0cf0d471bb65a56f83796707f9c58219554c5
d0551c8faa4c9a37a7a50d140a45727354a54fd2
refs/heads/master
2023-02-03T00:30:38.628257
2020-12-22T11:17:30
2020-12-22T11:17:30
323,603,887
0
0
null
null
null
null
UTF-8
Python
false
false
6,440
py
import numpy as np import random import sys import utils from five_stone_game import main_process as five_stone_game import time num2char = {0: "a", 1: "b", 2: "c", 3: "d", 4: "e", 5: "f", 6: "g", 7: "h", 8: "i", 9: "j", 10: "k", 11: "l", 12: "m", 13: "n", 14: "o", 15: "p", 16: "q", 17: "r", 18: "s", 19: "t", 20: "u"} class edge: def __init__(self, action, parent_node, priorP, MCTS_pointer): self.action = action self.counter = 1.0 self.parent_node = parent_node self.priorP = priorP self.MCTS_ptr = MCTS_pointer self.child_node = None # self.search_and_get_child_node() self.action_value = 0.0 def backup(self, v): self.action_value += v self.counter += 1 self.parent_node.backup(-v) def get_child(self): if self.child_node is None: self.counter += 1 self.child_node = self.search_and_get_child_node() return self.child_node, True else: self.counter += 1 return self.child_node, False def search_and_get_child_node(self): new_state_name = utils.generate_new_state(self.parent_node.get_state(), self.action, self.parent_node.node_player) search_result = self.MCTS_ptr.search_node(new_state_name) if search_result: return search_result else: new_node = node(new_state_name, self, -self.parent_node.node_player, self.MCTS_ptr) self.MCTS_ptr.add_node(new_node) return new_node def UCB_value(self): if self.action_value: Q = self.action_value / self.counter else: Q = 0 return Q + utils.Cpuct * self.priorP * np.sqrt(self.parent_node.counter) / (1 + self.counter) class node: def __init__(self, state, parent, player, MCTS_pointer): self.state_name = state self.parent = parent self.value = 0.0 self.counter = 0.0 self.child = {} self.MCTS_pointer = MCTS_pointer self.node_player = player def get_state(self): return self.state_name def add_child(self, action, priorP): action_name = utils.move_to_str(action) self.child[action_name] = edge(action=action, parent_node=self, priorP=priorP, MCTS_pointer=self.MCTS_pointer) def get_child(self, action): return self.child[action].child_node def eval_or_not(self): return len(self.child)==0 def backup(self, v): self.value += v self.counter += 1 if self.parent: self.parent.backup(v) def get_distribution(self): ## used to get the step distribution of current for key in self.child.keys(): self.MCTS_pointer.distribution_calculater.push(key, self.child[key].counter) return self.MCTS_pointer.distribution_calculater.get() def UCB_sim(self): UCB_max = -sys.maxsize UCB_max_key = None for key in self.child.keys(): if self.child[key].UCB_value() > UCB_max: UCB_max_key = key UCB_max = self.child[key].UCB_value() this_node, expand = self.child[UCB_max_key].get_child() return this_node, expand, self.child[UCB_max_key].action class MCTS: def __init__(self, board_size=11, simulation_per_step=400, neural_network=None, init_state="", init_node=None): self.board_size = board_size self.s_per_step = simulation_per_step self.database = {0: {"":node(init_state, init_node, 1, self)}} # here we haven't complete a whole database that can be self.current_node = self.database[0][""] # used to search the exist node self.NN = neural_network self.game_process = five_stone_game(board_size=board_size) self.simulate_game = five_stone_game(board_size=board_size) self.distribution_calculater = utils.distribution_calculater(self.board_size) def renew(self): self.database = {0: {"": node("", None, 1, self)}} self.current_node = self.database[0][""] self.game_process.renew() def search_node(self, node_name): if len(node_name) in self.database.keys(): if node_name in self.database[len(node_name)].keys(): return self.database[len(node_name)][node_name] return None def add_node(self, node): if len(node.state_name) in self.database.keys(): self.database[len(node.state_name)][node.state_name] = node else: self.database[len(node.state_name)] = {node.state_name:node} def MCTS_step(self, action): next_node = self.current_node.get_child(action) return next_node def simulation(self): for _ in range(self.s_per_step): expand, game_continue = False, True this_node = self.current_node self.simulate_game.simulate_reset(self.game_process.current_board_state(True)) state = self.simulate_game.current_board_state() while game_continue and not expand: if this_node.eval_or_not(): state_prob, _ = self.NN.eval(state) valid_move = utils.valid_move(state) for move in valid_move: this_node.add_child(action=move, priorP=state_prob[0, move[0]*self.board_size + move[1]]) this_node, expand, action = this_node.UCB_sim() game_continue, state = self.simulate_game.step(action) if not game_continue: this_node.backup(1) elif expand: _, state_v = self.NN.eval(state) this_node.backup(state_v) def game(self): game_continue = True game_record = [] begin_time = int(time.time()) while game_continue: self.simulation() action, distribution = self.current_node.get_distribution() game_continue, state = self.game_process.step(utils.str_to_move(action)) self.current_node = self.MCTS_step(action) game_record.append({"distribution": distribution, "action":action}) self.renew() end_time = int(time.time()) min = int((end_time - begin_time)/60) second = (end_time - begin_time) % 60 print("In last game, we cost {}:{}".format(min, second)) return game_record
[ "18710113218@163.com" ]
18710113218@163.com
4438a410880898850073b4bc83f77e73ca792121
eadd15064aa74811e7a3718b617636627ef4fd47
/web/migrations/0020_rename_index_indexpage.py
8efb4605a936d00a29a46a2c95ef6c4263e63c65
[]
no_license
topsai/plasrefine_backstage
262f7bb032daa4d018aac1519e1139cb060c3f91
1eb34dd0b13ebdc2a42dd6ed1aaa2d08c18ab5fb
refs/heads/master
2023-04-12T13:24:22.710108
2021-05-08T14:16:41
2021-05-08T14:16:41
361,993,024
0
0
null
null
null
null
UTF-8
Python
false
false
326
py
# Generated by Django 3.2 on 2021-05-02 08:03 from django.db import migrations class Migration(migrations.Migration): dependencies = [ ('web', '0019_auto_20210502_1558'), ] operations = [ migrations.RenameModel( old_name='Index', new_name='IndexPage', ), ]
[ "hurte@foxmail.com" ]
hurte@foxmail.com
c67e75c1252e6c67b7433b5b12e21fc96a3a4abf
59a4a1409fabdd29c6db85775a549c00b5a7b3b7
/小型FTP/服务端/server.py
9d417fab71d1189dbf72fcb676df0e9a7689d76d
[]
no_license
laijin/mini_FTP
5be28ebe457dac57eef05d48c88e81c01fd0d1da
01c5777135a019efacddce0566dca1df757b9568
refs/heads/master
2020-04-01T17:08:10.404816
2018-10-17T07:37:31
2018-10-17T07:37:31
153,414,733
0
0
null
null
null
null
UTF-8
Python
false
false
1,614
py
import socket import os import subprocess import struct import json share_dir=r'D:\python3.7\chengxu\luffy\ch-5\小型FTP\客户端\share' def get(cmds,conn): filename = cmds[1] # 以读的方式打开文件,读取文件内容,发送给客户端 header_dic = { 'filename': filename, 'md5': 'graga', 'file_size': os.path.getsize(r'%s/%s' % (share_dir, filename)) } header_json = json.dumps(header_dic) header_bytes = header_json.encode('utf-8') header_len = struct.pack('i', len(header_bytes)) # 2、把报头发送给客户端 conn.send(header_len) conn.send(header_bytes) # 3、发送数据 f = open('%s/%s' % (share_dir, filename), 'rb') for line in f: conn.send(line) def server(): phone=socket.socket(socket.AF_INET,socket.SOCK_STREAM) phone.setsockopt(socket.SOL_SOCKET,socket.SO_REUSEADDR,1) phone.bind(('127.0.0.1',8081)) phone.listen(5) while True: conn,client_addr=phone.accept() while True: try: data=conn.recv(1024) if not data:break print('客户端的数据为',data) # 1、conn.send(data.upper()) #解析命令,提取相应的命令参数 cmds=data.decode('utf-8').split() if cmds[0] == 'get': get(cmds,conn) except ConnectionResetError: break conn.close() phone.close() if __name__ == '__main__': server()
[ "noreply@github.com" ]
noreply@github.com
a6f5732bacf0acd92492666f0a0fa0795679cf6c
e19ddbc01a0a8f6295a106ff969da9f4a9ffd408
/MIT_OCW/Textbook Finger Exercises/2-4-2_Largeset_Odd_Out_of_10.py
cfb1449cf4720b493d02eb99650255d70ff9f6f2
[]
no_license
Nmeece/MIT_OpenCourseware_Python
8860049c48090646fc9e98e469047cc5a0b171f1
69aecd3e9b448bc5ce2b98e46d72d65b35e212b5
refs/heads/master
2023-04-24T07:48:31.216771
2021-05-02T17:38:10
2021-05-02T17:38:10
318,314,220
0
1
null
null
null
null
UTF-8
Python
false
false
877
py
# -*- coding: utf-8 -*- """ Created on Fri Nov 13 11:44:55 2020 Second Finger Exercise of section 2.4 of the Introduction to Computation and Programming Using Python textbook. Prompt: Write a program that asks the user to input 10 integers, and then prints the largest odd number that was entered. If no odd number was entered, it should print a message to that effect. @author: Nygel """ numInts = 0 List = [] Big = 0 evenCount = 0 while numInts != 10: numIn = int(input("Enter value number: ")) List.append(numIn) numInts += 1 for n in List: if n % 2 != 0 and n > Big: Big = n elif n % 2 == 0: evenCount += 1 else: break if evenCount == 10: print("All Even") else: print("The largest odd number is:", Big)
[ "nygelmeece8@gmail.com" ]
nygelmeece8@gmail.com
e1136957d862e88999704359ee05863d30e39f15
4c99cd9846d6c9f196c7cbfa06f0576fb221a104
/virtuellenv ihkb/lib/python2.7/site-packages/whoosh/fields.py
0c6789cffdaa2519cb35c068b74592287435b576
[]
no_license
Gitlena/ichhabkeinblog
4622637773d5c4f3647455659058aab90d5bd050
1cfd7a030f9d41279a120724d29e1f48aa511562
refs/heads/master
2016-09-08T01:56:39.943817
2014-03-22T19:52:57
2014-03-22T19:52:57
null
0
0
null
null
null
null
UTF-8
Python
false
false
51,659
py
# Copyright 2007 Matt Chaput. All rights reserved. # # Redistribution and use in source and binary forms, with or without # modification, are permitted provided that the following conditions are met: # # 1. Redistributions of source code must retain the above copyright notice, # this list of conditions and the following disclaimer. # # 2. Redistributions in binary form must reproduce the above copyright # notice, this list of conditions and the following disclaimer in the # documentation and/or other materials provided with the distribution. # # THIS SOFTWARE IS PROVIDED BY MATT CHAPUT ``AS IS'' AND ANY EXPRESS OR # IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF # MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO # EVENT SHALL MATT CHAPUT OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, # INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT # LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, # OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF # LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING # NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, # EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. # # The views and conclusions contained in the software and documentation are # those of the authors and should not be interpreted as representing official # policies, either expressed or implied, of Matt Chaput. """ Contains functions and classes related to fields. """ import datetime, fnmatch, re, struct, sys from array import array from decimal import Decimal from whoosh import analysis, columns, formats from whoosh.compat import u, b, PY3 from whoosh.compat import with_metaclass from whoosh.compat import itervalues, xrange from whoosh.compat import bytes_type, string_type, integer_types, text_type from whoosh.system import emptybytes from whoosh.system import pack_byte, unpack_byte from whoosh.util.numeric import to_sortable, from_sortable from whoosh.util.numeric import typecode_max, NaN from whoosh.util.text import utf8encode, utf8decode from whoosh.util.times import datetime_to_long, long_to_datetime # Exceptions class FieldConfigurationError(Exception): pass class UnknownFieldError(Exception): pass # Field Types class FieldType(object): """Represents a field configuration. The FieldType object supports the following attributes: * format (formats.Format): the storage format for the field's contents. * analyzer (analysis.Analyzer): the analyzer to use to turn text into terms. * vector (formats.Format): the storage format for the field's vectors (forward index), or None if the field should not store vectors. * scorable (boolean): whether searches against this field may be scored. This controls whether the index stores per-document field lengths for this field. * stored (boolean): whether the content of this field is stored for each document. For example, in addition to indexing the title of a document, you usually want to store the title so it can be presented as part of the search results. * unique (boolean): whether this field's value is unique to each document. For example, 'path' or 'ID'. IndexWriter.update_document() will use fields marked as 'unique' to find the previous version of a document being updated. * multitoken_query is a string indicating what kind of query to use when a "word" in a user query parses into multiple tokens. The string is interpreted by the query parser. The strings understood by the default query parser are "first" (use first token only), "and" (join the tokens with an AND query), "or" (join the tokens with OR), "phrase" (join the tokens with a phrase query), and "default" (use the query parser's default join type). The constructor for the base field type simply lets you supply your own configured field format, vector format, and scorable and stored values. Subclasses may configure some or all of this for you. """ analyzer = format = vector = scorable = stored = unique = None indexed = True multitoken_query = "default" sortable_typecode = None spelling = False column_type = None def __init__(self, format, analyzer, vector=None, scorable=False, stored=False, unique=False, multitoken_query="default", sortable=False): assert isinstance(format, formats.Format) self.format = format self.analyzer = analyzer self.vector = vector self.scorable = scorable self.stored = stored self.unique = unique self.multitoken_query = multitoken_query self.set_sortable(sortable) def __repr__(self): temp = "%s(format=%r, vector=%r, scorable=%s, stored=%s, unique=%s)" return temp % (self.__class__.__name__, self.format, self.vector, self.scorable, self.stored, self.unique) def __eq__(self, other): return all((isinstance(other, FieldType), (self.format == other.format), (self.vector == other.vector), (self.scorable == other.scorable), (self.stored == other.stored), (self.unique == other.unique), (self.column_type == other.column_type))) def __ne__(self, other): return not(self.__eq__(other)) # Column methods def set_sortable(self, sortable): if sortable: if isinstance(sortable, columns.Column): self.column_type = sortable else: self.column_type = self.default_column() else: self.column_type = None def default_column(self): return columns.VarBytesColumn() # Methods for converting input into indexing information def index(self, value, **kwargs): """Returns an iterator of (btext, frequency, weight, encoded_value) tuples for each unique word in the input value. The default implementation uses the ``analyzer`` attribute to tokenize the value into strings, then encodes them into bytes using UTF-8. """ if not self.format: raise Exception("%s field %r cannot index without a format" % (self.__class__.__name__, self)) if not isinstance(value, (text_type, list, tuple)): raise ValueError("%r is not unicode or sequence" % value) assert isinstance(self.format, formats.Format) if "mode" not in kwargs: kwargs["mode"] = "index" word_values = self.format.word_values ana = self.analyzer for tstring, freq, wt, vbytes in word_values(value, ana, **kwargs): yield (utf8encode(tstring)[0], freq, wt, vbytes) def process_text(self, qstring, mode='', **kwargs): """Analyzes the given string and returns an iterator of token texts. >>> field = fields.TEXT() >>> list(field.process_text("The ides of March")) ["ides", "march"] """ if not self.format: raise Exception("%s field has no format" % self) return (t.text for t in self.tokenize(qstring, mode=mode, **kwargs)) def tokenize(self, value, **kwargs): """Analyzes the given string and returns an iterator of Token objects (note: for performance reasons, actually the same token yielded over and over with different attributes). """ if not self.analyzer: raise Exception("%s field has no analyzer" % self.__class__) return self.analyzer(value, **kwargs) def to_bytes(self, value): """Returns a bytes representation of the given value, appropriate to be written to disk. The default implementation assumes a unicode value and encodes it using UTF-8. """ if isinstance(value, (list, tuple)): value = value[0] if not isinstance(value, bytes_type): value = utf8encode(value)[0] return value def to_column_value(self, value): """Returns an object suitable to be inserted into the document values column for this field. The default implementation simply calls ``self.to_bytes(value)``. """ return self.to_bytes(value) def from_column_value(self, value): return self.from_bytes(value) def from_bytes(self, bs): return utf8decode(bs)[0] # Methods related to query parsing def self_parsing(self): """Subclasses should override this method to return True if they want the query parser to call the field's ``parse_query()`` method instead of running the analyzer on text in this field. This is useful where the field needs full control over how queries are interpreted, such as in the numeric field type. """ return False def parse_query(self, fieldname, qstring, boost=1.0): """When ``self_parsing()`` returns True, the query parser will call this method to parse basic query text. """ raise NotImplementedError(self.__class__.__name__) def parse_range(self, fieldname, start, end, startexcl, endexcl, boost=1.0): """When ``self_parsing()`` returns True, the query parser will call this method to parse range query text. If this method returns None instead of a query object, the parser will fall back to parsing the start and end terms using process_text(). """ return None # Methods related to sortings def sortable_terms(self, ixreader, fieldname): """Returns an iterator of the "sortable" tokens in the given reader and field. These values can be used for sorting. The default implementation simply returns all tokens in the field. This can be overridden by field types such as NUMERIC where some values in a field are not useful for sorting. """ return ixreader.lexicon(fieldname) # Methods related to spelling def separate_spelling(self): """Returns True if this field requires special handling of the words that go into the field's word graph. The default behavior is to return True if the field is "spelled" but not indexed, or if the field is indexed but the analyzer has morphological transformations (e.g. stemming). Exotic field types may need to override this behavior. This method should return False if the field does not support spelling (i.e. the ``spelling`` attribute is False). """ return self.spelling and self.analyzer.has_morph() def spellable_words(self, value): """Returns an iterator of each unique word (in sorted order) in the input value, suitable for inclusion in the field's word graph. The default behavior is to call the field analyzer with the keyword argument ``no_morph=True``, which should make the analyzer skip any morphological transformation filters (e.g. stemming) to preserve the original form of the words. Exotic field types may need to override this behavior. """ if isinstance(value, (list, tuple)): words = value else: words = [token.text for token in self.analyzer(value, no_morph=True)] return iter(sorted(set(words))) def has_morph(self): """Returns True if this field by default performs morphological transformations on its terms, e.g. stemming. """ if self.analyzer: return self.analyzer.has_morph() else: return False # Methods related to the posting/vector formats def supports(self, name): """Returns True if the underlying format supports the given posting value type. >>> field = TEXT() >>> field.supports("positions") True >>> field.supports("characters") False """ return self.format.supports(name) def clean(self): """Clears any cached information in the field and any child objects. """ if self.format and hasattr(self.format, "clean"): self.format.clean() if self.vector and hasattr(self.vector, "clean"): self.vector.clean() # Event methods def on_add(self, schema, fieldname): pass def on_remove(self, schema, fieldname): pass class ID(FieldType): """Configured field type that indexes the entire value of the field as one token. This is useful for data you don't want to tokenize, such as the path of a file. """ __inittypes__ = dict(stored=bool, unique=bool, field_boost=float) def __init__(self, stored=False, unique=False, field_boost=1.0, spelling=False, sortable=False, analyzer=None): """ :param stored: Whether the value of this field is stored with the document. """ self.analyzer = analyzer or analysis.IDAnalyzer() self.format = formats.Existence(field_boost=field_boost) self.stored = stored self.unique = unique self.spelling = spelling self.set_sortable(sortable) class IDLIST(FieldType): """Configured field type for fields containing IDs separated by whitespace and/or punctuation (or anything else, using the expression param). """ __inittypes__ = dict(stored=bool, unique=bool, expression=bool, field_boost=float) def __init__(self, stored=False, unique=False, expression=None, field_boost=1.0, spelling=False): """ :param stored: Whether the value of this field is stored with the document. :param unique: Whether the value of this field is unique per-document. :param expression: The regular expression object to use to extract tokens. The default expression breaks tokens on CRs, LFs, tabs, spaces, commas, and semicolons. """ expression = expression or re.compile(r"[^\r\n\t ,;]+") self.analyzer = analysis.RegexAnalyzer(expression=expression) self.format = formats.Existence(field_boost=field_boost) self.stored = stored self.unique = unique self.spelling = spelling class NUMERIC(FieldType): """Special field type that lets you index integer or floating point numbers in relatively short fixed-width terms. The field converts numbers to sortable bytes for you before indexing. You specify the numeric type of the field (``int`` or ``float``) when you create the ``NUMERIC`` object. The default is ``int``. For ``int``, you can specify a size in bits (``32`` or ``64``). For both ``int`` and ``float`` you can specify a ``signed`` keyword argument (default is ``True``). >>> schema = Schema(path=STORED, position=NUMERIC(int, 64, signed=False)) >>> ix = storage.create_index(schema) >>> with ix.writer() as w: ... w.add_document(path="/a", position=5820402204) ... You can also use the NUMERIC field to store Decimal instances by specifying a type of ``int`` or ``long`` and the ``decimal_places`` keyword argument. This simply multiplies each number by ``(10 ** decimal_places)`` before storing it as an integer. Of course this may throw away decimal prcesision (by truncating, not rounding) and imposes the same maximum value limits as ``int``/``long``, but these may be acceptable for certain applications. >>> from decimal import Decimal >>> schema = Schema(path=STORED, position=NUMERIC(int, decimal_places=4)) >>> ix = storage.create_index(schema) >>> with ix.writer() as w: ... w.add_document(path="/a", position=Decimal("123.45") ... """ def __init__(self, numtype=int, bits=32, stored=False, unique=False, field_boost=1.0, decimal_places=0, shift_step=4, signed=True, sortable=False, default=None): """ :param numtype: the type of numbers that can be stored in this field, either ``int``, ``float``. If you use ``Decimal``, use the ``decimal_places`` argument to control how many decimal places the field will store. :param bits: When ``numtype`` is ``int``, the number of bits to use to store the number: 8, 16, 32, or 64. :param stored: Whether the value of this field is stored with the document. :param unique: Whether the value of this field is unique per-document. :param decimal_places: specifies the number of decimal places to save when storing Decimal instances. If you set this, you will always get Decimal instances back from the field. :param shift_steps: The number of bits of precision to shift away at each tiered indexing level. Values should generally be 1-8. Lower values yield faster searches but take up more space. A value of `0` means no tiered indexing. :param signed: Whether the numbers stored in this field may be negative. """ # Allow users to specify strings instead of Python types in case # docstring isn't clear if numtype == "int": numtype = int if numtype == "float": numtype = float # Raise an error if the user tries to use a type other than int or # float if numtype is Decimal: numtype = int if not decimal_places: raise TypeError("To store Decimal instances, you must set the " "decimal_places argument") elif numtype not in (int, float): raise TypeError("Can't use %r as a type, use int or float" % numtype) # Sanity check if numtype is float and decimal_places: raise Exception("A float type and decimal_places argument %r are " "incompatible" % decimal_places) intsizes = [8, 16, 32, 64] intcodes = ["B", "H", "I", "Q"] # Set up field configuration based on type and size if numtype is float: bits = 64 # Floats are converted to 64 bit ints else: if bits not in intsizes: raise Exception("Invalid bits %r, use 8, 16, 32, or 64" % bits) # Type code for the *sortable* representation self.sortable_typecode = intcodes[intsizes.index(bits)] self._struct = struct.Struct(">" + self.sortable_typecode) self.numtype = numtype self.bits = bits self.stored = stored self.unique = unique self.decimal_places = decimal_places self.shift_step = shift_step self.signed = signed self.analyzer = analysis.IDAnalyzer() self.format = formats.Existence(field_boost=field_boost) self.min_value, self.max_value = self._min_max() # Column configuration if default is None: if numtype is int: default = typecode_max[self.sortable_typecode] else: default = NaN elif not self.is_valid(default): raise Exception("The default %r is not a valid number for this " "field" % default) self.default = default self.set_sortable(sortable) def __getstate__(self): d = self.__dict__.copy() if "_struct" in d: del d["_struct"] return d def __setstate__(self, d): self.__dict__.update(d) self._struct = struct.Struct(">" + self.sortable_typecode) if "min_value" not in d: d["min_value"], d["max_value"] = self._min_max() def _min_max(self): numtype = self.numtype bits = self.bits signed = self.signed # Calculate the minimum and maximum possible values for error checking min_value = from_sortable(numtype, bits, signed, 0) max_value = from_sortable(numtype, bits, signed, 2 ** bits - 1) return min_value, max_value def default_column(self): return columns.NumericColumn(self.sortable_typecode, default=self.default) def is_valid(self, x): try: x = self.to_bytes(x) except ValueError: return False except OverflowError: return False return True def index(self, num, **kwargs): # If the user gave us a list of numbers, recurse on the list if isinstance(num, (list, tuple)): for n in num: for item in self.index(n): yield item return # word, freq, weight, valuestring if self.shift_step: for shift in xrange(0, self.bits, self.shift_step): yield (self.to_bytes(num, shift), 1, 1.0, emptybytes) else: yield (self.to_bytes(num), 1, 1.0, emptybytes) def prepare_number(self, x): if x == emptybytes or x is None: return x dc = self.decimal_places if dc and isinstance(x, (string_type, Decimal)): x = Decimal(x) * (10 ** dc) elif isinstance(x, Decimal): raise TypeError("Can't index a Decimal object unless you specified " "decimal_places on the field") try: x = self.numtype(x) except OverflowError: raise ValueError("Value %r overflowed number type %r" % (x, self.numtype)) if x < self.min_value or x > self.max_value: raise ValueError("Numeric field value %s out of range [%s, %s]" % (x, self.min_value, self.max_value)) return x def unprepare_number(self, x): dc = self.decimal_places if dc: s = str(x) x = Decimal(s[:-dc] + "." + s[-dc:]) return x def to_column_value(self, x): if isinstance(x, (list, tuple, array)): x = x[0] x = self.prepare_number(x) return to_sortable(self.numtype, self.bits, self.signed, x) def from_column_value(self, x): x = from_sortable(self.numtype, self.bits, self.signed, x) return self.unprepare_number(x) def to_bytes(self, x, shift=0): # Try to avoid re-encoding; this sucks because on Python 2 we can't # tell the difference between a string and encoded bytes, so we have # to require the user use unicode when they mean string if isinstance(x, bytes_type): return x if x == emptybytes or x is None: return self.sortable_to_bytes(0) x = self.prepare_number(x) x = to_sortable(self.numtype, self.bits, self.signed, x) return self.sortable_to_bytes(x, shift) def sortable_to_bytes(self, x, shift=0): if shift: x >>= shift return pack_byte(shift) + self._struct.pack(x) def from_bytes(self, bs): x = self._struct.unpack(bs[1:])[0] x = from_sortable(self.numtype, self.bits, self.signed, x) x = self.unprepare_number(x) return x def process_text(self, text, **kwargs): return (self.to_bytes(text),) def self_parsing(self): return True def parse_query(self, fieldname, qstring, boost=1.0): from whoosh import query from whoosh.qparser.common import QueryParserError if qstring == "*": return query.Every(fieldname, boost=boost) if not self.is_valid(qstring): raise QueryParserError("%r is not a valid number" % qstring) token = self.to_bytes(qstring) return query.Term(fieldname, token, boost=boost) def parse_range(self, fieldname, start, end, startexcl, endexcl, boost=1.0): from whoosh import query from whoosh.qparser.common import QueryParserError if start is not None: if not self.is_valid(start): raise QueryParserError("Range start %r is not a valid number" % start) start = self.prepare_number(start) if end is not None: if not self.is_valid(end): raise QueryParserError("Range end %r is not a valid number" % end) end = self.prepare_number(end) return query.NumericRange(fieldname, start, end, startexcl, endexcl, boost=boost) def sortable_terms(self, ixreader, fieldname): zero = b("\x00") for token in ixreader.lexicon(fieldname): if token[0:1] != zero: # Only yield the full-precision values break yield token class DATETIME(NUMERIC): """Special field type that lets you index datetime objects. The field converts the datetime objects to sortable text for you before indexing. Since this field is based on Python's datetime module it shares all the limitations of that module, such as the inability to represent dates before year 1 in the proleptic Gregorian calendar. However, since this field stores datetimes as an integer number of microseconds, it could easily represent a much wider range of dates if the Python datetime implementation ever supports them. >>> schema = Schema(path=STORED, date=DATETIME) >>> ix = storage.create_index(schema) >>> w = ix.writer() >>> w.add_document(path="/a", date=datetime.now()) >>> w.commit() """ __inittypes__ = dict(stored=bool, unique=bool) def __init__(self, stored=False, unique=False, sortable=False): """ :param stored: Whether the value of this field is stored with the document. :param unique: Whether the value of this field is unique per-document. """ super(DATETIME, self).__init__(int, 64, stored=stored, unique=unique, shift_step=8, sortable=sortable) def prepare_datetime(self, x): from whoosh.util.times import floor if isinstance(x, text_type): # For indexing, support same strings as for query parsing -- # convert unicode to datetime object x = self._parse_datestring(x) x = floor(x) # this makes most sense (unspecified = lowest) if isinstance(x, datetime.datetime): return datetime_to_long(x) elif isinstance(x, bytes_type): return x else: raise Exception("%r is not a datetime" % (x,)) def to_column_value(self, x): if isinstance(x, bytes_type): raise Exception("%r is not a datetime" % (x,)) if isinstance(x, (list, tuple)): x = x[0] return self.prepare_datetime(x) def from_column_value(self, x): return long_to_datetime(x) def to_bytes(self, x, shift=0): x = self.prepare_datetime(x) return NUMERIC.to_bytes(self, x, shift=shift) def from_bytes(self, bs): x = NUMERIC.from_bytes(self, bs) return long_to_datetime(x) def _parse_datestring(self, qstring): # This method parses a very simple datetime representation of the form # YYYY[MM[DD[hh[mm[ss[uuuuuu]]]]]] from whoosh.util.times import adatetime, fix, is_void qstring = qstring.replace(" ", "").replace("-", "").replace(".", "") year = month = day = hour = minute = second = microsecond = None if len(qstring) >= 4: year = int(qstring[:4]) if len(qstring) >= 6: month = int(qstring[4:6]) if len(qstring) >= 8: day = int(qstring[6:8]) if len(qstring) >= 10: hour = int(qstring[8:10]) if len(qstring) >= 12: minute = int(qstring[10:12]) if len(qstring) >= 14: second = int(qstring[12:14]) if len(qstring) == 20: microsecond = int(qstring[14:]) at = fix(adatetime(year, month, day, hour, minute, second, microsecond)) if is_void(at): raise Exception("%r is not a parseable date" % qstring) return at def parse_query(self, fieldname, qstring, boost=1.0): from whoosh import query from whoosh.util.times import is_ambiguous try: at = self._parse_datestring(qstring) except: e = sys.exc_info()[1] return query.error_query(e) if is_ambiguous(at): startnum = datetime_to_long(at.floor()) endnum = datetime_to_long(at.ceil()) return query.NumericRange(fieldname, startnum, endnum) else: return query.Term(fieldname, at, boost=boost) def parse_range(self, fieldname, start, end, startexcl, endexcl, boost=1.0): from whoosh import query if start is None and end is None: return query.Every(fieldname, boost=boost) if start is not None: startdt = self._parse_datestring(start).floor() start = datetime_to_long(startdt) if end is not None: enddt = self._parse_datestring(end).ceil() end = datetime_to_long(enddt) return query.NumericRange(fieldname, start, end, boost=boost) class BOOLEAN(FieldType): """Special field type that lets you index boolean values (True and False). The field converts the boolean values to text for you before indexing. >>> schema = Schema(path=STORED, done=BOOLEAN) >>> ix = storage.create_index(schema) >>> w = ix.writer() >>> w.add_document(path="/a", done=False) >>> w.commit() """ bytestrings = (b("f"), b("t")) trues = frozenset(u("t true yes 1").split()) falses = frozenset(u("f false no 0").split()) __inittypes__ = dict(stored=bool, field_boost=float) def __init__(self, stored=False, field_boost=1.0): """ :param stored: Whether the value of this field is stored with the document. """ self.stored = stored self.field_boost = field_boost self.format = formats.Existence(field_boost=field_boost) def _obj_to_bool(self, x): # We special case strings such as "true", "false", "yes", "no", but # otherwise call bool() on the query value. This lets you pass objects # as query values and do the right thing. if isinstance(x, string_type) and x.lower() in self.trues: x = True elif isinstance(x, string_type) and x.lower() in self.falses: x = False else: x = bool(x) return x def to_bytes(self, x): if isinstance(x, bytes_type): return x elif isinstance(x, string_type): x = x.lower() in self.trues else: x = bool(x) bs = self.bytestrings[int(x)] return bs def index(self, bit, **kwargs): if isinstance(bit, string_type): bit = bit.lower() in self.trues else: bit = bool(bit) # word, freq, weight, valuestring return [(self.bytestrings[int(bit)], 1, 1.0, emptybytes)] def self_parsing(self): return True def parse_query(self, fieldname, qstring, boost=1.0): from whoosh import query if qstring == "*": return query.Every(fieldname, boost=boost) return query.Term(fieldname, self._obj_to_bool(qstring), boost=boost) class STORED(FieldType): """Configured field type for fields you want to store but not index. """ indexed = False stored = True def __init__(self): pass class COLUMN(FieldType): """Configured field type for fields you want to store as a per-document value column but not index. """ indexed = False stored = False def __init__(self, columnobj=None): if columnobj is None: columnobj = columns.VarBytesColumn() if not isinstance(columnobj, columns.Column): raise TypeError("%r is not a column object" % (columnobj,)) self.column_type = columnobj def to_bytes(self, v): return v def from_bytes(self, b): return b class KEYWORD(FieldType): """Configured field type for fields containing space-separated or comma-separated keyword-like data (such as tags). The default is to not store positional information (so phrase searching is not allowed in this field) and to not make the field scorable. """ __inittypes__ = dict(stored=bool, lowercase=bool, commas=bool, scorable=bool, unique=bool, field_boost=float) def __init__(self, stored=False, lowercase=False, commas=False, vector=None, scorable=False, unique=False, field_boost=1.0, spelling=False, sortable=False): """ :param stored: Whether to store the value of the field with the document. :param comma: Whether this is a comma-separated field. If this is False (the default), it is treated as a space-separated field. :param scorable: Whether this field is scorable. """ self.analyzer = analysis.KeywordAnalyzer(lowercase=lowercase, commas=commas) self.format = formats.Frequency(field_boost=field_boost) self.scorable = scorable self.stored = stored self.unique = unique self.spelling = spelling if vector: if type(vector) is type: vector = vector() elif isinstance(vector, formats.Format): pass else: vector = self.format else: vector = None self.vector = vector if sortable: self.column_type = self.default_column() class TEXT(FieldType): """Configured field type for text fields (for example, the body text of an article). The default is to store positional information to allow phrase searching. This field type is always scorable. """ __inittypes__ = dict(analyzer=analysis.Analyzer, phrase=bool, vector=object, stored=bool, field_boost=float) def __init__(self, analyzer=None, phrase=True, chars=False, vector=None, stored=False, field_boost=1.0, multitoken_query="default", spelling=False, sortable=False, lang=None): """ :param analyzer: The analysis.Analyzer to use to index the field contents. See the analysis module for more information. If you omit this argument, the field uses analysis.StandardAnalyzer. :param phrase: Whether the store positional information to allow phrase searching. :param chars: Whether to store character ranges along with positions. If this is True, "phrase" is also implied. :param vector: A :class:`whoosh.formats.Format` object to use to store term vectors, or ``True`` to store vectors using the same format as the inverted index, or ``None`` or ``False`` to not store vectors. By default, fields do not store term vectors. :param stored: Whether to store the value of this field with the document. Since this field type generally contains a lot of text, you should avoid storing it with the document unless you need to, for example to allow fast excerpts in the search results. :param spelling: Whether to generate word graphs for this field to make spelling suggestions much faster. :param sortable: If True, make this field sortable using the default column type. If you pass a :class:`whoosh.columns.Column` instance instead of True, the field will use the given column type. :param lang: automaticaly configure a :class:`whoosh.analysis.LanguageAnalyzer` for the given language. This is ignored if you also specify an ``analyzer``. """ if analyzer: self.analyzer = analyzer elif lang: self.analyzer = analysis.LanguageAnalyzer(lang) else: self.analyzer = analysis.StandardAnalyzer() if chars: formatclass = formats.Characters elif phrase: formatclass = formats.Positions else: formatclass = formats.Frequency self.format = formatclass(field_boost=field_boost) if vector: if type(vector) is type: vector = vector() elif isinstance(vector, formats.Format): pass else: vector = formatclass() else: vector = None self.vector = vector if sortable: if isinstance(sortable, columns.Column): self.column_type = sortable else: self.column_type = columns.VarBytesColumn() else: self.column_type = None self.multitoken_query = multitoken_query self.scorable = True self.stored = stored self.spelling = spelling class NGRAM(FieldType): """Configured field that indexes text as N-grams. For example, with a field type NGRAM(3,4), the value "hello" will be indexed as tokens "hel", "hell", "ell", "ello", "llo". This field type chops the entire text into N-grams, including whitespace and punctuation. See :class:`NGRAMWORDS` for a field type that breaks the text into words first before chopping the words into N-grams. """ __inittypes__ = dict(minsize=int, maxsize=int, stored=bool, field_boost=float, queryor=bool, phrase=bool) scorable = True def __init__(self, minsize=2, maxsize=4, stored=False, field_boost=1.0, queryor=False, phrase=False, sortable=False): """ :param minsize: The minimum length of the N-grams. :param maxsize: The maximum length of the N-grams. :param stored: Whether to store the value of this field with the document. Since this field type generally contains a lot of text, you should avoid storing it with the document unless you need to, for example to allow fast excerpts in the search results. :param queryor: if True, combine the N-grams with an Or query. The default is to combine N-grams with an And query. :param phrase: store positions on the N-grams to allow exact phrase searching. The default is off. """ formatclass = formats.Frequency if phrase: formatclass = formats.Positions self.analyzer = analysis.NgramAnalyzer(minsize, maxsize) self.format = formatclass(field_boost=field_boost) self.stored = stored self.queryor = queryor self.set_sortable(sortable) def self_parsing(self): return True def parse_query(self, fieldname, qstring, boost=1.0): from whoosh import query terms = [query.Term(fieldname, g) for g in self.process_text(qstring, mode='query')] cls = query.Or if self.queryor else query.And return cls(terms, boost=boost) class NGRAMWORDS(NGRAM): """Configured field that chops text into words using a tokenizer, lowercases the words, and then chops the words into N-grams. """ __inittypes__ = dict(minsize=int, maxsize=int, stored=bool, field_boost=float, tokenizer=analysis.Tokenizer, at=str, queryor=bool) scorable = True def __init__(self, minsize=2, maxsize=4, stored=False, field_boost=1.0, tokenizer=None, at=None, queryor=False, sortable=False): """ :param minsize: The minimum length of the N-grams. :param maxsize: The maximum length of the N-grams. :param stored: Whether to store the value of this field with the document. Since this field type generally contains a lot of text, you should avoid storing it with the document unless you need to, for example to allow fast excerpts in the search results. :param tokenizer: an instance of :class:`whoosh.analysis.Tokenizer` used to break the text into words. :param at: if 'start', only takes N-grams from the start of the word. If 'end', only takes N-grams from the end. Otherwise the default is to take all N-grams from each word. :param queryor: if True, combine the N-grams with an Or query. The default is to combine N-grams with an And query. """ self.analyzer = analysis.NgramWordAnalyzer(minsize, maxsize, tokenizer, at=at) self.format = formats.Frequency(field_boost=field_boost) self.stored = stored self.queryor = queryor self.set_sortable(sortable) # Schema class class MetaSchema(type): def __new__(cls, name, bases, attrs): super_new = super(MetaSchema, cls).__new__ if not any(b for b in bases if isinstance(b, MetaSchema)): # If this isn't a subclass of MetaSchema, don't do anything special return super_new(cls, name, bases, attrs) # Create the class special_attrs = {} for key in list(attrs.keys()): if key.startswith("__"): special_attrs[key] = attrs.pop(key) new_class = super_new(cls, name, bases, special_attrs) fields = {} for b in bases: if hasattr(b, "_clsfields"): fields.update(b._clsfields) fields.update(attrs) new_class._clsfields = fields return new_class def schema(self): return Schema(**self._clsfields) class Schema(object): """Represents the collection of fields in an index. Maps field names to FieldType objects which define the behavior of each field. Low-level parts of the index use field numbers instead of field names for compactness. This class has several methods for converting between the field name, field number, and field object itself. """ def __init__(self, **fields): """ All keyword arguments to the constructor are treated as fieldname = fieldtype pairs. The fieldtype can be an instantiated FieldType object, or a FieldType sub-class (in which case the Schema will instantiate it with the default constructor before adding it). For example:: s = Schema(content = TEXT, title = TEXT(stored = True), tags = KEYWORD(stored = True)) """ self._fields = {} self._dyn_fields = {} for name in sorted(fields.keys()): self.add(name, fields[name]) def copy(self): """Returns a shallow copy of the schema. The field instances are not deep copied, so they are shared between schema copies. """ return self.__class__(**self._fields) def __eq__(self, other): return (other.__class__ is self.__class__ and list(self.items()) == list(other.items())) def __ne__(self, other): return not(self.__eq__(other)) def __repr__(self): return "<%s: %r>" % (self.__class__.__name__, self.names()) def __iter__(self): """Returns the field objects in this schema. """ return iter(self._fields.values()) def __getitem__(self, name): """Returns the field associated with the given field name. """ if name in self._fields: return self._fields[name] for expr, fieldtype in itervalues(self._dyn_fields): if expr.match(name): return fieldtype raise KeyError("No field named %r" % (name,)) def __len__(self): """Returns the number of fields in this schema. """ return len(self._fields) def __contains__(self, fieldname): """Returns True if a field by the given name is in this schema. """ # Defined in terms of __getitem__ so that there's only one method to # override to provide dynamic fields try: field = self[fieldname] return field is not None except KeyError: return False def items(self): """Returns a list of ("fieldname", field_object) pairs for the fields in this schema. """ return sorted(self._fields.items()) def names(self, check_names=None): """Returns a list of the names of the fields in this schema. :param check_names: (optional) sequence of field names to check whether the schema accepts them as (dynamic) field names - acceptable names will also be in the result list. Note: You may also have static field names in check_names, that won't create duplicates in the result list. Unsupported names will not be in the result list. """ fieldnames = set(self._fields.keys()) if check_names is not None: check_names = set(check_names) - fieldnames fieldnames.update(fieldname for fieldname in check_names if fieldname in self) return sorted(fieldnames) def clean(self): for field in self: field.clean() def add(self, name, fieldtype, glob=False): """Adds a field to this schema. :param name: The name of the field. :param fieldtype: An instantiated fields.FieldType object, or a FieldType subclass. If you pass an instantiated object, the schema will use that as the field configuration for this field. If you pass a FieldType subclass, the schema will automatically instantiate it with the default constructor. """ # Check field name if name.startswith("_"): raise FieldConfigurationError("Field names cannot start with an " "underscore") if " " in name: raise FieldConfigurationError("Field names cannot contain spaces") if name in self._fields or (glob and name in self._dyn_fields): raise FieldConfigurationError("Schema already has a field %r" % name) # If the user passed a type rather than an instantiated field object, # instantiate it automatically if type(fieldtype) is type: try: fieldtype = fieldtype() except: e = sys.exc_info()[1] raise FieldConfigurationError("Error: %s instantiating field " "%r: %r" % (e, name, fieldtype)) if not isinstance(fieldtype, FieldType): raise FieldConfigurationError("%r is not a FieldType object" % fieldtype) if glob: expr = re.compile(fnmatch.translate(name)) self._dyn_fields[name] = (expr, fieldtype) else: fieldtype.on_add(self, name) self._fields[name] = fieldtype def remove(self, fieldname): if fieldname in self._fields: self._fields[fieldname].on_remove(self, fieldname) del self._fields[fieldname] elif fieldname in self._dyn_fields: del self._dyn_fields[fieldname] else: raise KeyError("No field named %r" % fieldname) def has_vectored_fields(self): """Returns True if any of the fields in this schema store term vectors. """ return any(ftype.vector for ftype in self) def has_scorable_fields(self): return any(ftype.scorable for ftype in self) def stored_names(self): """Returns a list of the names of fields that are stored. """ return [name for name, field in self.items() if field.stored] def scorable_names(self): """Returns a list of the names of fields that store field lengths. """ return [name for name, field in self.items() if field.scorable] def vector_names(self): """Returns a list of the names of fields that store vectors. """ return [name for name, field in self.items() if field.vector] def separate_spelling_names(self): """Returns a list of the names of fields that require special handling for generating spelling graphs... either because they store graphs but aren't indexed, or because the analyzer is stemmed. """ return [name for name, field in self.items() if field.spelling and field.separate_spelling()] class SchemaClass(with_metaclass(MetaSchema, Schema)): """Allows you to define a schema using declarative syntax, similar to Django models.py:: class MySchema(SchemaClass): path = ID date = DATETIME content = TEXT You can use inheritance to share common fields between schemas:: class Parent(SchemaClass): path = ID(stored=True) date = DATETIME class Child1(Parent): content = TEXT(positions=False) class Child2(Parent): tags = KEYWORD This class overrides ``__new__`` so instantiating your sub-class always results in an instance of ``Schema``. >>> class MySchema(SchemaClass): ... title = TEXT(stored=True) ... content = TEXT ... >>> s = MySchema() >>> type(s) <class 'whoosh.fields.Schema'> """ def __new__(cls, *args, **kwargs): obj = super(Schema, cls).__new__(Schema) kw = getattr(cls, "_clsfields", {}) kw.update(kwargs) obj.__init__(*args, **kw) return obj def ensure_schema(schema): if isinstance(schema, type) and issubclass(schema, Schema): schema = schema.schema() if not isinstance(schema, Schema): raise FieldConfigurationError("%r is not a Schema" % schema) return schema def merge_fielddict(d1, d2): keyset = set(d1.keys()) | set(d2.keys()) out = {} for name in keyset: field1 = d1.get(name) field2 = d2.get(name) if field1 and field2 and field1 != field2: raise Exception("Inconsistent field %r: %r != %r" % (name, field1, field2)) out[name] = field1 or field2 return out def merge_schema(s1, s2): schema = Schema() schema._fields = merge_fielddict(s1._fields, s2._fields) schema._dyn_fields = merge_fielddict(s1._dyn_fields, s2._dyn_fields) return schema def merge_schemas(schemas): schema = schemas[0] for i in xrange(1, len(schemas)): schema = merge_schema(schema, schemas[i]) return schema
[ "ubunicorn@googlemail.com" ]
ubunicorn@googlemail.com
e3ea3a268a625b1c2669d93b5c29caa0a984cda4
f63e53c9d7a23c03ecbf488e9ea156c5c98f3f33
/bookstore_oop.py
a6eeea598ec9a34570aa17f2e95f7d001407c3bb
[]
no_license
matramir/bookstore
957841664becc9bf00113e51eaed97373d8c8920
c0a19c767a784d9c1c2fafb6267dbdfaf3ff9b03
refs/heads/master
2020-03-27T16:19:10.401750
2018-08-30T17:44:12
2018-08-30T17:44:12
146,773,602
0
0
null
null
null
null
UTF-8
Python
false
false
3,386
py
from tkinter import * from backend_oop import Database """ Program that stores Title, Author, Year, ISBN User can View, Search, Add, Update, Delete, close """ database = Database("books.db") def get_selected_row(event): try: global selected_tuple index=list1.curselection()[0] selected_tuple = list1.get(index) e1.delete(0,END) e1.insert(END,selected_tuple[1]) e2.delete(0,END) e2.insert(END,selected_tuple[2]) e3.delete(0,END) e3.insert(END,selected_tuple[3]) e4.delete(0,END) e4.insert(END,selected_tuple[4]) except: pass def view_all(): for row in database.view(): list1.delete(0,END) list1.insert(END, row) def search_entry(): try: list1.delete(0, END) for row in database.search(title_text.get(), author_text.get(), year_text.get(), isbn_text.get()): list1.insert(END, row) except: list1.delete(0, END) list.insert(END, ("Search Error",)) def add_entry(): try: database.insert(title_text.get(), author_text.get(), year_text.get(), isbn_text.get()) list1.delete(0, END) list.insert(END,(title_text.get(), author_text.get(), year_text.get(), isbn_text.get())) except: list1.delete(0, END) list.insert(END, "Add Error") def update_entry(): try: database.update(selected_tuple[0],title_text.get(), author_text.get(), year_text.get(), isbn_text.get()) except: list1.delete(0, END) list.insert(END, "Update Error") def delete_data(): try: database.delete(selected_tuple[0]) except: list1.delete(0, END) list.insert(END, "Delete Error") #def close_app(): window = Tk() window.title("Bookstore") #labels lb1 = Label(window,text="Title") lb2 = Label(window,text="Year") lb3 = Label(window,text="Author") lb4 = Label(window,text="ISBN") #entry values title_text = StringVar() year_text = StringVar() author_text = StringVar() isbn_text = StringVar() #entries e1 = Entry(window, textvariable=title_text) e2 = Entry(window, textvariable=year_text) e3 = Entry(window, textvariable=author_text) e4 = Entry(window, textvariable=isbn_text) #buttons b1 = Button(window, width=12, text="View All", command=view_all) b2 = Button(window, width=12, text="Search Entry", command=search_entry) b3 = Button(window, width=12, text="Add Entry", command=add_entry) b4 = Button(window, width=12, text="Update Selected", command=update_entry) b5 = Button(window, width=12, text="Delete Selected", command=delete_data) b6 = Button(window, width=12, text="Close",command=window.destroy) #fist entry column lb1.grid(row=0, column=0) e1.grid(row=0, column=1) lb2.grid(row=1, column=0) e2.grid(row=1, column=1) #second entry column lb3.grid(row=0, column=2) e3.grid(row=0, column=3) lb4.grid(row=1, column=2) e4.grid(row=1, column=3) #button column b1.grid(row=2, column=3) b2.grid(row=3, column=3) b3.grid(row=4, column=3) b4.grid(row=5, column=3) b5.grid(row=6, column=3) b6.grid(row=7, column=3) #listbox list1=Listbox(window, height=6,width=35) list1.grid(row=2,column=0,rowspan=6,columnspan=2) #scrollbar sb1=Scrollbar(window) sb1.grid(row=2,column=2,rowspan=6) list1.configure(yscrollcommand=sb1.set) sb1.configure(command=list1.yview) list1.bind('<<ListboxSelect>>',get_selected_row) window.mainloop()
[ "matias.ramirez.parisi@gmail.com" ]
matias.ramirez.parisi@gmail.com
9443edd0436310cbd47d8b6a75e21c106482fb9c
1eee09ce02d848a8e2653f199b7171f74cb7d1c9
/service/api.py
6496e828b7afe578a9d14aa5a0c97a8185b15a63
[ "MIT" ]
permissive
etalab/csv_detective_api
7416cde112fcd4ed70e703231569bce5901c2161
7c96f497374d842226a95a26cb6627ac22cd799b
refs/heads/master
2023-01-12T09:41:37.519013
2020-06-15T14:09:34
2020-06-15T14:09:34
193,232,263
3
3
MIT
2023-01-04T08:20:45
2019-06-22T12:35:26
Jupyter Notebook
UTF-8
Python
false
false
7,487
py
#!flask/bin/python import os import sys from collections import defaultdict sys.path.append("./csv_detective_ml") # horrible hack to load my features class to load my ML pipeline :/ from flask import Flask from flask import request from flask import jsonify from flask_restplus import Api, Resource, fields from flask_cors import CORS from tempfile import NamedTemporaryFile import logging import json from joblib import load from utils.reference_matcher import link_reference_datasets from utils.parsers import file_upload from csv_detective_ml.analyze_csv_cli import analyze_csv logger = logging.getLogger() logger.setLevel(logging.ERROR) logger.addHandler(logging.StreamHandler()) app = Flask(__name__) CORS(app) api = Api(app=app, version="0.1", title="CSV Detective API", description="Get info about the data contained in a DGF CSV file.") ns_csv_detective = api.namespace('csv_detective', description='Get data from DGF CSVs') resource_model = api.model('Analysis parameters', {'resource_id': fields.String(required=True, description="DGF Resource ID or CSV path", help="Resource ID cannot be blank") }) type_model = api.model('Type analysis parameters', {'target_type': fields.String(required=True, description="Target type to find among resources/datasets", help="Resource ID cannot be blank") }) DATASET_CSV_INFO = {} TYPE_CSV_INFO = defaultdict(lambda: defaultdict(dict)) ML_PIPELINE = None def load_ml_model(): global ML_PIPELINE logger.info("Loading ML model...") ML_PIPELINE = load('./csv_detective_ml/models/model.joblib') return ML_PIPELINE load_ml_model() @ns_csv_detective.route("/dataset_id") class CSVDetectiveDataset(Resource): @api.expect(resource_model) def get(self): global DATASET_CSV_INFO try: resource_id = request.args.get('resource_id') if resource_id in DATASET_CSV_INFO: response = DATASET_CSV_INFO[resource_id] response = reformat_response(response) response = link_reference_datasets(response) return jsonify(response) else: logger.info("Resource id not found in 'database'.") return jsonify({"error": "ID {} not found".format(resource_id)}) except Exception as e: return jsonify({"error": str(e)}) @ns_csv_detective.route("/resource_id") class CSVDetectiveResource(Resource): @api.expect(resource_model) def get(self): global DATASET_CSV_INFO try: resource_id = request.args.get('resource_id') if resource_id in DATASET_CSV_INFO: response = DATASET_CSV_INFO[resource_id] response = reformat_response(response) response = link_reference_datasets(response) return jsonify(response) else: logger.info("Resource id not found in 'database'.") return jsonify({"error": "ID {} not found".format(resource_id)}) except Exception as e: return jsonify({"error": str(e)}) @ns_csv_detective.expect(file_upload) def post(self): args = file_upload.parse_args() if "resource_csv" in args and args["resource_csv"].mimetype != "text/csv": return jsonify({"error": "No uploaded file or the file seems to not be a CSV."}) if ML_PIPELINE is None: analysis_type = "rule" else: analysis_type = "both" uploaded_csv = args["resource_csv"] tmp = NamedTemporaryFile(delete=False) try: tmp.write(uploaded_csv.read()) tmp.close() _, response = analyze_csv(tmp.name, analysis_type=analysis_type, pipeline=ML_PIPELINE, num_rows=500) finally: os.remove(tmp.name) response = reformat_response(response) response = link_reference_datasets(response) return jsonify(response) @ns_csv_detective.route("/type") class CSVDetectiveType(Resource): @api.expect(type_model) def get(self): global TYPE_CSV_INFO try: target_type = request.args.get('target_type') if target_type in TYPE_CSV_INFO: response = TYPE_CSV_INFO[target_type] return jsonify(response) else: logger.info("Type not found in 'database'.") return jsonify({"error": "Type {} not found".format(target_type)}) except Exception as e: return jsonify({"error": str(e)}) @ns_csv_detective.route("/isAlive") class IsAlive(Resource): def get(self): return "True" def reformat_response(response): response = dict(response) new_response = {} if "columns_rb" in response: reformatted_rb = {k: v[0] for k, v in response["columns_rb"].items()} new_response["columns_rb"] = reformatted_rb response.pop("columns_rb") if "columns_ml" in response: reformatted_ml = {k: v[0] for k, v in response["columns_ml"].items()} new_response["columns_ml"] = reformatted_ml response.pop("columns_ml") new_response["metadata"] = dict(response) return new_response def load_result_dict(): global DATASET_CSV_INFO try: with open("./data/interim/2019-10-25-11_59_dgf_friendly.json", "r") as filo: logger.info("Loading JSON file with csv info...") DATASET_CSV_INFO = json.load(filo) except Exception as e: logger.error("Error reading JSON data file: {0}".format(str(e))) exit(1) return DATASET_CSV_INFO def crate_type_index(dataset_csv_info): """ Invert the results dict to have a mapping of types --> dataset (and resource). Something like this: { type1: { datasetID1: { resourceID1 : {csv_detective results}, {...} } } :return: """ results_keynames = ["columns_rb", "columns_ml"] def extract_types_detected(csv_detective_results): detected_types = set([]) for res in results_keynames: if res not in csv_detective_results: continue detected_types.update([f[0] for f in csv_detective_results[res].values()]) return detected_types for dataset_id, resources in dataset_csv_info.items(): for resource_id, csv_detective_result in resources.items(): if not any([f in csv_detective_result for f in results_keynames]): continue for type_detected in extract_types_detected(csv_detective_result): TYPE_CSV_INFO[type_detected][dataset_id][resource_id] = csv_detective_result return TYPE_CSV_INFO if __name__ == '__main__': # load csv_detective info json DATASET_CSV_INFO = load_result_dict() TYPE_CSV_INFO = crate_type_index(DATASET_CSV_INFO) if 'ENVIRONMENT' in os.environ: if os.environ['ENVIRONMENT'] == 'production': app.run(port=80, host='0.0.0.0') if os.environ['ENVIRONMENT'] == 'local': app.run(port=5000, host='0.0.0.0') else: app.run(port=5000, host='0.0.0.0')
[ "pavel.soriano@data.gouv.fr" ]
pavel.soriano@data.gouv.fr
d1ecf38656a166455efc2f5666bd6b3ebcedb4e8
c124e35f3d8dbbdb0423b4b7c30d84baaa04a208
/python/decorators.py
4d8bcc0067b39bc9310c7754fa2809d787c3332c
[]
no_license
Tettipappus/harvard-cs50
29914c929743c27011c5205b6cbaca997f1e6295
c8fc42c874cc5989fd99ccf3ce59fd60d3c809b8
refs/heads/main
2023-06-25T08:54:55.932761
2021-07-19T13:58:49
2021-07-19T13:58:49
376,718,943
0
0
null
2021-06-14T09:20:41
2021-06-14T06:02:01
HTML
UTF-8
Python
false
false
209
py
def announce(f): def wrapper(): print("About to run a funtion...") f() print("Done with the function.") return wrapper @announce def hello(): print("Hello, world!") hello()
[ "sig.d.skjerve@gmail.com" ]
sig.d.skjerve@gmail.com
75747144a3735564364a1962ef1c25bd322f6760
ed2234df3d7e1eecd1b198df66f4fd7263145d18
/MergeSort.py
5e0ed44e686d16540544f188564254a64609b9b5
[]
no_license
Manohar-Gunturu/AlgoPython
bb887f1508c24798c5ae4b1d55fc75d41d674254
5800336886b1cde4deb577f4dc671ce993a0e7f4
refs/heads/master
2021-07-06T10:31:44.984945
2020-07-18T15:45:03
2020-07-18T15:45:27
132,955,184
1
0
null
null
null
null
UTF-8
Python
false
false
1,201
py
arr = [0, 3, 1, 5, 2, 4, 6, 1, 3, 0] # arr = [5, 2, 4, 6, 1, 3] def merge(array, begin, mid, end): # size of first array [begin, mid] n1 = mid - begin + 1 # +1 as index starts at 0 # size of first array [mid+1, end] n2 = end - mid # temp array to hold left and right leftarr = [0] * n1 rightarr = [0] * n2 for i2 in range(0, n1): leftarr[i2] = array[i2 + begin] for i1 in range(0, n2): rightarr[i1] = array[i1 + mid + 1] # leftarr[n1] = -1 # rightarr[n2] = -1 i = 0 j = 0 k = begin while i < n1 and j < n2: if leftarr[i] <= rightarr[j]: array[k] = leftarr[i] i = i + 1 else: array[k] = rightarr[j] j = j + 1 k = k + 1 while i < n1: array[k] = leftarr[i] i = i + 1 k = k + 1 while j < n2: array[k] = rightarr[j] j = j + 1; k = k + 1; def mergesort(array, l, r): if l < r: mid = (l + r) >> 1 mergesort(array, l, mid) mergesort(array, mid + 1, r) merge(array, l, mid, r) mergesort(arr, 0, len(arr) - 1) for i in range(0, len(arr)): print(arr[i])
[ "gunturumanohar2@gmail.com" ]
gunturumanohar2@gmail.com
ea69fdc5fcb293b108694f4a74dd68b98284def9
c744b461256a699d2f039edee0d09344523e86ae
/src/bloch_plot_state.py
a772c653ab022806f6e410d7ee7455d904acab47
[]
no_license
rajeshn76/Blochsphere_Demo
672e348991d8918d858464ac2ee7c2827e38b837
46371dca1b390a639da25d60bce2fce4174c8874
refs/heads/master
2020-04-01T15:54:21.442509
2018-10-26T14:59:45
2018-10-26T14:59:45
153,356,924
0
0
null
null
null
null
UTF-8
Python
false
false
610
py
import numpy as np from qiskit import QuantumCircuit, QuantumRegister, ClassicalRegister, execute, Aer from qiskit.tools.visualization import plot_state, plot_histogram q = QuantumRegister(1) c = ClassicalRegister(1) qc = QuantumCircuit(q, c) qc.s(q).inverse() qc.h(q) qc.rz(np.pi/4, q) qc.rx(-np.pi/4, q) qc.barrier() backend = Aer.get_backend('statevector_simulator') job = execute(qc, backend, shots=1024) result = job.result() data = np.round(result.get_data(qc)['statevector'], 5) print(data) plot_state(data, 'city') plot_state(data, 'bloch') plot_state(data, 'qsphere') plot_state(data, 'paulivec')
[ "rajeshn76@gmail.com" ]
rajeshn76@gmail.com
df79b390cc42bfa1f4ab91ea70a0d961c8fa1131
c4f0e34324d5a44895212c9c24c6ed57e62cf9f9
/MinorProject/settings.py
5d5b9498599232e03e6175b4075b4e1f0e074a09
[]
no_license
Shinde-Hmn/MINDSHUFFLER
7fce564d567200dea01319cebce15f3f9d405bbc
0bf22df015d9e62e9a5622cea40f6915edf1b500
refs/heads/master
2022-04-26T20:10:31.581050
2020-04-30T16:57:09
2020-04-30T16:57:09
260,267,649
0
0
null
null
null
null
UTF-8
Python
false
false
3,900
py
""" Django settings for mysite project. Generated by 'django-admin startproject' using Django 3.0.4. For more information on this file, see https://docs.djangoproject.com/en/3.0/topics/settings/ For the full list of settings and their values, see https://docs.djangoproject.com/en/3.0/ref/settings/ """ import os # Build paths inside the project like this: os.path.join(BASE_DIR, ...) BASE_DIR = os.path.dirname(os.path.dirname(os.path.abspath(__file__))) # Quick-start development settings - unsuitable for production # See https://docs.djangoproject.com/en/3.0/howto/deployment/checklist/ # SECURITY WARNING: keep the secret key used in production secret! SECRET_KEY = '@w!k6l-+(6s^j8-gxsqyo(5^s5cns=(_ubf#xht=q31ts^f_(#' # SECURITY WARNING: don't run with debug turned on in production! DEBUG = True ALLOWED_HOSTS = ['localhost','127.0.0.1' ] # Application definition INSTALLED_APPS = [ 'rest_framework', 'blog.apps.BlogConfig', 'users.apps.UsersConfig', #'django_leaderboard.apps.DjangoLeaderboardConfig', 'crispy_forms', 'django.contrib.admin', 'django.contrib.auth', 'django.contrib.contenttypes', 'django.contrib.sessions', 'django.contrib.messages', 'django.contrib.staticfiles', 'django.contrib.sites', # allauth 'allauth', 'allauth.account', 'allauth.socialaccount', #providers 'allauth.socialaccount.providers.facebook', 'allauth.socialaccount.providers.google', ] MIDDLEWARE = [ 'django.middleware.security.SecurityMiddleware', 'django.contrib.sessions.middleware.SessionMiddleware', 'django.middleware.common.CommonMiddleware', 'django.middleware.csrf.CsrfViewMiddleware', 'django.contrib.auth.middleware.AuthenticationMiddleware', 'django.contrib.messages.middleware.MessageMiddleware', 'django.middleware.clickjacking.XFrameOptionsMiddleware', ] ROOT_URLCONF = 'MinorProject.urls' TEMPLATES = [ { 'BACKEND': 'django.template.backends.django.DjangoTemplates', 'DIRS': [], 'APP_DIRS': True, 'OPTIONS': { 'context_processors': [ 'django.template.context_processors.debug', 'django.template.context_processors.request', 'django.contrib.auth.context_processors.auth', 'django.contrib.messages.context_processors.messages', ], }, }, ] WSGI_APPLICATION = 'MinorProject.wsgi.application' # Database # https://docs.djangoproject.com/en/3.0/ref/settings/#databases DATABASES = { 'default': { 'ENGINE': 'django.db.backends.sqlite3', 'NAME': os.path.join(BASE_DIR, 'db.sqlite3'), } } # Password validation # https://docs.djangoproject.com/en/3.0/ref/settings/#auth-password-validators AUTH_PASSWORD_VALIDATORS = [ { 'NAME': 'django.contrib.auth.password_validation.UserAttributeSimilarityValidator', }, { 'NAME': 'django.contrib.auth.password_validation.MinimumLengthValidator', }, { 'NAME': 'django.contrib.auth.password_validation.CommonPasswordValidator', }, { 'NAME': 'django.contrib.auth.password_validation.NumericPasswordValidator', }, ] # Internationalization # https://docs.djangoproject.com/en/3.0/topics/i18n/ LANGUAGE_CODE = 'en-us' TIME_ZONE = 'UTC' USE_I18N = True USE_L10N = True USE_TZ = True # Static files (CSS, JavaScript, Images) # https://docs.djangoproject.com/en/3.0/howto/static-files/ STATIC_URL = '/static/' MEDIA_ROOT = os.path.join(BASE_DIR,'media') MEDIA_URL = '/media/' CRISPY_TEMPLATE_PACK = 'bootstrap4' LOGIN_REDIRECT_URL = 'blog-home' LOGIN_URL = 'login' SITE_ID =2 EMAIL_BACKEND = 'django.core.mail.backends.smtp.EmailBackend' EMAIL_HOST = 'smtp.gmail.com' EMAIL_PORT = 587 EMAIL_USE_TLS = True EMAIL_HOST_USER = os.environ.get('EMAIL_USER') EMAIL_HOST_PASSWORD = os.environ.get('EMAIL_PASS')
[ "himanshushinde1997812@gmail.com" ]
himanshushinde1997812@gmail.com
6f5a178c8d1ba0fb6bb65c7f38002457ca8ef23a
e3565e1ce607f60745f2a045aae8026661a6b99b
/resources/Onyx-1.0.511/py/onyx/util/rocutils.py
48dd8b474c58afe853ee45475fc9479842d375ed
[ "Apache-2.0" ]
permissive
eternity668/speechAD
4c08d953b2ed06b3357b1c39d8709dd088a2471c
f270a1be86372b7044615e4fd82032029e123bc1
refs/heads/master
2021-01-12T22:10:33.358500
2014-02-03T16:03:28
2014-02-03T16:03:28
null
0
0
null
null
null
null
UTF-8
Python
false
false
6,676
py
########################################################################### # # File: rocutils.py (directory: ./py/onyx/util) # Date: Mon 10 Mar 2008 18:34 # Author: Ken Basye # Description: Utility code for generating ROC and DET curves # # This file is part of Onyx http://onyxtools.sourceforge.net # # Copyright 2008, 2009 The Johns Hopkins University # # Licensed under the Apache License, Version 2.0 (the "License"). # You may not use this file except in compliance with the License. # You may obtain a copy of the License at # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or # implied. See the License for the specific language governing # permissions and limitations under the License. # ########################################################################### """ Utilities for generating ROC and DET curves """ import StringIO def _uniquify_preserving_first(iterable, eq_pred): item = iterable.next() while 1: try: next_item = iterable.next() except: yield item break if not eq_pred(item, next_item): yield item item = next_item def _uniquify_preserving_last(iterable, eq_pred): item = iterable.next() while 1: try: next_item = iterable.next() except: yield item break if not eq_pred(item, next_item): yield item item = next_item else: item = next_item def make_ROC_data(reference, ratios): """ reference is a list of 0/1 values which are the correct classifications values is a parallel list of numeric values, with higher values intending to map toward classifications of 1. Returns data for a ROC curve in the form of a list of triples, where each triple contains an interesting threshold value, the fraction of correct identifications (true positives) as a percent, and the fraction of false positives, at that threshold. The triples are ordered by threshold from lowest (fewest false positives) to highest (most true positives) Note that a typical ROC curve would plot false_pos on the X axis and true_pos on the Y axis using a linear scale. >>> ref = [0,0,0,0,0,1,1,1,1,1] >>> values = [2, 3, 4, 9, 4, 5, 6, 9, 9, 3] >>> res = make_ROC_data(ref, values) >>> res [(0.0, 0.0, 9), (20.0, 80.0, 4), (80.0, 100.0, 2)] """ det_data = make_DET_data(reference, ratios) roc_data = [(fp, 100-miss, t) for (fp, miss, t) in det_data] return roc_data def make_DET_data(reference, ratios): """ reference is a list of 0/1 values which are the correct classifications values is a parallel list of numeric values, with higher values intending to map toward classifications of 1. Returns data for a DET curve in the form of a list of triples, where each triple contains the fraction of false positives as a percent, the fraction of false negatives, and the threshold value that generated those rates. The triples are ordered by threshold from lowest (fewest false positives) to highest (fewest misses) Note that a typical DET curve would plot false_pos on the X axis and false_neg on the Y axis, oftentimes with a normal deviate scale. >>> ref = [0,0,0,0,0,1,1,1,1,1] >>> values = [2, 3, 4, 9, 4, 5, 6, 9, 9, 3] >>> res = make_DET_data(ref, values) >>> res [(0.0, 100.0, 9), (20.0, 19.999999999999996, 4), (80.0, 0.0, 2)] """ assert( len(reference) == len(ratios) ) num_pos = reference.count(1) num_neg = reference.count(0) assert( num_pos + num_neg == len(reference)) full_result = [] # Find the list of interesting threshholds, which is any value in # the list of ratios # Seems like there should be an easier way to uniquify a list all_threshes = set(ratios) all_threshes = list(all_threshes) all_threshes.sort() def count_values_over_thresh(value, ref, ratios, t): result = 0 for (i, r) in enumerate(ratios): if ref[i] == value and r > t: result += 1 return result # Now find precision and recall at each threshold for thresh in all_threshes: num_neg_accepted = count_values_over_thresh(0, reference, ratios, thresh) num_pos_accepted = count_values_over_thresh(1, reference, ratios, thresh) full_result.append((100 * float(num_neg_accepted) / num_neg, # false positives 100 * (1 - float(num_pos_accepted) / num_pos), # misses thresh)) def eq0(x,y): return x[0] == y[0] def eq1(x,y): return x[1] == y[1] iter1 = _uniquify_preserving_first(iter(full_result), eq0) ret = list(_uniquify_preserving_last(iter1, eq1)) ret.reverse() return ret def write_data_as_csv(data, stream, header_type = "DET"): """ Write either ROC or DET data as comma-separated text, suitable for import into a spreadsheet or other tool. Writes DET header fields be default, use header_type of "ROC" or None for ROC headers or no headers, respectively. >>> ref = [0,0,0,0,0,1,1,1,1,1] >>> values = [2, 3, 4, 9, 4, 5, 6, 9, 9, 3] >>> res = make_DET_data(ref, values) >>> s = StringIO.StringIO() >>> write_data_as_csv(res, s) >>> out = s.getvalue() >>> print out False Alarm Rate, Miss Rate, Threshold 0.0, 100.0, 9 20.0, 20.0, 4 80.0, 0.0, 2 <BLANKLINE> >>> s.seek(0) >>> res = make_ROC_data(ref, values) >>> write_data_as_csv(res, s, header_type="ROC") >>> out = s.getvalue() >>> print out False Pos Rate, True Pos Rate, Threshold 0.0, 0.0, 9 20.0, 80.0, 4 80.0, 100.0, 2 <BLANKLINE> >>> s.close() """ if header_type == "DET": stream.write("False Alarm Rate, Miss Rate, Threshold") elif header_type == "ROC": stream.write("False Pos Rate, True Pos Rate, Threshold") [stream.write("\n%s, %s, %s" % triple) for triple in data] stream.write("\n") def _test0(): ref = [0,0,0,0,0,1,1,1,1,1] values = [2, 3, 4, 9, 4, 5, 6, 9, 9, 3] res = make_DET_data(ref, values) s = open("foo_csv.txt", "w") write_data_as_csv(res, s) s.close() if __name__ == '__main__': from onyx import onyx_mainstartup onyx_mainstartup() # _test0()
[ "nassos@n12mavra.cs.ntua.gr" ]
nassos@n12mavra.cs.ntua.gr
e81f2da8791b183158fe5fb31295415705244e0d
15fbec197e9e8508bce3189b0d6d6f0b2fc317ee
/core/migrations/0001_initial.py
74785cf38e45b6aa551bf1a24a0c5aecd7f27a42
[]
no_license
douglasbastos/minhatabela
d698071b23ecd2c880af790bd1d85be3a04c7a46
df8499e46b00d45b9cf7ffaf236b1635fa6e2b85
refs/heads/master
2021-01-09T20:29:02.844595
2016-07-04T04:30:45
2016-07-04T04:30:45
62,531,533
0
0
null
null
null
null
UTF-8
Python
false
false
4,487
py
# -*- coding: utf-8 -*- # Generated by Django 1.9.7 on 2016-07-04 04:06 from __future__ import unicode_literals from django.conf import settings from django.db import migrations, models import django.db.models.deletion class Migration(migrations.Migration): initial = True dependencies = [ migrations.swappable_dependency(settings.AUTH_USER_MODEL), ] operations = [ migrations.CreateModel( name='Competition', fields=[ ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), ('name', models.CharField(max_length=255, verbose_name=b'Nome')), ('logo', models.ImageField(upload_to=b'', verbose_name=b'Logo')), ('creator_user', models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to=settings.AUTH_USER_MODEL)), ], options={ 'verbose_name': 'Competi\xe7\xe3o', 'verbose_name_plural': 'Competi\xe7\xf5es', }, ), migrations.CreateModel( name='CompetitionHasTeam', fields=[ ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), ('competition', models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='core.Competition')), ], options={ 'verbose_name': 'Time/competi\xe7\xe3o', 'verbose_name_plural': 'Times/competi\xe7\xf5es', }, ), migrations.CreateModel( name='Game', fields=[ ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), ('result_home', models.PositiveSmallIntegerField(blank=True, null=True, verbose_name=b'Resultado Mandante')), ('result_visitor', models.PositiveSmallIntegerField(blank=True, null=True, verbose_name=b'Resultado Visitante')), ('local', models.CharField(blank=True, max_length=255, null=True, verbose_name=b'Local')), ('date', models.DateTimeField(blank=True, null=True, verbose_name=b'Data e Hora')), ], options={ 'verbose_name': 'Jogo', 'verbose_name_plural': 'Jogos', }, ), migrations.CreateModel( name='Rounds', fields=[ ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), ('number', models.PositiveSmallIntegerField()), ('competition', models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='core.Competition')), ], options={ 'verbose_name': 'Rodada', 'verbose_name_plural': 'Rodadas', }, ), migrations.CreateModel( name='Team', fields=[ ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), ('name', models.CharField(max_length=255, verbose_name=b'Nome')), ('flag', models.ImageField(upload_to=b'', verbose_name=b'Escudo')), ], options={ 'verbose_name': 'Time', 'verbose_name_plural': 'Times', }, ), migrations.AddField( model_name='game', name='rounds', field=models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='core.Rounds'), ), migrations.AddField( model_name='game', name='team_home', field=models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, related_name='Mandante', to='core.Team'), ), migrations.AddField( model_name='game', name='team_visitor', field=models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, related_name='Visitante', to='core.Team'), ), migrations.AddField( model_name='competitionhasteam', name='team', field=models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='core.Team'), ), migrations.AddField( model_name='competition', name='teams', field=models.ManyToManyField(through='core.CompetitionHasTeam', to='core.Team'), ), ]
[ "douglashsb@gmail.com" ]
douglashsb@gmail.com
d69078e714be2c5e36cd35abb4d251b6fb236980
16582b82a7cb119f218b6d4db025c338263852fc
/download.py
6246ed7d49a2d7362dc394241198fcef86f9194f
[]
no_license
a1004123217/music
849921ffbeea264e258ea53d9b4c773b21a68a9d
9ef35653c0872c70463c9802ea6fe318f726cfcf
refs/heads/master
2022-12-12T11:50:49.233881
2020-09-14T10:42:29
2020-09-14T10:42:29
292,523,572
0
0
null
null
null
null
UTF-8
Python
false
false
1,735
py
from urllib.request import urlretrieve import os # 解决 import ssl ssl._create_default_https_context = ssl._create_unverified_context ''' 通过txt网址文件,现在图片到本地 ''' def download(): categories = ['music'] for category in categories: # 新建存储ladder文件夹存储图片 os.makedirs('/Users/wzy/PycharmProjects/pythonProject/music/data/%s' % category, exist_ok=True) # 读取txt文件 with open('/Users/wzy/PycharmProjects/pythonProject/music/beatanddownbeat/uploadfile.txt', 'r') as file: urls = file.readlines() print(urls) #计算链接地址条数 n_urls = len(urls) # 遍历链接地址下载图片 for i, url in enumerate(urls): #print(url.strip()) #print(url.strip().split('/')[-1]) tmp = url.strip().split('/')[-1] tmp2 = tmp.strip().split('_') str1 = '' for i in range(len(tmp2)): if i == 2: str1 = tmp2[i]+tmp2[3][:] #print(t) print("str1",str1) #print("tmp2",tmp[2]) # print(tmp[2]) print(str1) try: # 请求下载图片,并截取最后链接第一最后一节字段命名图片 urlretrieve(url.strip(), '/Users/wzy/PycharmProjects/pythonProject/music/data/json/%s' % str1) print('%s %i/%i' % (category, i, n_urls)) except: print('%s %i/%i' % (category, i, n_urls), 'no image') if __name__ == '__main__': download()
[ "wzy@wangzhenyudeMacBook-Pro.local" ]
wzy@wangzhenyudeMacBook-Pro.local
241ac7b70c2142fba7ff196677ed61d5e0910d2f
587290dbd33c5fb60a154eacd2155e681a3f9ecf
/js/gatsby/index.cgi
db257841537410c023fb93f34bb0b3e0a10dcd00
[]
no_license
jaredly/prog
b6408db52c16e9d3c322933f0624c23663d33ce0
e8fe82ccd1abe42371adbb3f317576facac546ca
refs/heads/master
2021-01-20T09:12:48.931999
2013-08-29T04:24:36
2013-08-29T04:24:36
null
0
0
null
null
null
null
UTF-8
Python
false
false
2,523
cgi
#!/usr/bin/python print 'Content-type:text/html\n' import cgi,cgitb,os,sys,re cgitb.enable() def load_chapter(chap): return open('chapter%s.txt'%chap).read() def load_chapters(): d=[] for i in range(1,10): d.append([i,load_chapter(i)]) return d def print_entry(at,chap,pageat,item): if at==1:nd="<sup>st</sup>" elif at==2:nd="<sup>nd</sup>" elif at==3:nd="<sup>rd</sup>" else:nd="<sup>th</sup>" return "<br><br><br><b>%s%s</b> paragraph in chapter <b>%s</b> (around page %s)<br><br>\n"%(at, nd, chap, pageat)+item form = cgi.FieldStorage() print """ <html><head><title>Great Gatsby Search</title></head><body> <style> span { font-weight: bold; font-size: 1.1em; color: black; background-color: #ccc; } h2 { text-align:center; } div.searchform { background-color:#BBFFAA; border:2px solid green; padding:15px; position:absolute; right:0px; top:0px; } form { margin: 0px; } </style> <h1>Search the Great Gatsby</h1> <div class="searchform"> <form method="GET"> Search For: <input name="s" value="%s"> <input type="checkbox" name="whole" value="1"> Whole word <input type="submit" value="Search"> </form> </div> <br>"""%(form.has_key("s") and form["s"].value or "") pages = [1, 23, 39, 61, 81, 97, 113, 147, 163, 180 ] ## None ## [3, 16, 26, 39, 52, 62, 93, 103] retr = "" num = 0 if form.has_key('s'): term = form['s'].value.strip() iterm=term if form.has_key('whole'): term='(?<=\W)'+term+'(?=\W)' for chapter,text in load_chapters(): for i,body in enumerate(text.split('\n')): all = re.search(term,body,re.I|re.S) if pages: pchap = pages[chapter-1] #print text.find(body),len(text) pat = int(round(float(pages[chapter]-pchap)* (text.find(body)/float(len(text)))+pchap)) else: pat = "" rgx = re.compile(term,re.I) bdy = rgx.sub(lambda x:'<span>'+x.group()+'</span>', body)+'<br><br>' ## bdy = re.sub(term, lambda x:'<span>'+x.group()+'</span>', body)+'<br><br>' if all: ## print (text.find(body)/float(len(text))),float(pages[chapter]-pchap) ## print float(pages[chapter]-pchap)*(text.find(body)/float(len(text)))+pchap retr += print_entry(i,chapter,pat,bdy) num+=1 print "<h3>Found %d results for %s</h3>"%(num,iterm) print retr
[ "jared@jaredforsyth.com" ]
jared@jaredforsyth.com
996aac45cc3fff5b7b5a9eb0567f864fdb8f7981
8c4af05e0257661195c95b0b9e0873eeb6391dab
/packages/python-packages/apiview-gpt/src/_models.py
85d435a1aa1dd9ca713a5cccc403df55f16ebc0d
[ "MIT", "LicenseRef-scancode-generic-cla" ]
permissive
Azure/azure-sdk-tools
6d171054800807fcbe7b8b878c5903a202d31faa
2dce521dedc3f5169007d4c481ae8ec077be4450
refs/heads/main
2023-09-01T00:00:32.662190
2023-08-31T22:21:44
2023-08-31T22:21:44
170,592,186
113
174
MIT
2023-09-14T21:53:41
2019-02-13T22:55:46
C#
UTF-8
Python
false
false
744
py
from pydantic import BaseModel, Field from typing import List, Optional class Violation(BaseModel): rule_ids: List[str] = Field(description="unique rule ID or IDs that were violated.") line_no: Optional[int] = Field(description="the line number of the violation.") bad_code: str = Field(description="the original code that was bad, cited verbatim.") suggestion: str = Field(description="the suggested fix for the bad code.") comment: str = Field(description="a comment about the violation.") class GuidelinesResult(BaseModel): status: str = Field(description="Succeeded if the request has no violations. Error if there are violations.") violations: List[Violation] = Field(description="list of violations if any")
[ "noreply@github.com" ]
noreply@github.com
def4e0c9060cbb6946a984f723129a8064a91715
16679038c7a0b75097ffdd2d5b6be28ae8dae68f
/test/utilities/test_catch_exceptions.py
74b112680efb4b09be050e92f0ab2d6cc4bcdc2b
[ "MIT" ]
permissive
elifesciences/profiles
d98e5c2391630f9877e0585e07143d7904f1e777
9cd2e523f9dfa864891511e6525381f191951b24
refs/heads/develop
2023-08-31T03:09:08.723797
2023-08-25T06:54:55
2023-08-25T06:54:55
94,993,646
2
0
MIT
2023-06-21T01:15:37
2017-06-21T10:43:52
Python
UTF-8
Python
false
false
959
py
import logging from logging import Handler, Logger, Manager from logging.handlers import BufferingHandler from pytest import fixture from profiles.utilities import catch_exceptions @fixture def logger(handler: Handler) -> Logger: logger = Logger('logger', logging.DEBUG) logger.addHandler(handler) logger.manager = Manager('root') return logger @fixture def handler() -> Handler: return BufferingHandler(100) def test_it_catches_and_logs_exceptions(logger: Logger, handler: BufferingHandler): @catch_exceptions(logger) def my_function(): raise Exception('My exception') result = my_function() assert result is None assert len(handler.buffer) == 1 def test_it_does_nothing_when_no_exception(logger: Logger, handler: BufferingHandler): @catch_exceptions(logger) def my_function(): return True result = my_function() assert result is True assert len(handler.buffer) == 0
[ "noreply@github.com" ]
noreply@github.com
6278c357cec08deee02c0d6bb03990e43931a5d9
1f5ff597e967a63de3e005b876b644e25cd08dfc
/stage3.py
873501d6cef32b2731fb6e2a05ffb818309dec3e
[]
no_license
codefloww/BBA
713f7e1a96c8f291d917d5a44f8ae7f5e2d5e09e
2b0c4c1a408dd41960f736a2f20336db71995b82
refs/heads/main
2023-08-24T23:08:16.814240
2021-10-31T13:35:45
2021-10-31T13:35:45
422,621,197
0
0
null
2021-10-31T13:35:46
2021-10-29T15:14:34
Python
UTF-8
Python
false
false
1,753
py
import main import os import pygame import physics pygame.mixer.init() def stage(): stage3 = main.Window(1920, 1080, "Enter.mp3", "background3.jpg") stage3.play_audio("start") running = True objects = [] moveable = [] # Each moveable object, basically mobs and player player = main.Wolf(100, 100, 100, 10, "Wolf") moveable.append(player) objects.append(player) soil1 = main.Road(100, 500, "ground", 900, 50) soil2 = main.Road(300, 400, "ground", 100, 100) objects.append(soil1) objects.append(soil2) objects_unmoveable = objects.copy() objects_unmoveable.remove(player) clock = pygame.time.Clock() while running: clock.tick(120) for event in pygame.event.get(): if event.type == pygame.QUIT: running = False if event.type == pygame.KEYDOWN: if event.key == pygame.K_s: stage3.play_audio("stop") if event.key == pygame.K_p: stage3.play_audio("play") for entity in moveable: stage3.update_screen(objects) # physics.gravity(entity, objects) stand = 0 wall_collision = physics.stutter(entity, objects) while stand == 0 and entity.alive: stand = physics.gravity(entity, objects, stage3) entity.direction = None stage3.update_screen(objects) keys_pressed = pygame.key.get_pressed() player.movement_handle(keys_pressed, objects, stand, stage3, wall_collision) stage3.update_screen(objects) if not player.alive: running = False return player.alive if __name__ == "__main__": stage()
[ "kryvpaul@gmail.com" ]
kryvpaul@gmail.com
f8fdac6b1a2846a9f74f7db1f038fed9022ab0a4
5dd03f9bd8886f02315c254eb2569e4b6d368849
/3rdparty/python/GitPython-0.3.1-py2.6.egg/git/__init__.py
500d053f7729d7172f300870e30b00ae7a523f09
[ "LicenseRef-scancode-warranty-disclaimer", "Apache-2.0" ]
permissive
adamsxu/commons
9e1bff8be131f5b802d3aadc9916d5f3a760166c
9fd5a4ab142295692994b012a2a2ef3935d35c0b
refs/heads/master
2021-01-17T23:13:51.478337
2012-03-11T17:30:24
2012-03-11T17:30:24
null
0
0
null
null
null
null
UTF-8
Python
false
false
1,171
py
# __init__.py # Copyright (C) 2008, 2009 Michael Trier (mtrier@gmail.com) and contributors # # This module is part of GitPython and is released under # the BSD License: http://www.opensource.org/licenses/bsd-license.php import os import sys import inspect __version__ = '0.3.1' #{ Initialization def _init_externals(): """Initialize external projects by putting them into the path""" sys.path.append(os.path.join(os.path.dirname(__file__), 'ext', 'gitdb')) try: import gitdb except ImportError: raise ImportError("'gitdb' could not be found in your PYTHONPATH") #END verify import #} END initialization ################# _init_externals() ################# #{ Imports from git.config import GitConfigParser from git.objects import * from git.refs import * from git.diff import * from git.exc import * from git.db import * from git.cmd import Git from git.repo import Repo from git.remote import * from git.index import * from git.util import ( LockFile, BlockingLockFile, Stats, Actor ) #} END imports __all__ = [ name for name, obj in locals().items() if not (name.startswith('_') or inspect.ismodule(obj)) ]
[ "jsirois@twitter.com" ]
jsirois@twitter.com
05e6de66f7d42a5e5cc0ec5d44e586790873c9b5
46cbc45de7cd3ad8c565603918dcec8f52eeab82
/11.12 Selenium单元测试.py
fdbf1a76c6ecbf251ecacdab7aa61d0b2dc057ea
[]
no_license
Patrickctyyx/WebScrapers
8bb946a2b60e21426e1b3048bb4752d71c79b401
02ebd9bef8242223b1e73ade0a4290235bebf668
refs/heads/master
2020-05-29T08:52:18.017387
2016-11-16T14:05:52
2016-11-16T14:05:52
69,369,962
1
0
null
null
null
null
UTF-8
Python
false
false
340
py
from selenium import webdriver # 如果通过了就没提示 # 没通过就报错 driver = webdriver.PhantomJS(executable_path='/home/patrick/phantomjs-2.1.1-linux-x86_64/bin/phantomjs') driver.get('http://en.wikipedia.org/wiki/Kevin_Durant') # 这不就是系统自带的断言么... assert 'Kevin Durant' in driver.title driver.close()
[ "873948000@qq.com" ]
873948000@qq.com
39ecdda194fb35100d42e0cdad25360099fde4b9
665089f28faa90d31a3ba4d62246c8323e4de9f1
/communityid/algo.py
366748f6056a748a8ccb618adde26f5c7098b04f
[ "BSD-2-Clause" ]
permissive
corelight/pycommunityid
64b7385609bdd5b329d95b99a1389ddb5257033a
64a6fa18ed2dc9ea1f6fed04e670660a9854a632
refs/heads/master
2022-04-29T07:09:53.287643
2022-03-05T00:45:39
2022-03-05T00:45:53
196,440,073
22
5
BSD-3-Clause
2022-03-05T00:43:57
2019-07-11T17:40:14
Python
UTF-8
Python
false
false
17,482
py
""" This module implements Community ID network flow hashing. """ import abc import base64 import collections import hashlib import logging import socket import string import struct from communityid import error from communityid import compat from communityid import icmp from communityid import icmp6 from . import LOG # Proper enums here would be nice, but this aims to support Python # 2.7+ and while there are ways to get "proper" enums pre-3.0, it just # seems overkill. --cpk PROTO_ICMP = 1 PROTO_TCP = 6 PROTO_UDP = 17 PROTO_ICMP6 = 58 PROTO_SCTP = 132 # The set of protocols we explicitly support as port-enabled: # Community ID computations on those protocols should be based on a # five-tuple. PORT_PROTOS = set([PROTO_ICMP, PROTO_TCP, PROTO_UDP, PROTO_ICMP6, PROTO_SCTP]) class FlowTuple: """ Tuples of network flow endpoints, used as input for the Community ID computation. These tuple objects are flexible regarding the input data types -- for the addresses you can use NBO byte-strings or ASCII, for example. They usually are 5-tuples of address & port pairs, plus IP protocol number, but port-less tuples are supported for less common IP payloads. """ Data = collections.namedtuple( 'Data', ['proto', 'saddr', 'daddr', 'sport', 'dport']) def __init__(self, proto, saddr, daddr, sport=None, dport=None, is_one_way=False): """Tuple initializer. The proto argument is a non-negative integer and represents an IP protocol number, e.g. 6 for TCP. You can use the PROTO_* constants if convenient, and communityid.get_proto() to help convert to integer. The saddr and daddr arguments are source & destination IP addresses, either IPv4 or IPv6. Multiple data types are supported, including bytes (as str in older Pythons, or the explicit bytes type), IPv4Address, IPv6Address, and string representations. The sport and dport arguments are numeric port numbers, either provided as ints or in packed 16-bit network byte order, of type "bytes". When the protocol number is one of PORT_PROTOS (TCP, UDP, etc), they are required. For other IP protocols they are optional. The optional Boolean is_one_way argument indicates whether the tuple captures a bidirectional flow (the default) or not. Setting this to true means that the computation will consider the tuple directional and not try to pair up with flipped-endpoint tuples. Normally you don't need to pass this. This can raise FlowTupleErrors when the input is inconsistent. """ self.proto = proto self.saddr, self.daddr = saddr, daddr self.sport, self.dport = sport, dport if proto is None or type(proto) != int: raise error.FlowTupleError('Need numeric protocol number') if saddr is None or daddr is None: raise error.FlowTupleError('Need source and destination address') if not self.is_ipaddr(saddr): raise error.FlowTupleError('Unsupported format for source IP address "%s"' % saddr) if not self.is_ipaddr(daddr): raise error.FlowTupleError('Unsupported format for destination IP address "%s"' % daddr) if ((sport is None and dport is not None) or (dport is None and sport is not None)): raise error.FlowTupleError('Need either both or no port numbers') if sport is not None and not self.is_port(sport): raise error.FlowTupleError('Source port "%s" invalid' % sport) if dport is not None and not self.is_port(dport): raise error.FlowTupleError('Destination port "%s" invalid' % dport) if proto in PORT_PROTOS and sport is None: raise error.FlowTupleError('Need port numbers for port-enabled protocol %s' % proto) # Our ICMP handling directly mirrors that of Zeek, since it # tries hardest to map ICMP into traditional 5-tuples. For # this, it evaluates the message type & code to identify # whether the notion of two-way communication applies. If not, # tuple-flipping isn't an option either. The following flag # stores this result, assuming by default we're bidirectional. self.is_one_way = is_one_way # The rest of the constructor requires ports. if sport is None or dport is None: return # If we're explicitly told this is a one-way flow-tuple, we # don't need to consider directionality further. And, testing # directionality only makes sense when the ports are integers, # not lower-level NBO representations. Throughout we need to # keep track of the types of the ports, since the ICMP logic # works only with regular ints. if self.is_one_way is False: if self.proto == PROTO_ICMP: sport, dport, self.is_one_way = icmp.get_port_equivalents( self._port_to_int(sport), self._port_to_int(dport)) self.sport = self._port_to_same(sport, self.sport) self.dport = self._port_to_same(dport, self.dport) elif self.proto == PROTO_ICMP6: sport, dport, self.is_one_way = icmp6.get_port_equivalents( self._port_to_int(sport), self._port_to_int(dport)) self.sport = self._port_to_same(sport, self.sport) self.dport = self._port_to_same(dport, self.dport) def __repr__(self): data = self.get_data() ordered = 'ordered' if self.is_ordered() else 'flipped' if data.sport is None or data.dport is None: return '%s -> %s, proto %s, %s' % ( data.saddr, data.daddr, data.proto, ordered) return '%s %s -> %s %s, proto %s, %s' % ( data.saddr, data.sport, data.daddr, data.dport, data.proto, ordered) def get_data(self): """ Returns a FlowTuple.Data namedtuple with this flow tuple's data. The protocol is an integer number (e.g. 6 for TCP), saddr and daddr are ASCII-rendered/unpacked, and the ports are integers or None, if absent. """ # Absent good types, make it best-effort to get these # renderable. If all characters are printable, we assume this # in not NBO. saddr, daddr, sport, dport = self.saddr, self.daddr, self.sport, self.dport if compat.have_real_bytes_type() and isinstance(saddr, bytes): saddr = self._addr_to_ascii(saddr) elif compat.is_ipaddress_type(saddr): saddr = saddr.exploded elif not all(c in string.printable for c in saddr): saddr = self._addr_to_ascii(saddr) if compat.have_real_bytes_type() and isinstance(daddr, bytes): daddr = self._addr_to_ascii(daddr) elif compat.is_ipaddress_type(daddr): daddr = daddr.exploded elif not all(c in string.printable for c in daddr): daddr = self._addr_to_ascii(daddr) if sport is not None and not isinstance(sport, int): sport = struct.unpack('!H', sport)[0] if dport is not None and not isinstance(dport, int): dport = struct.unpack('!H', dport)[0] return FlowTuple.Data(self.proto, saddr, daddr, sport, dport) def is_ordered(self): """ Predicate, returns True when this flow tuple is ordered. A flow tuple is ordered when any of the following are true: - It's marked as a one-way flow. - Its source IP address is smaller than its dest IP address, both in network byte order (NBO). - The IP addresses are equal and the source port is smaller than the dest port, in NBO. """ nbo = self.in_nbo() return (nbo.is_one_way or nbo.saddr < nbo.daddr or (nbo.saddr == nbo.daddr and nbo.sport is not None and nbo.dport is not None and nbo.sport < nbo.dport)) def has_ports(self): """ Predicate, returns True when this tuple features port numbers. """ return self.sport is not None and self.dport is not None def in_order(self): """ Returns a copy of this tuple that is ordered canonically. Ie, regardless of the src/dest IP addresses and ports, the returned tuple will be be the same: the source side will contain the smaller endpoint (see FlowTuple.is_ordered() for details). """ if self.is_ordered(): return FlowTuple(self.proto, self.saddr, self.daddr, self.sport, self.dport, self.is_one_way) return FlowTuple(self.proto, self.daddr, self.saddr, self.dport, self.sport, self.is_one_way) def in_nbo(self): """ Returns a copy of this tuple where the addresses and port are rendered into NBO byte strings. """ saddr = self._addr_to_nbo(self.saddr) daddr = self._addr_to_nbo(self.daddr) if isinstance(self.sport, int): sport = struct.pack('!H', self.sport) else: sport = self.sport if isinstance(self.dport, int): dport = struct.pack('!H', self.dport) else: dport = self.dport return FlowTuple(self.proto, saddr, daddr, sport, dport, self.is_one_way) @staticmethod def is_ipaddr(val): return (FlowTuple.addr_is_text(val) or FlowTuple.addr_is_packed(val) or FlowTuple.addr_is_ipaddress_type(val)) @staticmethod def addr_is_text(addr): for family in (socket.AF_INET, socket.AF_INET6): try: socket.inet_pton(family, addr) return True except (socket.error, TypeError): pass return False @staticmethod def addr_is_packed(addr): for family in (socket.AF_INET, socket.AF_INET6): try: socket.inet_ntop(family, addr) return True except (socket.error, ValueError, TypeError): pass return False @staticmethod def addr_is_ipaddress_type(addr): return compat.is_ipaddress_type(addr) @staticmethod def is_port(val): if isinstance(val, bytes): try: port = struct.unpack('!H', val)[0] return 0 <= port <= 65535 except (struct.error, IndexError, TypeError): pass if isinstance(val, int): return 0 <= val <= 65535 return False @staticmethod def _port_to_int(port): """Convert a port number to regular integer.""" if isinstance(port, int): return port # Assume it's two bytes in NBO: return struct.unpack('!H', port)[0] @staticmethod def _port_to_nbo(port): """Convert a port number to 2-byte NBO.""" if isinstance(port, int): return struct.pack('!H', port) # Assume it's two bytes in NBO return port @staticmethod def _port_to_same(port, sample): """Convert a port number to the same type as that of another instance.""" if isinstance(sample, int): return FlowTuple._port_to_int(port) return FlowTuple._port_to_nbo(port) @staticmethod def _addr_to_ascii(addr): if compat.is_ipaddress_type(addr): return addr.exploded for family in (socket.AF_INET, socket.AF_INET6): try: return socket.inet_ntop(family, addr) except (socket.error, ValueError, TypeError): pass return addr @staticmethod def _addr_to_nbo(addr): if compat.is_ipaddress_type(addr): return addr.packed for family in (socket.AF_INET, socket.AF_INET6): try: return socket.inet_pton(family, addr) except (socket.error, TypeError): pass return addr # Convenience wrappers for making protocol-specific tuple instances. @classmethod def make_tcp(cls, saddr, daddr, sport, dport): return cls(PROTO_TCP, saddr, daddr, int(sport), int(dport)) @classmethod def make_udp(cls, saddr, daddr, sport, dport): return cls(PROTO_UDP, saddr, daddr, int(sport), int(dport)) @classmethod def make_sctp(cls, saddr, daddr, sport, dport): return cls(PROTO_SCTP, saddr, daddr, int(sport), int(dport)) @classmethod def make_icmp(cls, saddr, daddr, mtype, mcode): return cls(PROTO_ICMP, saddr, daddr, int(mtype), int(mcode)) @classmethod def make_icmp6(cls, saddr, daddr, mtype, mcode): return cls(PROTO_ICMP6, saddr, daddr, int(mtype), int(mcode)) @classmethod def make_ip(cls, saddr, daddr, proto): return cls(proto, saddr, daddr) class CommunityIDBase: __metaclass__ = abc.ABCMeta @abc.abstractmethod def get_error(self): """ Error handler. After something fails during the ID computation, this method should return an explanation why. """ return None @abc.abstractmethod def calc(self, tpl): """ Entrypoint to the ID computation, given a FlowTuple instance. Returns a string containing the Community ID value, or None on error. """ return None @abc.abstractmethod def hash(self, tpl): """ The tuple-hashing part of the computation. Returns hashlib algorithm instance ready for digesting, or None on error. """ return None @abc.abstractmethod def render(self, hashstate): """ The rendering part of the computation. Receives a hashlib algorithm instance and returns a string containing the community ID value according to this instance's configuration, or None on error. """ return None class CommunityID(CommunityIDBase): """ An algorithm object that computes Community IDs on FlowTuple instances. """ def __init__(self, seed=0, use_base64=True): self._version = 1 self._seed = seed self._use_base64 = use_base64 self._err = None def __repr__(self): return 'CommunityID(v=%s,seed=%s,base64=%s)' \ % (self._version, self._seed, self._use_base64) def get_error(self): """ Returns an error string when problems came up during the computation. This is only valid directly after calc() returned None, i.e., something went wrong during the calculation. """ return self._err def calc(self, tpl): """ The biggie: given a FlowTuple instance, returns a string containing the Community ID. In case of problems, returns None. In that case consider get_error() to learn more about what happened. """ LOG.info('CommunityID for %s:' % tpl) tpl = tpl.in_nbo().in_order() return self.render(self.hash(tpl)) def hash(self, tpl): hashstate = hashlib.sha1() def hash_update(data, context): # Handy for troubleshooting: show the hashed byte sequence # when verbosity level is at least INFO (-vv): if LOG.isEnabledFor(logging.INFO): # On Python 2.7 the data comes as a string, later it's bytes. # The difference matters for rendering the byte hex values. if isinstance(data, str): hexbytes = ':'.join('%02x' % ord(b) for b in data) else: hexbytes = ':'.join('%02x' % b for b in data) LOG.info('| %-7s %s' % (context, hexbytes)) hashstate.update(data) return len(data) try: dlen = hash_update(struct.pack('!H', self._seed), 'seed') # 2-byte seed dlen += hash_update(tpl.saddr, 'ipaddr') # 4 bytes (v4 addr) or 16 bytes (v6 addr) dlen += hash_update(tpl.daddr, 'ipaddr') # 4 bytes (v4 addr) or 16 bytes (v6 addr) dlen += hash_update(struct.pack('B', tpl.proto), 'proto') # 1 byte for transport proto dlen += hash_update(struct.pack('B', 0), 'padding') # 1 byte padding if tpl.has_ports(): dlen += hash_update(tpl.sport, 'port') # 2 bytes dlen += hash_update(tpl.dport, 'port') # 2 bytes except struct.error as err: self._err = 'Could not pack flow tuple: %s' % err LOG.warning(self._err) return None # The data structure we hash should always align on 32-bit # boundaries. if dlen % 4 != 0: self._err = 'Unexpected hash input length: %s' % dlen LOG.warning(self._err) return None return hashstate def render(self, hashstate): if hashstate is None: return None # Unless the user disabled the feature, base64-encode the # (binary) hash digest. Otherwise, print the ASCII digest. if self._use_base64: return str(self._version) + ':' + base64.b64encode(hashstate.digest()).decode('ascii') return str(self._version) + ':' + hashstate.hexdigest()
[ "christian@corelight.com" ]
christian@corelight.com
ad74fb1924b3b8c48c2d09b0bbc03cd105920500
af97a81a69d100e684b8e9f02bc1b117069b9895
/Testcases_Copilot/CWE-730/codeql_2.py
1c59968d10ace977410229a13e67c6c9188734a4
[]
no_license
s2e-lab/SecurityEval
8de3e2f78fffa7058d79cce84fbcd8ff29c8468d
f637f67b8e8d52d641ab15903e46a41ead7e30a5
refs/heads/main
2023-06-17T22:55:08.047769
2023-06-16T16:54:54
2023-06-16T16:54:54
509,664,226
28
7
null
2023-05-27T04:55:05
2022-07-02T05:10:17
Python
UTF-8
Python
false
false
294
py
from flask import request import re @app.route("/direct") def direct(): ''' get the regular expression pattern from the get request, comipile it and then use it for search ''' pattern = request.args.get('pattern') regex = re.compile(pattern) return regex.search(text)
[ "lsiddiqsunny@gmail.com" ]
lsiddiqsunny@gmail.com
433233f2557ef53958b800aa05a1f64706518639
7170b86508060488b07b07738ee6742ac9a3bc3c
/zeusci/zeus/tests/test_api_builds.py
6b2e3ea7484adedd9b2c399a3b2cfaf30bcd9cf7
[]
no_license
lukaszb/zeusci
3bb98ca8c35633613c40dc8ea1af010487b8e155
8a0d2e99ef4f2d12c9d90a9d3aad82cd4ade7d8c
refs/heads/master
2020-05-18T16:36:19.349260
2013-10-20T08:29:29
2013-10-20T08:29:29
null
0
0
null
null
null
null
UTF-8
Python
false
false
5,784
py
from .test_api_base import BaseApiTestCase from django.core.urlresolvers import reverse from django.core.cache import cache from zeus.models import Build from zeus.models import Buildset from zeus.models import Command from zeus.models import Output from zeus.models import Project from zeus.models import Status import datetime class TestBuildApi(BaseApiTestCase): maxDiff = None def setUp(self): zeus = Project.objects.create( name='zeus', url='https://github.com/lukaszb/zeus', repo_url='git://github.com/lukaszb/zeus.git', ) self.buildset = Buildset.objects.create( project=zeus, number=1, ) self.build1 = Build.objects.create( buildset=self.buildset, number=1, ) self.build1_cmd1 = Command.objects.create( number=1, build=self.build1, title='Step 1 -- Configuration', cmd=['./configure'], ) output = Output.objects.create(output='Configured') self.build1_cmd1.command_output = output delta = datetime.timedelta(seconds=2) self.build1_cmd1.started_at = self.build1_cmd1.created_at + delta self.build1_cmd1.finished_at = self.build1_cmd1.started_at + delta self.build1_cmd1.returncode = 0 self.build1_cmd1.status = Status.PASSED self.build1_cmd1.save() self.build1_cmd2 = Command.objects.create( number=2, build=self.build1, title='Step 2 -- Build', cmd=['make', 'all'], ) output = Output.objects.create(output='Build in progress ...') self.build1_cmd2.command_output = output self.build1_cmd2.started_at = self.build1_cmd2.created_at self.build1_cmd2.status = Status.RUNNING self.build1_cmd2.save() dt = datetime.datetime(2013, 7, 2, 22, 8) self.build2 = Build( buildset=self.buildset, number=2, created_at=dt, finished_at=(dt + datetime.timedelta(seconds=3)), ) self.build2.save() cache.clear() def test_build_detail(self): url_params = {'name': 'zeus', 'buildset_no': 1, 'build_no': 1} url = reverse('zeus_api_build_detail', kwargs=url_params) response = self.client.get(url) expected = { 'uri': self.make_api_build_detail_url('zeus', 1, 1), 'url': self.build1.get_absolute_url(), 'number': 1, 'created_at': self.build1.created_at, 'finished_at': self.build1.finished_at, 'status': 'running', 'commands': [ { 'number': 1, 'title': 'Step 1 -- Configuration', 'cmd': './configure', 'output': 'Configured', 'started_at': self.build1_cmd1.started_at, 'finished_at': self.build1_cmd1.finished_at, 'status': 'passed', 'returncode': 0, }, { 'number': 2, 'title': 'Step 2 -- Build', 'cmd': 'make all', 'output': 'Build in progress ...', 'started_at': self.build1_cmd2.started_at, 'finished_at': None, 'status': 'running', 'returncode': None, }, ], } self.assertDictEqual(response.data, expected) url_params = {'name': 'zeus', 'buildset_no': 1, 'build_no': 2} url = reverse('zeus_api_build_detail', kwargs=url_params) response = self.client.get(url) expected = { 'uri': self.make_api_build_detail_url('zeus', 1, 2), 'url': self.build2.get_absolute_url(), 'number': 2, 'created_at': self.build2.created_at, 'finished_at': self.build2.finished_at, 'status': 'pending', 'commands': [], } self.assertDictEqual(response.data, expected) def test_build_restart_fails_if_build_is_still_running(self): url_params = {'name': 'zeus', 'buildset_no': 1, 'build_no': 1} url = reverse('zeus_api_build_detail', kwargs=url_params) response = self.client.put(url) self.assertEqual(response.status_code, 409) def test_build_restart(self): url_params = {'name': 'zeus', 'buildset_no': 1, 'build_no': 1} url = reverse('zeus_api_build_detail', kwargs=url_params) self.build1.commands.update(status=Status.FAILED) response = self.client.put(url) self.assertEqual(response.status_code, 200) self.assertEqual(response.data['status'], Status.PENDING) self.assertIsNone(response.data['finished_at']) self.assertEqual(self.build1.commands.count(), 0) def test_build_status_is_changing_correctly(self): self.assertEqual(self.build1.status, Status.RUNNING) self.build1_cmd1.status = Status.FAILED self.build1_cmd1.save() self.assertEqual(self.build1.status, Status.FAILED) self.build1.commands.update(status=Status.PASSED) self.assertEqual(self.build1.status, Status.PASSED) self.build1.commands.all().delete() self.assertEqual(self.build1.status, Status.PENDING) # Create a command and make sure status us RUNNING once again Command.objects.create( number=1, build=self.build1, title='Step 1 -- Configuration', cmd=['./configure'], status=Status.RUNNING, ) self.assertEqual(self.build1.status, Status.RUNNING)
[ "lukaszbalcerzak@gmail.com" ]
lukaszbalcerzak@gmail.com
d193d711f2be24fe4204a34d2b1a3b14eda09afd
d40ab8694389d1a0d80013a2b0ecd8c426e6e8f8
/graphs/scc.py
5847ec8cbff74c175a28bd22a6d879601af33ceb
[]
no_license
lukebiggerstaff/Stanford-Algorithms-MOOC
b5b34c8d8ff7725461fd03bb3aac505c87a1012e
382a30f27dff6ca9d30c071a4d3418ff6333f4c3
refs/heads/master
2021-01-01T17:02:26.883373
2017-11-14T19:06:14
2017-11-14T19:06:14
97,980,984
0
0
null
null
null
null
UTF-8
Python
false
false
2,171
py
import sys import re import resource from collections import defaultdict sys.setrecursionlimit(10 ** 6) resource.setrlimit(resource.RLIMIT_STACK, (2 ** 29, 2 ** 30)) def dfsfirstpass(graph): visited = set() stack = list() for i in graph.keys(): start = str(i) if start in graph: dfsfirstpassrecursive(graph, start, stack, visited) return stack def dfsfirstpassrecursive(graph, start, stack, visited): if start not in visited: visited.add(start) if start in graph: for edge in graph[start]: if edge not in visited: dfsfirstpassrecursive(graph, edge, stack, visited) stack.append(start) def dfssecondpass(rgraph, stack): visited = set() leaderlist = defaultdict(list) while stack: start = stack.pop() if start not in visited: visited.add(start) leader = start leaderlist[leader] += [start] for edge in set(rgraph[start]) - visited: dfsrecursive(rgraph, edge, visited, leaderlist, leader) return leaderlist def dfsrecursive(graph, start, visited, leaderlist, leader): visited.add(start) leaderlist[leader] += [start] for edge in set(graph[start]) - visited: dfsrecursive(graph, edge, visited, leaderlist, leader) def return_top_five_scc(leaderlist): sccsizelist = list() for key in leaderlist.keys(): size = len(leaderlist[key]) sccsizelist.append(size) sccsizelist.sort() return sccsizelist[-5:] def kosaraju(graph, rgraph): stack = dfsfirstpass(rgraph) #print(f'stack is {stack}') leaderdict = dfssecondpass(graph, stack) #print(f'graph is {graph}\n' #f'leader is {leaderdict}\n') top5 = return_top_five_scc(leaderdict) return top5 if __name__ == '__main__': graph = defaultdict(list) rgraph = defaultdict(list) with open(sys.argv[1]) as f: for line in f: line_lst = re.findall(r'(\d+|\w+)',line) graph[line_lst[0]] += [line_lst[1]] rgraph[line_lst[1]] += [line_lst[0]] print(kosaraju(graph,rgraph))
[ "luke.biggerstaff@gmail.com" ]
luke.biggerstaff@gmail.com
f9881fe5f317247e0325b463ffcc17864d748ccf
882e97d3392f11692d8cae815439cd2609c3b96b
/StacksAndQueues/Stack.py
9bf9a1a0179c83472be858b75092b24b8ce7ba3c
[]
no_license
springrolldippedinsoysauce/pythonshit
af0a48ee99838df35da4cdb9d1990ae41ccb07c4
340f9d9aab0f75914d3306694f73b821808702e2
refs/heads/master
2020-07-18T17:28:42.122560
2019-09-04T09:42:24
2019-09-04T09:42:24
206,284,550
0
0
null
null
null
null
UTF-8
Python
false
false
954
py
from StacksAndQueues import Exceptions as EXP class Stack: def __init__(self, capacity): self._capacity = capacity self._stack = [] self._count = 0 def get_count(self): return self._count def is_empty(self): empty = (self._count == 0) return empty def is_full(self): full = (self._count == self._capacity) return full def push(self, value): if self.is_full() is True: raise EXP.StackOverflowError("Stack is full!") else: self._stack.append(value) self._count += 1 def pop(self): top_val = self.top() self._count -= 1 return top_val def top(self): if self.is_empty() is True: raise EXP.StackUnderflowError("Stack is already empty!") else: top_val = self._stack[self._count - 1] return top_val
[ "noreply@github.com" ]
noreply@github.com
aeff8452dda485b474d6f78dec9db2f3f258e6ff
80d1f1f50b9f7f6be17302df7695154f54aa7a76
/pdp8/core.py
3691cb7deeb3ab2b53a7ab5d3ba6877bf6d1f838
[ "MIT" ]
permissive
romilly/pdp8
af7f7df0136b31df03a4f53b709869a59f25ca2c
68e1025c5e4c6b6fa92a62cc81117d841c214137
refs/heads/master
2020-03-13T15:24:12.955602
2019-06-19T05:58:50
2019-06-19T05:58:50
131,176,107
4
1
null
2018-05-16T07:13:14
2018-04-26T15:31:01
HTML
UTF-8
Python
false
false
7,983
py
from io import StringIO from pdp8.tracing import NullTracer def octal(string): return int(string, 8) OPR_GROUP1 = octal('0400') OPR_GROUP2 = octal('0001') CLA1 = octal('0200') CLL = octal('0100') CMA = octal('0040') CML = octal('0020') RAR = octal('0010') RAL = octal('0004') RTR = octal('0012') RTL = octal('0006') IAC = octal('0001') HALT = octal('0002') BIT8 = octal('0010') Z_BIT = octal('0200') I_BIT = octal('0400') class PDP8: # TODO simplify these, use constants rather than calculating? W_BITS = 12 # number of bits in a word W_MASK = 2 ** W_BITS - 1 # word mask OP_BITS = 3 # 3 bits in the opcode V_BITS = 7 # 7 bits for the value part of an instruction OP_MASK = (2 ** OP_BITS - 1) << W_BITS - OP_BITS V_MASK = 2 ** V_BITS - 1 # mask for instruction data MAX = 2 ** (V_BITS - 1) def __init__(self): self.memory = 2 ** self.W_BITS * [0] self.pc = 0 self.accumulator = 0 self.link = 0 self.running = False self.debugging = False self.stepping = False self.ia = None self.instruction = None self.tape = StringIO('') self.READER1 = 0o03 self.PUNCH1 = 0o04 self.punchflag = 0 self.output = '' self.tracer = None self.ops = [self.andi, self.tad, self.isz, self.dca, self.jms, self.jmp, self.iot, self.opr] def __getitem__(self, address): return self.memory[address] & self.W_MASK # only 12 bits retrieved def is_group1(self): return 0 == self.i_mask(OPR_GROUP1) def i_mask(self, mask): return self.instruction & mask def is_iac(self): return 0 != self.i_mask(IAC) def is_group2(self): return (not self.is_group1()) and 0 == self.i_mask(OPR_GROUP2) # Group 2 def is_halt(self): return self.i_mask(HALT) def __setitem__(self, address, contents): self.memory[address] = contents & self.W_MASK # only 12 bits stored if self.debugging: self.tracer.setting(address, contents) def run(self, debugging=False, start=None, tape='', stepping=None, tracer=None): self.running = True if tracer is not None: self.tracer = tracer else: if self.tracer is None: self.tracer = NullTracer() if start: self.pc = start # TODO: smarter tape creation to cope with text and binary tapes. self.tape = StringIO(tape) if stepping is not None: self.stepping = stepping self.debugging = debugging while self.running: self.execute() if self.stepping: self.running = False def execute(self): old_pc = self.pc # for debugging self.instruction = self[self.pc] self.ia = self.instruction_address() op = self.opcode() self.pc += 1 self.ops[op]() if self.debugging: self.tracer.instruction(old_pc, self.instruction, self.accumulator, self.link, self.pc) def opcode(self): bits = self.i_mask(self.OP_MASK) code = bits >> self.W_BITS - self.OP_BITS return code def andi(self): self.accumulator &= self[self.ia] def tad(self): self.add_12_bits(self[self.ia]) def add_12_bits(self, increment): self.accumulator += increment total = self.accumulator self.accumulator &= octal('7777') if self.accumulator == total: self.link = 0 else: self.link = 1 def isz(self): contents = self[self.ia] contents += 1 self[self.ia] = contents # forces 12-bit value if self[self.ia] == 0: self.pc += 1 # skip def dca(self): self[self.ia] = self.accumulator self.accumulator = 0 def jmp(self): self.pc = self.ia def jms(self): self[self.ia] = self.pc self.pc = self.ia + 1 def iot(self): device = (self.instruction & 0o0770) >> 3 io_op = self.instruction & 0o0007 if device == self.READER1: self.reader(io_op) elif device == self.PUNCH1: self.punch(io_op) else: raise ValueError('uknown device') def opr(self): if self.is_group1(): self.group1() return if self.is_group2(): self.group2() return raise ValueError('Unknown opcode in instruction 0o%o at %d(%o)' % (self.instruction, self.pc-1, self.pc-1) ) def instruction_address(self): o = self.i_mask(self.V_MASK) if not self.i_mask(Z_BIT): o += self.pc & 0o7600 if self.i_mask(I_BIT): o = self[o] return o def cla(self): self.accumulator = 0 def cll(self): self.link = 0 def cma(self): self.accumulator ^= 0o7777 def cml(self): self.link = 1-self.link def rr(self): self.rar(0 < self.i_mask(2)) def rar(self, flag): count = 2 if flag else 1 for i in range(count): new_link = self.accumulator & 0o0001 self.accumulator = self.accumulator >> 1 if self.link: self.accumulator |= 0o4000 self.link = new_link def rl(self): self.ral(self.i_mask(2)) def ral(self, flag): count = 2 if flag else 1 for i in range(count): new_link = 1 if self.accumulator & 0o4000 else 0 self.accumulator = 0o7777 & self.accumulator << 1 if self.link: self.accumulator |= 0o0001 self.link = new_link def iac(self): self.add_12_bits(1) def halt(self): if self.debugging: print('Halted') self.tracer.halt(self.pc) self.running = False def group1(self): for (mask, ins) in zip([ CLA1, CLL, CMA, CML, IAC, RAR, RAL], [self.cla, self.cll, self.cma, self.cml, self.iac,self.rr, self.rl]): if self.i_mask(mask): ins() def is_or_group(self): return not self.i_mask(BIT8) def is_and_group(self): return self.i_mask(BIT8) def group2(self): if self.is_or_group() and (self.sma() or self.sza() or self.snl()): self.pc += 1 if self.is_and_group() and self.spa() and self.sna() and self.szl(): self.pc += 1 if self.is_cla2(): self.cla() if self.is_halt(): self.halt() def sma(self): return self.accumulator_is_negative() and (self.i_mask(octal('0100'))) def accumulator_is_negative(self): return self.accumulator & octal('4000') def sza(self): return self.accumulator == 0 and (self.i_mask(octal('0040'))) def snl(self): return self.link == 1 and (self.i_mask(octal('0020'))) def spa(self): return self.accumulator_is_positive() or not (self.i_mask(octal('0100'))) def accumulator_is_positive(self): return not self.accumulator_is_negative() def sna(self): return self.accumulator != 0 or not (self.i_mask(octal('0040'))) def szl(self): return self.link == 0 or not (self.i_mask(octal('0020'))) def reader(self, io_op): pass def punch(self, io_op): if (io_op & 1) and self.punchflag: self.pc += 1 if io_op & 2: self.punchflag = 0 if io_op & 4: if self.accumulator != 0: self.output += str(chr(self.accumulator)) self.punchflag = 1 def is_cla2(self): return self.instruction & octal('0200')
[ "romilly.cocking@gmail.com" ]
romilly.cocking@gmail.com
91c8298a8f35841bf72996c47795505cf4afd03a
65c001b5f572a6b0ca09dd9821016d628b745009
/frappe-bench/env/lib/python2.7/site-packages/watchdog/observers/polling.py
3039ceb3678ce611aeccc6d88d0586c3f632a5e2
[ "MIT" ]
permissive
ibrahmm22/library-management
666dffebdef1333db122c2a4a99286e7c174c518
b88a2129a5a2e96ce1f945ec8ba99a0b63b8c506
refs/heads/master
2022-10-30T17:53:01.238240
2020-06-11T18:36:41
2020-06-11T18:36:41
271,620,992
0
1
MIT
2022-10-23T05:04:57
2020-06-11T18:36:21
CSS
UTF-8
Python
false
false
4,687
py
#!/usr/bin/env python # -*- coding: utf-8 -*- # # Copyright 2011 Yesudeep Mangalapilly <yesudeep@gmail.com> # Copyright 2012 Google, Inc. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """ :module: watchdog.observers.polling :synopsis: Polling emitter implementation. :author: yesudeep@google.com (Yesudeep Mangalapilly) Classes ------- .. autoclass:: PollingObserver :members: :show-inheritance: .. autoclass:: PollingObserverVFS :members: :show-inheritance: :special-members: """ from __future__ import with_statement import os import threading from functools import partial from watchdog.utils import stat as default_stat from watchdog.utils.dirsnapshot import DirectorySnapshot, DirectorySnapshotDiff from watchdog.observers.api import ( EventEmitter, BaseObserver, DEFAULT_OBSERVER_TIMEOUT, DEFAULT_EMITTER_TIMEOUT ) from watchdog.events import ( DirMovedEvent, DirDeletedEvent, DirCreatedEvent, DirModifiedEvent, FileMovedEvent, FileDeletedEvent, FileCreatedEvent, FileModifiedEvent ) class PollingEmitter(EventEmitter): """ Platform-independent emitter that polls a directory to detect file system changes. """ def __init__(self, event_queue, watch, timeout=DEFAULT_EMITTER_TIMEOUT, stat=default_stat, listdir=os.listdir): EventEmitter.__init__(self, event_queue, watch, timeout) self._snapshot = None self._lock = threading.Lock() self._take_snapshot = lambda: DirectorySnapshot( self.watch.path, self.watch.is_recursive, stat=stat, listdir=listdir) def queue_events(self, timeout): if not self._snapshot: self._snapshot = self._take_snapshot() # We don't want to hit the disk continuously. # timeout behaves like an interval for polling emitters. if self.stopped_event.wait(timeout): return with self._lock: if not self.should_keep_running(): return # Get event diff between fresh snapshot and previous snapshot. # Update snapshot. new_snapshot = self._take_snapshot() events = DirectorySnapshotDiff(self._snapshot, new_snapshot) self._snapshot = new_snapshot # Files. for src_path in events.files_deleted: self.queue_event(FileDeletedEvent(src_path)) for src_path in events.files_modified: self.queue_event(FileModifiedEvent(src_path)) for src_path in events.files_created: self.queue_event(FileCreatedEvent(src_path)) for src_path, dest_path in events.files_moved: self.queue_event(FileMovedEvent(src_path, dest_path)) # Directories. for src_path in events.dirs_deleted: self.queue_event(DirDeletedEvent(src_path)) for src_path in events.dirs_modified: self.queue_event(DirModifiedEvent(src_path)) for src_path in events.dirs_created: self.queue_event(DirCreatedEvent(src_path)) for src_path, dest_path in events.dirs_moved: self.queue_event(DirMovedEvent(src_path, dest_path)) class PollingObserver(BaseObserver): """ Platform-independent observer that polls a directory to detect file system changes. """ def __init__(self, timeout=DEFAULT_OBSERVER_TIMEOUT): BaseObserver.__init__(self, emitter_class=PollingEmitter, timeout=timeout) class PollingObserverVFS(BaseObserver): """ File system independent observer that polls a directory to detect changes. """ def __init__(self, stat, listdir, polling_interval=1): """ :param stat: stat function. See ``os.stat`` for details. :param listdir: listdir function. See ``os.listdir`` for details. :type polling_interval: float :param polling_interval: interval in seconds between polling the file system. """ emitter_cls = partial(PollingEmitter, stat=stat, listdir=listdir) BaseObserver.__init__(self, emitter_class=emitter_cls, timeout=polling_interval)
[ "iabouelftouh@trudoc24x7.com" ]
iabouelftouh@trudoc24x7.com
e0ec0a0e6219b4c6d9bf344fbe2feece9e5e101f
b034c20e2cc4bb480060c2a2a3db78ad8c760818
/3.7-predicting-house-prices.py
a7d57877dd4f4e3feb808ac138229e876cdf5987
[]
no_license
practice-more/try-keras
e84c61d396d34f73743cbb40b8a105d1dbb72986
4921c0c38e04dada327d994d3a1b531a7d032cd8
refs/heads/master
2020-09-05T03:25:14.372434
2019-11-15T10:14:34
2019-11-15T10:14:34
219,967,715
0
0
null
null
null
null
UTF-8
Python
false
false
3,255
py
from keras.datasets import boston_housing from keras import models from keras import layers import numpy as np import matplotlib.pyplot as plt import os os.environ["CUDA_VISIBLE_DEVICES"]="-1" (train_data, train_targets), (test_data, test_targets) = boston_housing.load_data() print(train_data.shape) print(train_data[0,:]) # train data is value of the following classes #1. Per capita crime rate. #2. Proportion of residential land zoned for lots over 25,000 square feet. #3. Proportion of non-retail business acres per town. #4. Charles River dummy variable (= 1 if tract bounds river; 0 otherwise). #5. Nitric oxides concentration (parts per 10 million). #6. Average number of rooms per dwelling. #7. Proportion of owner-occupied units built prior to 1940. #8. Weighted distances to five Boston employment centres. #9. Index of accessibility to radial highways. #10. Full-value property-tax rate per $10,000. #11. Pupil-teacher ratio by town. #12. 1000 * (Bk - 0.63) ** 2 where Bk is the proportion of Black people by town. #13. % lower status of the population. # train target is the price mean = train_data.mean(axis=0) train_data -= mean std = train_data.std(axis=0) train_data /= std test_data -= mean test_data /= std def build_model(): model = models.Sequential() model.add(layers.Dense(64, activation='relu', input_shape=(train_data.shape[1], ))) model.add(layers.Dense(64, activation='relu')) model.add(layers.Dense(1)) model.compile(optimizer='rmsprop', loss='mse', metrics=['mae']) return model # use k-fold validation k = 4 num_val_samples = len(train_data) // k num_epochs = 100 all_scores = [] all_mae_history = [] for i in range(k): print("process folder #", i) val_data = train_data[i * num_val_samples: (i+1) * num_val_samples] val_target = train_targets[i * num_val_samples: (i+1) * num_val_samples] partial_train_data = np.concatenate( [ train_data[:i * num_val_samples], train_data[(i+1) * num_val_samples:] ], axis=0 ) partial_train_targets = np.concatenate( [ train_targets[:i * num_val_samples], train_targets[(i+1) * num_val_samples:] ], axis=0 ) model = build_model() history = model.fit( partial_train_data, partial_train_targets, validation_data=(val_data, val_target), epochs=num_epochs, batch_size=1, verbose=0) mae_history = history.history['val_mae'] all_mae_history.append(mae_history) #val_mes, val_mae = model.evaluate(val_data, val_target, verbose=0) #all_scores.append(val_mae) average_mae_history = [ np.mean([x[i] for x in all_mae_history]) for i in range(num_epochs) ] def smooth_curve(points, factor=0.9): smoothed_points = [] for point in points: if smoothed_points: previous = smoothed_points[-1] smoothed_points.append(previous * factor + point * (1 - factor)) else: smoothed_points.append(point) return smoothed_points average_mae_history = smooth_curve(average_mae_history[10:]) plt.plot(range(1, len(average_mae_history) + 1), average_mae_history) plt.xlabel("Epochs") plt.ylabel("Validation MAE") plt.show()
[ "btan@esri.com" ]
btan@esri.com
e47bd5ee0609619cf412769fd1666431740a27f0
4929b707e07efedcff122241f33ff05e25004d10
/orders/migrations/0007_auto_20160411_1257.py
2f2327ec6d0a1e617b5cd122315c69e90c712d4f
[ "MIT" ]
permissive
loafbaker/django_ecommerce2
be06f00ae637dc9228ba3b5830f8ecef043922d0
7f1e95cf4082960861c3090ed47480eefb9593b9
refs/heads/master
2021-01-21T04:44:46.496191
2016-07-29T05:40:53
2016-07-29T05:41:44
53,693,350
0
1
null
null
null
null
UTF-8
Python
false
false
786
py
# -*- coding: utf-8 -*- # Generated by Django 1.9.4 on 2016-04-11 12:57 from __future__ import unicode_literals from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('orders', '0006_usercheckout_braintree_id'), ] operations = [ migrations.AddField( model_name='order', name='transaction_id', field=models.CharField(blank=True, max_length=20, null=True), ), migrations.AlterField( model_name='order', name='status', field=models.CharField(choices=[('created', 'Created'), ('paid', 'Paid'), ('shipped', 'Shipped'), ('refunded', 'Refunded'), ('completed', 'Completed')], default='created', max_length=120), ), ]
[ "loafbaker@hotmail.com" ]
loafbaker@hotmail.com
7d471f712e0a1b1cc63bcd73298eac954a746430
2a3bd5db7f3db99b6b098ba569655a591b673aac
/10_Sony_Select_Service/Documents/02_Design/05_Module_Specification_Design/V5.6/05_BackupRestore/03_BackupRestoreAppListManager/source/conf.py
8fbb8ae9311a6706e70f7a0820eeb30175fbcbdd
[]
no_license
imace/service-dev
3c6ce201b51bbc481484ee83aa338ee2af31e67f
a2ee6b8bf844cde34b7f4b5915e2465d84e6258f
refs/heads/master
2021-01-17T06:01:05.093541
2014-12-08T06:37:22
2014-12-08T06:37:22
null
0
0
null
null
null
null
UTF-8
Python
false
false
8,323
py
# -*- coding: utf-8 -*- # # Pro_Name documentation build configuration file, created by # sphinx-quickstart on Wed Apr 18 22:33:41 2012. # # This file is execfile()d with the current directory set to its containing dir. # # Note that not all possible configuration values are present in this # autogenerated file. # # All configuration values have a default; values that are commented out # serve to show the default. import sys, os # If extensions (or modules to document with autodoc) are in another directory, # add these directories to sys.path here. If the directory is relative to the # documentation root, use os.path.abspath to make it absolute, like shown here. #sys.path.insert(0, os.path.abspath('.')) # -- General configuration ----------------------------------------------------- # If your documentation needs a minimal Sphinx version, state it here. #needs_sphinx = '1.0' # Add any Sphinx extension module names here, as strings. They can be extensions # coming with Sphinx (named 'sphinx.ext.*') or your custom ones. extensions = [] # Add any paths that contain templates here, relative to this directory. templates_path = ['_templates'] # The suffix of source filenames. source_suffix = '.rst' # The encoding of source files. #source_encoding = 'utf-8-sig' # The master toctree document. master_doc = 'index' # General information about the project. project = u'PRD_DOC' copyright = u'2012, Service_Dev' # The version info for the project you're documenting, acts as replacement for # |version| and |release|, also used in various other places throughout the # built documents. # # The short X.Y version. version = '1.1.0' # The full version, including alpha/beta/rc tags. release = '1.1.0' # The language for content autogenerated by Sphinx. Refer to documentation # for a list of supported languages. language = 'zh_CN' # There are two options for replacing |today|: either, you set today to some # non-false value, then it is used: #today = '' # Else, today_fmt is used as the format for a strftime call. #today_fmt = '%B %d, %Y' # List of patterns, relative to source directory, that match files and # directories to ignore when looking for source files. exclude_patterns = [] # The reST default role (used for this markup: `text`) to use for all documents. default_role = "Service_Dev" # If true, '()' will be appended to :func: etc. cross-reference text. #add_function_parentheses = True # If true, the current module name will be prepended to all description # unit titles (such as .. function::). #add_module_names = True # If true, sectionauthor and moduleauthor directives will be shown in the # output. They are ignored by default. #show_authors = False # The name of the Pygments (syntax highlighting) style to use. pygments_style = 'sphinx' # A list of ignored prefixes for module index sorting. #modindex_common_prefix = [] # -- Options for HTML output --------------------------------------------------- # The theme to use for HTML and HTML Help pages. See the documentation for # a list of builtin themes. html_theme = 'basic' html_style = 'adctheme.css' # Theme options are theme-specific and customize the look and feel of a theme # further. For a list of options available for each theme, see the # documentation. html_theme_options = { #"stickysidebar":"true", #"sidebarbgcolor":"#eceae7", #"sidebartextcolor":"#000000", #"relbartextcolor":"black", #"textcolor":"black", #"headtextcolor":"black", #"footertextcolor":"black", } # Add any paths that contain custom themes here, relative to this directory. #html_theme_path = [] # The name for this set of Sphinx documents. If None, it defaults to # "<project> v<release> documentation". html_title = u'SonySelect Service' # A shorter title for the navigation bar. Default is the same as html_title. html_short_title = "Service Design" # The name of an image file (relative to this directory) to place at the top # of the sidebar. #html_logo = "./images/select_icon.png" # The name of an image file (within the static path) to use as favicon of the # docs. This file should be a Windows icon file (.ico) being 16x16 or 32x32 # pixels large. html_favicon = "./images/select_icon.png" # Add any paths that contain custom static files (such as style sheets) here, # relative to this directory. They are copied after the builtin static files, # so a file named "default.css" will overwrite the builtin "default.css". html_static_path = ['_static'] # If not '', a 'Last updated on:' timestamp is inserted at every page bottom, # using the given strftime format. #html_last_updated_fmt = '%b %d, %Y' # If true, SmartyPants will be used to convert quotes and dashes to # typographically correct entities. #html_use_smartypants = True # Custom sidebar templates, maps document names to template names. #html_sidebars = {} # Additional templates that should be rendered to pages, maps page names to # template names. #html_additional_pages = {} # If false, no module index is generated. #html_domain_indices = True # If false, no index is generated. html_use_index = True # If true, the index is split into individual pages for each letter. html_split_index = False # If true, links to the reST sources are added to the pages. #html_show_sourcelink = True # If true, "Created using Sphinx" is shown in the HTML footer. Default is True. #html_show_sphinx = True # If true, "(C) Copyright ..." is shown in the HTML footer. Default is True. #html_show_copyright = True # If true, an OpenSearch description file will be output, and all pages will # contain a <link> tag referring to it. The value of this option must be the # base URL from which the finished HTML is served. #html_use_opensearch = '' # This is the file name suffix for HTML files (e.g. ".xhtml"). #html_file_suffix = None # Output file base name for HTML help builder. htmlhelp_basename = u'Service_Dev' # -- Options for LaTeX output -------------------------------------------------- latex_elements = { # The paper size ('letterpaper' or 'a4paper'). #'papersize': 'letterpaper', # The font size ('10pt', '11pt' or '12pt'). #'pointsize': '10pt', # Additional stuff for the LaTeX preamble. #'preamble': '', 'preamble': ''' \usepackage{xeCJK} \usepackage{indentfirst} \setlength{\parindent}{2em} \setCJKmainfont[BoldFont=SimHei, ItalicFont=KaiTi_GB2312]{SimSun} \setCJKmonofont[Scale=0.9]{Droid Sans Mono} \setCJKfamilyfont{song}[BoldFont=SimSun]{SimSun} \setCJKfamilyfont{sf}[BoldFont=SimSun]{SimSun} ''', } # Grouping the document tree into LaTeX files. List of tuples # (source start file, target name, title, author, documentclass [howto/manual]). latex_documents = [ ('index', 'Pro_Name.tex', u'Pro\\_Name Documentation', u'ServiceDev', 'manual'), ] # The name of an image file (relative to this directory) to place at the top of # the title page. #latex_logo = None # For "manual" documents, if this is true, then toplevel headings are parts, # not chapters. #latex_use_parts = False # If true, show page references after internal links. #latex_show_pagerefs = False # If true, show URL addresses after external links. #latex_show_urls = False # Documents to append as an appendix to all manuals. #latex_appendices = [] # If false, no module index is generated. #latex_domain_indices = True # -- Options for manual page output -------------------------------------------- # One entry per manual page. List of tuples # (source start file, name, description, authors, manual section). man_pages = [ ('index', 'pro_name', u'Pro_Name Documentation', [u'Robert Guo'], 1) ] # If true, show URL addresses after external links. #man_show_urls = False # -- Options for Texinfo output ------------------------------------------------ # Grouping the document tree into Texinfo files. List of tuples # (source start file, target name, title, author, # dir menu entry, description, category) texinfo_documents = [ ('index', 'Pro_Name', u'Pro_Name Documentation', u'ServiceDev', 'Pro_Name', 'One line description of project.', 'Miscellaneous'), ] # Documents to append as an appendix to all manuals. #texinfo_appendices = [] # If false, no module index is generated. #texinfo_domain_indices = True # How to display URL addresses: 'footnote', 'no', or 'inline'. #texinfo_show_urls = 'footnote'
[ "devin.wei@sonymobile.com" ]
devin.wei@sonymobile.com
ff0f2048594d475b598b860002a8ba4d6b8f0047
e1e59ba5f60886be8b1f587f274d4fe177982e35
/familiarize_project/apps/users_app/urls.py
c0a27ca2522a0c48a2afd59e929931634f7d45cf
[]
no_license
betalantz/CodingDojo_LearningProjects
0b2d82a7e32e8e22a07de79d50a46a61bb8fbeca
4188a0c1df42da3ce5d29f2bdeb92865ea2c6cd1
refs/heads/master
2021-01-01T19:26:16.879191
2019-04-23T00:48:12
2019-04-23T00:48:12
98,582,797
1
0
null
null
null
null
UTF-8
Python
false
false
877
py
"""familiarize_project URL Configuration The `urlpatterns` list routes URLs to views. For more information please see: https://docs.djangoproject.com/en/1.10/topics/http/urls/ Examples: Function views 1. Add an import: from my_app import views 2. Add a URL to urlpatterns: url(r'^$', views.home, name='home') Class-based views 1. Add an import: from other_app.views import Home 2. Add a URL to urlpatterns: url(r'^$', Home.as_view(), name='home') Including another URLconf 1. Import the include() function: from django.conf.urls import url, include 2. Add a URL to urlpatterns: url(r'^blog/', include('blog.urls')) """ from django.conf.urls import url from . import views urlpatterns = [ url(r'^users/$', views.index), url(r'^users/new/$', views.register), url(r'^register/$', views.register), url(r'^login/$', views.login), ]
[ "noreply@github.com" ]
noreply@github.com
a655da1503c403cc2cfe9a6e9f8fb672253bd71f
a7efab7b8555d75349917f9a23368a87e9d98726
/venv/Scripts/pip-script.py
1d65201c8efab4a9201514b89c1303150ae235b0
[]
no_license
text007/python
0779d43c9e2a235939117a8d86901a1d9878bf0d
9c39f8876953ad7d2065e06c96ceb28b1bb43085
refs/heads/master
2022-11-24T13:32:48.217181
2020-07-24T02:45:06
2020-07-24T02:45:06
265,194,028
0
0
null
null
null
null
UTF-8
Python
false
false
383
py
#!F:\python2\venv\Scripts\python.exe # EASY-INSTALL-ENTRY-SCRIPT: 'pip==19.0.3','console_scripts','pip' __requires__ = 'pip==19.0.3' import re import sys from pkg_resources import load_entry_point if __name__ == '__main__': sys.argv[0] = re.sub(r'(-script\.pyw?|\.exe)?$', '', sys.argv[0]) sys.exit( load_entry_point('pip==19.0.3', 'console_scripts', 'pip')() )
[ "604573679@qq.com" ]
604573679@qq.com
a7ca9c17124ca0830796ae2a129f45c0f8d02b13
458495651f84b6ef033d084fa6caad15c2154077
/book/views.py
8b57fc286b811c4742788b028b40bfdc25d4fead
[]
no_license
waldemarantypov/sprint17_django_Forms
85c5333d4f39ed998c844b0056193ad0fc20f84a
15061b558a4b5e385d6a538ff89d664f6255223c
refs/heads/main
2023-04-04T21:53:17.876812
2021-04-04T14:52:44
2021-04-04T14:52:44
353,795,486
0
0
null
null
null
null
UTF-8
Python
false
false
300
py
from django.shortcuts import render from django.http import HttpResponse from book.models import Book # Create your views here. def show_books(request): randomBooks = Book.objects.all().order_by('?')[:3] context = {"allBooks": randomBooks} return render(request, 'index.html', context)
[ "waldemar.antypov@gmail.com" ]
waldemar.antypov@gmail.com
4d057261b30773fd2e2100f2226e925e29b70202
7e48f2d3592dc815e06956bfaba78adbda58626c
/sibu/sample_programs/python/Python/37~48강/alzio07.py
da2f5c72d175385629b371779f69b2799919125c
[]
no_license
joonholee95/Sibu-taikai
bec247a1a1d480f6b659d885b1b7e8dc89cdfd39
75ca616daab89f2016b75d340dfe8b49ec3af3c8
refs/heads/master
2020-03-22T15:25:34.664224
2018-10-01T02:30:15
2018-10-01T02:30:15
140,251,655
0
0
null
null
null
null
UTF-8
Python
false
false
3,831
py
# coding: utf-8 # In[1]: from pandas import Series, DataFrame import pandas as pd import numpy as np import matplotlib.pyplot as plt from matplotlib import font_manager, rc font_name = font_manager.FontProperties(fname="c:/Windows/Fonts/malgun.ttf").get_name() rc('font', family=font_name) # # pandas로 데이터 다루기 # ## 변형 # - 중복 제거 # - 매핑 # - 치환 # - 특이값 다루기 # ### 중복 제거 # In[2]: df = DataFrame({'key1': list("가가가나나다다다"), 'key2': list("13122414")}) df # In[3]: df.duplicated() # In[4]: df.drop_duplicates() # In[5]: df.drop_duplicates(keep='last') # In[6]: df.drop_duplicates('key1') # In[7]: df.drop_duplicates('key2') # ### 매핑 # In[8]: data = { '이름': list("가나다라마"), '점수': np.arange(5) * 10 } df = DataFrame(data) df # In[9]: gender = { "나": "여", "마": "여", "가": "남", "라": "남", "다": "여", } # In[10]: df['성별'] = df['이름'].map(gender) df # In[11]: df['점수'] = df['점수'].map("{}점".format) df # ### 치환 # In[12]: sr = Series([2000, -1, 2001, 2002, -1, 0, 2003]) sr # In[13]: sr.replace(-1, np.nan) # In[14]: sr.replace([-1, 0], np.nan) # In[15]: sr.replace({-1: np.nan, 0: 2000}) # ### 특이값 다루기 # In[16]: np.random.seed(1) df = DataFrame(np.random.randn(1000, 5)) df.describe() # In[17]: plt.hist(df[2], bins=100) plt.show(); # In[18]: abs(df[2]) > 2.5 # In[19]: df[abs(df[2]) > 2.5] # In[20]: df[abs(df[2]) > 2.5] = np.sign(df) * 2.5 # In[21]: plt.hist(df[2], bins=100) plt.show(); # # matplotlib로 그래프 그리기 # In[22]: from numpy.random import randn # ## 기본 # In[23]: plt.plot(randn(50).cumsum()) plt.show(); # ## 서브플롯 # In[24]: fig, axes = plt.subplots(2, 2) axes[0, 0].hist(randn(100), color='k') axes[0, 1].plot(randn(100).cumsum(), c='red') axes[1, 0].scatter(randn(100), randn(100), color='green') axes[1, 1].bar(np.arange(100), np.random.randint(1, 100, size=100), color='blue') plt.show(); # ## 서브플롯간 간격 조절 # In[25]: fig, axes = plt.subplots(2, 2, sharex=True, sharey=True) axes[0, 0].hist(randn(1000), color='k') axes[1, 0].hist(randn(1000), color='r') axes[0, 1].hist(randn(1000), color='g') axes[1, 1].hist(randn(1000), color='b') plt.subplots_adjust(wspace=0, hspace=0) plt.show(); # ## 크기, 마커, 선 스타일 # In[26]: plt.figure(figsize=(12, 6)) plt.plot(randn(50).cumsum(), linestyle='-', color='k', marker='*') plt.plot(randn(50).cumsum(), linestyle='--', color='r', marker='o') plt.plot(randn(50).cumsum(), linestyle='-.', color='g', marker='v') plt.plot(randn(50).cumsum(), linestyle=':', color='b', marker='D') plt.show(); # ## 눈금 # In[27]: plt.plot(randn(500).cumsum()) plt.show(); # In[28]: plt.plot(randn(500).cumsum()) plt.xticks([0, 250, 500]) plt.show(); # In[29]: plt.plot(randn(500).cumsum()) plt.xticks([0, 250, 500], ["2000년 1월 1일", "7월 1일", "2001년 1월 1일"], rotation=45) plt.show(); # ## 라벨, 범례 # In[30]: plt.plot(randn(50).cumsum(), linestyle='-', color='k', label="검정") plt.plot(randn(50).cumsum(), linestyle='--', color='r', label="빨강") plt.plot(randn(50).cumsum(), linestyle='-.', color='g', label="초록") plt.xlabel("스텝") plt.ylabel("값") plt.title("제목") plt.legend(loc="lower left") plt.show(); # ## 저장 # In[32]: fig = plt.figure() plt.plot(randn(50).cumsum(), linestyle='-', color='k', label="검정") plt.plot(randn(50).cumsum(), linestyle='--', color='r', label="빨강") plt.plot(randn(50).cumsum(), linestyle='-.', color='g', label="초록") plt.xlabel("스텝") plt.ylabel("값") plt.title("제목") plt.legend(loc="lower left") plt.savefig("test.jpg", dpi=300); plt.show(); # ![](test.jpg)
[ "joonho.lee@human.ait.kyushu-u.ac.jp" ]
joonho.lee@human.ait.kyushu-u.ac.jp
9cf3d784947858edaf604e6c180fff1007ff9612
0b01cb61a4ae4ae236a354cbfa23064e9057e434
/alipay/aop/api/request/AlipayExscUserFirstsignGetRequest.py
e639edfe18817ae9d75d814b5ee4329e27910ec7
[ "Apache-2.0" ]
permissive
hipacloud/alipay-sdk-python-all
e4aec2869bf1ea6f7c6fb97ac7cc724be44ecd13
bdbffbc6d5c7a0a3dd9db69c99443f98aecf907d
refs/heads/master
2022-11-14T11:12:24.441822
2020-07-14T03:12:15
2020-07-14T03:12:15
277,970,730
0
0
Apache-2.0
2020-07-08T02:33:15
2020-07-08T02:33:14
null
UTF-8
Python
false
false
3,664
py
#!/usr/bin/env python # -*- coding: utf-8 -*- import json from alipay.aop.api.FileItem import FileItem from alipay.aop.api.constant.ParamConstants import * class AlipayExscUserFirstsignGetRequest(object): def __init__(self, biz_model=None): self._biz_model = biz_model self._alipay_id = None self._version = "1.0" self._terminal_type = None self._terminal_info = None self._prod_code = None self._notify_url = None self._return_url = None self._udf_params = None self._need_encrypt = False @property def biz_model(self): return self._biz_model @biz_model.setter def biz_model(self, value): self._biz_model = value @property def alipay_id(self): return self._alipay_id @alipay_id.setter def alipay_id(self, value): self._alipay_id = value @property def version(self): return self._version @version.setter def version(self, value): self._version = value @property def terminal_type(self): return self._terminal_type @terminal_type.setter def terminal_type(self, value): self._terminal_type = value @property def terminal_info(self): return self._terminal_info @terminal_info.setter def terminal_info(self, value): self._terminal_info = value @property def prod_code(self): return self._prod_code @prod_code.setter def prod_code(self, value): self._prod_code = value @property def notify_url(self): return self._notify_url @notify_url.setter def notify_url(self, value): self._notify_url = value @property def return_url(self): return self._return_url @return_url.setter def return_url(self, value): self._return_url = value @property def udf_params(self): return self._udf_params @udf_params.setter def udf_params(self, value): if not isinstance(value, dict): return self._udf_params = value @property def need_encrypt(self): return self._need_encrypt @need_encrypt.setter def need_encrypt(self, value): self._need_encrypt = value def add_other_text_param(self, key, value): if not self.udf_params: self.udf_params = dict() self.udf_params[key] = value def get_params(self): params = dict() params[P_METHOD] = 'alipay.exsc.user.firstsign.get' params[P_VERSION] = self.version if self.biz_model: params[P_BIZ_CONTENT] = json.dumps(obj=self.biz_model.to_alipay_dict(), ensure_ascii=False, sort_keys=True, separators=(',', ':')) if self.alipay_id: if hasattr(self.alipay_id, 'to_alipay_dict'): params['alipay_id'] = json.dumps(obj=self.alipay_id.to_alipay_dict(), ensure_ascii=False, sort_keys=True, separators=(',', ':')) else: params['alipay_id'] = self.alipay_id if self.terminal_type: params['terminal_type'] = self.terminal_type if self.terminal_info: params['terminal_info'] = self.terminal_info if self.prod_code: params['prod_code'] = self.prod_code if self.notify_url: params['notify_url'] = self.notify_url if self.return_url: params['return_url'] = self.return_url if self.udf_params: params.update(self.udf_params) return params def get_multipart_params(self): multipart_params = dict() return multipart_params
[ "liuqun.lq@alibaba-inc.com" ]
liuqun.lq@alibaba-inc.com
afbea8d9ea26a5aca621937e13f4651df94afc0f
51e61e09fc0440acc56e94446873a686e5f19a9b
/sentiment_analysis.py
580864d55f9c0138a144fed68b75040afa71a50c
[]
no_license
Yunhan0816/band-lyrics-analysis
f6d3c1981e0718dc53a1cd2908f4f19d28a93371
e4e0b5cca125dfa70ea1b1f699d595956e2c7b97
refs/heads/master
2020-06-23T09:46:56.817147
2019-08-15T09:02:15
2019-08-15T09:02:15
198,588,897
0
0
null
null
null
null
UTF-8
Python
false
false
1,056
py
import nltk nltk.download('vader_lexicon') from nltk.sentiment.vader import SentimentIntensityAnalyzer import config import matplotlib.pyplot as plt import pandas as pd artists = config.artists df = pd.DataFrame(columns = ('artist', 'positive', 'neautral', 'negative')) s = SentimentIntensityAnalyzer() i=0 for artist in artists: pos = 0 neg = 0 neu = 0 f = open('lyrics/' + artist + '_cleaned', 'rb') for sentence in f.readlines(): this_sentence = sentence.decode('utf-8') comp = s.polarity_scores(this_sentence) comp = comp['compound'] if comp >= 0.5: pos += 1 elif comp > -0.5 and comp < 0.5: neu += 1 else: neg += 1 num_total = pos + neu + neg percent_negative = (neg / float(num_total)) * 100 percent_neutral = (neu / float(num_total)) * 100 percent_positive = (pos / float(num_total)) * 100 df.loc[i] = (artist, percent_positive, percent_neutral, percent_negative) i+=1 df.plot.bar(x='artist', stacked=True) plt.show()
[ "yunhanh@bu.edu" ]
yunhanh@bu.edu
aeb4299c83b8a609952126db23c62ac404ef1759
b422f6330d969887db3e6f7cafa52b681bd181de
/bank_system/decorators/client_decorator.py
ee7bd500164ccc60d932da471a15587876bf2d0e
[]
no_license
weslleymberg/eispatterns-examples
0346f78cdfcb0729fae72465662d9b797a0aa4b9
ae1a1759886fb1e05067ea26589e0e229fc646f0
refs/heads/master
2021-01-17T16:49:34.464541
2011-12-06T17:21:10
2011-12-06T17:21:10
1,979,529
0
0
null
null
null
null
UTF-8
Python
false
false
746
py
#!/usr/bin/env python # -*- coding: utf-8 -*- from should_dsl import should from domain.base.decorator import Decorator from domain.node.person import Person from domain.resource.operation import operation from domain.supportive.rule import rule from domain.supportive.association_error import AssociationError class ClientDecorator(Decorator): '''A general porpuse Client decorator''' decoration_rules = ['should_be_instance_of_person'] def __init__(self): Decorator.__init__(self) self.description = "Supplies the basis for representing clients" self.accounts = [] def generate_register(self, register): ''' generates the register number for the client''' self.register = register
[ "weslleym.lisboa@gmail.com" ]
weslleym.lisboa@gmail.com
bd6a9831ec6c286db6ed408f8d0f498356771f0c
24c8065f5508126675029cdb41cfa22ec1d70539
/venv/lib/python3.6/site-packages/haystackbrowser/models.py
5e8d9bf2e45c5c418883af394de4584f3d9d0dd1
[]
no_license
Shawnmhy/BEProject
3b2eaff1737a8bc08af87e6d1e078912449531dd
06bff81c5b8942c3b713263b0d417c4d19f1d7b3
refs/heads/master
2020-03-10T03:18:53.341055
2018-05-23T21:51:16
2018-05-23T21:51:16
129,161,516
0
0
null
null
null
null
UTF-8
Python
false
false
12,794
py
# -*- coding: utf-8 -*- import logging from copy import deepcopy try: from urllib import quote_plus except ImportError: # > Python 3 from django.utils.six.moves.urllib import parse quote_plus = parse.quote_plus from operator import itemgetter from itertools import groupby from collections import namedtuple from django.db import models try: from django.utils.encoding import force_text except ImportError: # < Django 1.5 from django.utils.encoding import force_unicode as force_text from django.utils.safestring import mark_safe from django.utils.html import strip_tags from django.urls import NoReverseMatch, reverse from django.utils.translation import ugettext_lazy as _ from django.conf import settings logger = logging.getLogger(__name__) class HaystackResults(models.Model): """ Our fake model, used for mounting :py:class:`~haystackbrowser.admin.HaystackResultsAdmin` onto the appropriate AdminSite. .. note:: the model is marked as unmanaged, so will never get created via ``syncdb``. """ class Meta: managed = False verbose_name = _('Search result') verbose_name_plural = _('Search results') class SearchResultWrapper(object): """Value object which consumes a standard Haystack SearchResult, and the current admin site, and exposes additional methods and attributes for displaying the data appropriately. :param obj: the item to be wrapped. :type obj: object :param admin_site: the parent site instance. :type admin_site: AdminSite object """ def __init__(self, obj, admin_site=None): self.admin = admin_site self.object = obj if getattr(self.object, 'searchindex', None) is None: # < Haystack 1.2 from haystack import site self.object.searchindex = site.get_index(self.object.model) def __repr__(self): return '<%(module)s.%(cls)s [%(app)s.%(model)s pk=%(pk)r]>' % { 'module': self.__class__.__module__, 'cls': self.__class__.__name__, 'obj': self.object, 'app': self.object.app_label, 'model': self.object.model_name, 'pk': self.object.pk, } def get_app_url(self): """Resolves a given object's app into a link to the app administration. .. warning:: This link may return a 404, as pretty much anything may be reversed and fit into the ``app_list`` urlconf. :return: string or None """ try: return reverse('%s:app_list' % self.admin, kwargs={ 'app_label': self.object.app_label, }) except NoReverseMatch: return None def get_model_url(self): """Generates a link to the changelist for a specific Model in the administration. :return: string or None """ try: parts = (self.admin, self.object.app_label, self.object.model_name) return reverse('%s:%s_%s_changelist' % parts) except NoReverseMatch: return None def get_pk_url(self): """Generates a link to the edit page for a specific object in the administration. :return: string or None """ try: parts = (self.admin, self.object.app_label, self.object.model_name) return reverse('%s:%s_%s_change' % parts, args=(self.object.pk,)) except NoReverseMatch: return None def get_detail_url(self): try: urlname = '%s:haystackbrowser_haystackresults_change' % self.admin return reverse(urlname, kwargs={ 'content_type': '.'.join([self.object.app_label, self.object.model_name]), 'pk': self.object.pk}) except NoReverseMatch: return None def get_model_attrs(self): outfields = {} try: fields = self.object.searchindex.fields except: fields = {} else: for key, field in fields.items(): has_model_attr = getattr(field, 'model_attr', None) if has_model_attr is not None: outfields[key] = force_text(has_model_attr) return outfields def get_stored_fields(self): stored_fields = {} model_attrs = self.get_model_attrs() for key, value in self.object.get_stored_fields().items(): safe_value = force_text(value).strip() stored_fields[key] = { 'raw': safe_value, 'safe': mark_safe(strip_tags(safe_value)) } if key in model_attrs: stored_fields[key].update(model_attr=model_attrs.get(key)) return stored_fields def get_additional_fields(self): """Find all fields in the Haystack SearchResult which have not already appeared in the stored fields. :return: dictionary of field names and values. """ additional_fields = {} stored_fields = self.get_stored_fields().keys() model_attrs = self.get_model_attrs() for key, value in self.object.get_additional_fields().items(): if key not in stored_fields: safe_value = force_text(value).strip() additional_fields[key] = { 'raw': safe_value, 'safe': mark_safe(strip_tags(safe_value)) } if key in model_attrs: additional_fields[key].update(model_attr=model_attrs.get(key)) return additional_fields def get_content_field(self): """Find the name of the main content field in the Haystack SearchIndex for this object. :return: string representing the attribute name. """ return self.object.searchindex.get_content_field() def get_content(self): """Given the name of the main content field in the Haystack Search Index for this object, get the named attribute on this object. :return: whatever is in ``self.object.<content_field_name>`` """ return getattr(self.object, self.get_content_field()) def get_stored_field_count(self): """ Provides mechanism for finding the number of stored fields stored on this Search Result. :return: the count of all stored fields. :rtype: integer """ return len(self.object.get_stored_fields().keys()) def get_additional_field_count(self): """ Provides mechanism for finding the number of stored fields stored on this Search Result. :return: the count of all stored fields. :rtype: integer """ return len(self.get_additional_fields().keys()) def __getattr__(self, attr): return getattr(self.object, attr) class FacetWrapper(object): """ A simple wrapper around `sqs.facet_counts()` to filter out things with 0, and re-arrange the data in such a way that the template can handle it. """ __slots__ = ('dates', 'fields', 'queries', '_total_count', '_querydict') def __init__(self, facet_counts, querydict): self.dates = facet_counts.get('dates', {}) self.fields = facet_counts.get('fields', {}) self.queries = facet_counts.get('queries', {}) self._total_count = len(self.dates) + len(self.fields) + len(self.queries) # querydict comes from the cleaned form data ... page_key = 'p' if querydict is not None and page_key in querydict: querydict.pop(page_key) self._querydict = querydict def __repr__(self): return '<%(module)s.%(cls)s fields=%(fields)r dates=%(dates)r ' \ 'queries=%(queries)r>' % { 'module': self.__class__.__module__, 'cls': self.__class__.__name__, 'fields': self.fields, 'dates': self.dates, 'queries': self.queries, } def get_facets_from(self, x): if x not in ('dates', 'queries', 'fields'): raise AttributeError('Wrong field, silly.') for field, items in getattr(self, x).items(): for content, count in items: content = content.strip() if count > 0 and content: yield {'field': field, 'value': content, 'count': count, 'fieldvalue': quote_plus('%s:%s' % (field, content)), 'facet': Facet(field, querydict=self._querydict)} def get_grouped_facets_from(self, x): data = sorted(self.get_facets_from(x), key=itemgetter('field')) #return data results = ({'grouper': Facet(key), 'list': list(val)} for key, val in groupby(data, key=itemgetter('field'))) return results def get_field_facets(self): return self.get_grouped_facets_from('fields') def get_date_facets(self): return self.get_grouped_facets_from('dates') def get_query_facets(self): return self.get_grouped_facets_from('queries') def __bool__(self): """ Used for doing `if facets: print(facets)` - this is the Python 2 magic method; __nonzero__ is the equivalent thing in Python 3 """ return self._total_count > 0 __nonzero__ = __bool__ def __len__(self): """ For checking things via `if len(facets) > 0: print(facets)` """ return self._total_count class AppliedFacet(namedtuple('AppliedFacet', 'field value querydict')): __slots__ = () def title(self): return self.value @property def facet(self): """ a richer object """ return Facet(self.raw) @property def raw(self): """ the original data, rejoined """ return '%s:%s' % (self.field, self.value) @property def narrow(self): """ returns a string format value """ return '{0}:"{{cleaned_value}}"'.format(self.field) def link(self): """ link to just this facet """ new_qd = self.querydict.copy() page_key = 'p' if page_key in new_qd: new_qd.pop(page_key) new_qd['selected_facets'] = self.raw new_qd['possible_facets'] = self.field return '?%s' % new_qd.urlencode() def remove_link(self): new_qd = self.querydict.copy() # remove page forcibly ... page_key = 'p' if page_key in new_qd: new_qd.pop(page_key) # remove self from the existing querydict/querystring ... key = 'selected_facets' if key in new_qd and self.raw in new_qd.getlist(key): new_qd.getlist(key).remove(self.raw) return '?%s' % new_qd.urlencode() class AppliedFacets(object): __slots__ = ('_applied',) def __init__(self, querydict): self._applied = {} selected = () if 'selected_facets' in querydict: selected = querydict.getlist('selected_facets') for raw_facet in selected: if ":" not in raw_facet: continue field, value = raw_facet.split(":", 1) to_add = AppliedFacet(field=field, value=value, querydict=querydict) self._applied[raw_facet] = to_add def __iter__(self): return iter(self._applied.values()) def __len__(self): return len(self._applied) def __contains__(self, item): return item in self._applied def __repr__(self): raw = tuple(v.raw for k, v in self._applied.items()) return '<{cls!s}.{name!s} selected_facets={raw}>'.format( cls=self.__class__.__module__, name=self.__class__.__name__, raw=raw) def __str__(self): raw = [v.facet.get_display() for k, v in self._applied.items()] return '{name!s} {raw!s}'.format(name=self.__class__.__name__, raw=raw) class Facet(object): """ Takes a facet field name, like `thing_exact` """ __slots__ = ('fieldname', '_querydict') def __init__(self, fieldname, querydict=None): self.fieldname = fieldname self._querydict = querydict def __repr__(self): return '<%(module)s.%(cls)s - %(field)s>' % { 'module': self.__class__.__module__, 'cls': self.__class__.__name__, 'field': self.fieldname, } def link(self): qd = self._querydict if qd is not None: return '?%s' % qd.urlencode() return '?' def get_display(self): return self.fieldname.replace('_', ' ').title() def choices(self): return (self.fieldname, self.get_display())
[ "shawnmhy@gmail.com" ]
shawnmhy@gmail.com
64bef1b8d66e25515d68a737b143f8d15d5675ce
7790e3a3f2de068fef343585ec856983591997a2
/bank/migrations/0021_followlawtype.py
67c1e5c0f8ed434aeb042dbf4b3e27f516602279
[]
no_license
mehdi1361/tadbir
ce702a9a02672826f0bf06e8d5cf0644efe31949
c0a67710099f713cf96930e25df708625de89a6f
refs/heads/master
2021-06-04T07:35:37.624372
2018-07-23T05:25:04
2018-07-23T05:25:04
148,870,028
0
0
null
2019-10-22T21:40:28
2018-09-15T04:40:26
HTML
UTF-8
Python
false
false
1,147
py
# -*- coding: utf-8 -*- # Generated by Django 1.11 on 2018-05-11 16:29 from __future__ import unicode_literals from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('bank', '0020_auto_20180510_1351'), ] operations = [ migrations.CreateModel( name='FollowLawType', fields=[ ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), ('created_at', models.DateTimeField(auto_now_add=True, verbose_name='تاریخ ایجاد')), ('update_at', models.DateTimeField(auto_now=True, verbose_name='تاریخ بروزرسانی')), ('type', models.CharField(max_length=100, verbose_name='نوع پیگیری')), ('enable', models.BooleanField(default=False, verbose_name='فعال')), ], options={ 'verbose_name': 'پیگیری حقوقی', 'verbose_name_plural': 'پیگیری های حقوقی', 'db_table': 'follow_low_type', }, ), ]
[ "mhd.mosavi@gmail.com" ]
mhd.mosavi@gmail.com
5a6ea935188b2c941b1c2a18e79fce894cd1916c
ab77b9ee5bed093016ec44ab2b47e055f7f1e0d1
/Python/mysql/full_friends/bin/pip2.7
d99268cba2cf8698d7f0b6eceef55b75de7ba898
[]
no_license
minokuchi/coding_dojo
f07737642228ca069e82894496ff2c0aed1f735f
0f6d8262c093dc9c6f3825667df5e67b3810c80d
refs/heads/master
2021-09-07T04:59:10.116505
2018-02-17T17:41:51
2018-02-17T17:41:51
109,316,639
0
0
null
2018-02-17T17:41:52
2017-11-02T20:41:00
Python
UTF-8
Python
false
false
275
7
#!/Users/marki/Desktop/Coding_Dojo/DojoAssignments/Python/mysql/full_friends/bin/python # -*- coding: utf-8 -*- import re import sys from pip import main if __name__ == '__main__': sys.argv[0] = re.sub(r'(-script\.pyw?|\.exe)?$', '', sys.argv[0]) sys.exit(main())
[ "minokuchi@gmail.com" ]
minokuchi@gmail.com
21bc2d0fbd981fbefdd919c846357da41182c5ac
e48eac671ea8335f696ec5fd0511b12800b0fca0
/accounts/models.py
6625a130b79945d6d4613bcc9a0b047c32339541
[]
no_license
linusidom/django-basic-reservation-system
8fba10708cebd42d5ad308c0ef838a0fe1ac8778
5dd5db4832fe83e862424af18aae3aad6cf3f1ed
refs/heads/master
2020-03-26T08:37:15.052362
2018-08-14T11:34:19
2018-08-14T11:34:19
144,711,607
2
0
null
null
null
null
UTF-8
Python
false
false
359
py
from django.db import models # Create your models here. from django.contrib.auth.models import AbstractUser from django.shortcuts import reverse class Profile(AbstractUser): ideal_weight = models.IntegerField(default=185) def __str__(self): return self.email def get_absolute_url(self): return reverse('accounts:profile_detail', kwargs={'pk':pk})
[ "linusidom@gmail.com" ]
linusidom@gmail.com
8180cf64858edff78040580e57e99bca34304e13
bd649f51496a24a55a2327e658f31d6e03e2f602
/InvTL/lm_py/py/bin/py.which
f2b30cfff3ced4c92ec7655ae03a05a1f0b7818b
[ "MIT" ]
permissive
mickg10/DARLAB
6507530231f749e8fc1647f3a9bec22a20bebe46
0cd8d094fcaf60a48a3b32f15e836fcb48d93e74
refs/heads/master
2020-04-15T20:39:23.403215
2019-01-10T06:54:50
2019-01-10T06:54:50
16,510,433
4
1
null
null
null
null
UTF-8
Python
false
false
65
which
#!/usr/bin/env python from _findpy import py py.cmdline.pywhich()
[ "root@darlab1.mickg.net" ]
root@darlab1.mickg.net
4d797ba65df470752fb7054a0f9213d8436220cc
ccfd41c1f6f90608f4c2596963a28303ab0bdee3
/test_standalone/test_rpc_client.py
ca8e4062e9ba10e622c152fed40fb4932da79473
[]
no_license
jsmoyam/zserver
4e8ffb4543f663a70ca96f1988958dba66b79d49
d838937cf92262176faec6a6a833a602dbcd868e
refs/heads/master
2022-12-23T01:21:29.200727
2019-04-30T11:05:53
2019-04-30T11:05:53
184,247,118
0
0
null
2022-12-12T10:22:38
2019-04-30T11:05:41
Python
UTF-8
Python
false
false
165
py
import Pyro4 # obj = Pyro4.Proxy("PYRONAME:test_module") # obj.example_method() # obj.initialize() obj = Pyro4.Proxy("PYRONAME:mytestserver") print(obj.m1('hola'))
[ "jsmoya@one-esecurity.com" ]
jsmoya@one-esecurity.com
65522eea8575a938acfbe8a6f92ea0e0fbd28d33
d36ea3f3a03299d0572914dd31c63a74c9da2c45
/listapar.py
217a2e0645c11b0ef1054b356536875010eb7a4c
[]
no_license
BrauCamacho/Mis_Trabajos_python
a05b3954751e37e5b2cbede0d8746faa774a7e1d
e6e0168b33e87c9b8924cf36441d0b585c3aee76
refs/heads/master
2022-11-30T11:42:43.520350
2020-08-15T04:52:05
2020-08-15T04:52:05
null
0
0
null
null
null
null
UTF-8
Python
false
false
541
py
lista = [] for i in range(2,21,2): lista.append(i) print(lista) #lo mismo pero comprimido lista2= [val for val in range(2,21,2)] print(lista2) valor = 0 #if ternario valor = 10 Edad = 10 if valor < 20 else 0 print (Edad) #strings nombre = " Juan " print(nombre[0:2]) print(nombre[:3]) print(nombre[-1]) print(nombre[::-1]) print(lista[::-1]) #print(nombre.suorted()) print(sorted(nombre)) print(nombre.upper()) print(nombre.lower()) print(nombre.split("a")) print(nombre.replace("a","A")) print(nombre.strip())
[ "noreply@github.com" ]
noreply@github.com
de2749d93e0cc76fcd54f2d8c178db2908200109
b92db3e3eae34b69dce099f78ebe8c0446439faf
/models.py
30e97d1a6a4e3ae8d3438702123b4f5d29f413e0
[]
no_license
gayathri44/Portfolio-project
06834ab0ca261d96f56e80809c4c138c1332dd69
260bf23700888fc18fcd0d808d57dd0d8d6650f9
refs/heads/main
2023-03-30T22:24:46.645209
2021-03-29T17:39:27
2021-03-29T17:39:27
352,729,848
0
0
null
null
null
null
UTF-8
Python
false
false
269
py
from django.db import models class Project(models.Model): title = models.CharField(max_length=100) description = models.TextField() technology = models.CharField(max_length=20) image = models.FilePathField(path="/img") # Create your models here.
[ "noreply@github.com" ]
noreply@github.com
ca7db0a3b82cc1c33311d1481cc649686e98ba8e
b8c8af33c8a1a7130c62b42cf22f2eeaf3d81427
/譌ァ/PUBLIC.py
ea18217cffe68043903b90b91aad46320da0bc0b
[]
no_license
natsukaze4545/nnnn
57e7a020373569d3ccf26e2e86e210c7aa4d3182
d0dc58d1fb8981c625aed204aa865992114e0144
refs/heads/master
2020-03-22T08:44:38.124591
2018-07-05T03:41:02
2018-07-05T03:41:02
139,787,776
0
0
null
null
null
null
UTF-8
Python
false
false
10,402
py
#SRSU LINE SELF BOT v7.0.3 # -*- coding: utf-8 -*- import LAG from LAG.lib.Gen.ttypes import * from datetime import datetime from LAG.lib.Gen.ttypes import * from datetime import datetime from LAPI.main import qr from threading import Thread import time,random,sys,re,os,json,subprocess,codecs,threading,glob,requests,string ############################################## cl = LAG.LINE() cl.login(token="Eq0xkbFlQkZuAuJZ9j70.3So3iW8wB5sEssSoxCD4ea.MRWAvyDMfS7ZWlOAMr9je7YP48464T4J80Xc/OOwWLg=") cl.loginResult() ############################################## reload(sys) sys.setdefaultencoding('utf-8') ############################################## helpMessage ="""PUBLIC BOT help ...コマンド一覧を送信します test ...動いているか確認します speed ...bot speed mid ...midを送信します csend: ...midから連絡先を送信します ginfo ...グループ情報を送信します g: ...googleで検索します search: ...検索 srsu.weebly.com""" ############################################## mid = cl.getProfile().mid profile = cl.getProfile() ############################################## wait2 = { 'readPoint':{}, 'readMember':{}, 'setTime':{}, 'ROM':{} } setTime = {} setTime = wait2['setTime'] ############################################## def Cmd(string, commands): #/XXX, >XXX, ;XXX, ^XXX, %XXX, $XXX... tex = [""] for texX in tex: for command in commands: if string ==texX + command: return True return False def sendMessage(to, text, contentMetadata={}, contentType=0): mes = Message() mes.to, mes.from_ = to, profile.mid mes.text = text mes.contentType, mes.contentMetadata = contentType, contentMetadata if to not in messageReq: messageReq[to] = -1 messageReq[to] += 1 def summon(to, nama): aa = "" bb = "" strt = int(14) akh = int(14) nm = nama for mm in nm: akh = akh + 2 aa += """{"S":"""+json.dumps(str(strt))+""","E":"""+json.dumps(str(akh))+""","M":"""+json.dumps(mm)+"},""" strt = strt + 6 akh = akh + 4 bb += "\xe2\x95\xa0 @x \n" aa = (aa[:int(len(aa)-1)]) x = Message() x.to = to x.text = "\xe2\x95\x94\xe2\x95\x90\xe2\x95\x90\xe2\x95\x90\xe2\x95\x90\xe2\x95\x90\xe2\x95\x90\xe2\x95\x90\xe2\x95\x90\xe2\x95\x90\xe2\x95\x90\n"+bb+"\xe2\x95\x9a\xe2\x95\x90\xe2\x95\x90\xe2\x95\x90\xe2\x95\x90\xe2\x95\x90\xe2\x95\x90\xe2\x95\x90\xe2\x95\x90\xe2\x95\x90\xe2\x95\x90" x.contentMetadata ={'MENTION':'{"MENTIONEES":['+aa+']}','EMTVER':'4'} try: cl.sendMessage(x) except Exception as error: print error ############################################## def bot(op): try: if op.type == 0: return if op.type == 26: x = op.message if x.contentType == 13: return if x.contentType == 0: if x.text == None: return ### command ################################## elif x.text.lower() == '.help': cl.sendText(x.to,helpMessage) elif x.text in [".test"]: cl.sendText(x.to,"ok") elif x.text in [".speed"]: start = time.time() cl.sendText(x.to, "...") elapsed_time = time.time() - start cl.sendText(x.to, "%s" % (elapsed_time)) elif x.text in [".version"]: cl.sendText(x.to,"PUBLIC") ############################################## elif x.text in [".mid"]: cl.sendText(x.to,mid) elif ".csend:" in x.text: mmid = x.text.replace("csend:","") x.contentType = 13 x.contentMetadata = {"mid":mmid} cl.sendMessage(x) ############################################## elif ".ginfo" == x.text: if x.toType == 2: g = cl.getGroup(x.to) try: gCreator = g.creator.displayName except: gCreator = "不明" if g.invitee is None: sinvitee = "0" else: sinvitee = str(len(g.invitee)) if g.preventJoinByTicket == True: QR = "拒否" else: QR = "許可" cl.sendText(x.to,"グループ名 : " + str(g.name) + "\ngid : " + x.to + "\n作成者 : " + gCreator + "\n\n招待URL :" + QR + "\nグループ画像 : http://dl.profile.line.naver.jp/" + g.pictureStatus + "\nメンバー数 : " + str(len(g.members)) + "\n招待数 : " + sinvitee) #全員をメンション(上が使えなかったらこっち) /allmention elif x.text in ["/all mention"]: group = cl.getGroup(x.to) nama = [contact.mid for contact in group.members] nm1, nm2, nm3, nm4, nm5, jml = [], [], [], [], [], len(nama) if jml <= 100: summon(x.to, nama) if jml > 100 and jml < 200: for i in range(0, 99): nm1 += [nama[i]] summon(x.to, nm1) for j in range(100, len(nama)-1): nm2 += [nama[j]] summon(x.to, nm2) if jml > 200 and jml < 500: for i in range(0, 99): nm1 += [nama[i]] summon(x.to, nm1) for j in range(100, 199): nm2 += [nama[j]] summon(x.to, nm2) for k in range(200, 299): nm3 += [nama[k]] summon(x.to, nm3) for l in range(300, 399): nm4 += [nama[l]] summon(x.to, nm4) for m in range(400, len(nama)-1): nm5 += [nama[m]] summon(x.to, nm5) if jml > 500: print "Terlalu Banyak Men 500+" cnt = Message() cnt.text = "メンションした人数:\n" + str(jml) + " 人" cnt.to = x.to cl.sendMessage(cnt) ############################################## elif ".g:" in x.text: s = x.text.replace("g:","") cl.sendText(x.to,"google.com/search?q=" + s.replace("g:","")) elif ".search:" in x.text: s = x.text.replace("search:","") cl.sendText(x.to,"\ngoogle : " + "google.com/search?q=" + s.replace("search:","") + "\ngoo : " + "search.goo.ne.jp/web.jsp?MT=" + s.replace("search:","") + "&mode=0&sbd=goo001&IE=UTF-8&OE=UTF-8&from=gootop&PT=TOP" + "\nbing : " + "bing.com/search?scope=web&q=" + s.replace("search:","") + "\nexcite : " + "websearch.excite.co.jp/?q=" + s.replace("search:","") + "\nnever : " + "matome.naver.jp/search?q=" + s.replace("search:","") + "\n楽天 : " + "websearch.rakuten.co.jp/WebIS?col=OW&svx=100610&nc=1&lg=all&svp=SEEK&enc=UTF-8&qt=" + s.replace("search:","") + "\nBIGLOBE : " + "cgi.search.biglobe.ne.jp/cgi-bin/search2-b?search=検索&q=" + s.replace("search:","") + "\nlivedoor : " + "search.livedoor.com/search?ie=utf-8&q=%E6%A4%9C%E7%B4%A2&search_btn=1" + s.replace("search:","") + "\nnifty : " + "search.nifty.com/websearch/search?select=2&ss=nifty_top_tp&cflg=検索&q=" + s.replace("search:","") + "\nyahoo : " + "search.yahoo.co.jp/search;_ylt=A2RA5ONbn7NaBjEAmGCJBtF7?p=" + s.replace("search:","")) #-------------------------------------------------------------------------------#lastseen elif x.text == "ぽいんと": cl.sendText(x.to, "既読ポイントを設定しました。確認したい場合は「ちぇっく」と送信してください。") try: del wait2['readPoint'][x.to] del wait2['readMember'][x.to] except: pass now2 = datetime.now() wait2['readPoint'][x.to] = x.id wait2['readMember'][x.to] = "" wait2['setTime'][x.to] = datetime.strftime(now2,"%H:%M") wait2['ROM'][x.to] = {} print wait2 elif x.text == "ちぇっく": if x.to in wait2['readPoint']: if wait2["ROM"][x.to].items() == []: chiya = "" else: chiya = "" for rom in wait2["ROM"][x.to].items(): print rom chiya += rom[1] + "\n" cl.sendText(x.to,"きどく %s\nきどくむし\n%s\nせっていじこく:\n[%s]" % (wait2['readMember'][x.to],chiya,setTime[x.to])) else: cl.sendText(x.to,"既読ポイントが設定されていません。「ぽいんと」と送信して既読ポイントを設定してください") #-------------------------------------------------------#search if op.type == 55: if op.param1 in wait2['readPoint']: Name = cl.getContact(op.param2).displayName if Name in wait2['readMember'][op.param1]: pass else: wait2['readMember'][op.param1] += "\n・" + Name wait2['ROM'][op.param1][op.param2] = "・" + Name else: cl.sendText if op.type == 59: print op except Exception as error: print error while True: try: Ops = cl.fetchOps(cl.Poll.rev, 5) except EOFError: raise Exception("It might be wrong revision\n" + str(cl.Poll.rev)) for Op in Ops: if (Op.type != OpType.END_OF_OPERATION): cl.Poll.rev = max(cl.Poll.rev, Op.revision) bot(Op)
[ "114514kusai@usako.net" ]
114514kusai@usako.net
eb78acba4bf3287ebe00fb95622aa68a63eb3407
5e5dfefcd80df20efa07bf101d202902921ba40d
/scripts/bam2fna
0c7af193443d72b98092cab452115f4a0076e792
[]
no_license
veg/BioExt
ba8539361dffd6a366cf174b938c00e28c6649c7
5d0e283553dcc0016567182dae59e4b4994b505f
refs/heads/master
2023-08-31T11:29:56.342017
2023-06-13T20:56:40
2023-06-13T20:56:40
20,779,973
1
11
null
2023-08-11T21:55:30
2014-06-12T19:20:46
Python
UTF-8
Python
false
false
1,120
#!/usr/bin/env python3 import signal from Bio import SeqIO from BioExt.io import BamIO def main(bam_file, out_handle): try: signal.signal(signal.SIGPIPE, signal.SIG_DFL) except ValueError: pass SeqIO.write(BamIO.parse(bam_file), out_handle, 'fasta') return 0 if __name__ == '__main__': import sys import argparse parser = argparse.ArgumentParser( description='convert a BAM file to a FASTA file' ) parser.add_argument( 'input', metavar='INPUT', type=argparse.FileType('rb'), help='input BAM file' ) parser.add_argument( 'output', default=sys.stdout, metavar='OUTPUT', type=argparse.FileType('w'), help='output FASTA file' ) args = None retcode = -1 try: args = parser.parse_args() bam_file = args.input.name args.input.close() retcode = main(bam_file, args.output) finally: if args is not None: if args.output != sys.stdout: args.output.close() sys.exit(retcode)
[ "nlhepler@gmail.com" ]
nlhepler@gmail.com
5497eed0b98d3d44dc25ed39c7376e7800f9fcaa
350cb6c7c7a7842e80aa06ee32bfffc5bc35ee03
/programming/language/python/python-pillow/actions.py
09179eb1424c0d20883d92c49aeb6480d96ba765
[]
no_license
LimeLinux/Packages-2
f41d11343e8b39274ccd85b9850d0f4e76830031
356975df129f2097f12dbed3bc2604cadb5a6c64
refs/heads/master
2021-04-30T23:25:31.121967
2017-01-21T21:46:54
2017-01-21T21:46:54
79,139,920
0
2
null
2017-01-21T21:46:55
2017-01-16T17:02:37
Python
UTF-8
Python
false
false
823
py
#!/usr/bin/python # -*- coding: utf-8 -*- # # Licensed under the GNU General Public License, version 3. # See the file http://www.gnu.org/licenses/gpl.txt from pisi.actionsapi import pisitools from pisi.actionsapi import pythonmodules from pisi.actionsapi import shelltools from pisi.actionsapi import get #WorkDir="Imaging-%s" % get.srcVERSION() def install(): pisitools.dosed("_imagingft.c", "<freetype/freetype.h>", "<freetype2/freetype.h>") pisitools.dosed("_imagingft.c", "<freetype/fterrors.h>", "<freetype2/fterrors.h>") pythonmodules.install() #shelltools.cd("Sane") #pythonmodules.install() #shelltools.cd("..") for header in ["Imaging.h","ImPlatform.h"]: pisitools.insinto("/usr/include/%s" % get.curPYTHON(), "libImaging/%s" % header) pisitools.dodoc("README.rst")
[ "ergunsalman@hotmail.com" ]
ergunsalman@hotmail.com
0593e05afd2470b385d6c0f363d7d4ef5f18a772
b770761d348794b86ac0d87b16750625eadc176d
/est_imp_from_file.py
84100f6ca75cc3a32dc41421fcdfe72d6e81bb47
[ "MIT" ]
permissive
fujiyuer-fujiyu/imp_tsp
51855f722cea1864ca32adf66dbec42c3d4bcc8e
4d19a28fc0dea4d81c38dcfb0ddd2800513257cb
refs/heads/master
2020-06-20T19:40:05.935343
2018-03-02T03:12:41
2018-03-02T03:12:41
74,744,644
0
0
null
2016-11-25T09:34:19
2016-11-25T09:34:18
null
UTF-8
Python
false
false
1,413
py
from scipy.io import wavfile from scipy import signal as sig import numpy as np import matplotlib.pyplot as plt #input file (observed signal) (rate,data)=wavfile.read("tsp_out.16.wav") #(rate,data)=wavfile.read("./20180228/ref/ec_tsp.wav") i_length=data.shape[0] print "lenght of Observed signal and number of channels is {0}".format(str(data.shape)) #inverse of TSP signal (rate,data2)=wavfile.read("./itsp_out.wav") print "length of Original TSP signal is {}".format(str(data2.shape)) length = data2.shape[0] #number of chanel if len(data.shape) > 1: nc = data.shape[1] else : nc = 1 print "Number of channel of Observed signal is {}".format(nc) ## synchronous addition if nc > 1: data_s = np.zeros((length,nc)) else: data_s = np.zeros(length) print data_s.shape for s in range(0,int(i_length/length)): print str(s)+"th addition.. " for i in range(0,length): if nc > 1: data_s[i][nc] += data[i+s*length][nc] else: data_s[i] += data[i+s*length] # calc inpulse response data4_f=np.fft.fft(data_s,length)*np.fft.fft(data2,length) data4=np.fft.ifft(data4_f) data4_i = np.zeros(length, dtype=np.int16) # normalize and convert to integer data4_i = 9000*data4.real/np.max(data4.real) data4_i = data4_i.astype(np.int16) wavfile.write("./IR-from-FFT.wav",rate,data4_i) plt.figure() plt.title("IR: estimated from FFT") plt.plot(data4_i) plt.show()
[ "yuyfujit@mbp-15uas-012.yjoffice.local" ]
yuyfujit@mbp-15uas-012.yjoffice.local
c63f6c71799ea453d1f3eec67be2aff4089d9177
bc41457e2550489ebb3795f58b243da74a1c27ae
/fabfile.py
36e45dab0037e8a64b682e70626dadcb3e9d14de
[]
no_license
SEL-Columbia/ss_sql_views
28a901d95fe779b278d2a51aec84d6bf51245c02
d146fd96849a4d165f3dc3f197aadda804a2f60a
refs/heads/master
2021-01-01T19:35:18.999147
2012-05-10T18:43:36
2012-05-10T18:43:36
3,020,367
0
0
null
null
null
null
UTF-8
Python
true
false
1,649
py
''' fabfile for offline gateway tasks ''' import datetime as dt from fabric.api import local, lcd, run, env env.hosts = ['gateway.sharedsolar.org'] env.user = 'root' def sync_db(): time = dt.datetime.now().strftime('%y%m%d') file = 'gateway.' + time + '.sql.zip' url = 'root@gateway.sharedsolar.org' path = 'var/lib/postgresql/backups/' local('mkdir temp') with lcd('temp'): download_db(url, path, file) load_db(path, file) create_views() local('rm -rf temp') show_disk_space() def download_db(url, path, file): # create local temp folder print 'Creating temporary folder ./temp' # create timestamp # create string for getting database # scp database print 'Downloading database from gateway' local('scp ' + url + ':/' + path + file + ' .') # locally unzip database print 'Expanding database' local('unzip ' + file) def load_db(path, file): # if database exists, dropdb local('dropdb gateway') # create db local('createdb gateway') # load database print 'Loading database' local('psql -d gateway -f ' + path + file[:-4]) def create_views(): print 'Executing create_views' # execute all sql files local('psql -d gateway -f views/create_view_primary_log.sql') local('psql -d gateway -f views/create_view_midnight.sql') local('psql -d gateway -f views/create_view_meter.sql') local('psql -d gateway -f views/create_view_alarms.sql') local('psql -d gateway -f views/create_view_solar.sql') local('psql -d gateway -f views/create_view_recharge.sql') def show_disk_space(): run('df -h')
[ "danielrsoto@gmail.com" ]
danielrsoto@gmail.com
e08c3ff7f96aee19729d67accd05c2b3f7ffbdf7
49fb950bafeeac754107a6c962f8ea1a1b5212d2
/app/database.py
98872f91a5824945704474bf2c4430fc9e34d0fc
[ "LicenseRef-scancode-warranty-disclaimer" ]
no_license
VOC-Electronics/WiFiDomo_Manager
5a50fcaf59e88172dc15b478b06deca3e35feee3
b7f319fa0d8f5c7328948a40962cbed6abffcf25
refs/heads/master
2023-08-10T09:02:17.785967
2016-08-30T09:40:30
2016-08-30T09:40:30
60,178,446
0
1
null
2016-08-30T09:40:30
2016-06-01T13:17:25
Python
UTF-8
Python
false
false
7,259
py
#!/usr/bin/python # -*- coding: utf-8 -*- __author__ = 'Martijn van Leeuwen' __email__ = 'info@voc-electronics.com' ''' # =[ DISCLAIMER ]=============================================================== # # THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR # IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, # FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE # AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER # LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, # OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN # THE SOFTWARE. # # ============================================================================== # # Todo: # # Get data from input parameters. # Check connections # Login # Update # ============================================================================== # Imports # ============================================================================== ''' from datetime import datetime from sqlalchemy import create_engine, Column, Integer, String, DateTime, Boolean, \ ForeignKey, event from sqlalchemy.orm import scoped_session, sessionmaker, backref, relation from sqlalchemy.ext.declarative import declarative_base from werkzeug import cached_property, http_date from flask import url_for, Markup from app.wifidomo_manager import app, db from modules.config import config # =========================================================================== # Global settings # =========================================================================== engine = create_engine(app.config[ 'DATABASE_URI' ], convert_unicode=True, **app.config[ 'DATABASE_CONNECT_OPTIONS' ]) db_session = scoped_session(sessionmaker(autocommit=False, autoflush=False, bind=engine)) def init_db(): Base.metadata.create_all(bind=engine) Base = declarative_base(name='Base') Base.query = db_session.query_property() class Person(Base): __tablename__ = 'person' # Here we define columns for the table person # Notice that each column is also a normal Python instance attribute. id = Column(Integer, primary_key=True) surname = Column(String, nullable=False) lastname = Column(String, nullable=False) fullname = Column(String, nullable=False) loginid = Column(String, nullable=True) password = Column(String, nullable=False) email = Column(String, nullable=True) created = Column(DateTime, default=datetime.utcnow, onupdate=datetime.utcnow) updated_on = Column(DateTime, default=datetime.utcnow, onupdate=datetime.utcnow) # authenticated = Column(db.Boolean, default=False) def __init__(self, surname, lastname, fullname, password): self.fullname = fullname self.surname = surname self.lastname = lastname self.created = datetime.utcnow() self.password = password def __repr__(self): return '<User %r>' % (self.name) def is_active(self): """True, as all users are active.""" return True def get_id(self): """Return the email address to satisfy Flask-Login's requirements.""" return self.email def get_email(self): return self.email def get_lastname(self): return self.lastname def get_surname(self): return self.surname def is_authenticated(self): """Return True if the user is authenticated.""" return self.authenticated def is_anonymous(self): """False, as anonymous users aren't supported.""" return False class WiFiDomo(Base): __tablename__ = 'wifidomo' id = Column(Integer, primary_key=True) name = Column(String, index=True, nullable=True) MAC = Column(String, nullable=True) locationid = Column(Integer, nullable=True) ip4 = Column(String(16), nullable=True) ip6 = Column(String, nullable=True) port = Column(Integer, nullable=True, default=80) fqdn = Column(String, nullable=True) status = Column(Boolean, default=False) powerstatus = Column(Boolean, default=False) last_used_r = Column(Integer) last_used_g = Column(Integer) last_used_b = Column(Integer) last_used_preset = Column(Integer) created = Column(DateTime, default=datetime.utcnow, onupdate=datetime.utcnow) updated_on = Column(DateTime, default=datetime.utcnow, onupdate=datetime.utcnow) def __init__(self, name, MAC, location_id, fqdn, status, ip4, ip6=0, port=80): self.name = name self.MAC = MAC self.locationid = location_id self.fqdn = fqdn self.ip4 = ip4 self.ip6 = ip6 self.port = port self.status = status self.powerstatus = False self.created = datetime.utcnow() self.updated_on = datetime.utcnow() def to_json(self): return dict( name=self.name, MAC=self.MAC, status=self.status, powerstatus=self.powerstatus, fqdn=self.fqdn, ip4=self.ip4, ip6=self.ip6, port=self.port, last_used_rgb=self.last_used_rgb ) @cached_property def count(self): return self.wifidomo.count() class WiFiNetworks(Base): __tablename__ = 'wifinetworks' id = Column(Integer, primary_key=True) wifi_sid = Column(String(128), nullable=True) wifi_loc = Column(String(128), nullable=True) created = Column(DateTime) def __init__(self, wifi_sid, wifi_loc): self.wifi_sid = wifi_sid self.wifi_loc = wifi_loc self.created = datetime.utcnow() def to_json(self): return dict(wifi_sid=self.wifi_sid, wifi_loc=self.wifi_loc) class Locations(Base): __tablename__ = 'locations' id = Column(Integer, primary_key=True) location_name = Column(String, unique=True, nullable=False) location_code = Column(Integer, default=0) location_description = Column(String, nullable=True) created = Column(DateTime) def __init__(self, location_name, location_code, location_description): self.location_name = location_name self.location_code = location_code self.location_description = location_description self.created = datetime.utcnow() def to_json(self): return dict(name=self.location_name, code=self.location_code, description=self.location_description) class Loginlog(Base): __tablename__ = 'loginlog' id = Column(Integer, primary_key=True) loginby = Column(String, nullable=True) logindate = Column(DateTime) class Preset(Base): __tablename__ = 'preset' id = Column(Integer, primary_key=True) name = Column(String, nullable=False) r_code = Column(Integer, nullable=False) g_code = Column(Integer, nullable=False) b_code = Column(Integer, nullable=False) created = Column(DateTime) def __init__(self, name, r_code, g_code, b_code): self.name = name self.r_code = r_code self.g_code = g_code self.b_code = b_code self.created = datetime.utcnow() def to_json(self): return dict(name=self.name, r_code=self.r_code, g_code=self.g_code, b_code=self.b_code) class Pattern(Base): __tablename__ = 'pattern' id = Column(Integer, primary_key=True) name = Column(String(200), nullable=False) def __init__(self, name): self.name = name
[ "info@voc-electronics.com" ]
info@voc-electronics.com
ca725619d876cdd4db8f6f079b26a3b27c7f51e9
e16a00326983750985f0d308b2b8cb156d2b7048
/DesignPatterns/Strategy.py
241764acd6a53fea462f8e5394fa31ed73a4ded5
[]
no_license
farabbit/learning_area
ff6b6cee6de5551a58068dd73b34f427903d3973
e553186ececb7ef2149356465044138cd84588c5
refs/heads/master
2020-05-22T22:45:58.115383
2019-09-27T08:23:13
2019-09-27T08:23:13
186,551,926
0
0
null
null
null
null
UTF-8
Python
false
false
2,838
py
import abc """ Traditional way """ # every strategy is as class that inherits base stategy print("Traditional: ") class Strategy: """ Base Strategy """ @abc.abstractmethod def operation(self, context): pass # concrate stategies class StrategyAdd(Strategy): def operation(self, context): return context.intA + context.intB class StrategyMultiple(Strategy): def operation(self, context): return context.intA * context.intB class Context: """ class that using Strategy """ def __init__(self, intA, intB): self.intA, self.intB = intA, intB def setStrategy(self, strategy): self.strategy = strategy def executeStrategy(self): return self.strategy.operation(self) context = Context(2, 3) context.setStrategy(StrategyAdd()) # using add print("ADD:", context.executeStrategy()) # ADD: 5 context.setStrategy(StrategyMultiple()) # using multiple print("MUL:", context.executeStrategy()) # MUL: 6 """ Restruct using functional programming """ # concrete stragegies had no internal status -> performanced like functions print("Restructed: ") def StrategyAdd_res(context): return context.intA + context.intB def StrategyMultiple_res(context): return context.intA * context.intB class Context_res: """ class that using Strategy """ def __init__(self, intA, intB): self.intA, self.intB = intA, intB def setStrategy(self, strategy): # syntax keeps the same but strategy here is function now, not a class self.strategy = strategy def executeStrategy(self): return self.strategy(self) context_res = Context_res(2, 3) context_res.setStrategy(StrategyAdd_res) # using add print("ADD:", context_res.executeStrategy()) # ADD: 5 context_res.setStrategy(StrategyMultiple_res) # using multiple print("MUL:", context_res.executeStrategy()) # MUL: 6 """ best strategy """ # additional functionalities print("Choose best strategy among all: ") strategies = (StrategyAdd_res,StrategyMultiple_res) def bestStrategy(context): """ choose a best strategy from strategy list """ return max(strategy(context) for strategy in strategies) print("MAX(ADD(2,3), MUL(2,3)):", bestStrategy(Context_res(2,3))) # 6 """ Find all strategies """ # using globals(): returns every global variable that in current module (module that defines that function/method) strategies = (globals()[name] for name in globals() if name.endswith('_res') and name.startswith('Strategy')) print(list(strategies)) # [StrategyAdd_res,StrategyMultiple_res] """out: Traditional ADD: 5 MUL: 6 Restructed: ADD: 5 MUL: 6 Choose strategy among all MAX(ADD(2,3), MUL(2,3)): 6 [<function StrategyAdd_res at 0x00000299A4A33E18>, <function StrategyMultiple_res at 0x00000299A4D5B378>] """
[ "noreply@github.com" ]
noreply@github.com
c33d59f6b5de8274dc27c8aaca734bc82148e904
6a6ca78209adc4e342529d9d3d748f664b2fb511
/run.py
a7a9504931696e6ad18b1b4a377b020d113d6ccd
[]
no_license
feibl/geometa_search
88b641a38699bf9ab3af4ef1153c2ee57356ef79
a4afca92e6639617bcb905ebe3351f0fc3fb440f
refs/heads/master
2021-01-02T08:57:26.283601
2015-01-30T10:58:29
2015-01-30T10:58:29
26,922,245
0
1
null
null
null
null
UTF-8
Python
false
false
643
py
from geometa_search.factory import create_app as create_frontend from search_rex.factory import create_app as create_backend from search_rex.recommendations import create_recommender_system from werkzeug.wsgi import DispatcherMiddleware from werkzeug.serving import run_simple backend = create_backend('config.DevelopmentConfig') create_recommender_system(backend) frontend = create_frontend('config.DevelopmentConfig') app = DispatcherMiddleware( frontend, { '/recommender': backend } ) if __name__ == '__main__': run_simple( 'localhost', 5000, app, use_reloader=True, use_debugger=True, threaded=True)
[ "fabian.senn@gmail.com" ]
fabian.senn@gmail.com
c67befeec81e23a656e21cce799a95202898726a
51d602577affebc8d91ffe234f926469d389dc75
/lis/specimen/lab_result_item/migrations/0001_initial.py
69d94bf24a380776b43692bb1088907fc36bac4f
[]
no_license
botswana-harvard/lis
5ac491373f74eaf3855f173580b000539d7f4740
48dc601ae05e420e8f3ebb5ea398f44f02b2e5e7
refs/heads/master
2020-12-29T01:31:07.821681
2018-06-24T06:06:57
2018-06-24T06:06:57
35,820,860
0
0
null
null
null
null
UTF-8
Python
false
false
40,300
py
# encoding: utf-8 import datetime from south.db import db from south.v2 import SchemaMigration from django.db import models class Migration(SchemaMigration): def forwards(self, orm): # Adding model 'ResultItemAudit' db.create_table('bhp_lab_core_resultitem_audit', ( ('created', self.gf('django.db.models.fields.DateTimeField')(default=datetime.datetime.now, blank=True)), ('modified', self.gf('django.db.models.fields.DateTimeField')(default=datetime.datetime.now, blank=True)), ('user_created', self.gf('django.db.models.fields.CharField')(default='', max_length=250)), ('user_modified', self.gf('django.db.models.fields.CharField')(default='', max_length=250)), ('hostname_created', self.gf('django.db.models.fields.CharField')(default='home', max_length=50, blank=True)), ('hostname_modified', self.gf('django.db.models.fields.CharField')(default='home', max_length=50, blank=True)), ('id', self.gf('django.db.models.fields.CharField')(max_length=36, blank=True)), ('result', self.gf('django.db.models.fields.related.ForeignKey')(related_name='_audit_resultitem', to=orm['lab_result.Result'])), ('test_code', self.gf('django.db.models.fields.related.ForeignKey')(related_name='_audit_resultitem', to=orm['lab_test_code.TestCode'])), ('result_item_value', self.gf('django.db.models.fields.CharField')(max_length=25, db_index=True)), ('result_item_quantifier', self.gf('django.db.models.fields.CharField')(default='=', max_length=25)), ('result_item_datetime', self.gf('django.db.models.fields.DateTimeField')(db_index=True)), ('result_item_operator', self.gf('django.db.models.fields.CharField')(db_index=True, max_length=50, null=True, blank=True)), ('validation_status', self.gf('django.db.models.fields.CharField')(default='P', max_length=10, db_index=True)), ('validation_datetime', self.gf('django.db.models.fields.DateTimeField')(db_index=True, null=True, blank=True)), ('validation_username', self.gf('django.db.models.fields.CharField')(db_index=True, max_length=50, null=True, blank=True)), ('validation_reference', self.gf('django.db.models.fields.CharField')(max_length=50, null=True, blank=True)), ('comment', self.gf('django.db.models.fields.CharField')(max_length=50, null=True, blank=True)), ('result_item_source', self.gf('django.db.models.fields.related.ForeignKey')(related_name='_audit_resultitem', to=orm['lab_result.ResultSource'])), ('result_item_source_reference', self.gf('django.db.models.fields.CharField')(max_length=50, null=True, blank=True)), ('error_code', self.gf('django.db.models.fields.CharField')(max_length=50, null=True, blank=True)), ('_audit_id', self.gf('django.db.models.fields.AutoField')(primary_key=True)), ('_audit_timestamp', self.gf('django.db.models.fields.DateTimeField')(auto_now_add=True, db_index=True, blank=True)), ('_audit_change_type', self.gf('django.db.models.fields.CharField')(max_length=1)), )) db.send_create_signal('lab_result_item', ['ResultItemAudit']) # Adding model 'ResultItem' db.create_table('bhp_lab_core_resultitem', ( ('created', self.gf('django.db.models.fields.DateTimeField')(default=datetime.datetime.now, blank=True)), ('modified', self.gf('django.db.models.fields.DateTimeField')(default=datetime.datetime.now, blank=True)), ('user_created', self.gf('django.db.models.fields.CharField')(default='', max_length=250)), ('user_modified', self.gf('django.db.models.fields.CharField')(default='', max_length=250)), ('hostname_created', self.gf('django.db.models.fields.CharField')(default='home', max_length=50, blank=True)), ('hostname_modified', self.gf('django.db.models.fields.CharField')(default='home', max_length=50, blank=True)), ('id', self.gf('django.db.models.fields.CharField')(max_length=36, primary_key=True)), ('result', self.gf('django.db.models.fields.related.ForeignKey')(to=orm['lab_result.Result'])), ('test_code', self.gf('django.db.models.fields.related.ForeignKey')(to=orm['lab_test_code.TestCode'])), ('result_item_value', self.gf('django.db.models.fields.CharField')(max_length=25, db_index=True)), ('result_item_quantifier', self.gf('django.db.models.fields.CharField')(default='=', max_length=25)), ('result_item_datetime', self.gf('django.db.models.fields.DateTimeField')(db_index=True)), ('result_item_operator', self.gf('django.db.models.fields.CharField')(db_index=True, max_length=50, null=True, blank=True)), ('validation_status', self.gf('django.db.models.fields.CharField')(default='P', max_length=10, db_index=True)), ('validation_datetime', self.gf('django.db.models.fields.DateTimeField')(db_index=True, null=True, blank=True)), ('validation_username', self.gf('django.db.models.fields.CharField')(db_index=True, max_length=50, null=True, blank=True)), ('validation_reference', self.gf('django.db.models.fields.CharField')(max_length=50, null=True, blank=True)), ('comment', self.gf('django.db.models.fields.CharField')(max_length=50, null=True, blank=True)), ('result_item_source', self.gf('django.db.models.fields.related.ForeignKey')(to=orm['lab_result.ResultSource'])), ('result_item_source_reference', self.gf('django.db.models.fields.CharField')(max_length=50, null=True, blank=True)), ('error_code', self.gf('django.db.models.fields.CharField')(max_length=50, null=True, blank=True)), )) db.send_create_signal('lab_result_item', ['ResultItem']) def backwards(self, orm): # Deleting model 'ResultItemAudit' db.delete_table('bhp_lab_core_resultitem_audit') # Deleting model 'ResultItem' db.delete_table('bhp_lab_core_resultitem') models = { 'bhp_research_protocol.fundingsource': { 'Meta': {'ordering': "['name']", 'object_name': 'FundingSource'}, 'description': ('django.db.models.fields.TextField', [], {'max_length': '500'}), 'id': ('django.db.models.fields.AutoField', [], {'primary_key': 'True'}), 'name': ('django.db.models.fields.CharField', [], {'max_length': '25'}), 'reference': ('django.db.models.fields.CharField', [], {'max_length': '25', 'blank': 'True'}) }, 'bhp_research_protocol.location': { 'Meta': {'ordering': "['name']", 'object_name': 'Location'}, 'id': ('django.db.models.fields.AutoField', [], {'primary_key': 'True'}), 'name': ('django.db.models.fields.CharField', [], {'unique': 'True', 'max_length': '25'}) }, 'bhp_research_protocol.protocol': { 'Meta': {'ordering': "['protocol_identifier']", 'object_name': 'Protocol'}, 'date_opened': ('django.db.models.fields.DateField', [], {}), 'date_registered': ('django.db.models.fields.DateField', [], {}), 'description': ('django.db.models.fields.TextField', [], {'max_length': '500'}), 'funding_source': ('django.db.models.fields.related.ManyToManyField', [], {'to': "orm['bhp_research_protocol.FundingSource']", 'symmetrical': 'False'}), 'id': ('django.db.models.fields.AutoField', [], {'primary_key': 'True'}), 'local_title': ('django.db.models.fields.CharField', [], {'max_length': '25', 'blank': 'True'}), 'protocol_identifier': ('django.db.models.fields.CharField', [], {'max_length': '25', 'null': 'True'}), 'research_title': ('django.db.models.fields.TextField', [], {'max_length': '250'}), 'short_title': ('django.db.models.fields.CharField', [], {'max_length': '25'}), 'site_name_fragment': ('django.db.models.fields.CharField', [], {'max_length': '25'}) }, 'bhp_research_protocol.site': { 'Meta': {'ordering': "['site_identifier']", 'object_name': 'Site'}, 'id': ('django.db.models.fields.AutoField', [], {'primary_key': 'True'}), 'location': ('django.db.models.fields.related.ForeignKey', [], {'to': "orm['bhp_research_protocol.Location']"}), 'name': ('django.db.models.fields.CharField', [], {'unique': 'True', 'max_length': '25'}), 'site_identifier': ('django.db.models.fields.CharField', [], {'unique': 'True', 'max_length': '25'}) }, 'lab_account.account': { 'Meta': {'ordering': "['account_name']", 'object_name': 'Account', 'db_table': "'bhp_lab_registration_account'"}, 'account_closedate': ('django.db.models.fields.DateField', [], {}), 'account_holder': ('django.db.models.fields.related.ForeignKey', [], {'to': "orm['lab_account.AccountHolder']"}), 'account_name': ('django.db.models.fields.CharField', [], {'unique': 'True', 'max_length': '25'}), 'account_opendate': ('django.db.models.fields.DateField', [], {}), 'comment': ('django.db.models.fields.CharField', [], {'max_length': '250', 'blank': 'True'}), 'created': ('django.db.models.fields.DateTimeField', [], {'default': 'datetime.datetime.now', 'blank': 'True'}), 'hostname_created': ('django.db.models.fields.CharField', [], {'default': "'home'", 'max_length': '50', 'blank': 'True'}), 'hostname_modified': ('django.db.models.fields.CharField', [], {'default': "'home'", 'max_length': '50', 'blank': 'True'}), 'id': ('django.db.models.fields.CharField', [], {'max_length': '36', 'primary_key': 'True'}), 'modified': ('django.db.models.fields.DateTimeField', [], {'default': 'datetime.datetime.now', 'blank': 'True'}), 'user_created': ('django.db.models.fields.CharField', [], {'default': "''", 'max_length': '250'}), 'user_modified': ('django.db.models.fields.CharField', [], {'default': "''", 'max_length': '250'}) }, 'lab_account.accountholder': { 'Meta': {'ordering': "['last_name', 'first_name']", 'unique_together': "(['last_name', 'first_name'],)", 'object_name': 'AccountHolder', 'db_table': "'bhp_lab_registration_accountholder'"}, 'comment': ('django.db.models.fields.TextField', [], {'max_length': '100', 'blank': 'True'}), 'created': ('django.db.models.fields.DateTimeField', [], {'default': 'datetime.datetime.now', 'blank': 'True'}), 'first_name': ('django.db.models.fields.CharField', [], {'max_length': '50'}), 'hostname_created': ('django.db.models.fields.CharField', [], {'default': "'home'", 'max_length': '50', 'blank': 'True'}), 'hostname_modified': ('django.db.models.fields.CharField', [], {'default': "'home'", 'max_length': '50', 'blank': 'True'}), 'id': ('django.db.models.fields.CharField', [], {'max_length': '36', 'primary_key': 'True'}), 'initials': ('django.db.models.fields.CharField', [], {'max_length': '3'}), 'last_name': ('django.db.models.fields.CharField', [], {'max_length': '50'}), 'modified': ('django.db.models.fields.DateTimeField', [], {'default': 'datetime.datetime.now', 'blank': 'True'}), 'user_created': ('django.db.models.fields.CharField', [], {'default': "''", 'max_length': '250'}), 'user_modified': ('django.db.models.fields.CharField', [], {'default': "''", 'max_length': '250'}) }, 'lab_aliquot.aliquot': { 'Meta': {'object_name': 'Aliquot', 'db_table': "'bhp_lab_core_aliquot'"}, 'aliquot_datetime': ('django.db.models.fields.DateTimeField', [], {'default': 'datetime.datetime(2011, 8, 30, 13, 11, 14, 896689)'}), 'aliquot_identifier': ('django.db.models.fields.CharField', [], {'unique': 'True', 'max_length': '25'}), 'aliquot_type': ('django.db.models.fields.related.ForeignKey', [], {'to': "orm['lab_aliquot.AliquotType']"}), 'comment': ('django.db.models.fields.CharField', [], {'max_length': '50', 'null': 'True', 'blank': 'True'}), 'condition': ('django.db.models.fields.related.ForeignKey', [], {'default': '10', 'to': "orm['lab_aliquot.AliquotCondition']"}), 'count': ('django.db.models.fields.IntegerField', [], {'null': 'True'}), 'created': ('django.db.models.fields.DateTimeField', [], {'default': 'datetime.datetime.now', 'blank': 'True'}), 'current_measure': ('django.db.models.fields.DecimalField', [], {'default': "'5.00'", 'max_digits': '10', 'decimal_places': '2'}), 'hostname_created': ('django.db.models.fields.CharField', [], {'default': "'home'", 'max_length': '50', 'blank': 'True'}), 'hostname_modified': ('django.db.models.fields.CharField', [], {'default': "'home'", 'max_length': '50', 'blank': 'True'}), 'id': ('django.db.models.fields.CharField', [], {'max_length': '36', 'primary_key': 'True'}), 'measure_units': ('django.db.models.fields.CharField', [], {'default': "'mL'", 'max_length': '25'}), 'medium': ('django.db.models.fields.CharField', [], {'default': "'TUBE'", 'max_length': '25'}), 'modified': ('django.db.models.fields.DateTimeField', [], {'default': 'datetime.datetime.now', 'blank': 'True'}), 'original_measure': ('django.db.models.fields.DecimalField', [], {'default': "'5.00'", 'max_digits': '10', 'decimal_places': '2'}), 'parent_identifier': ('django.db.models.fields.related.ForeignKey', [], {'to': "orm['lab_aliquot.Aliquot']", 'to_field': "'aliquot_identifier'", 'null': 'True', 'blank': 'True'}), 'receive': ('django.db.models.fields.related.ForeignKey', [], {'to': "orm['lab_receive.Receive']"}), 'status': ('django.db.models.fields.CharField', [], {'default': "'available'", 'max_length': '25'}), 'user_created': ('django.db.models.fields.CharField', [], {'default': "''", 'max_length': '250'}), 'user_modified': ('django.db.models.fields.CharField', [], {'default': "''", 'max_length': '250'}) }, 'lab_aliquot.aliquotcondition': { 'Meta': {'ordering': "['short_name']", 'object_name': 'AliquotCondition', 'db_table': "'bhp_lab_core_aliquotcondition'"}, 'created': ('django.db.models.fields.DateTimeField', [], {'default': 'datetime.datetime.now', 'blank': 'True'}), 'display_index': ('django.db.models.fields.IntegerField', [], {'default': '0'}), 'field_name': ('django.db.models.fields.CharField', [], {'max_length': '25', 'null': 'True', 'blank': 'True'}), 'hostname_created': ('django.db.models.fields.CharField', [], {'default': "'home'", 'max_length': '50', 'blank': 'True'}), 'hostname_modified': ('django.db.models.fields.CharField', [], {'default': "'home'", 'max_length': '50', 'blank': 'True'}), 'id': ('django.db.models.fields.AutoField', [], {'primary_key': 'True'}), 'modified': ('django.db.models.fields.DateTimeField', [], {'default': 'datetime.datetime.now', 'blank': 'True'}), 'name': ('django.db.models.fields.CharField', [], {'unique': 'True', 'max_length': '250'}), 'short_name': ('django.db.models.fields.CharField', [], {'unique': 'True', 'max_length': '250'}), 'user_created': ('django.db.models.fields.CharField', [], {'default': "''", 'max_length': '250'}), 'user_modified': ('django.db.models.fields.CharField', [], {'default': "''", 'max_length': '250'}), 'version': ('django.db.models.fields.CharField', [], {'default': "'1.0'", 'max_length': '35'}) }, 'lab_aliquot.aliquottype': { 'Meta': {'ordering': "['name']", 'object_name': 'AliquotType', 'db_table': "'bhp_lab_core_aliquottype'"}, 'alpha_code': ('django.db.models.fields.CharField', [], {'unique': 'True', 'max_length': '15'}), 'created': ('django.db.models.fields.DateTimeField', [], {'default': 'datetime.datetime.now', 'blank': 'True'}), 'dmis_reference': ('django.db.models.fields.IntegerField', [], {}), 'hostname_created': ('django.db.models.fields.CharField', [], {'default': "'home'", 'max_length': '50', 'blank': 'True'}), 'hostname_modified': ('django.db.models.fields.CharField', [], {'default': "'home'", 'max_length': '50', 'blank': 'True'}), 'id': ('django.db.models.fields.AutoField', [], {'primary_key': 'True'}), 'modified': ('django.db.models.fields.DateTimeField', [], {'default': 'datetime.datetime.now', 'blank': 'True'}), 'name': ('django.db.models.fields.CharField', [], {'max_length': '50'}), 'numeric_code': ('django.db.models.fields.CharField', [], {'unique': 'True', 'max_length': '2'}), 'user_created': ('django.db.models.fields.CharField', [], {'default': "''", 'max_length': '250'}), 'user_modified': ('django.db.models.fields.CharField', [], {'default': "''", 'max_length': '250'}) }, 'lab_order.order': { 'Meta': {'object_name': 'Order', 'db_table': "'bhp_lab_core_order'"}, 'aliquot': ('django.db.models.fields.related.ForeignKey', [], {'to': "orm['lab_aliquot.Aliquot']"}), 'comment': ('django.db.models.fields.CharField', [], {'max_length': '150', 'null': 'True', 'blank': 'True'}), 'created': ('django.db.models.fields.DateTimeField', [], {'default': 'datetime.datetime.now', 'blank': 'True'}), 'dmis_reference': ('django.db.models.fields.IntegerField', [], {'null': 'True', 'blank': 'True'}), 'hostname_created': ('django.db.models.fields.CharField', [], {'default': "'home'", 'max_length': '50', 'blank': 'True'}), 'hostname_modified': ('django.db.models.fields.CharField', [], {'default': "'home'", 'max_length': '50', 'blank': 'True'}), 'id': ('django.db.models.fields.CharField', [], {'max_length': '36', 'primary_key': 'True'}), 'modified': ('django.db.models.fields.DateTimeField', [], {'default': 'datetime.datetime.now', 'blank': 'True'}), 'order_datetime': ('django.db.models.fields.DateTimeField', [], {'db_index': 'True'}), 'order_identifier': ('django.db.models.fields.CharField', [], {'max_length': '25', 'db_index': 'True'}), 'panel': ('django.db.models.fields.related.ForeignKey', [], {'to': "orm['lab_panel.Panel']"}), 'user_created': ('django.db.models.fields.CharField', [], {'default': "''", 'max_length': '250'}), 'user_modified': ('django.db.models.fields.CharField', [], {'default': "''", 'max_length': '250'}) }, 'lab_panel.panel': { 'Meta': {'object_name': 'Panel', 'db_table': "'bhp_lab_core_panel'"}, 'account': ('django.db.models.fields.related.ManyToManyField', [], {'to': "orm['lab_account.Account']", 'symmetrical': 'False'}), 'aliquot_type': ('django.db.models.fields.related.ManyToManyField', [], {'to': "orm['lab_aliquot.AliquotType']", 'symmetrical': 'False'}), 'comment': ('django.db.models.fields.CharField', [], {'max_length': '250', 'blank': 'True'}), 'created': ('django.db.models.fields.DateTimeField', [], {'default': 'datetime.datetime.now', 'blank': 'True'}), 'dmis_panel_identifier': ('django.db.models.fields.CharField', [], {'max_length': '25', 'null': 'True', 'blank': 'True'}), 'hostname_created': ('django.db.models.fields.CharField', [], {'default': "'home'", 'max_length': '50', 'blank': 'True'}), 'hostname_modified': ('django.db.models.fields.CharField', [], {'default': "'home'", 'max_length': '50', 'blank': 'True'}), 'id': ('django.db.models.fields.AutoField', [], {'primary_key': 'True'}), 'modified': ('django.db.models.fields.DateTimeField', [], {'default': 'datetime.datetime.now', 'blank': 'True'}), 'name': ('django.db.models.fields.CharField', [], {'unique': 'True', 'max_length': '50', 'db_index': 'True'}), 'panel_group': ('django.db.models.fields.related.ForeignKey', [], {'to': "orm['lab_panel.PanelGroup']"}), 'test_code': ('django.db.models.fields.related.ManyToManyField', [], {'to': "orm['lab_test_code.TestCode']", 'symmetrical': 'False'}), 'user_created': ('django.db.models.fields.CharField', [], {'default': "''", 'max_length': '250'}), 'user_modified': ('django.db.models.fields.CharField', [], {'default': "''", 'max_length': '250'}) }, 'lab_panel.panelgroup': { 'Meta': {'object_name': 'PanelGroup', 'db_table': "'bhp_lab_core_panelgroup'"}, 'created': ('django.db.models.fields.DateTimeField', [], {'default': 'datetime.datetime.now', 'blank': 'True'}), 'hostname_created': ('django.db.models.fields.CharField', [], {'default': "'home'", 'max_length': '50', 'blank': 'True'}), 'hostname_modified': ('django.db.models.fields.CharField', [], {'default': "'home'", 'max_length': '50', 'blank': 'True'}), 'id': ('django.db.models.fields.AutoField', [], {'primary_key': 'True'}), 'modified': ('django.db.models.fields.DateTimeField', [], {'default': 'datetime.datetime.now', 'blank': 'True'}), 'name': ('django.db.models.fields.CharField', [], {'unique': 'True', 'max_length': '25'}), 'user_created': ('django.db.models.fields.CharField', [], {'default': "''", 'max_length': '250'}), 'user_modified': ('django.db.models.fields.CharField', [], {'default': "''", 'max_length': '250'}) }, 'lab_patient.patient': { 'Meta': {'ordering': "['subject_identifier']", 'unique_together': "(['subject_identifier'],)", 'object_name': 'Patient', 'db_table': "'bhp_lab_registration_patient'"}, 'account': ('django.db.models.fields.related.ManyToManyField', [], {'symmetrical': 'False', 'to': "orm['lab_account.Account']", 'null': 'True', 'blank': 'True'}), 'art_status': ('django.db.models.fields.CharField', [], {'default': "'UNKNOWN'", 'max_length': '10'}), 'comment': ('django.db.models.fields.CharField', [], {'max_length': '250', 'blank': 'True'}), 'created': ('django.db.models.fields.DateTimeField', [], {'default': 'datetime.datetime.now', 'blank': 'True'}), 'dob': ('django.db.models.fields.DateField', [], {}), 'gender': ('django.db.models.fields.CharField', [], {'max_length': '3'}), 'hiv_status': ('django.db.models.fields.CharField', [], {'default': "'UNKNOWN'", 'max_length': '10'}), 'hostname_created': ('django.db.models.fields.CharField', [], {'default': "'home'", 'max_length': '50', 'blank': 'True'}), 'hostname_modified': ('django.db.models.fields.CharField', [], {'default': "'home'", 'max_length': '50', 'blank': 'True'}), 'id': ('django.db.models.fields.CharField', [], {'max_length': '36', 'primary_key': 'True'}), 'initials': ('django.db.models.fields.CharField', [], {'max_length': '3'}), 'is_dob_estimated': ('django.db.models.fields.CharField', [], {'max_length': '25'}), 'modified': ('django.db.models.fields.DateTimeField', [], {'default': 'datetime.datetime.now', 'blank': 'True'}), 'simple_consent': ('django.db.models.fields.related.ManyToManyField', [], {'symmetrical': 'False', 'to': "orm['lab_patient.SimpleConsent']", 'null': 'True', 'blank': 'True'}), 'subject_identifier': ('django.db.models.fields.CharField', [], {'unique': 'True', 'max_length': '25', 'db_index': 'True'}), 'user_created': ('django.db.models.fields.CharField', [], {'default': "''", 'max_length': '250'}), 'user_modified': ('django.db.models.fields.CharField', [], {'default': "''", 'max_length': '250'}) }, 'lab_patient.simpleconsent': { 'Meta': {'ordering': "['consent_startdate']", 'object_name': 'SimpleConsent', 'db_table': "'bhp_lab_registration_simpleconsent'"}, 'consent_enddate': ('django.db.models.fields.DateField', [], {'null': 'True', 'blank': 'True'}), 'consent_site': ('django.db.models.fields.related.ForeignKey', [], {'to': "orm['bhp_research_protocol.Site']"}), 'consent_startdate': ('django.db.models.fields.DateField', [], {}), 'created': ('django.db.models.fields.DateTimeField', [], {'default': 'datetime.datetime.now', 'blank': 'True'}), 'hostname_created': ('django.db.models.fields.CharField', [], {'default': "'home'", 'max_length': '50', 'blank': 'True'}), 'hostname_modified': ('django.db.models.fields.CharField', [], {'default': "'home'", 'max_length': '50', 'blank': 'True'}), 'id': ('django.db.models.fields.CharField', [], {'max_length': '36', 'primary_key': 'True'}), 'may_store_samples': ('django.db.models.fields.CharField', [], {'max_length': '3'}), 'modified': ('django.db.models.fields.DateTimeField', [], {'default': 'datetime.datetime.now', 'blank': 'True'}), 'protocol': ('django.db.models.fields.related.ForeignKey', [], {'to': "orm['bhp_research_protocol.Protocol']"}), 'user_created': ('django.db.models.fields.CharField', [], {'default': "''", 'max_length': '250'}), 'user_modified': ('django.db.models.fields.CharField', [], {'default': "''", 'max_length': '250'}) }, 'lab_receive.receive': { 'Meta': {'object_name': 'Receive', 'db_table': "'bhp_lab_core_receive'"}, 'clinician_initials': ('django.db.models.fields.CharField', [], {'max_length': '3'}), 'created': ('django.db.models.fields.DateTimeField', [], {'default': 'datetime.datetime.now', 'blank': 'True'}), 'datetime_drawn': ('django.db.models.fields.DateTimeField', [], {'db_index': 'True'}), 'dmis_reference': ('django.db.models.fields.IntegerField', [], {}), 'hostname_created': ('django.db.models.fields.CharField', [], {'default': "'home'", 'max_length': '50', 'blank': 'True'}), 'hostname_modified': ('django.db.models.fields.CharField', [], {'default': "'home'", 'max_length': '50', 'blank': 'True'}), 'id': ('django.db.models.fields.CharField', [], {'max_length': '36', 'primary_key': 'True'}), 'modified': ('django.db.models.fields.DateTimeField', [], {'default': 'datetime.datetime.now', 'blank': 'True'}), 'patient': ('django.db.models.fields.related.ForeignKey', [], {'to': "orm['lab_patient.Patient']"}), 'protocol': ('django.db.models.fields.related.ForeignKey', [], {'to': "orm['bhp_research_protocol.Protocol']"}), 'receive_datetime': ('django.db.models.fields.DateTimeField', [], {'default': 'datetime.datetime(2011, 8, 30, 13, 11, 14, 891588)', 'db_index': 'True'}), 'receive_identifier': ('django.db.models.fields.CharField', [], {'max_length': '25', 'null': 'True', 'db_index': 'True'}), 'site': ('django.db.models.fields.related.ForeignKey', [], {'to': "orm['bhp_research_protocol.Site']"}), 'user_created': ('django.db.models.fields.CharField', [], {'default': "''", 'max_length': '250'}), 'user_modified': ('django.db.models.fields.CharField', [], {'default': "''", 'max_length': '250'}), 'visit': ('django.db.models.fields.CharField', [], {'max_length': '25'}) }, 'lab_result.result': { 'Meta': {'ordering': "['result_identifier', 'order', 'result_datetime']", 'object_name': 'Result', 'db_table': "'bhp_lab_core_result'"}, 'comment': ('django.db.models.fields.CharField', [], {'max_length': '50', 'null': 'True', 'blank': 'True'}), 'created': ('django.db.models.fields.DateTimeField', [], {'default': 'datetime.datetime.now', 'blank': 'True'}), 'dmis_result_guid': ('django.db.models.fields.CharField', [], {'max_length': '36', 'null': 'True', 'blank': 'True'}), 'hostname_created': ('django.db.models.fields.CharField', [], {'default': "'home'", 'max_length': '50', 'blank': 'True'}), 'hostname_modified': ('django.db.models.fields.CharField', [], {'default': "'home'", 'max_length': '50', 'blank': 'True'}), 'id': ('django.db.models.fields.CharField', [], {'max_length': '36', 'primary_key': 'True'}), 'modified': ('django.db.models.fields.DateTimeField', [], {'default': 'datetime.datetime.now', 'blank': 'True'}), 'order': ('django.db.models.fields.related.ForeignKey', [], {'to': "orm['lab_order.Order']"}), 'release_datetime': ('django.db.models.fields.DateTimeField', [], {'db_index': 'True', 'null': 'True', 'blank': 'True'}), 'release_status': ('django.db.models.fields.CharField', [], {'default': "'NEW'", 'max_length': '25', 'db_index': 'True'}), 'release_username': ('django.db.models.fields.CharField', [], {'db_index': 'True', 'max_length': '50', 'null': 'True', 'blank': 'True'}), 'result_datetime': ('django.db.models.fields.DateTimeField', [], {'db_index': 'True'}), 'result_identifier': ('django.db.models.fields.CharField', [], {'max_length': '25', 'db_index': 'True'}), 'user_created': ('django.db.models.fields.CharField', [], {'default': "''", 'max_length': '250'}), 'user_modified': ('django.db.models.fields.CharField', [], {'default': "''", 'max_length': '250'}) }, 'lab_result.resultsource': { 'Meta': {'object_name': 'ResultSource', 'db_table': "'bhp_lab_core_resultsource'"}, 'created': ('django.db.models.fields.DateTimeField', [], {'default': 'datetime.datetime.now', 'blank': 'True'}), 'display_index': ('django.db.models.fields.IntegerField', [], {'default': '0'}), 'field_name': ('django.db.models.fields.CharField', [], {'max_length': '25', 'null': 'True', 'blank': 'True'}), 'hostname_created': ('django.db.models.fields.CharField', [], {'default': "'home'", 'max_length': '50', 'blank': 'True'}), 'hostname_modified': ('django.db.models.fields.CharField', [], {'default': "'home'", 'max_length': '50', 'blank': 'True'}), 'id': ('django.db.models.fields.AutoField', [], {'primary_key': 'True'}), 'modified': ('django.db.models.fields.DateTimeField', [], {'default': 'datetime.datetime.now', 'blank': 'True'}), 'name': ('django.db.models.fields.CharField', [], {'unique': 'True', 'max_length': '250'}), 'short_name': ('django.db.models.fields.CharField', [], {'unique': 'True', 'max_length': '250'}), 'user_created': ('django.db.models.fields.CharField', [], {'default': "''", 'max_length': '250'}), 'user_modified': ('django.db.models.fields.CharField', [], {'default': "''", 'max_length': '250'}), 'version': ('django.db.models.fields.CharField', [], {'default': "'1.0'", 'max_length': '35'}) }, 'lab_result_item.resultitem': { 'Meta': {'object_name': 'ResultItem', 'db_table': "'bhp_lab_core_resultitem'"}, 'comment': ('django.db.models.fields.CharField', [], {'max_length': '50', 'null': 'True', 'blank': 'True'}), 'created': ('django.db.models.fields.DateTimeField', [], {'default': 'datetime.datetime.now', 'blank': 'True'}), 'error_code': ('django.db.models.fields.CharField', [], {'max_length': '50', 'null': 'True', 'blank': 'True'}), 'hostname_created': ('django.db.models.fields.CharField', [], {'default': "'home'", 'max_length': '50', 'blank': 'True'}), 'hostname_modified': ('django.db.models.fields.CharField', [], {'default': "'home'", 'max_length': '50', 'blank': 'True'}), 'id': ('django.db.models.fields.CharField', [], {'max_length': '36', 'primary_key': 'True'}), 'modified': ('django.db.models.fields.DateTimeField', [], {'default': 'datetime.datetime.now', 'blank': 'True'}), 'result': ('django.db.models.fields.related.ForeignKey', [], {'to': "orm['lab_result.Result']"}), 'result_item_datetime': ('django.db.models.fields.DateTimeField', [], {'db_index': 'True'}), 'result_item_operator': ('django.db.models.fields.CharField', [], {'db_index': 'True', 'max_length': '50', 'null': 'True', 'blank': 'True'}), 'result_item_quantifier': ('django.db.models.fields.CharField', [], {'default': "'='", 'max_length': '25'}), 'result_item_source': ('django.db.models.fields.related.ForeignKey', [], {'to': "orm['lab_result.ResultSource']"}), 'result_item_source_reference': ('django.db.models.fields.CharField', [], {'max_length': '50', 'null': 'True', 'blank': 'True'}), 'result_item_value': ('django.db.models.fields.CharField', [], {'max_length': '25', 'db_index': 'True'}), 'test_code': ('django.db.models.fields.related.ForeignKey', [], {'to': "orm['lab_test_code.TestCode']"}), 'user_created': ('django.db.models.fields.CharField', [], {'default': "''", 'max_length': '250'}), 'user_modified': ('django.db.models.fields.CharField', [], {'default': "''", 'max_length': '250'}), 'validation_datetime': ('django.db.models.fields.DateTimeField', [], {'db_index': 'True', 'null': 'True', 'blank': 'True'}), 'validation_reference': ('django.db.models.fields.CharField', [], {'max_length': '50', 'null': 'True', 'blank': 'True'}), 'validation_status': ('django.db.models.fields.CharField', [], {'default': "'P'", 'max_length': '10', 'db_index': 'True'}), 'validation_username': ('django.db.models.fields.CharField', [], {'db_index': 'True', 'max_length': '50', 'null': 'True', 'blank': 'True'}) }, 'lab_result_item.resultitemaudit': { 'Meta': {'ordering': "['-_audit_timestamp']", 'object_name': 'ResultItemAudit', 'db_table': "'bhp_lab_core_resultitem_audit'"}, '_audit_change_type': ('django.db.models.fields.CharField', [], {'max_length': '1'}), '_audit_id': ('django.db.models.fields.AutoField', [], {'primary_key': 'True'}), '_audit_timestamp': ('django.db.models.fields.DateTimeField', [], {'auto_now_add': 'True', 'db_index': 'True', 'blank': 'True'}), 'comment': ('django.db.models.fields.CharField', [], {'max_length': '50', 'null': 'True', 'blank': 'True'}), 'created': ('django.db.models.fields.DateTimeField', [], {'default': 'datetime.datetime.now', 'blank': 'True'}), 'error_code': ('django.db.models.fields.CharField', [], {'max_length': '50', 'null': 'True', 'blank': 'True'}), 'hostname_created': ('django.db.models.fields.CharField', [], {'default': "'home'", 'max_length': '50', 'blank': 'True'}), 'hostname_modified': ('django.db.models.fields.CharField', [], {'default': "'home'", 'max_length': '50', 'blank': 'True'}), 'id': ('django.db.models.fields.CharField', [], {'max_length': '36', 'blank': 'True'}), 'modified': ('django.db.models.fields.DateTimeField', [], {'default': 'datetime.datetime.now', 'blank': 'True'}), 'result': ('django.db.models.fields.related.ForeignKey', [], {'related_name': "'_audit_resultitem'", 'to': "orm['lab_result.Result']"}), 'result_item_datetime': ('django.db.models.fields.DateTimeField', [], {'db_index': 'True'}), 'result_item_operator': ('django.db.models.fields.CharField', [], {'db_index': 'True', 'max_length': '50', 'null': 'True', 'blank': 'True'}), 'result_item_quantifier': ('django.db.models.fields.CharField', [], {'default': "'='", 'max_length': '25'}), 'result_item_source': ('django.db.models.fields.related.ForeignKey', [], {'related_name': "'_audit_resultitem'", 'to': "orm['lab_result.ResultSource']"}), 'result_item_source_reference': ('django.db.models.fields.CharField', [], {'max_length': '50', 'null': 'True', 'blank': 'True'}), 'result_item_value': ('django.db.models.fields.CharField', [], {'max_length': '25', 'db_index': 'True'}), 'test_code': ('django.db.models.fields.related.ForeignKey', [], {'related_name': "'_audit_resultitem'", 'to': "orm['lab_test_code.TestCode']"}), 'user_created': ('django.db.models.fields.CharField', [], {'default': "''", 'max_length': '250'}), 'user_modified': ('django.db.models.fields.CharField', [], {'default': "''", 'max_length': '250'}), 'validation_datetime': ('django.db.models.fields.DateTimeField', [], {'db_index': 'True', 'null': 'True', 'blank': 'True'}), 'validation_reference': ('django.db.models.fields.CharField', [], {'max_length': '50', 'null': 'True', 'blank': 'True'}), 'validation_status': ('django.db.models.fields.CharField', [], {'default': "'P'", 'max_length': '10', 'db_index': 'True'}), 'validation_username': ('django.db.models.fields.CharField', [], {'db_index': 'True', 'max_length': '50', 'null': 'True', 'blank': 'True'}) }, 'lab_test_code.testcode': { 'Meta': {'ordering': "['name']", 'object_name': 'TestCode', 'db_table': "'bhp_lab_test_code_testcode'"}, 'code': ('django.db.models.fields.CharField', [], {'unique': 'True', 'max_length': '15'}), 'created': ('django.db.models.fields.DateTimeField', [], {'default': 'datetime.datetime.now', 'blank': 'True'}), 'display_decimal_places': ('django.db.models.fields.IntegerField', [], {'null': 'True', 'blank': 'True'}), 'formula': ('django.db.models.fields.CharField', [], {'max_length': "'50'", 'null': 'True', 'blank': 'True'}), 'hostname_created': ('django.db.models.fields.CharField', [], {'default': "'home'", 'max_length': '50', 'blank': 'True'}), 'hostname_modified': ('django.db.models.fields.CharField', [], {'default': "'home'", 'max_length': '50', 'blank': 'True'}), 'id': ('django.db.models.fields.AutoField', [], {'primary_key': 'True'}), 'is_absolute': ('django.db.models.fields.CharField', [], {'default': "'absolute'", 'max_length': "'15'"}), 'modified': ('django.db.models.fields.DateTimeField', [], {'default': 'datetime.datetime.now', 'blank': 'True'}), 'name': ('django.db.models.fields.CharField', [], {'max_length': '50'}), 'test_code_group': ('django.db.models.fields.related.ForeignKey', [], {'to': "orm['lab_test_code.TestCodeGroup']"}), 'units': ('django.db.models.fields.CharField', [], {'max_length': '25'}), 'user_created': ('django.db.models.fields.CharField', [], {'default': "''", 'max_length': '250'}), 'user_modified': ('django.db.models.fields.CharField', [], {'default': "''", 'max_length': '250'}) }, 'lab_test_code.testcodegroup': { 'Meta': {'ordering': "['code']", 'object_name': 'TestCodeGroup', 'db_table': "'bhp_lab_test_code_testcodegroup'"}, 'code': ('django.db.models.fields.CharField', [], {'max_length': '3'}), 'created': ('django.db.models.fields.DateTimeField', [], {'default': 'datetime.datetime.now', 'blank': 'True'}), 'hostname_created': ('django.db.models.fields.CharField', [], {'default': "'home'", 'max_length': '50', 'blank': 'True'}), 'hostname_modified': ('django.db.models.fields.CharField', [], {'default': "'home'", 'max_length': '50', 'blank': 'True'}), 'id': ('django.db.models.fields.AutoField', [], {'primary_key': 'True'}), 'modified': ('django.db.models.fields.DateTimeField', [], {'default': 'datetime.datetime.now', 'blank': 'True'}), 'name': ('django.db.models.fields.CharField', [], {'max_length': '25', 'null': 'True', 'blank': 'True'}), 'user_created': ('django.db.models.fields.CharField', [], {'default': "''", 'max_length': '250'}), 'user_modified': ('django.db.models.fields.CharField', [], {'default': "''", 'max_length': '250'}) } } complete_apps = ['lab_result_item']
[ "ckgathi@gmail.com" ]
ckgathi@gmail.com
13e0e56617747cebd4d1b90a7eb5a82697e973ae
b32224af384173f41732539dd216d9312ffb521e
/rabbitmq/routing/emit_log_direct.py
1d732cbd9bff728a0821c952e9d5dfd296bf7fc0
[]
no_license
4179e1/demo
2140e63cdecb0f5df4d5d435f482c3da49d63cfa
8cc84a64c34e2b822fe4114db1ca6531e647affa
refs/heads/master
2023-08-17T12:16:06.002630
2021-09-18T03:27:27
2021-09-18T03:27:27
407,744,907
0
0
null
null
null
null
UTF-8
Python
false
false
916
py
#!/usr/bin/env python import pika import sys connection = pika.BlockingConnection( pika.ConnectionParameters(host='localhost')) channel = connection.channel() channel.exchange_declare(exchange='direct_logs', exchange_type='direct') # If message cannot be routed to a queue, we need confirm_delivery() to get noticed. # see https://www.rabbitmq.com/amqp-0-9-1-reference.html#basic.publish # https://pika.readthedocs.io/en/latest/examples/blocking_publish_mandatory.html channel.confirm_delivery() severity = sys.argv[1] if len(sys.argv) > 1 else 'info' message = ' '.join(sys.argv[2:]) or 'Hello World!' try: res = channel.basic_publish( exchange='direct_logs', routing_key=severity, mandatory=True, body=message ) print(" [x] Sent %r:%r" % (severity, message)) except pika.exceptions.UnroutableError: print("Message was returned") connection.close()
[ "lyre@poetpalace.org" ]
lyre@poetpalace.org
1a7c77ff09ef224b7bba1da2aa9c3122854922d9
1c1269eff0161d23845444ce0254d2ed195d6fb4
/app.py
d01b14ee1672d92ed61693f1e504dc1dfedcda9b
[]
no_license
dich1123/tasks
abf5aa9cdf17be0ac245eb27e99e3cc894238ec1
3c9aeccb172c9272f573b5831c8c6b8b0c7f2ba6
refs/heads/master
2021-06-27T01:15:48.574125
2019-11-20T11:11:09
2019-11-20T11:11:09
222,911,800
0
0
null
2021-03-20T02:13:02
2019-11-20T10:29:34
Python
UTF-8
Python
false
false
1,756
py
from flask import Flask, render_template, url_for, request, redirect from flask_sqlalchemy import SQLAlchemy from datetime import datetime app = Flask(__name__) app.config['SQLALCHEMY_DATABASE_URI'] = 'sqlite:///test.db' db = SQLAlchemy(app) class Todo(db.Model): id = db.Column(db.Integer, primary_key=True) content = db.Column(db.String(200), nullable=False) date_created = db.Column(db.DateTime, default=datetime.utcnow()) def __repr__(self): return f'Task {self.id}' @app.route('/', methods=['POST', 'GET']) def index(): if request.method == 'POST': task_content = request.form['content'] new_task = Todo(content=task_content) try: db.session.add(new_task) db.session.commit() return redirect('/') except: return 'There were troubles with adding your task' else: tasks = Todo.query.order_by(Todo.date_created).all() return render_template('index.html', tasks=tasks) @app.route('/delete/<int:id>') def delete(id): task_to_delete = Todo.query.get_or_404(id) try: db.session.delete(task_to_delete) db.session.commit() return redirect('/') except: return 'There were troubles with deleting your task' @app.route('/update/<int:id>', methods=['GET', 'POST']) def update(id): task = Todo.query.get_or_404(id) if request.method == 'POST': task.content = request.form['content'] try: db.session.commit() return redirect('/') except: return 'There were troubles with updating your task' else: return render_template('update.html', task=task) if __name__ == '__main__': app.run(debug=True)
[ "cherenkov1123@gmail.com" ]
cherenkov1123@gmail.com
be4bda3f34c623e89442a9c1b39b272bef2b621d
65b3d3230cd6b828eeb664429997468cd39db92c
/GUI/venv/bin/pip3.6
e8ca7da5cd04e5723cc80121403f3f1dbfb27657
[]
no_license
shaniphankar/IR-Assignments
6bbd0c5e9d52704e67a8c49ee2551d2f693f6578
db9120861c8dca83d11340c57240e3dc53f660b1
refs/heads/master
2021-06-05T09:49:16.734160
2020-02-04T15:30:24
2020-02-04T15:30:24
147,077,391
1
3
null
2018-10-23T13:35:59
2018-09-02T11:01:01
Python
UTF-8
Python
false
false
222
6
#!/home/cross/GUI/venv/bin/python3 # -*- coding: utf-8 -*- import re import sys from pip import main if __name__ == '__main__': sys.argv[0] = re.sub(r'(-script\.pyw?|\.exe)?$', '', sys.argv[0]) sys.exit(main())
[ "vishalcross36@gmail.com" ]
vishalcross36@gmail.com
491e15f5ff3feba66ba8f3a0cb4ca00a15def0fe
58b37bcd7f406034274bfaf64fafe446ca7d07bd
/Asset_Management/Asset_Management/settings.py
6822e9cc1a8f9552726787cca43a7ce874115c1c
[]
no_license
jmarigondon/django-repo
cfb06c6e18f85334f15150085df6748633590d56
48a8271c76fca27a3a095a19586d79d684afe9c2
refs/heads/master
2021-01-20T12:44:59.729698
2013-10-15T11:04:48
2013-10-15T11:04:48
null
0
0
null
null
null
null
UTF-8
Python
false
false
5,317
py
# Django settings for Asset_Management project. DEBUG = True TEMPLATE_DEBUG = DEBUG ADMINS = ( # ('Your Name', 'your_email@example.com'), ) MANAGERS = ADMINS DATABASES = { 'default': { 'ENGINE': 'django.db.backends.sqlite3', # Add 'postgresql_psycopg2', 'postgresql', 'mysql', 'sqlite3' or 'oracle'. 'NAME': '/home/jmarigondon/workspace/Asset_Management/sqlite.db', # Or path to database file if using sqlite3. 'USER': 'jmarigondon', # Not used with sqlite3. 'PASSWORD': '12345Ye@h', # Not used with sqlite3. 'HOST': '', # Set to empty string for localhost. Not used with sqlite3. 'PORT': '', # Set to empty string for default. Not used with sqlite3. } } # Hosts/domain names that are valid for this site; required if DEBUG is False # See https://docs.djangoproject.com/en/{{ docs_version }}/ref/settings/#allowed-hosts ALLOWED_HOSTS = [] # Local time zone for this installation. Choices can be found here: # http://en.wikipedia.org/wiki/List_of_tz_zones_by_name # although not all choices may be available on all operating systems. # On Unix systems, a value of None will cause Django to use the same # timezone as the operating system. # If running in a Windows environment this must be set to the same as your # system time zone. TIME_ZONE = 'America/Chicago' # Language code for this installation. All choices can be found here: # http://www.i18nguy.com/unicode/language-identifiers.html LANGUAGE_CODE = 'en-us' SITE_ID = 1 # If you set this to False, Django will make some optimizations so as not # to load the internationalization machinery. USE_I18N = True # If you set this to False, Django will not format dates, numbers and # calendars according to the current locale USE_L10N = True # Absolute filesystem path to the directory that will hold user-uploaded files. # Example: "/home/media/media.lawrence.com/media/" MEDIA_ROOT = '' # URL that handles the media served from MEDIA_ROOT. Make sure to use a # trailing slash. # Examples: "http://media.lawrence.com/media/", "http://example.com/media/" MEDIA_URL = '' # Absolute path to the directory static files should be collected to. # Don't put anything in this directory yourself; store your static files # in apps' "static/" subdirectories and in STATICFILES_DIRS. # Example: "/home/media/media.lawrence.com/static/" STATIC_ROOT = '' # URL prefix for static files. # Example: "http://media.lawrence.com/static/" STATIC_URL = '/static/' # URL prefix for admin static files -- CSS, JavaScript and images. # Make sure to use a trailing slash. # Examples: "http://foo.com/static/admin/", "/static/admin/". ADMIN_MEDIA_PREFIX = '/static/admin/' # Additional locations of static files STATICFILES_DIRS = ( # Put strings here, like "/home/html/static" or "C:/www/django/static". # Always use forward slashes, even on Windows. # Don't forget to use absolute paths, not relative paths. ) # List of finder classes that know how to find static files in # various locations. STATICFILES_FINDERS = ( 'django.contrib.staticfiles.finders.FileSystemFinder', 'django.contrib.staticfiles.finders.AppDirectoriesFinder', # 'django.contrib.staticfiles.finders.DefaultStorageFinder', ) # Make this unique, and don't share it with anybody. SECRET_KEY = '@$id$eos_^=iz645bj_gb#)tq)!$-15d&i59io5kcku5#a=%xl' # List of callables that know how to import templates from various sources. TEMPLATE_LOADERS = ( 'django.template.loaders.filesystem.Loader', 'django.template.loaders.app_directories.Loader', # 'django.template.loaders.eggs.Loader', ) MIDDLEWARE_CLASSES = ( 'django.middleware.common.CommonMiddleware', 'django.contrib.sessions.middleware.SessionMiddleware', 'django.middleware.csrf.CsrfViewMiddleware', 'django.contrib.auth.middleware.AuthenticationMiddleware', 'django.contrib.messages.middleware.MessageMiddleware', ) ROOT_URLCONF = 'Asset_Management.urls' TEMPLATE_DIRS = ( # Put strings here, like "/home/html/django_templates" or "C:/www/django/templates". # Always use forward slashes, even on Windows. # Don't forget to use absolute paths, not relative paths. ) INSTALLED_APPS = ( 'django.contrib.auth', 'django.contrib.contenttypes', 'django.contrib.sessions', 'django.contrib.sites', 'django.contrib.messages', 'django.contrib.staticfiles', # Uncomment the next line to enable the admin: # 'django.contrib.admin', # Uncomment the next line to enable admin documentation: # 'django.contrib.admindocs', ) # A sample logging configuration. The only tangible logging # performed by this configuration is to send an email to # the site admins on every HTTP 500 error. # See http://docs.djangoproject.com/en/dev/topics/logging for # more details on how to customize your logging configuration. LOGGING = { 'version': 1, 'disable_existing_loggers': False, 'handlers': { 'mail_admins': { 'level': 'ERROR', 'class': 'django.utils.log.AdminEmailHandler' } }, 'loggers': { 'django.request': { 'handlers': ['mail_admins'], 'level': 'ERROR', 'propagate': True, }, } }
[ "jbmarigondon@gmail.com" ]
jbmarigondon@gmail.com
712213e339f5e2cbd3b35fc7dbce59681eaeb484
3c968361997cc8555bf69f8dcb2283831ac99832
/pubmed_prepro/train_test_splits.py
a303ab18a02a896f2d53af5c4ab517f90ff903ed
[]
no_license
yellajaswanth/Pubset
01014c4b99ae32561741c14c8c638f60daa36a8a
650bd925c82e11aa44ae7be2dc8c10f4401ec074
refs/heads/main
2023-03-21T07:06:32.503514
2021-03-05T22:01:49
2021-03-05T22:01:49
340,588,695
0
1
null
null
null
null
UTF-8
Python
false
false
1,082
py
#%% import shelve from tqdm import tqdm import json from sklearn.model_selection import train_test_split db = shelve.open('/home/aniljegga1/bigdataserver/pubmed_docs/abstracts/pubmed_small/pubmed.db', 'r') keys = list(db.keys()) all_indices = [] for key in tqdm(keys): if 'bert' in db[key]: num_sents = len(db[key]['bert']) all_indices.extend([(key, i) for i in range(num_sents-6)]) train, test = train_test_split(all_indices, test_size=0.25) train, val = train_test_split(train, test_size=0.15) #%% with open('/home/aniljegga1/bigdataserver/pubmed_docs/abstracts/pubmed_small/all_indices.json', 'w') as fin: json.dump(all_indices, fin) with open('/home/aniljegga1/bigdataserver/pubmed_docs/abstracts/pubmed_small/train_indices.json', 'w') as fin: json.dump(train, fin) with open('/home/aniljegga1/bigdataserver/pubmed_docs/abstracts/pubmed_small/test_indices.json', 'w') as fin: json.dump(test, fin) with open('/home/aniljegga1/bigdataserver/pubmed_docs/abstracts/pubmed_small/val_indices.json', 'w') as fin: json.dump(val, fin)
[ "yellajaswanth@gmail.com" ]
yellajaswanth@gmail.com
5c33f8c1cfcbce6f594edaa5d44c32aebcdf7169
0fe2847bf222a3df0847a08de244000207514d05
/src/seabreeze/pyseabreeze/features/introspection.py
3422b402411c9583f102c62d5d08ac6c55d658e4
[ "MIT" ]
permissive
asenchristov/python-seabreeze
3656161eb2bf2be082839700f021a5957b81f00b
573bae1d9de4e819611b2f5b9c66f98d7d0fe066
refs/heads/master
2022-12-01T09:39:46.079901
2020-08-18T09:07:30
2020-08-18T09:07:30
288,403,712
0
0
MIT
2020-08-18T08:49:58
2020-08-18T08:49:57
null
UTF-8
Python
false
false
669
py
from seabreeze.pyseabreeze.features._base import SeaBreezeFeature # Definition # ========== # # TODO: This feature needs to be implemented for pyseabreeze # class SeaBreezeIntrospectionFeature(SeaBreezeFeature): identifier = "introspection" def number_of_pixels(self): raise NotImplementedError("implement in derived class") def get_active_pixel_ranges(self): raise NotImplementedError("implement in derived class") def get_optical_dark_pixel_ranges(self): raise NotImplementedError("implement in derived class") def get_electric_dark_pixel_ranges(self): raise NotImplementedError("implement in derived class")
[ "andreas@poehlmann.io" ]
andreas@poehlmann.io
6ad6148e7378b35ec5dbeeb2a493dbad852d7119
781e2692049e87a4256320c76e82a19be257a05d
/all_data/exercism_data/python/bob/61f29ff81b664630acba807a6f4f14e9.py
b34cc9defb4248705933ec5d41f5347dc783be44
[]
no_license
itsolutionscorp/AutoStyle-Clustering
54bde86fe6dbad35b568b38cfcb14c5ffaab51b0
be0e2f635a7558f56c61bc0b36c6146b01d1e6e6
refs/heads/master
2020-12-11T07:27:19.291038
2016-03-16T03:18:00
2016-03-16T03:18:42
59,454,921
4
0
null
2016-05-23T05:40:56
2016-05-23T05:40:56
null
UTF-8
Python
false
false
286
py
# # Skeleton file for the Python "Bob" exercise. # def hey(what): if what.upper() == what and any(c.isalpha() for c in what): return "Whoa, chill out!" if what != '' and what[-1] == '?': return "Sure." if len(what) < 7: return "Fine. Be that way!" else: return "Whatever."
[ "rrc@berkeley.edu" ]
rrc@berkeley.edu
4979767ebde362092d71c72285624162e5330f9b
ba269272596a2427d46c21d3cebe0270775557c7
/setup.py
0b9e961c8d7687192ef0bc78f6f9eccb8dd148e7
[ "MIT" ]
permissive
renatoliveira/fixer-cli
e5e65177e7af38c5f478c181f2f151292ca716ed
d1a7a9d96fbc28392f600342e201f3098a6aa94a
refs/heads/master
2020-12-02T07:40:33.844985
2017-07-22T18:31:28
2017-07-22T18:31:28
96,712,040
0
0
null
null
null
null
UTF-8
Python
false
false
501
py
"""Setup module""" from setuptools import setup def readfile(filename): """Get readfile""" with open(filename, 'r+') as readmefile: return readmefile.read() setup( name="fixer", version="3.1.0", description="", long_description=readfile('README.md'), author="Renato O.", url="", py_modules=['fixer', 'history'], license=readfile('LICENSE'), entry_points={ 'console_scripts': [ 'fixer = fixer:handle_options' ] }, )
[ "ren811@gmail.com" ]
ren811@gmail.com
80cff87d5c4547d5aa8707328225342560d7d397
7620eb38a3744d33842ef867a4dda40748f370ba
/src/apiwrappers/exceptions.py
0537acd446dca141a49012ba99dcb5bb4827d2f4
[ "MIT" ]
permissive
unmade/apiwrappers
f0f57d2f2d8a02ee9de2d52575b3e9b48809abfa
eb004fd86cbbc1d468030544472851a262b196dc
refs/heads/master
2022-09-10T10:27:23.233400
2022-01-15T15:27:39
2022-01-15T15:27:39
231,653,904
18
1
MIT
2022-08-29T07:24:51
2020-01-03T19:43:47
Python
UTF-8
Python
false
false
220
py
class DriverError(Exception): """Base class for driver-specific errors.""" class ConnectionFailed(DriverError): """A Connection error occurred.""" class Timeout(DriverError): """The request timed out."""
[ "lesha.maslakov@gmail.com" ]
lesha.maslakov@gmail.com
64e6333fe55a93a5646cccf205ff50466adc18b0
8ed27d2b651e07d7d9cb489abe9686b7d158741e
/tasks.py
ceffdc72dbddf8e1f82cfa923555c1c7ddb68a3c
[]
no_license
Seven4ME/tgBot
6874b887849b0ae6ecceb6fae62b561d199dd1a3
1f51e41fd0c5c826e315c4f1f5e01b428b1cd528
refs/heads/master
2022-12-16T09:03:35.252717
2020-08-26T21:11:36
2020-08-26T21:11:36
163,940,738
0
0
null
2022-12-08T01:53:49
2019-01-03T07:53:21
Python
UTF-8
Python
false
false
155
py
from celery.schedules import crontab # https://api.telegram.org/bot1215880984:AAHuLxPx8vEuOVPIznPhOWBCKkBFUlZbKgs/sendMessage?chat_id=194650240&text=Hi
[ "kobun2341@gmail.com" ]
kobun2341@gmail.com
b620e42042438f0ddf82969a5e2f05dcf02a8e23
3922c05b9434bb5a96f7833a987c50c8e3e29107
/news/admin.py
6881fe98a61e98e3099d1a8b53bfb646d84da9fa
[ "MIT" ]
permissive
jasonmuchiri/moringa-tribune
e7769dca9aa2e7a9cdc62be56c3071104ba30f33
ceabe0cf9cc136b6eb5072253aef09f43bea7040
refs/heads/master
2020-05-23T23:19:55.111831
2019-05-18T21:32:44
2019-05-18T21:32:44
186,990,384
0
0
null
null
null
null
UTF-8
Python
false
false
239
py
from django.contrib import admin from .models import Article,tags # Register your models here. class ArticleAdmin(admin.ModelAdmin): filter_horizontal = ('tags',) admin.site.register(Article,ArticleAdmin) admin.site.register(tags)
[ "jasonmkinyua@gmail.com" ]
jasonmkinyua@gmail.com
6132ba796a9e7e8e7a7f159210569d88c3858189
eda48a0bb8322bb5bb107cea8324aafaf105bf46
/SVNDev/comments/setting.py
57e88967a9aeb8bb9c2c471c717f9eda2ae6cfe3
[]
no_license
sougannkyou/pyWorks
c171dfc2c213bd7d75dc168b987eb96cbcbfd7e1
2402b3da4521a17d7f828504c0745ff55cb031be
refs/heads/master
2020-04-04T06:23:28.678959
2017-05-02T09:06:04
2017-05-02T09:06:19
48,786,947
1
3
null
null
null
null
UTF-8
Python
false
false
377
py
# coding=utf-8 # MONGODB_SERVER = '127.0.0.1' # MONGODB_PORT = 27017 # REDIS_SERVER = 'redis://127.0.0.1/15' MONGODB_SERVER = '192.168.16.223' MONGODB_PORT = 37017 MONGODB_SERVER_DIRECT = '192.168.149.39' MONGODB_PORT_DIRECT = 37017 REDIS_SERVER = 'redis://192.168.187.55/15' CONFIG_ID = '37556' # 全站爬虫配置ID OK_PERCENT = 0.8 UNKOWN = '未知' PAGE_SIZE = 20
[ "syq314159" ]
syq314159