blob_id
stringlengths
40
40
directory_id
stringlengths
40
40
path
stringlengths
3
281
content_id
stringlengths
40
40
detected_licenses
listlengths
0
57
license_type
stringclasses
2 values
repo_name
stringlengths
6
116
snapshot_id
stringlengths
40
40
revision_id
stringlengths
40
40
branch_name
stringclasses
313 values
visit_date
timestamp[us]
revision_date
timestamp[us]
committer_date
timestamp[us]
github_id
int64
18.2k
668M
star_events_count
int64
0
102k
fork_events_count
int64
0
38.2k
gha_license_id
stringclasses
17 values
gha_event_created_at
timestamp[us]
gha_created_at
timestamp[us]
gha_language
stringclasses
107 values
src_encoding
stringclasses
20 values
language
stringclasses
1 value
is_vendor
bool
2 classes
is_generated
bool
2 classes
length_bytes
int64
4
6.02M
extension
stringclasses
78 values
content
stringlengths
2
6.02M
authors
listlengths
1
1
author
stringlengths
0
175
d1a3470fcaf15ed7dc2b2213d66baa0ad7186f38
0102b98c118cdc43dee4c849083bbbe6926512e2
/lemon/settings.py
24e03b9667b1bd2e1452344577c762d95ddfc86b
[]
no_license
solen0id/lemon
a201ea8c5822d47ae338bb41fd2d559e4e555c64
1bfa241a5ac318bae8e13351b5050d7366fe5241
refs/heads/master
2023-04-07T14:45:51.831355
2021-04-19T06:42:20
2021-04-19T08:35:37
359,386,561
0
0
null
null
null
null
UTF-8
Python
false
false
3,205
py
""" Django settings for lemon project. Generated by 'django-admin startproject' using Django 3.2. For more information on this file, see https://docs.djangoproject.com/en/3.2/topics/settings/ For the full list of settings and their values, see https://docs.djangoproject.com/en/3.2/ref/settings/ """ from pathlib import Path # Build paths inside the project like this: BASE_DIR / 'subdir'. BASE_DIR = Path(__file__).resolve().parent.parent # Quick-start development settings - unsuitable for production # See https://docs.djangoproject.com/en/3.2/howto/deployment/checklist/ # SECURITY WARNING: keep the secret key used in production secret! SECRET_KEY = "django-insecure-**z!4@h9!il_tq))m=j@@d$dcq9n1x=7)a(p)zoz_up-9^^yns" # SECURITY WARNING: don't run with debug turned on in production! DEBUG = True ALLOWED_HOSTS = ["*"] # Application definition INSTALLED_APPS = [ "django.contrib.admin", "django.contrib.auth", "django.contrib.contenttypes", "django.contrib.sessions", "django.contrib.messages", "django.contrib.staticfiles", "rest_framework", "drf_yasg", "order", ] MIDDLEWARE = [ "django.middleware.security.SecurityMiddleware", "django.contrib.sessions.middleware.SessionMiddleware", "django.middleware.common.CommonMiddleware", "django.middleware.csrf.CsrfViewMiddleware", "django.contrib.auth.middleware.AuthenticationMiddleware", "django.contrib.messages.middleware.MessageMiddleware", "django.middleware.clickjacking.XFrameOptionsMiddleware", ] ROOT_URLCONF = "urls" TEMPLATES = [ { "BACKEND": "django.template.backends.django.DjangoTemplates", "DIRS": [], "APP_DIRS": True, "OPTIONS": { "context_processors": [ "django.template.context_processors.debug", "django.template.context_processors.request", "django.contrib.auth.context_processors.auth", "django.contrib.messages.context_processors.messages", ] }, } ] WSGI_APPLICATION = "wsgi.application" # Database # https://docs.djangoproject.com/en/3.2/ref/settings/#databases DATABASES = { "default": {"ENGINE": "django.db.backends.sqlite3", "NAME": BASE_DIR / "db.sqlite3"} } # Password validation # https://docs.djangoproject.com/en/3.2/ref/settings/#auth-password-validators AUTH_PASSWORD_VALIDATORS = [ { "NAME": "django.contrib.auth.password_validation.UserAttributeSimilarityValidator" }, {"NAME": "django.contrib.auth.password_validation.MinimumLengthValidator"}, {"NAME": "django.contrib.auth.password_validation.CommonPasswordValidator"}, {"NAME": "django.contrib.auth.password_validation.NumericPasswordValidator"}, ] # Internationalization # https://docs.djangoproject.com/en/3.2/topics/i18n/ LANGUAGE_CODE = "en-us" TIME_ZONE = "UTC" USE_I18N = True USE_L10N = True USE_TZ = True # Static files (CSS, JavaScript, Images) # https://docs.djangoproject.com/en/3.2/howto/static-files/ STATIC_URL = "/static/" # Default primary key field type # https://docs.djangoproject.com/en/3.2/ref/settings/#default-auto-field DEFAULT_AUTO_FIELD = "django.db.models.BigAutoField"
[ "max.patzelt@legislayer.com" ]
max.patzelt@legislayer.com
b0c32aca6bcee164439fb7eba0c7676a3bb2499c
8928f04f1867358430df456bdef5086b88b43ccd
/snagging_parking.py
683fb7950cc374355322f68e65a4e697c08c6ae7
[]
no_license
MMSaiPrakash/parking_slots_assistant
83d8f76638f2f64e8fe3bf91d4abb3b0be86b5a2
944b775d955a65e041dda7960800b3bf887b9013
refs/heads/main
2023-05-06T22:18:03.437711
2021-05-22T06:48:27
2021-05-22T06:48:27
369,736,366
0
0
null
null
null
null
UTF-8
Python
false
false
5,901
py
import urllib import cv2 import numpy as np import time from firebase import firebase firebase = firebase.FirebaseApplication('https://snaggingparking-default-rtdb.firebaseio.com/') #Change the firebase url to yours and use this code. url = 'http://192.168.0.109:8080/shot.jpg' #Change this IP with your Camera's IP #cap = cv2.VideoCapture(0) freeframe1=0 freeframe3=0 freeframe5=0 freeframe7=0 freeframe2=0 freeframe4=0 freeframe6=0 freeframe8=0 while(1): free1 = False free3 = False free5 = False free7 = False free2 = False free4 = False free6 = False free8 = False #IPCam imgResp=urllib.urlopen(url) imgNp=np.array(bytearray(imgResp.read()),dtype=np.uint8) frame=cv2.imdecode(imgNp,-1) frame=cv2.resize(frame,None,fx=0.5,fy=0.5) #LapCam #ret, frame = cap.read() gray_img = cv2.cvtColor(frame, cv2.COLOR_BGR2GRAY) img = cv2.medianBlur(gray_img, 5) cimg = cv2.cvtColor(img,cv2.COLOR_GRAY2BGR) circles = cv2.HoughCircles(img,cv2.HOUGH_GRADIENT,1,120, param1=100,param2=30,minRadius=0,maxRadius=0) if circles is not None: circles = np.uint16(np.around(circles)) for i in circles[0,:]: if i[2]<70 and i[2]>50: cv2.circle(frame,(i[0],i[1]),i[2],(0,255,0),2) cv2.putText(frame, str(i[0])+','+str(i[1]), (int(i[0]+i[2]),int(i[1])), cv2.FONT_HERSHEY_SIMPLEX, 1, (0,0,0), 2, cv2.LINE_AA) print i[0],i[1] time.sleep(0.1) ####Configure the values of i[0], i[1] and i[2] (x,y and radius) using your ####camera at a fixed distance from the parking spcace if i[1]>70 and i[1]<100: if i[0]<80 and i[0]>50: free1 = True print '1 is available' firebase.put('/SLOT','P1','available') if i[0]<170 and i[0]>140: free3 = True print '3 is available' firebase.put('/SLOT','P3','available') if i[0]<255 and i[0]>220: free5 = True print '5 is available' firebase.put('/SLOT','P5','available') if i[0]<350 and i[0]>310: free7 = True print '7 is available' firebase.put('/SLOT','P7','available') elif i[1]<310 and i[1]>270: if i[0]<80 and i[0]>50: free2 = True print '2 is available' firebase.put('/SLOT','P2','available') if i[0]<170 and i[0]>140: free4 = True print '4 is available' firebase.put('/SLOT','P4','available') if i[0]<255 and i[0]>220: free6 = True print '6 is available' firebase.put('/SLOT','P6','available') if i[0]<350 and i[0]>310: free8 = True print '8 is available' firebase.put('/SLOT','P8','available') if free1==False: freeframe1 += 1 else: freeframe1=0 if free3==False: freeframe3 += 1 else: freeframe3=0 if free5==False: freeframe5 += 1 else: freeframe5=0 if free7==False: freeframe7 += 1 else: freeframe7=0 if free2==False: freeframe2 += 1 else: freeframe2=0 if free4==False: freeframe4 += 1 else: freeframe4=0 if free6==False: freeframe6 += 1 else: freeframe6=0 if free8==False: freeframe8 += 1 else: freeframe8=0 if freeframe1>20: print '1 is occupied' firebase.put('/SLOT','P1','occupied') cv2.circle(frame,(65,85),36,(0,0,255),3) if freeframe3>20: print '3 is occupied' firebase.put('/SLOT','P3','occupied') cv2.circle(frame,(155,85),36,(0,0,255),3) if freeframe5>20: print '5 is occupied' firebase.put('/SLOT','P5','occupied') cv2.circle(frame,(240,85),36,(0,0,255),3) if freeframe7>20: print '7 is occupied' firebase.put('/SLOT','P7','occupied') cv2.circle(frame,(330,85),36,(0,0,255),3) if freeframe2>20: print '2 is occupied' firebase.put('/SLOT','P2','occupied') cv2.circle(frame,(65,290),36,(0,0,255),3) if freeframe4>20: print '4 is occupied' firebase.put('/SLOT','P4','occupied') cv2.circle(frame,(155,290),36,(0,0,255),3) if freeframe6>20: print '6 is occupied' firebase.put('/SLOT','P6','occupied') cv2.circle(frame,(240,290),36,(0,0,255),3) if freeframe8>20: print '8 is occupied' cv2.circle(frame,(330,290),36,(0,0,255),3) firebase.put('/SLOT','P8','occupied') cv2.imshow("Output", frame) k=cv2.waitKey(10) & 0xFF if k==27: break cv2.destroyAllWindows() #cap.release()
[ "noreply@github.com" ]
noreply@github.com
fafe549264abbc0749503ca3d08dbbc62fe4299b
91d1a6968b90d9d461e9a2ece12b465486e3ccc2
/lambda_write_f/function-event-invoke-config_delete.py
15016e42d8d9604bacece336bdaec2f86610b9a1
[]
no_license
lxtxl/aws_cli
c31fc994c9a4296d6bac851e680d5adbf7e93481
aaf35df1b7509abf5601d3f09ff1fece482facda
refs/heads/master
2023-02-06T09:00:33.088379
2020-12-27T13:38:45
2020-12-27T13:38:45
318,686,394
0
0
null
null
null
null
UTF-8
Python
false
false
1,010
py
#!/usr/bin/python # -*- codding: utf-8 -*- import os import sys sys.path.append(os.path.dirname(os.path.abspath(os.path.dirname(__file__)))) from common.execute_command import write_parameter # url : https://awscli.amazonaws.com/v2/documentation/api/latest/reference/ec2/describe-instances.html if __name__ == '__main__': """ get-function-event-invoke-config : https://awscli.amazonaws.com/v2/documentation/api/latest/reference/lambda/get-function-event-invoke-config.html list-function-event-invoke-configs : https://awscli.amazonaws.com/v2/documentation/api/latest/reference/lambda/list-function-event-invoke-configs.html put-function-event-invoke-config : https://awscli.amazonaws.com/v2/documentation/api/latest/reference/lambda/put-function-event-invoke-config.html update-function-event-invoke-config : https://awscli.amazonaws.com/v2/documentation/api/latest/reference/lambda/update-function-event-invoke-config.html """ write_parameter("lambda", "delete-function-event-invoke-config")
[ "hcseo77@gmail.com" ]
hcseo77@gmail.com
c8c314c5d709f0fa4bdd77ab501e68a1e22ad86a
f05908fcf4236df817fe04fecff8dce670c28c57
/Classification/Extra/rmse_class_sigm.py
1af3ca6ad5adf3999db1212cc862046c6e75d211
[]
no_license
visionion/Neural-Networks
5db843232c7889cc9593c0097bf8ed93d5016e03
d60d486399b458ad6ad0cf0dce188591ddbedd68
refs/heads/master
2020-04-09T11:55:01.028022
2017-01-02T07:34:52
2017-01-02T07:37:00
null
0
0
null
null
null
null
UTF-8
Python
false
false
5,234
py
def featureNormalize(X): X_norm = np.divide((X-np.mean(X,axis=0)), np.std(X,axis=0)) return X_norm def activation(a): #here it is sigmoid function return np.divide(float(1),1+np.exp(-a)) def feed_forward(xtrain,weights1,bias1,weights2,bias2): a1= np.dot(xtrain,np.transpose(weights1))+bias1 h1=activation(a1) a2= np.dot(h1,np.transpose(weights2))+bias2 prd=activation(a2) return prd def convert_to_output(prd): prd=np.argmax(prd,axis=1) return prd def confusion_matrix(prd,y): cm=np.zeros((classes,classes)) for idx,i in enumerate(y): cm[i][prd[idx]]+=1 return cm def compute_error(xtrain,weights1,bias1,weights2,bias2,y): # here this is mean square error prd=(feed_forward(xtrain,weights1,bias1,weights2,bias2)) y=convert_to_output(y) prd=convert_to_output(prd) return ((float(np.sum(np.multiply((prd-y),(prd-y))))/(2*y.shape[0]))) def diff_act(a): #here sigmoid fucntion return np.multiply(activation(a),1-activation(a)) def accuracy_rmse(prd,y): y=convert_to_output(y) prd=convert_to_output(prd) print "confusion matrix :: " print confusion_matrix(prd,y) return float(100*np.sum(prd==y))/y.shape[0] def accuracy_average(prd,y): y=convert_to_output(y) prd=convert_to_output(prd) cls=np.zeros((classes,1)) for i in range(classes): cls[i]=np.sum(y==i) acc=0 for i in range(classes): if(cls[i] != 0): acc= acc + ( np.sum(prd==i)/cls[i]) return 100* np.divide(acc,(classes)) def accuracy_gm(prd,y): y=convert_to_output(y) prd=convert_to_output(prd) cls=np.zeros((classes,1)) for i in range(classes): cls[i]=np.sum(y==i) acc=1; cnf=confusion_matrix(prd,y) for i in range(classes): if(cls[i] != 0): a=np.sum(prd==i) acc= acc* (cnf[i,i]/float(cls[i])) return 100* np.power(acc,float(1/float(classes))) def back_prop(xtrain,weights1,bias1,weights2,bias2,ytrain,alpha,yval,xval): #prd=convert_to_output(feed_forward(xtrain,weights1,bias1,weights2,bias2)) #print compute_error(prd,ytrain) for i in range(0,epochs): a1= np.dot(xtrain,np.transpose(weights1))+bias1 h1=activation(a1) a2= np.dot(h1,np.transpose(weights2))+bias2 prd=activation(a2) delta2=np.multiply(np.multiply((prd-ytrain),prd),1- prd ) dW2= np.dot(np.transpose(delta2),h1) weights2=weights2 - alpha*dW2 bias2=bias2-alpha*np.sum(delta2) delta1=(np.multiply(np.multiply(np.dot(delta2,weights2),h1),(1-h1))) dW1= np.dot(np.transpose(delta1),xtrain) weights1=weights1 - alpha*dW1 bias1=bias1-alpha*np.sum(dW1) print compute_error(xtrain,weights1,bias1,weights2,bias2,ytrain) ta1= np.dot(test,np.transpose(weights1))+bias1 th1=activation(ta1) ta2= np.dot(th1,np.transpose(weights2))+bias2 tprd=activation(ta2) prd2=prd ytrain2=ytrain print "\n\nOverall accuracy ::\n" print "on training data ::" print accuracy_rmse(prd2,ytrain2) print "\n" print "on test data ::" print accuracy_rmse(tprd,answer) print "\n\nGeometric mean accuracy ::\n" print "on training data ::" print accuracy_gm(prd2,ytrain2) print "\n" print "on test data ::" print accuracy_gm(tprd,answer) print "\n" print "\n\nAverage mean accuracy ::\n" print "on training data ::" print accuracy_average(prd2,ytrain2) print "\n" print "on test data ::" print accuracy_average(tprd,answer) print "\n" feat=34 classes=2 epochs=1000 alpha=0.001 hidden=100 train=open("ION.tra") test=open("ION.tes") train=train.read() test=test.read() answer=open("ION.cla") answer=answer.read() train = [float(x) for x in train.split()] test = [float(x) for x in test.split()] answer = [float(x) for x in answer.split()] import numpy as np train=np.array(train) test=np.array(test) answer=np.array(answer) def change_shape(x,f): m=x.shape[0] return np.reshape(x,(m/f,f)) train=change_shape(train,feat+1) test=change_shape(test,feat) np.random.seed(25) xtrain=train[:,0:feat] ytrain=train[:,feat] total_train=xtrain.shape[0] total_val=int(total_train*0.1) #10-fold validation xval=xtrain[total_train-total_val:total_train,:] xtrain=xtrain[0:total_train-total_val,:] yval=ytrain[total_train-total_val:total_train] ytrain=ytrain[0:total_train-total_val] def one_hot(y): m=y.shape[0] narr=np.zeros((m,classes)) y=y.astype(int) for idx,i in enumerate(y): narr[idx][i-1]=1 return narr weights1=(np.random.rand(hidden,feat)-.5)/10 weights2=(np.random.rand(classes,hidden)-.5)/10 bias1=(np.random.rand(1,hidden)-.5)/10 bias2=(np.random.rand(1,classes)-.5)/10 ytrain=one_hot(ytrain) yval=one_hot(yval) answer=one_hot(answer) # feature normalization # the hero, savier in this code xtrain=featureNormalize(xtrain) xval=featureNormalize(xval) back_prop(xtrain,weights1,bias1,weights2,bias2,ytrain,alpha,yval,xval)
[ "vaibhav290797@gmail.com" ]
vaibhav290797@gmail.com
4766e7836e32da1470704406f4da46c5ba8fe537
1fabc88dcd905494baa9ef6797c9b0b512c95011
/Q12.py
df33329ecc3ca606977a358249dc1ec2e78e6e2b
[]
no_license
LYW-91118/python
f5847bef2c28f9f3daadf24f2decfc9aad7c3598
3798924f708752607242686398087dc3208b22fc
refs/heads/main
2023-04-16T22:50:49.669657
2021-04-28T06:53:13
2021-04-28T06:53:13
362,366,555
0
0
null
null
null
null
UTF-8
Python
false
false
237
py
num = input("輸入一整數序列:") s = num.split() sum = len(s) a = 0 for i in range(sum): if(s.count(s[i]) >= (sum/2)): print("過半元素為:%s" %(s[i])) a = 1 break if(a == 0 ): print("no")
[ "yoyo0928101611@gmail.com" ]
yoyo0928101611@gmail.com
a68861444e000bbcdbbae9937a8b09291c122dad
ceb23942414751fb5a14b46c6f1d90e40e41cb81
/Householder Method/QR_decomposition.py
c5dc0bf6d0326de5fe915c9b822a39446c269299
[ "MIT" ]
permissive
Kshitiz-Jain/Algorithms-1
fc06cb35b3621ab0efdef6c020109926fe2264eb
3039428ba48da4872777bc3b15f75447854a9642
refs/heads/master
2020-08-04T22:43:31.264046
2019-10-02T14:52:09
2019-10-02T14:52:09
212,302,229
0
0
MIT
2019-10-02T09:36:14
2019-10-02T09:36:14
null
UTF-8
Python
false
false
1,466
py
#Roll no.- Name #201651007 - Aman Yadav #201651029 - Mayank Pathela #201651060 - Anshuman Verma # Importing needful libraries import numpy as np # A function to evaluate Q & R for a given array def qr(A): m, n = A.shape Q = np.eye(m) for i in range(n - (m == n)): H = np.eye(m) # Computing matrix Hi H[i:, i:] = make_householder(A[i:, i]) # Finding orthogonal matrix Q = np.dot(Q, H) # Find new matrix A using Hi A = np.dot(H, A) return Q, A def make_householder(a): v = a / (a[0] + np.copysign(np.linalg.norm(a), a[0])) v[0] = 1 H = np.eye(a.shape[0]) H -= (2 / np.dot(v, v)) * np.dot(v[:, None], v[None, :]) return H def main(): # Taking in the number of rows m = int(input('\n Please enter the number of rows - ')) # Taking in the number of columns n = int(input('\n Please enter the number of columns - ')) if m < n: print('\nNumber of rows should be greater than or equal to number of columns') return # Creating a matrix of random values with dimensions m,n a = np.random.rand(m, n) print('\n Original Matrix\n', a) # Printing original matrix q, r = qr(a) # Evaluating Q and R print('\n Q:\n', q.round(6)) # Printing Q print('\n R:\n', r.round(6)) # Printing R # print(np.matmul(q,r)) #Sample print of multiplication of q,r to check the correctness if __name__ == '__main__': main()
[ "mp.pathela@gmail.com" ]
mp.pathela@gmail.com
af0fc5daf91f3b8d1cc064e06a4521d8bbcdd22e
e4d298905eaa0ba0d5fae062cfa6167669825a96
/Snake_game.py
670d25759fd6134d157292714aa855baa288c856
[]
no_license
Blockcipher123/Snake-Game
55f98cc58ff7d1f0005ee2ce513dae8412454786
e64e4dd8d7374292519bbd23a851656bd12377db
refs/heads/main
2023-08-11T08:17:39.600454
2021-09-26T04:55:50
2021-09-26T04:55:50
393,858,226
6
0
null
null
null
null
UTF-8
Python
false
false
6,013
py
import pygame import random import os pygame.mixer.init() pygame.init() # colors white = (255, 255,255) red = (255, 0, 0) black = (0,0, 0) green = (9, 237, 24) screen_width = 900 screen_hight = 600 # Creation Window gameWindow = pygame.display.set_mode((screen_width, screen_hight)) # Background Image bgimg = pygame.image.load("back2.jpg") bgimg = pygame.transform.scale(bgimg, (screen_width, screen_hight)).convert_alpha() # game over image gameing = pygame.image.load("firstintro.png") gameing = pygame.transform.scale(gameing, (screen_width, screen_hight)).convert_alpha() # Game title pygame.display.set_caption('Snakes_Game') pygame.display.update() clock = pygame.time.Clock() font = pygame.font.SysFont(None, 55) def text_screen(text, color, x, y): screen_text = font.render(text, True, color) gameWindow.blit(screen_text, [x,y]) def plot_snake(gameWindow, color, snk_list, snake_size): # print(snk_list) for x,y in snk_list: pygame.draw.rect(gameWindow,color,[x, y, snake_size, snake_size]) def welcome(): exit_game = False while not exit_game: gameWindow.fill((220,100,229)) text_screen("Welcome To Snake", black, 260, 250) text_screen("Press Space Bar To Play", black, 230, 290) for event in pygame.event.get(): if event.type == pygame.QUIT: exit_game = True if event.type == pygame.KEYDOWN: if event.key == pygame.K_SPACE: pygame.mixer.music.load("back.mp3") pygame.mixer.music.play() gameloop() pygame.display.update() clock.tick(50) # Game loop def gameloop(): # Game specific variable exit_game = False game_over = False snake_x = 45 snake_y = 55 velocity_x = 0 velocity_y = 0 snk_list = [] snk_length = 1 # check if highscore file exists if (not os.path.exists("")): with open("hiscore.txt", "w") as f: f.write("0") with open("highscore.txt", 'r') as f: highscore = f.read() apple_x = random.randint(20,screen_width/2) apple_y = random.randint(20,screen_hight/2) score = 0 init_velocity = 5 snake_size = 30 fps = 50 while not exit_game: if game_over: with open("highscore.txt", 'w') as f: f.write(str(highscore)) gameWindow.fill((0,0,0)) gameWindow.blit(gameing,(5,5)) text_screen(f"Your Score is {score}", red, 320, 400) # foont = text_screen(f'By Block_Cipher', green, 500, 500) # foont1(Font(20)) # if score > highscore: # text_screen(f"Great, Score is {score}", red, 320, 400) for event in pygame.event.get(): # print(event) if event.type==pygame.QUIT: exit_game = True if event.type == pygame.KEYDOWN: if event.key == pygame.K_RETURN: # gameloop() welcome() else: for event in pygame.event.get(): # print(event) if event.type==pygame.QUIT: exit_game = True if event.type == pygame.KEYDOWN: if event.key == pygame.K_RIGHT: velocity_x = init_velocity velocity_y = 0 if event.key == pygame.K_LEFT: velocity_x = - init_velocity velocity_y = 0 if event.key == pygame.K_UP: velocity_y = - init_velocity velocity_x = 0 if event.key == pygame.K_DOWN: velocity_y = init_velocity velocity_x = 0 # if event.click == pygame.C_RIGHT: # velocity_x = init_velocity # velocity_y = 0 if event.key == pygame.K_q: score += 10 snake_x = snake_x + velocity_x snake_y = snake_y + velocity_y if abs (snake_x - apple_x) <15 and abs(snake_y - apple_y) <15: score += 10 apple_x = random.randint(20,screen_width/2) apple_y = random.randint(20,screen_hight/2) snk_length += 5 # print(highscore) if score>int(highscore): highscore = score gameWindow.fill(white) gameWindow.blit(bgimg, (0,0)) text_screen("Score : " + str(score) + " Hiscore : " + str(highscore), green, 5 , 5 ) pygame.draw.rect(gameWindow, red, [apple_x, apple_y, snake_size, snake_size]) head = [] head.append(snake_x) head.append(snake_y) snk_list.append(head) if len(snk_list)>snk_length: del snk_list[0] if head in snk_list[:-1]: game_over = True pygame.mixer.music.load("gameover.mp3") pygame.mixer.music.play() if snake_x<0 or snake_x>screen_width or snake_y<0 or snake_y>screen_hight: game_over = True pygame.mixer.music.load("gameover.mp3") pygame.mixer.music.play() # print("Game over ! -") # pygame.draw.rect(gameWindow,black,[snake_x, snake_y, snake_size, snake_size]) plot_snake(gameWindow, black, snk_list, snake_size) pygame.display.update() clock.tick(fps) pygame.quit() quit() welcome()
[ "noreply@github.com" ]
noreply@github.com
793c15be2778bfa6a0852f657ea403fc51e685ba
a3f793a53361d08f3e0cdedc7fab9df40e201eef
/main.py
a53882b59400172fbcb656c830535363798e384d
[]
no_license
songshanshi/imoocc_py3
156db4f072bc956f45cbcc8c61fca964be8acfb9
6f3491ce857c541bf55d5ed8993265b7dd4dee09
refs/heads/master
2020-04-28T02:25:18.241155
2018-10-16T07:20:15
2018-10-16T07:20:15
null
0
0
null
null
null
null
UTF-8
Python
false
false
6,917
py
#!/usr/bin/env python # -*- coding:utf-8 -*- ######################################################################### # Author:Jeson # Email:jeson@imoocc.com import datetime import os import re import yaml PROJECT_ROOT = os.path.realpath(os.path.dirname(__file__)) # import sys os.environ["DJANGO_SETTINGS_MODULE"] = 'admin.settings.local_cj' import django import time django.setup() from scanhosts.models import HostLoginifo from scanhosts.util.nmap_all_server import NmapNet from scanhosts.util.nmap_all_server import NmapDocker from scanhosts.util.nmap_all_server import NmapKVM from scanhosts.util.nmap_all_server import NmapVMX from scanhosts.util.nmap_all_server import snmp_begin from scanhosts.util.j_filter import FilterRules from scanhosts.util.get_pv_relation import GetHostType from detail.models import PhysicalServerInfo,ConnectionInfo,OtherMachineInfo,StatisticsRecord from operations.models import MachineOperationsInfo from scanhosts.util.nmap_all_server import NetDevLogin from admin.settings.local_cj import BASE_DIR import logging logger = logging.getLogger("django") from apps.detail.utils.machines import Machines # def net_begin(): # ''' # 开始执行网络扫描 # :return: # ''' # nm = NmapNet(oid='1.3.6.1.2.1.1.5.0',Version=2) # nm_res = nm.query() # print "...................",nm_res def main(): ''' 读取扫描所需配置文件 :return: ''' s_conf = yaml.load(open('conf/scanhosts.yaml')) s_nets = s_conf['hostsinfo']['nets'] s_ports = s_conf['hostsinfo']['ports'] s_pass = s_conf['hostsinfo']['ssh_pass'] s_cmds = s_conf['hostsinfo']['syscmd_list'] s_keys = s_conf['hostsinfo']['ssh_key_file'] s_blacks = s_conf['hostsinfo']['black_list'] s_emails = s_conf['hostsinfo']['email_list'] n_sysname_oid = s_conf['netinfo']['sysname_oid'] n_sn_oid = s_conf['netinfo']['sn_oids'] n_commu = s_conf['netinfo']['community'] n_login_sw = s_conf['netinfo']['login_enable'] n_backup_sw = s_conf['netinfo']['backup_enable'] n_backup_sever = s_conf['netinfo']['tfp_server'] d_pass = s_conf['dockerinfo']['ssh_pass'] starttime = datetime.datetime.now() ''' 扫描主机信息 ''' for nmap_type in s_nets: unkown_list,key_not_login_list = snmp_begin(nmap_type,s_ports,s_pass,s_keys,s_cmds,s_blacks,s_emails) ''' 扫描网络信息 ''' nm = NmapNet(n_sysname_oid,n_sn_oid,n_commu) if key_not_login_list: for item in key_not_login_list: is_net = nm.query(item) if is_net[0] or is_net[1]: HostLoginifo.objects.update_or_create(ip=item,hostname=is_net[0],sn=is_net[1],mathine_type="Network device") else: HostLoginifo.objects.update_or_create(ip=item,ssh_port=key_not_login_list[item][0],ssh_status=0) other_sn = item.replace('.','') ob = OtherMachineInfo.objects.filter(sn_key=other_sn) if not ob: print(".........................OtherMachineInfo",item,other_sn) OtherMachineInfo.objects.create(ip=item,sn_key=other_sn,reson_str=u"SSH端口存活,无法登录",oth_cab_id=1) if unkown_list: for item in unkown_list: is_net = nm.query(item) if is_net[0] or is_net[1]: HostLoginifo.objects.update_or_create(ip=item,hostname=is_net,mathine_type="Network device") else: HostLoginifo.objects.update_or_create(ip=item,ssh_status=0) other_sn = item.replace('.','') ob = OtherMachineInfo.objects.filter(sn_key=other_sn) if not ob: OtherMachineInfo.objects.create(ip=item,sn_key=other_sn,reson_str=u"IP存活,非Linux服务器",oth_cab_id=1) # ''' # 网络设备备份或者登录功能 # ''' # net_login_dct = {} # with open("%s/conf/net_dev.pass"%BASE_DIR,'r') as f: # for item in f.readlines(): # ip,username,passwd,en_passwd = re.split("\s+",item)[:4] # net_login_dct[ip] = (username,passwd,en_passwd) # if n_login_sw == "True": # res = NetDevLogin(dev_ips=net_login_dct,backup_sw=n_backup_sw,back_server=n_backup_sever) ''' 规则:主机信息,去重、生成关系字典 ''' ft = FilterRules() key_ip_dic = ft.run() ''' 梳理虚拟服务器主机于服务器信息 ''' pv = GetHostType() p_relate_dic = pv.get_host_type(key_ip_dic) ''' 更新宿主机类型中表对应关系 ''' ip_key_dic = {v:k for k,v in key_ip_dic.items()} docker_p_list = p_relate_dic["docker-containerd"] kvm_p_list = p_relate_dic["qemu-system-x86_64"] vmware_p_list = p_relate_dic["vmx"] for item in docker_p_list: PhysicalServerInfo.objects.filter(conn_phy__sn_key=ip_key_dic[item]).update(vir_type="1") for item in kvm_p_list: PhysicalServerInfo.objects.filter(conn_phy__sn_key=ip_key_dic[item]).update(vir_type="0") for item in vmware_p_list: PhysicalServerInfo.objects.filter(conn_phy__sn_key=ip_key_dic[item]).update(vir_type="2") ''' 扫描docker的宿主机和虚拟服务的关系 ''' ds = NmapDocker(s_cmds,d_pass,ip_key_dic) ds.do_nmap(docker_p_list) ''' 扫描KVM的宿主机和虚拟服务的关系 # ''' ks = NmapKVM(ip_key_dic) ks.do_nmap(kvm_p_list) ''' 扫描ESXI虚拟机配置 ''' ne = NmapVMX(vmware_p_list,ip_key_dic) ne.dosnmp() ''' 更新状态表,用户信息表 ''' c_sn_lst = [item.sn_key for item in ConnectionInfo.objects.all()] o_sn_lst = [item.sn_key for item in OtherMachineInfo.objects.all()] old_sn_list = [item.sn_key for item in MachineOperationsInfo.objects.all()] new_sn_lst = c_sn_lst + o_sn_lst diff_sn_lst = set(new_sn_lst + old_sn_list) for item in diff_sn_lst: try: nsin = MachineOperationsInfo.objects.filter(sn_key=item) if not nsin: MachineOperationsInfo.objects.create(sn_key=item) except Exception as e: print("Error:SN:%s not insert into database,reason is:%s"%(item,e)) logger.error("Error:SN:%s not insert into database,reason is:%s"%(item,e)) ''' 统计总数 ''' info_dic = Machines().get_all_count() StatisticsRecord.objects.create(all_count=info_dic['all_c'],pyh_count=info_dic['pyh_c'],net_count=info_dic['net_c'], other_count=info_dic['other_c'],vmx_count=info_dic['vmx_c'],kvm_count=info_dic['kvm_c'],docker_count=info_dic['docker_c']) endtime = datetime.datetime.now() totaltime = (endtime - starttime).seconds logger.info("{Finish:Use time %s s}"%totaltime) print("{Finish:Use time %s s}"%totaltime) if __name__ == "__main__": main()
[ "gengming8859@icloud.com" ]
gengming8859@icloud.com
51372281f96d983816766c266f8a2e1c3e0a83f6
57206f40094fc5b0c8ff41ed8a4cec34bfd04ab5
/Vehicle detection/speedCal.py
31c276a7b812c787afb80524f9458de293bb93d1
[]
no_license
wuhaostudy1992/ECPS210
4a7b0b8e580b07997134fad4e6cd30d6e057acdd
d68493cbba25f800e9ec57619485b6231e8724bf
refs/heads/master
2020-03-09T09:53:55.633723
2018-05-11T04:28:17
2018-05-11T04:28:17
128,723,714
0
0
null
null
null
null
UTF-8
Python
false
false
1,795
py
# coding=utf-8 import numpy as np import cv2 cap = cv2.VideoCapture("high.flv") # Shi-Tomasi feature_params = dict(maxCorners=10, qualityLevel=0.1, minDistance=1, blockSize=9) # LK lk_params = dict(winSize=(30, 30), maxLevel=2, criteria=(cv2.TERM_CRITERIA_EPS | cv2.TERM_CRITERIA_COUNT, 10, 0.03)) # create random color color = np.random.randint(0, 255, (100, 3)) # get the first frame and turn to gery ret, old_frame = cap.read() old_gray = cv2.cvtColor(old_frame, cv2.COLOR_BGR2GRAY) # ST p0 = cv2.goodFeaturesToTrack(old_gray, mask=None, **feature_params) mask = np.zeros_like(old_frame) while 1: ret, frame = cap.read() if frame is None: cv2.waitKey(0) break else: frame_gray = cv2.cvtColor(frame, cv2.COLOR_BGR2GRAY) # 计算光流 p1, st, err = cv2.calcOpticalFlowPyrLK(old_gray, frame_gray, p0, None, **lk_params) # 选择好的特征点 if p1 is None: pass elif p0 is None: pass else: good_new = p1[st == 1] good_old = p0[st == 1] # 输出每一帧内特征点的坐标 # 坐标个数为之前指定的个数 #print(good_new) # 绘制轨迹 for i, (new, old) in enumerate(zip(good_new, good_old)): a, b = new.ravel() c, d = old.ravel() mask = cv2.line(mask, (a, b), (c, d), color[i].tolist(), 2) frame = cv2.circle(frame, (a, b), 5, color[i].tolist(), -1) img = cv2.add(frame, mask) cv2.imshow('frame', img) k = cv2.waitKey(30) & 0xff if k == 27: break # 更新上一帧以及特征点 old_gray = frame_gray.copy() p0 = good_new.reshape(-1, 1, 2) cv2.destroyAllWindows() cap.release()
[ "wuhaostudy1992@gmail.com" ]
wuhaostudy1992@gmail.com
fa4a0b305407775dbe2e6907373318daf7bef42b
7e2351db6ee3d9bf1edaf728855c7a5934e8b7dc
/examples/tutorials/lighting_tutorial.py
d2d0052338def553a03dbae1ec4a11415c4fcff2
[ "MIT" ]
permissive
alstar8/habitat-sim
b8fd764fb57b8cb678c7ac03ec1f95f4b3012d01
fc3261f89732c4712ca1db00b24487a983641d62
refs/heads/main
2023-08-14T11:15:50.555848
2021-10-03T11:39:03
2021-10-03T11:39:03
413,045,671
0
0
MIT
2021-10-03T10:31:02
2021-10-03T10:31:02
null
UTF-8
Python
false
false
7,095
py
# [setup] import math import os import magnum as mn import numpy as np from matplotlib import pyplot as plt import habitat_sim from habitat_sim.gfx import LightInfo, LightPositionModel from habitat_sim.utils.common import quat_from_angle_axis dir_path = os.path.dirname(os.path.realpath(__file__)) data_path = os.path.join(dir_path, "../../data") output_path = os.path.join(dir_path, "lighting_tutorial_output/") save_index = 0 def show_img(data, save): plt.figure(figsize=(12, 12)) plt.imshow(data, interpolation="nearest") plt.axis("off") plt.show(block=False) if save: global save_index plt.savefig( output_path + str(save_index) + ".jpg", bbox_inches="tight", pad_inches=0, quality=50, ) save_index += 1 plt.pause(1) def get_obs(sim, show, save): obs = sim.get_sensor_observations()["rgba_camera"] if show: show_img(obs, save) return obs def place_agent(sim): # place our agent in the scene agent_state = habitat_sim.AgentState() agent_state.position = [5.0, 0.0, 1.0] agent_state.rotation = quat_from_angle_axis( math.radians(70), np.array([0, 1.0, 0]) ) * quat_from_angle_axis(math.radians(-20), np.array([1.0, 0, 0])) agent = sim.initialize_agent(0, agent_state) return agent.scene_node.transformation_matrix() def make_configuration(): # simulator configuration backend_cfg = habitat_sim.SimulatorConfiguration() backend_cfg.scene_id = "data/scene_datasets/habitat-test-scenes/van-gogh-room.glb" backend_cfg.enable_physics = True # agent configuration sensor_cfg = habitat_sim.CameraSensorSpec() sensor_cfg.resolution = [1080, 960] agent_cfg = habitat_sim.agent.AgentConfiguration() agent_cfg.sensor_specifications = [sensor_cfg] return habitat_sim.Configuration(backend_cfg, [agent_cfg]) # [/setup] # This is wrapped such that it can be added to a unit test def main(show_imgs=True, save_imgs=False): if save_imgs and not os.path.exists(output_path): os.mkdir(output_path) # [default scene lighting] # create the simulator and render flat shaded scene cfg = make_configuration() sim = habitat_sim.Simulator(cfg) agent_transform = place_agent(sim) get_obs(sim, show_imgs, save_imgs) # [scene swap shader] # close the simulator and re-initialize with DEFAULT_LIGHTING_KEY: sim.close() cfg = make_configuration() cfg.sim_cfg.scene_light_setup = habitat_sim.gfx.DEFAULT_LIGHTING_KEY sim = habitat_sim.Simulator(cfg) agent_transform = place_agent(sim) get_obs(sim, show_imgs, save_imgs) # create and register new light setup: my_scene_lighting_setup = [ LightInfo(vector=[0.0, 2.0, 0.6, 0.0], model=LightPositionModel.Global) ] sim.set_light_setup(my_scene_lighting_setup, "my_scene_lighting") # reconfigure with custom key: new_cfg = make_configuration() new_cfg.sim_cfg.scene_light_setup = "my_scene_lighting" sim.reconfigure(new_cfg) agent_transform = place_agent(sim) get_obs(sim, show_imgs, save_imgs) # [/scene] # reset to default scene shading sim.close() cfg = make_configuration() sim = habitat_sim.Simulator(cfg) agent_transform = place_agent(sim) # noqa: F841 # [example 2] # get the rigid object attributes manager, which manages # templates used to create objects obj_template_mgr = sim.get_object_template_manager() # get the rigid object manager, which provides direct # access to objects rigid_obj_mgr = sim.get_rigid_object_manager() # load some object templates from configuration files sphere_template_id = obj_template_mgr.load_configs( str(os.path.join(data_path, "test_assets/objects/sphere")) )[0] chair_template_id = obj_template_mgr.load_configs( str(os.path.join(data_path, "test_assets/objects/chair")) )[0] # create a sphere and place it at a desired location obj_1 = rigid_obj_mgr.add_object_by_template_id(sphere_template_id) obj_1.translation = [3.2, 0.23, 0.03] get_obs(sim, show_imgs, save_imgs) # [/example 2] # [example 3] # create a custom light setup my_default_lighting = [ LightInfo(vector=[2.0, 2.0, 1.0, 0.0], model=LightPositionModel.Camera) ] # overwrite the default DEFAULT_LIGHTING_KEY light setup sim.set_light_setup(my_default_lighting) get_obs(sim, show_imgs, save_imgs) # [/example 3] # [example 4] # create a chair and place it at a location with a specified orientation obj_2 = rigid_obj_mgr.add_object_by_template_id(chair_template_id) obj_2.rotation = mn.Quaternion.rotation(mn.Deg(-115), mn.Vector3.y_axis()) obj_2.translation = [3.06, 0.47, 1.15] get_obs(sim, show_imgs, save_imgs) # [/example 4] # [example 5] light_setup_2 = [ LightInfo( vector=[2.0, 1.5, 5.0, 1.0], color=[0.0, 100.0, 100.0], model=LightPositionModel.Global, ) ] sim.set_light_setup(light_setup_2, "my_custom_lighting") # [/example 5] rigid_obj_mgr.remove_all_objects() # [example 6] # create and place 2 chairs with custom light setups chair_1 = rigid_obj_mgr.add_object_by_template_id( chair_template_id, light_setup_key="my_custom_lighting" ) chair_1.rotation = mn.Quaternion.rotation(mn.Deg(-115), mn.Vector3.y_axis()) chair_1.translation = [3.06, 0.47, 1.15] chair_2 = rigid_obj_mgr.add_object_by_template_id( chair_template_id, light_setup_key="my_custom_lighting" ) chair_2.rotation = mn.Quaternion.rotation(mn.Deg(50), mn.Vector3.y_axis()) chair_2.translation = [3.45927, 0.47, -0.624958] get_obs(sim, show_imgs, save_imgs) # [/example 6] # [example 7] existing_light_setup = sim.get_light_setup("my_custom_lighting") # create a new setup with an additional light new_light_setup = existing_light_setup + [ LightInfo( vector=[0.0, 0.0, 1.0, 0.0], color=[1.6, 1.6, 1.4], model=LightPositionModel.Camera, ) ] # register the old setup under a new name sim.set_light_setup(existing_light_setup, "my_old_custom_lighting") # [/example 7] # [example 8] # register the new setup overwriting the old one sim.set_light_setup(new_light_setup, "my_custom_lighting") get_obs(sim, show_imgs, save_imgs) # [/example 8] # [example 9] chair_1.set_light_setup(habitat_sim.gfx.DEFAULT_LIGHTING_KEY) get_obs(sim, show_imgs, save_imgs) # [/example 9] if __name__ == "__main__": import argparse parser = argparse.ArgumentParser() parser.add_argument("--no-show-images", dest="show_images", action="store_false") parser.add_argument("--no-save-images", dest="save_images", action="store_false") parser.set_defaults(show_images=True, save_images=True) args = parser.parse_args() main(show_imgs=args.show_images, save_imgs=args.save_images)
[ "noreply@github.com" ]
noreply@github.com
d8d13467a7799672f45585a90fa8be260b9dd1a7
028e16089ebbbd81adf41789a8da3ba3e25323d0
/문자열 마음데로 정렬하기.py
f4bbb55c432e09c2ad24557bce06c0be8279bb49
[]
no_license
FlowerLSH/Study1
18b840a1d3345e39f759661c88aa3fbc63a649f0
bedc273d77ad7eb9fd214967dadcf9e49f21f2c8
refs/heads/master
2023-01-31T13:32:34.374341
2020-12-16T15:03:33
2020-12-16T15:03:33
278,109,655
0
0
null
null
null
null
UTF-8
Python
false
false
590
py
def solution(strings, n): answer = [] index = dict() arr1 = [] for i in strings: index[i] = i[n] arr1.append(i[n]) arr = list(set(arr1)) arr.sort() for j in arr: emp = [] for k in index: if index[k] == j: emp.append(k) emp.sort() answer.extend(emp) return answer def solution(strings, n): answer = [] for i in range(len(strings)): strings[i] = strings[i][n] + strings[i] strings.sort() for j in strings: answer.append(j[1:]) return answer
[ "noreply@github.com" ]
noreply@github.com
b8b2d228ddc37132400ffc22205e20ffee445333
c9cf408754c54743b3ac2e4cb963726f5204dce5
/src/nc_service/constant_rate_server.py
235c3d850c4c157b87b5c69b6d71761ee2dd2810
[]
no_license
bocattelan/python-snc-mgf-Bruno
29066777bedb54ddbe12f1863807890067f70ddf
7009ecf46e6a68b206c252477fb76a70fe713b11
refs/heads/master
2020-04-13T02:27:36.597327
2018-12-18T20:06:43
2018-12-18T20:06:43
162,903,276
0
0
null
null
null
null
UTF-8
Python
false
false
805
py
"""Implemented service classes for different distributions""" from nc_service.service import Service from utils.exceptions import ParameterOutOfBounds class ConstantRate(Service): """Constant rate service""" def __init__(self, rate: float) -> None: self.rate = rate def sigma(self, theta=0.0) -> float: # TODO: remove the sign constraints later if theta <= 0: raise ParameterOutOfBounds(f"theta = {theta} must be > 0") return 0.0 def rho(self, theta: float) -> float: # TODO: remove the sign constraints later if theta <= 0: raise ParameterOutOfBounds(f"theta = {theta} must be > 0") return self.rate def to_value(self, number=1): return "rate{0}={1}".format(str(number), str(self.rate))
[ "nikolaus@cs.uni-kl.de" ]
nikolaus@cs.uni-kl.de
2bb1e7e593dfb67298aa570a9c0e2c150b0dc54b
d0bd9c3c5539141c74e0eeae2fa6b7b38af84ce2
/src/cogent3/parse/__init__.py
7559bc6dcc006e4be1bcd02096d3c56f55fc2512
[ "BSD-3-Clause" ]
permissive
KaneWh1te/cogent3
150c72e2f80a6439de0413b39c4c37c09c9966e3
115e9eb5700627fdb24be61441a7e3e155c02c61
refs/heads/master
2023-07-29T00:32:03.742351
2021-04-20T04:32:00
2021-04-20T04:32:00
null
0
0
null
null
null
null
UTF-8
Python
false
false
956
py
#!/usr/bin/env python __all__ = [ "blast", "cigar", "clustal", "dialign", "ebi", "fasta", "gcg", "genbank", "gff", "locuslink", "ncbi_taxonomy", "newick", "nexus", "paml", "paml_matrix", "phylip", "rdb", "record", "record_finder", "sequence", "table", "tinyseq", "tree", "tree_xml", "unigene", ] __author__ = "" __copyright__ = "Copyright 2007-2021, The Cogent Project" __credits__ = [ "Gavin Huttley", "Peter Maxwell", "Rob Knight", "Catherine Lozupone", "Jeremy Widmann", "Matthew Wakefield", "Sandra Smit", "Greg Caporaso", "Zongzhi Liu", "Micah Hamady", "Jason Carnes", "Raymond Sammut", "Hua Ying", "Andrew Butterfield", "Marcin Cieslik", ] __license__ = "BSD-3" __version__ = "2021.04.20a" __maintainer__ = "Gavin Huttley" __email__ = "Gavin.Huttley@anu.edu.au" __status__ = "Production"
[ "Gavin.Huttley@anu.edu.au" ]
Gavin.Huttley@anu.edu.au
4e42e313b4e8f4517cca59865a67badc6b525b39
200df6cda6e54d56a4c800e10e6d5f248d7d59f2
/02-算法思想/广度优先搜索/778.水位上升的泳池中游泳(H).py
0d5613ed1b63a75e2a20984da04124b0b0f7e70b
[]
no_license
jh-lau/leetcode_in_python
b9b9a47d0b3ce29c3c56836b39decc3ec4487777
1d1876620a55ff88af7bc390cf1a4fd4350d8d16
refs/heads/master
2023-04-17T15:01:49.925774
2021-04-24T01:17:39
2021-04-24T01:17:39
192,735,826
0
0
null
null
null
null
UTF-8
Python
false
false
6,600
py
""" @Author : liujianhan @Date : 20/9/26 19:31 @Project : leetcode_in_python @FileName : 778.水位上升的泳池中游泳(H).py @Description : 在一个 N x N 的坐标方格 grid 中,每一个方格的值 grid[i][j] 表示在位置 (i,j) 的平台高度。 现在开始下雨了。当时间为 t 时,此时雨水导致水池中任意位置的水位为 t 。你可以从一个平台游向四周相邻的任意一个平台, 但是前提是此时水位必须同时淹没这两个平台。假定你可以瞬间移动无限距离,也就是默认在方格内部游动是不耗时的。 当然,在你游泳的时候你必须待在坐标方格里面。 你从坐标方格的左上平台 (0,0) 出发。最少耗时多久你才能到达坐标方格的右下平台 (N-1, N-1)? 示例 1: 输入: [[0,2],[1,3]] 输出: 3 解释: 时间为0时,你位于坐标方格的位置为 (0, 0)。 此时你不能游向任意方向,因为四个相邻方向平台的高度都大于当前时间为 0 时的水位。 等时间到达 3 时,你才可以游向平台 (1, 1). 因为此时的水位是 3,坐标方格中的平台没有比水位 3 更高的,所以你可以游向坐标方格中的任意位置 示例2: 输入: [[0,1,2,3,4],[24,23,22,21,5],[12,13,14,15,16],[11,17,18,19,20],[10,9,8,7,6]] 输出: 16 解释: 0 1 2 3 4 24 23 22 21 5 12 13 14 15 16 11 17 18 19 20 10 9 8 7 6 最终的路线用加粗进行了标记。 我们必须等到时间为 16,此时才能保证平台 (0, 0) 和 (4, 4) 是连通的   提示: 2 <= N <= 50. grid[i][j] 位于区间 [0, ..., N*N - 1] 内。 """ import bisect import sys from typing import List class Solution: # 228ms, 14MB @staticmethod def swim_in_water(grid: List[List[int]]) -> int: """ 并查集 @param grid: @return: """ n = len(grid) p = [[(i, j) for j in range(n)] for i in range(n)] # 并查集二维数组初始化 h = sorted([[grid[i][j], i, j] for j in range(n) for i in range(n)]) # 按高度对点排序 def f(a, b): if (a, b) != p[a][b]: p[a][b] = f(*p[a][b]) # 二元并查集,元组传参要用*解包 return p[a][b] k = 0 for t in range(max(grid[0][0], grid[-1][-1]), h[-1][0]): # 起点是两个对角的最大值,终点是整个数据里的最大高度 while h[k][0] <= t: _, i, j = h[k] for x, y in [(i + 1, j), (i, j + 1), (i - 1, j), (i, j - 1)]: if 0 <= x < n and 0 <= y < n: if grid[i][j] <= t and grid[x][y] <= t: (pi, pj), (px, py) = f(i, j), f(x, y) if (pi, pj) != (px, py): # 让符合时间空间条件且不相同的集合合并 p[px][py] = (pi, pj) k += 1 if f(0, 0) == f(n - 1, n - 1): # 首末元素属于同一个集合就返回答案 return t return h[-1][0] # 172ms,, 13.8MB @staticmethod def swim_in_water_v2(grid: List[List[int]]) -> int: """ BFS @param grid: @return: """ n = len(grid) c = {(0, 0)} # 访问标记 for t in range(max(grid[0][0], grid[-1][-1]), sys.maxsize): # 从首末元素的最大时间作为最开始的判断条件 p = c.copy() # 宽搜队列初始化,每个时间点的初始状态是上一轮时间访问标记过的坐标 while p: q = set() # 下一批宽搜队列 for i, j in p: if i == j == n - 1: # 如果走到目标了就返回时间 return t for x, y in [(i + 1, j), (i, j + 1), (i - 1, j), (i, j - 1)]: if 0 <= x < n and 0 <= y < n and grid[x][y] <= t and (x, y) not in c: # 符合时空条件就扩散地图 q |= {(x, y)} c |= {(x, y)} p = q # 128ms, 13.8MB @staticmethod def swim_in_water_v3(grid: List[List[int]]) -> int: """ 升序队列 @param grid: @return: """ n = len(grid) b = {(0, 0)} # 访问标记 p = [[grid[0][0], 0, 0]] # 升序队列初始化 t = 0 # 途径最大时间标记 while True: h, i, j = p.pop(0) t = max(t, h) if i == j == n - 1: # 找到终点就就返回时间 return t for x, y in [(i + 1, j), (i, j + 1), (i - 1, j), (i, j - 1)]: if 0 <= x < n and 0 <= y < n and (x, y) not in b: bisect.insort(p, [grid[x][y], x, y]) # 二分插入 b |= {(x, y)} # 140ms, 13.7MB @staticmethod def swim_in_water_v4(grid: List[List[int]]) -> int: """ 双向升序队列 @param grid: @return: """ n = len(grid) b, e = {(0, 0)}, {(n - 1, n - 1)} # 双向访问标记 p, q = [[grid[0][0], 0, 0]], [[grid[-1][-1], n - 1, n - 1]] # 双向升序队列初始化 t = 0 # 途径最大时间标记 while True: h, i, j = p.pop(0) t = max(t, h) if (i, j) in e: # 如果找到的点已经存在于另一个队列里,就返回答案 return t for x, y in [(i + 1, j), (i, j + 1), (i - 1, j), (i, j - 1)]: if 0 <= x < n and 0 <= y < n and (x, y) not in b: bisect.insort(p, [grid[x][y], x, y]) b |= {(x, y)} h, i, j = q.pop(0) # 从这里开始都是对称的,调换p,q,b,e就行。 t = max(t, h) if (i, j) in b: return t for x, y in [(i + 1, j), (i, j + 1), (i - 1, j), (i, j - 1)]: if 0 <= x < n and 0 <= y < n and (x, y) not in e: bisect.insort(q, [grid[x][y], x, y]) e |= {(x, y)} if __name__ == '__main__': test_cases = [ [[0, 2], [1, 3]], [[0, 1, 2, 3, 4], [24, 23, 22, 21, 5], [12, 13, 14, 15, 16], [11, 17, 18, 19, 20], [10, 9, 8, 7, 6]], ] for tc in test_cases: print(Solution.swim_in_water(tc)) print(Solution.swim_in_water_v2(tc)) print(Solution.swim_in_water_v3(tc)) print(Solution.swim_in_water_v4(tc))
[ "lorgerd@163.com" ]
lorgerd@163.com
fc9f01838570e13f69a2601763e4edf059156fb3
f5a92dfca6e4428992f539ef2819f797c1ddf324
/flask329/run.py
7890a0a49e6f65cb675d7b64aab282fabd8a81f4
[ "MIT" ]
permissive
ZJUguquan/CRF_for_entity_identify
c553013ce3c3cf630adb82a37aa7ad9106946b2f
416e7c96854265a5873ce2aa23a8cfbe19358576
refs/heads/master
2021-06-16T10:48:01.825782
2017-05-17T08:34:01
2017-05-17T08:34:01
null
0
0
null
null
null
null
UTF-8
Python
false
false
14,241
py
#coding:utf-8 from flask import Flask, render_template from flask_bootstrap import Bootstrap from flask_moment import Moment from flask_wtf import FlaskForm from wtforms import TextAreaField, SubmitField from wtforms.validators import Required import numpy as np from numpy import * import pandas as pd import json import jieba import jieba.analyse import thulac import requests import re import os import sys reload(sys) sys.setdefaultencoding('utf-8') #地点关联相应函数 #将文本文件导入到列表 def loadDataSet(fileName): dataSet = [] fr = open(fileName) i = 0 for line in fr.readlines(): curLine = line.strip().split('\t') fltLine = map(float, curLine) dataSet.append(fltLine) i += 1 return dataSet #单位km def distSLC(vecA, vecB): # 将十进制度数转化为弧度 lng1, lat1 = map(radians, vecA) lng2, lat2 = map(radians, vecB) # haversine公式 dlng = lng2 - lng1 dlat = lat2 - lat1 a = sin(dlat / 2) ** 2 + cos(lat1) * cos(lat2) * sin(dlng / 2) ** 2 c = 2 * arcsin(sqrt(a)) r = 6371 # 地球平均半径,单位为公里 return c * r def geocodeB(address): response = requests.get('http://api.map.baidu.com/geocoder/v2/?address=' + address + '&output=json&ak=CkTBCTLMZjwfPmM0KvMqug4xdbVtS2BL') ans = response.json() if ans['status'] == 0: return ans['result']['location']['lng'],ans['result']['location']['lat'] else: return 'NULL', 'NULL' #返回与当前点距离较近的点的集合 #address:输入地址,maxNum:最多取附近的地点的个数,maxDist:最大的距离范围 def nearby(dataSet, address, maxNum, maxDist): #判断输入地址的格式,并作适当调整 if address.startswith('浙江') or address.startswith('绍兴') or (address.find('市') != -1 and address.find('市区') == -1): lng, lat = geocodeB(address) else: lng, lat = geocodeB('浙江绍兴' + address) result = [] if lng == 'NULL': result.append('there is something wrong in the address!') else: near = [] dataSize = shape(dataSet)[0] for i in range(dataSize): dist = distSLC(dataSet[i], [lng, lat]) near.append(dist) nplist = np.array(near) index = np.argsort(nplist) num = 0 result = [] distance = [] for i in index: if nplist[i] < maxDist and num < maxNum: num += 1 result.append(i) distance.append(nplist[i]) return result, distance thu1 = thulac.thulac(user_dict=u'E:/project/绍兴公安项目/容留吸毒/data/shaoxingInfo.txt') app = Flask(__name__) app.config['SECRET_KEY'] = 'hard to guess string' bootstrap = Bootstrap(app) moment = Moment(app) #全局变量 resultDf=pd.read_csv(u'E:/project/绍兴公安项目/容留吸毒/data/resultDf.csv') nrow = resultDf.shape[0] df = pd.read_csv(u'E:/project/绍兴公安项目/容留吸毒/data/rlaq_u2.txt',sep='\t',encoding='gbk') jyaq = df['jyaq'].values dataSet = loadDataSet(u'E:/project/绍兴公安项目/容留吸毒/data/jingweidu.txt') #表单输入的类 class NameForm(FlaskForm): content = TextAreaField(u'请输入案情:', validators=[Required()], render_kw={'rows': 5, 'placeholder': u'如:2012年5月8日08时30分,本所在对已抓获吸毒人员胡国松的审查中,发现该犯于2012年4月1日18时30分至2012年4月28日凌晨2时40分,在解放北路隆来堂8438房间,提供吸毒工具,毒品冰毒,先后三次容留金浩勇吸食冰毒,每次共同吸食冰毒约0.2克。'}) submit = SubmitField(u'提交') #导出json的类 class NumPyEncoder(json.JSONEncoder): def default(self, obj): if isinstance(obj, np.int64) or isinstance(obj, np.int32): return int(obj) if isinstance(obj, np.float64) or isinstance(obj, np.float32): return float(obj) return json.JSONEncoder.default(self, obj) @app.errorhandler(404) def page_not_found(e): return render_template('404.html'), 404 @app.errorhandler(500) def internal_server_error(e): return render_template('500.html'), 500 @app.route('/', methods=['GET', 'POST']) def index(): ###案情要素提取 content = None form = NameForm() #待输出的结果 result = {'CJ':[],'AF':[],'DD':[],'RL':[],'BRL':[],'DP':[],'GJ':[],'XY':[]} keywords = [] if form.validate_on_submit(): #分词 content = form.content.data content = re.sub('\r|\n|\t| ','',content) fw = open('C:/Users/guquan/Desktop/CRF++-0.58/test/testData2.txt', 'w') segs = thu1.cut(content, text=True).split(' ') for seg in segs: spl = seg.split('_') fw.write(spl[0] + ' ' + spl[1] + '\n') fw.write('\n') fw.close() #执行CRF os.system('C:/Users/guquan/Desktop/CRF++-0.58/crf_test.exe -m C:/Users/guquan/Desktop/CRF++-0.58/test/model C:/Users/guquan/Desktop/CRF++-0.58/test/testData2.txt > C:/Users/guquan/Desktop/CRF++-0.58/test/result012.txt') #读取预测结果 lines = open('C:/Users/guquan/Desktop/CRF++-0.58/test/result012.txt').readlines() tags = ['CJ','AF','DD','RL','BRL','DP','GJ','XY'] #处理预测结果 strsText = [] strsTag = [] for line in lines: if len(line) > 1: text = line.split('\t')[0] tag = line.split('\t')[2] strsText.append(text) strsTag.append(tag) else: #每一整条结果 string = '' for i in range(len(strsTag)): if strsTag[i][0] == 'B' or strsTag[i][0] == 'M': tag = strsTag[i][2:-1] string = string + strsText[i] #结束点处理 if i+1<len(strsTag) and strsTag[i+1][0] != 'M' and strsTag[i+1][0] != 'E': result[tag].append(string) result[tag] = list(set(result[tag])) string = '' continue elif strsTag[i][0] == 'E': string = string + strsText[i] result[tag].append(string) result[tag] = list(set(result[tag])) string = '' continue #后期处理 result['XY'] = list(set(result['XY']) - set(result['RL'])) result['BRL'] = list(set(result['BRL']) - set(result['RL'])) if len(result['RL'])==0 and len(result['XY'])>0: result['RL'], result['XY'] = result['XY'], [] if len(result['RL'])==0 and len(result['XY'])==0: result['XY'], result['BRL'] = result['BRL'], [] if len(result['RL'])>0 and len(set(result['XY']) - set(result['BRL']))>0: result['XY'] = list(set(result['XY']) - set(result['BRL'])) #print result strsText = [] strsTag = [] slct = result for key in result.keys(): result[key] = ','.join(result[key]).decode('utf8') #关键词提取 jieba.analyse.set_idf_path("C:/Users/guquan/Desktop/CRF++-0.58/IDF_rongliu.txt") pat = reduce(lambda x,y:x+y, result.values()) pat = ('('+ '|'.join(pat)+')').decode('utf8') remain = re.sub(pat, '', content) keywords = jieba.analyse.extract_tags(remain, topK=10, allowPOS=('a', 'ad', 'an', 'n', 'ns', 'nt', 'nz', 'v', 'vd', 'vn', 'i', 'I')) slct['KW'] = keywords keywords = ','.join(keywords).decode('utf8') ###社会网络分析 ##根据涉案人关联 #'\uff0c'表示逗号 rl = slct['RL'].replace('[','').replace(']','').replace(' ','').split(u'\uff0c') brl = slct['BRL'].replace('[','').replace(']','').replace(' ','').split(u'\uff0c') xy = slct['XY'].replace('[','').replace(']','').replace(' ','').split(u'\uff0c') people = list(set(rl + brl + xy)) people = [p for p in people if len(p)>0] pat = ('('+ '|'.join(people)+')') pat = pat.encode('utf8') # num1 = [i for i,v in enumerate(resultDf.RL) if len(re.findall(pat, v))>0] # num2 = [i for i,v in enumerate(resultDf.BRL) if len(re.findall(pat, v))>0] # num3 = [i for i,v in enumerate(resultDf.XY) if len(re.findall(pat, v))>0] # num = list(set(num1 + num2 +num3)) num1 = resultDf.RL.str.contains(pat) num2 = resultDf.BRL.str.contains(pat) num3 = resultDf.XY.str.contains(pat) num = [(num1[i] or num2[i] or num3[i]) for i in range(nrow)] num = [i for i,v in enumerate(num) if v==True] print num ##根据地点关联 d0 = slct['DD'].replace('[','').replace(']','').replace(' ','').split(u'\uff0c')[0] d0 = d0.decode('utf8') near = ['NULL', 'NULL'] if len(d0)>0: near = nearby(dataSet, d0,10, 1) if near[0] != 'NULL': num = list(sorted(set(num+near[0]))) ##关联结果 res = resultDf.loc[num, ] print res #生成nodes和edges nrow2 = res.shape[0] ajbh = res.AJBH.tolist() index = res.index.values nodes = {} edges = [] nodes[d0] = {'id':d0, 'size': 3, 'group': 4, 'class': u'案发地点', 'num': -1, 'dist': 0} if len(d0)==0: nodes[d0]['id'] = u'本案案发地(未知)' nodes[u'AJ本案'] = {'id': u'AJ本案', 'size': len(people), 'group': 0, 'class': u'案件', 'num': -1, 'wzbh': u'无案件编号'} edges.append({'source': u'AJ本案', 'target': u'本案案发地(未知)', 'value': 1}) for p in people: edges.append({'source': u'AJ本案', 'target': p, 'value': 1}) for i in range(nrow2): aj = 'AJ'+ajbh[i][-8:] nodes[aj] = {'id': aj, 'size': 0, 'group': 0, 'class': u'案件', 'num': index[i], 'wzbh': ajbh[i]} for p in res.RL.tolist()[i][1:-1].replace(' ','').split(','): if len(p)>0: nodes[aj]['size'] += 1 p = p.decode('utf8') if p in nodes.keys(): nodes[p]['size'] += 3 nodes[p]['group'] = 1 nodes[p]['class'] = u'容留者' else : nodes[p] = {'id': p, 'size': 3, 'group': 1, 'class': u'容留者', 'num': index[i]} edges.append({'source': aj, 'target': p, 'value': 1}) for p in res.BRL.tolist()[i][1:-1].replace(' ','').split(','): if len(p)>0: nodes[aj]['size'] += 1 p = p.decode('utf8') if p in nodes.keys(): nodes[p]['size'] += 1.5 else : nodes[p] = {'id': p, 'size': 1.5, 'group': 3, 'class': u'被容留', 'num': index[i]} edges.append({'source': aj, 'target':p, 'value': 1}) for p in res.XY.tolist()[i][1:-1].replace(' ','').split(','): if len(p)>0: nodes[aj]['size'] += 1 p = p.decode('utf8') if p in nodes.keys(): nodes[p]['size'] += 2 else : nodes[p] = {'id':p, 'size': 2, 'group': 2, 'class': u'其他嫌疑人', 'num': index[i]} edges.append({'source': aj, 'target': p, 'value': 1}) d = res.DD.tolist()[i][1:-1].replace(' ','').split(',')[0] if len(d)>0: nodes[aj]['size'] += 1 d = d.decode('utf8') edges.append({'source': aj, 'target': d, 'value': 1}) #判断是否为本案地点 if d!=d0: if d in nodes.keys(): nodes[d]['size'] += 1 else : nodes[d] = {'id':d, 'size': 2, 'group': 4, 'class': u'案发地点', 'num': index[i]} #判断是否为根据经纬度关联到的地点 if near[0] != 'NULL' and (index[i] in near[0]): nodes[d]['dist'] = near[1][near[0].index(index[i])] edges.append({'source': d0, 'target': d, 'value': 1}) #每个nodes的相应说明info info = {} for item in nodes.items(): if item[1]['class'] == u'案件': #判断是否‘AJ本案’ if item[1]['num'] == -1: info[item[0]] = {'name': item[0], u'案件编号': item[1]['wzbh'], u'简要案情': content} else: info[item[0]] = {'name': item[0], u'案件编号': item[1]['wzbh'], u'简要案情': jyaq[item[1]['num']]} elif item[1]['class'] == u'案发地点': if 'dist' in item[1].keys(): info[item[0]] = {'name': item[0], u'与本案距离': str(round(item[1]['dist']*1000))+u'米'} else: info[item[0]] = {'name': item[0]} else: info[item[0]] = {'name': item[0], u'性别': 'XX', u'出生日期': '19XX-XX-XX'} #导出json供d3调用 nodes1 = [] for value in nodes.values(): nodes1.append(value) global data1 global data2 data1 = {'nodes': nodes1, 'links': edges} data2 = info return render_template('index.html', form=form, name=content, result=result, keywords=keywords) @app.route('/data1') def data1(): return json.dumps(data1,ensure_ascii=False,indent=2,cls=NumPyEncoder,encoding='utf8') @app.route('/data2') def data2(): return json.dumps(data2,ensure_ascii=False,indent=2,cls=NumPyEncoder,encoding='utf8',sort_keys=True) if __name__ == '__main__': app.run(port=111)
[ "guquan291410772@126.com" ]
guquan291410772@126.com
648a45ddee93210901fcbaf778e75181ac7c38e8
fba6fca82bf1edd8e69ef94cede17193b845c48d
/planer.py
f8f6b7202193cf0d4658884a66e5d68ecea28ca3
[]
no_license
sweettpickle/planer
9da0cc9b34e62563979f2757b733a1612b749918
cd505e6aebc5e3dcfdc66fabadaaaa4d75170998
refs/heads/master
2022-05-14T17:53:30.819859
2020-04-24T14:23:35
2020-04-24T14:23:35
258,514,324
0
0
null
null
null
null
UTF-8
Python
false
false
3,969
py
import telebot from telebot import types # bot = telebot.TeleBot('%ваш токен%') token = '1010919676:AAFlETQiiF6PUzGctcTFtNZLzCb12aVJjt4' bot = telebot.TeleBot(token) # обработчик сообщений @bot.message_handler(commands=['start']) def welcome(message): # bot.reply_to(message, message.text) # bot.send_message(message.chat.id, "Привет!") menu = types.ReplyKeyboardMarkup(resize_keyboard=True, row_width=1) buttom1 = types.KeyboardButton("Список привычек") buttom2 = types.KeyboardButton("Добавить привычку") buttom3 = types.KeyboardButton("Удалить привычку") menu.add(buttom1, buttom2, buttom3) bot.send_message(message.chat.id, "Выберите действие:", reply_markup=menu) done = "\u274c" not_done = "\u2b55\ufe0f" key = '' def create_progress(n): lst = [] for i in range(n): lst.append(not_done) return lst track = { "Спорт": create_progress(21), "Чтение 30 минут": create_progress(21) } @bot.message_handler(content_types=['text']) def get_message(message): if message.text == "Список привычек": inline = types.InlineKeyboardMarkup(row_width=1) for key in track.keys(): inline.add(types.InlineKeyboardButton(key, callback_data=key)) bot.send_message(message.chat.id, "Ваш список привычек:", reply_markup=inline) if message.text == "Добавить привычку": bot.register_next_step_handler(message, add_tracker) bot.send_message(message.chat.id, "Введите название:") if message.text == "Удалить привычку": bot.register_next_step_handler(message, del_tracker) bot.send_message(message.chat.id, "Введите название:") def add_tracker(message): if message.text in track: bot.send_message(message.chat.id, "Привычка с таким названием уже есть") else: global key key = message.text bot.register_next_step_handler(message, add_tracker2) bot.send_message(message.chat.id, "Введите количество дней:") def add_tracker2(message): track[key] = create_progress(int(message.text)) bot.send_message(message.chat.id, "Привычка добавлена") def del_tracker(message): if message.text in track: track.pop(message.text) bot.send_message(message.chat.id, "Привычка удалена") else: bot.send_message(message.chat.id, "Такой привычки нет") @bot.callback_query_handler(func=lambda call: True) def callback_inline(call): if call.data in track: global key key = call.data inline = types.InlineKeyboardMarkup(row_width=1) but = types.InlineKeyboardButton(''.join(track[key]), callback_data="check") inline.add(but) bot.edit_message_text(chat_id=call.message.chat.id, message_id=call.message.message_id, text=key, reply_markup=inline) elif call.data == "check": check(key) bot.edit_message_text(chat_id=call.message.chat.id, message_id=call.message.message_id, text=key, reply_markup=None) inline = types.InlineKeyboardMarkup(row_width=1) but = types.InlineKeyboardButton(''.join(track[key]), callback_data="check") inline.add(but) bot.edit_message_text(chat_id=call.message.chat.id, message_id=call.message.message_id, text=key, reply_markup=inline) bot.answer_callback_query(call.id, text="Отмечено") def check(key): # def check(key, id): lst = track.get(key) # lst = users[id].get(key) for i in range(len(lst)): if lst[i] == not_done: lst[i] = done break track[key] = lst bot.polling(none_stop=True)
[ "noreply@github.com" ]
noreply@github.com
18cbb1000278e07122b15b44881c258ca0146e5d
2ec96d507e800a8c669b1507a1bfc7b5aee697ac
/test/test_logging.py
b32bd591ce2603f9644f63853170ceb6f57d3feb
[ "Apache-2.0" ]
permissive
leizhen10000/rainbrow
719e83ac3702d6b2309c15bd26d4d4c5e49be199
6a61ed3550e9fc6d2ff48eb49fed0fb81f4a41c9
refs/heads/master
2020-03-19T04:57:47.181019
2019-07-09T10:02:40
2019-07-09T10:02:40
79,789,459
0
0
null
null
null
null
UTF-8
Python
false
false
976
py
#!/usr/bin/env python # -*- coding: utf-8 -*- """ # @Time : 2019-07-03 17:58 # @Author : Lei Zhen # @Contract: leizhen8080@gmail.com # @File : test_logging.py # @Software: PyCharm # code is far away from bugs with the god animal protecting I love animals. They taste delicious. ┏┓ ┏┓ ┏┛┻━━━┛┻┓ ┃ ☃ ┃ ┃ ┳┛ ┗┳ ┃ ┃ ┻ ┃ ┗━┓ ┏━┛ ┃ ┗━━━┓ ┃ 神兽保佑 ┣┓ ┃ 永无BUG┏┛ ┗┓┓┏━┳┓┏┛ ┃┫┫ ┃┫┫ ┗┻┛ ┗┻┛ """ from common.log_util import logger def test_logger(): logger.warning("Don't use this method") def test_error(): try: result = 10 / 0 except Exception: logger.error('Failed to get result', exc_info=True) logger.info('Finish')
[ "leizhen8080@gmail.com" ]
leizhen8080@gmail.com
d4a5b1ba6b6f1f3a11524fac579af53d35e04cf7
b18660ec434f8ebafeb5397690aa1b4c0a1cb528
/train_ALL_LSTM.py
528fe8ecf051ad46dbbf40705f292c601be2e094
[]
no_license
wp0517/pytorch_SRU
5d46956406c7b64431b736981f4565264ca9aa29
96be5b4f4f0b73a4e0532bb18d726655af0fdb50
refs/heads/master
2020-04-09T12:06:47.847348
2018-06-17T00:53:05
2018-06-17T00:53:05
null
0
0
null
null
null
null
UTF-8
Python
false
false
8,866
py
import os import sys import torch import torch.autograd as autograd import torch.nn.functional as F import torch.nn.utils as utils import torch.optim.lr_scheduler as lr_scheduler import shutil import random import hyperparams import time torch.manual_seed(hyperparams.seed_num) random.seed(hyperparams.seed_num) def train(train_iter, dev_iter, test_iter, model, args): if args.cuda: model = model.cuda() if args.Adam is True: print("Adam Training......") optimizer = torch.optim.Adam(model.parameters(), lr=args.lr, weight_decay=args.init_weight_decay) elif args.SGD is True: print("SGD Training.......") optimizer = torch.optim.SGD(model.parameters(), lr=args.lr, weight_decay=args.init_weight_decay, momentum=args.momentum_value) elif args.Adadelta is True: print("Adadelta Training.......") optimizer = torch.optim.Adadelta(model.parameters(), lr=args.lr, weight_decay=args.init_weight_decay) ''' lambda1 = lambda epoch: epoch // 30 lambda2 = lambda epoch: 0.99 ** epoch print("lambda1 {} lambda2 {} ".format(lambda1, lambda2)) scheduler = lr_scheduler.LambdaLR(optimizer, lr_lambda=[lambda2]) scheduler = lr_scheduler.StepLR(optimizer, step_size=1, gamma=0.9) ''' # scheduler = lr_scheduler.ReduceLROnPlateau(optimizer, 'min') lambda2 = lambda epoch: args.learning_rate_decay ** epoch scheduler = lr_scheduler.LambdaLR(optimizer, lr_lambda=[lambda2]) steps = 0 model_count = 0 model.train() time_list = [] for epoch in range(1, args.epochs+1): print("\n## 第{} 轮迭代,共计迭代 {} 次 !##\n".format(epoch, args.epochs)) scheduler.step() # print("now lr is {} \n".format(scheduler.get_lr())) print("now lr is {} \n".format(optimizer.param_groups[0].get("lr"))) for batch in train_iter: feature, target = batch.text, batch.label # feature.data.t_() target.data.sub_(1) # batch first, index align if args.cuda: feature, target = feature.cuda(), target.cuda() # print(feature) # target = autograd.Variable(target) # question 1 optimizer.zero_grad() model.zero_grad() model.hidden = model.init_hidden(args.lstm_num_layers, args.batch_size) if feature.size(1) != args.batch_size: # continue model.hidden = model.init_hidden(args.lstm_num_layers, feature.size(1)) # start_time = time.time() logit = model(feature) loss = F.cross_entropy(logit, target) start_time = time.time() loss.backward() end_time = time.time() time_list.append(end_time - start_time) # print("Backward Time is {} ".format(end_time - start_time)) if args.init_clip_max_norm is not None: # print("aaaa {} ".format(args.init_clip_max_norm)) utils.clip_grad_norm(model.parameters(), max_norm=args.init_clip_max_norm) optimizer.step() steps += 1 if steps % args.log_interval == 0: train_size = len(train_iter.dataset) # print("sadasd", torch.max(logit, 0)) corrects = (torch.max(logit, 1)[1].view(target.size()).data == target.data).sum() accuracy = float(corrects)/batch.batch_size * 100.0 sys.stdout.write( '\rBatch[{}/{}] - loss: {:.6f} acc: {:.4f}%({}/{})'.format(steps, train_size, loss.data[0], accuracy, corrects, batch.batch_size)) if steps % args.test_interval == 0: eval(dev_iter, model, args, scheduler) if steps % args.save_interval == 0: if not os.path.isdir(args.save_dir): os.makedirs(args.save_dir) save_prefix = os.path.join(args.save_dir, 'snapshot') save_path = '{}_steps{}.pt'.format(save_prefix, steps) torch.save(model, save_path) print("\n", save_path, end=" ") test_model = torch.load(save_path) model_count += 1 test_eval(test_iter, test_model, save_path, args, model_count) sum = 0 for index, value in enumerate(time_list): if index != 0: sum += value avg = sum / len(time_list) print("Time is {} ".format(avg)) return model_count def eval(data_iter, model, args, scheduler): model.eval() corrects, avg_loss = 0, 0 for batch in data_iter: feature, target = batch.text, batch.label target.data.sub_(1) # feature, target = batch.text, batch.label.data.sub_(1) if args.cuda is True: feature, target = feature.cuda(), target.cuda() model.hidden = model.init_hidden(args.lstm_num_layers, args.batch_size) if feature.size(1) != args.batch_size: model.hidden = model.init_hidden(args.lstm_num_layers, feature.size(1)) logit = model(feature) loss = F.cross_entropy(logit, target, size_average=False) avg_loss += loss.data[0] corrects += (torch.max(logit, 1)[1].view(target.size()).data == target.data).sum() size = len(data_iter.dataset) avg_loss = loss.data[0]/size accuracy = float(corrects)/size * 100.0 model.train() print('\nEvaluation - loss: {:.6f} acc: {:.4f}%({}/{}) \n'.format(avg_loss, accuracy, corrects, size)) def test_eval(data_iter, model, save_path, args, model_count): # print(save_path) model.eval() corrects, avg_loss = 0, 0 for batch in data_iter: feature, target = batch.text, batch.label target.data.sub_(1) if args.cuda: feature, target = feature.cuda(), target.cuda() # feature.data.t_() # target.data.sub_(1) # batch first, index align # target = autograd.Variable(target) if args.cuda: feature, target = feature.cuda(), target.cuda() model.hidden = model.init_hidden(args.lstm_num_layers, args.batch_size) if feature.size(1) != args.batch_size: # continue model.hidden = model.init_hidden(args.lstm_num_layers, feature.size(1)) logit = model(feature) loss = F.cross_entropy(logit, target, size_average=False) avg_loss += loss.data[0] corrects += (torch.max(logit, 1) [1].view(target.size()).data == target.data).sum() size = len(data_iter.dataset) avg_loss = loss.data[0]/size accuracy = float(corrects)/size * 100.0 model.train() print('\nEvaluation - loss: {:.6f} acc: {:.4f}%({}/{}) \n'.format(avg_loss, accuracy, corrects, size)) print("model_count {}".format(model_count)) # test result if os.path.exists("./Test_Result.txt"): file = open("./Test_Result.txt", "a") else: file = open("./Test_Result.txt", "w") file.write("model " + save_path + "\n") file.write("Evaluation - loss: {:.6f} acc: {:.4f}%({}/{}) \n".format(avg_loss, accuracy, corrects, size)) file.write("model_count {} \n".format(model_count)) file.write("\n") file.close() # calculate the best score in current file resultlist = [] if os.path.exists("./Test_Result.txt"): file = open("./Test_Result.txt") for line in file.readlines(): if line[:10] == "Evaluation": resultlist.append(float(line[34:41])) result = sorted(resultlist) file.close() file = open("./Test_Result.txt", "a") file.write("\nThe Current Best Result is : " + str(result[len(result) - 1])) file.write("\n\n") file.close() shutil.copy("./Test_Result.txt", "./snapshot/" + args.mulu + "/Test_Result.txt") # whether to delete the model after test acc so that to save space if os.path.isfile(save_path) and args.rm_model is True: os.remove(save_path)
[ "bamtercelboo@163.com" ]
bamtercelboo@163.com
41a00bab3f061077909f54d74dc574355af1929d
1b77eaf078321b1320d72aa36a4357568101e4ca
/江南大学教务处/test.py
93ac06b18e5699d2285b3f417e63ee409aaa3bec
[]
no_license
BEE-JN/python_homework
92ffc1216a380d124901fd64cc541f70813847dc
8ba4ea79cbd422f40e6f9f1cc5fed4d75715d207
refs/heads/master
2020-03-23T08:02:47.863607
2018-07-17T15:30:21
2018-07-17T15:30:21
141,305,118
1
0
null
null
null
null
UTF-8
Python
false
false
187
py
import requests import time url = "https://survey.zkeycc/pku/xsdc/?dm=bk" if __name__=='__main__': while 1: r=requests.get(url) print(r.content) time.sleep(1)
[ "41156190+GCS-CN@users.noreply.github.com" ]
41156190+GCS-CN@users.noreply.github.com
3fc886a7c14ade2e0cb83e4bcde5765fa6f76294
9144f651ca34ba324270bbcd216c6f8396aa4602
/pro4_2.py
56371678cf000e81ff854ba1890b524610a9cdec
[]
no_license
NishantJindal41/guvi3
1a26bf132e4c660b287b42e8b538d2496d8468f4
3d1ba6626b4684abacbbd0f9c0a3e2d44e88d85c
refs/heads/master
2020-04-15T06:19:28.831298
2019-04-16T16:24:22
2019-04-16T16:24:22
164,456,708
0
0
null
null
null
null
UTF-8
Python
false
false
271
py
n, m = map(int, input().split()) A = [] for i in range(n); A.append(list(map(int,input().split()))) for i in range(len(A)); A[i]=sorted(A[i]) A = [list(i) for i in zip(*x)] for i in range(len(A)); A[i]=sorted(A[i]) A = [list(i) for i in zip(*A)]; print(A); }
[ "noreply@github.com" ]
noreply@github.com
5b3b646c4113d5b24b5038e64dcdf1fcd7ee035b
fa572b453270fd688e91cbed75d488c24b86cb12
/lists/tests/test_forms.py
7f7270b28c3a1678dd2f22d5451771c213844173
[]
no_license
XOyarz/TDD-with-Python
d3bfea9ac4b0391058a6b8b2d759cde8c53e759c
d2350e70cd77691255a667cbff60910b36a30cc3
refs/heads/master
2021-01-21T13:26:52.011789
2017-09-11T18:39:45
2017-09-11T18:39:45
102,126,072
0
1
null
2017-09-11T18:39:46
2017-09-01T15:19:02
Python
UTF-8
Python
false
false
929
py
from django.test import TestCase from lists.forms import ItemForm, EMPTY_ITEM_ERROR from lists.models import Item, List class ItemFormTest(TestCase): def test_form_item_input_has_placeholder_and_css_classes(self): form = ItemForm() self.assertIn('placeholder="Enter a to-do item"', form.as_p()) self.assertIn('class="form-control input-lg"', form.as_p()) def test_form_validation_for_blank_items(self): form = ItemForm(data={'text':''}) self.assertFalse(form.is_valid()) self.assertEqual(form.errors['text'], [EMPTY_ITEM_ERROR]) def test_form_save_handles_saving_to_a_list(self): list_ = List.objects.create() form = ItemForm(data={'text': 'do me'}) new_item = form.save(for_list=list_) self.assertEqual(new_item, Item.objects.first()) self.assertEqual(new_item.text, 'do me') self.assertEqual(new_item.list, list_)
[ "xavier982@hotmail.com" ]
xavier982@hotmail.com
2efc3b8d4f8b6993091cf4bfe85bdb9711ec2a74
602ae5fca1a1d25d70cc3e1a84759d0caf124b57
/Dash Basics/dash_core_components_example.py
a3bf2c3dec79cc947119eba68e52c15102ca1de5
[]
no_license
harryschaefer93/DashAppPractice
540f05380575a8afea8f8db2c9f5ef70588ec25f
aa4a144df94a32e55a206d99dd89d431baa77ccf
refs/heads/master
2023-07-31T22:19:22.413393
2021-09-19T21:09:57
2021-09-19T21:09:57
321,131,982
0
0
null
null
null
null
UTF-8
Python
false
false
918
py
import dash import dash_html_components as html import dash_core_components as dcc app = dash.Dash() app.layout = html.Div([ html.Label('Dropdown Component:'), dcc.Dropdown( options=[ {'label': 'Houston', 'value': 'HTX'}, {'label': 'Austin', 'value': 'ATX'}, {'label': 'Dallas', 'value': 'DTX'}], value='HTX'), html.P(html.Label('Slider Component:')), #html.P inserts linebreak so labels aren't on top of eachother dcc.Slider( min=0, max=9, marks={i: '{}'.format(i) for i in range(10)}, value=5), html.P(html.Label('Radio Items Component')), dcc.RadioItems( options=[ {'label': 'Houston', 'value': 'HTX'}, {'label': 'Austin', 'value': 'ATX'}, {'label': 'Dallas', 'value': 'DTX'}], value='HTX') ]) if __name__ == '__main__': app.run_server()
[ "harryschaefer1993@gmail.com" ]
harryschaefer1993@gmail.com
59750fad06e56331b012fa6b1c81a5617cb2f861
27df339496f93c8ac45878130e774ae7ef8d1f61
/csv_file/csv_demo.py
ba634edff5b3a8db253f2636b7987edded93beaa
[]
no_license
vutukurus/class_project
9ce2df421f44c9b90c04997ff41e19dade3d6cb7
d714b4edafc02dcbb59fab4ba78480e1c0657b66
refs/heads/master
2021-01-12T05:25:33.411576
2017-03-12T14:27:28
2017-03-12T14:27:28
77,926,207
0
0
null
null
null
null
UTF-8
Python
false
false
446
py
import csv #CSV file reading in python.. f = open('demo_file.csv',"r") file_content = csv.reader(f) #it will read csv file contents.. for i in file_content: print i ''' #logic for printing salaries grater than 15k c=0 for i in file_content: if c > 0: if int(i[-1]) > 15000: print i[-1] c=c+1 ''' f_write = open('demo_emp.csv',"w") write_content = csv.writer(f_write) for i in file_content: del i[1] write_content.writerow(i)
[ "vutukurus@gmail.com" ]
vutukurus@gmail.com
c7eee2a22a1efb7ddb4b5278189a0424acae6d63
14ebcf98e7c64505839e0b7bbab89e32af7abe1e
/deep4rec/datasets/census.py
4d58743074fe1b8d1eb3004a4a18b2d8469a1f62
[ "Apache-2.0" ]
permissive
Luiz-FS/Deep4Rec
b021fbc36e377e1055e46e9a52f68c32018894f8
78c5ca74f0e0d06a9f4bb2f267817b69abd40d1d
refs/heads/master
2020-09-03T14:27:45.828007
2019-11-18T01:10:30
2019-11-18T01:10:30
219,484,411
0
0
Apache-2.0
2019-11-18T01:10:31
2019-11-04T11:22:00
Python
UTF-8
Python
false
false
6,110
py
# Copyright 2017 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # ============================================================================== """Dataset interface for Census dataset. Census dataset: https://archive.ics.uci.edu/ml/machine-learning-databases/adult """ import os import urllib.request import numpy as np import pandas as pd from sklearn.preprocessing import OrdinalEncoder from sklearn.preprocessing import OneHotEncoder import tensorflow as tf from deep4rec.datasets.dataset import Dataset import deep4rec.utils as utils _CSV_COLUMNS = [ "age", "workclass", "fnlwgt", "education", "education_num", "marital_status", "occupation", "relationship", "race", "gender", "capital_gain", "capital_loss", "hours_per_week", "native_country", "income_bracket", ] _CSV_COLUMN_DEFAULTS = [ [0], [""], [0], [""], [0], [""], [""], [""], [""], [""], [0], [0], [0], [""], [""], ] class CensusDataset(Dataset): url = "https://archive.ics.uci.edu/ml/machine-learning-databases/adult" def __init__(self, dataset_name, output_dir, *args, **kwargs): super().__init__(dataset_name, output_dir, *args, **kwargs) self.train_filename = "adult.data" self.test_filename = "adult.test" self.train_url = os.path.join(self.url, self.train_filename) self.test_url = os.path.join(self.url, self.test_filename) self.train_path = os.path.join(self.output_dir, self.train_filename) self.test_path = os.path.join(self.output_dir, self.test_filename) self.preprocessed_path = os.path.join(self.output_dir, self.dataset_name) self._ord_encoder = OrdinalEncoder() self._occupation_ord_encoder = OrdinalEncoder() self._one_hot_encoder = OneHotEncoder(sparse=False) def _download_and_clean_file(self, url, filename): """Downloads data from url, and makes changes to match the CSV format.""" temp_file, _ = urllib.request.urlretrieve(url) with tf.gfile.Open(temp_file, "r") as temp_eval_file: with tf.gfile.Open(filename, "w") as eval_file: for line in temp_eval_file: line = line.strip() line = line.replace(", ", ",") if not line or "," not in line: continue if line[-1] == ".": line = line[:-1] line += "\n" eval_file.write(line) tf.gfile.Remove(temp_file) def download(self): if not os.path.exists(self.output_dir): os.mkdir(self.output_dir) self._download_and_clean_file(self.train_url, self.train_path) self._download_and_clean_file(self.test_url, self.test_path) def check_downloaded(self): return os.path.exists(self.train_path) and os.path.exists(self.test_path) def check_preprocessed(self): return False def _preprocess(self, filename, train_data=False): df = pd.read_csv(filename, names=_CSV_COLUMNS) # Categorical columns df_base_columns = df[ ["education", "marital_status", "relationship", "workclass"] ] if train_data: base_columns = self._ord_encoder.fit_transform(df_base_columns.values) occupation_column = self._occupation_ord_encoder.fit_transform( df["occupation"].values.reshape(-1, 1) ) one_hot_base_columns = self._one_hot_encoder.fit_transform( df_base_columns.values ) else: base_columns = self._ord_encoder.transform(df_base_columns.values) occupation_column = self._occupation_ord_encoder.transform( df["occupation"].values.reshape(-1, 1) ) one_hot_base_columns = self._one_hot_encoder.transform( df_base_columns.values ) # Age buckets buckets = [0, 18, 25, 30, 35, 40, 45, 50, 55, 60, 65, 200] age_buckets = np.array( pd.cut(df["age"], buckets, labels=range(len(buckets) - 1)).values ) wide_columns = np.concatenate( (base_columns, age_buckets.reshape(-1, 1)), axis=1 ) numerical_columns = df[ ["age", "education_num", "capital_gain", "capital_loss", "hours_per_week"] ].values deep_columns = np.concatenate((one_hot_base_columns, numerical_columns), axis=1) labels = np.where(df["income_bracket"].values == ">50K", 1, 0) return wide_columns, deep_columns, occupation_column, labels def preprocess(self): self.train_wide_data, self.train_deep_data, self.train_embedding_data, self.train_y = self._preprocess( self.train_path, train_data=True ) self.test_wide_data, self.test_deep_data, self.test_embedding_data, self.test_y = self._preprocess( self.test_path, train_data=False ) @property def train_size(self): return len(self.train_wide_data) @property def train_features(self): return [self.train_embedding_data, self.train_wide_data, self.train_deep_data] @property def test_features(self): return [self.test_embedding_data, self.test_wide_data, self.test_deep_data] @property def num_features_one_hot(self): return len(np.unique(self.train_embedding_data)) @property def num_features(self): return 1
[ "mariannelinharesm@gmail.com" ]
mariannelinharesm@gmail.com
950368a1376a80cc13f1e58217778e6f36f4931f
b5bbbed97f1c52180751cde5cc187158ae98cec3
/football_api/urls.py
465ee7a3e2df25f2d2d502b84d4abd6ea0d93e1a
[ "MIT" ]
permissive
king-tomi/total-football-api
d4066fd4005ba71df445edf46ccaead5140fa126
39f8efbd8b658a5a2e52458dc594f8354d28da04
refs/heads/main
2023-07-18T12:48:17.648402
2021-09-07T23:42:42
2021-09-07T23:42:42
404,150,481
1
0
null
null
null
null
UTF-8
Python
false
false
1,584
py
"""football_api URL Configuration The `urlpatterns` list routes URLs to views. For more information please see: https://docs.djangoproject.com/en/3.2/topics/http/urls/ Examples: Function views 1. Add an import: from my_app import views 2. Add a URL to urlpatterns: path('', views.home, name='home') Class-based views 1. Add an import: from other_app.views import Home 2. Add a URL to urlpatterns: path('', Home.as_view(), name='home') Including another URLconf 1. Import the include() function: from django.urls import include, path 2. Add a URL to urlpatterns: path('blog/', include('blog.urls')) """ from django.contrib import admin from django.urls import path from django.urls.conf import include from rest_framework.documentation import include_docs_urls from rest_framework.schemas import get_schema_view API_TITLE = "Football API" DESCRIPTION = "This is an API that lets you collects, update and mutate information about football clubs, players and fixtures." view = get_schema_view(title=API_TITLE, version='1.0.0', url='https://footballapi.herokuapp.com') urlpatterns = [ path('admin/', admin.site.urls), path('api/v1/', include('football.urls')), path('api_auth/', include('rest_framework.urls')), path('api/v1/rest_auth/', include('rest_auth.urls')), path('api/v1/rest_auth/registration/', include('rest_auth.registration.urls')), path('docs/', include_docs_urls(title = API_TITLE, description=DESCRIPTION)), path('schema/', view) ]
[ "noreply@github.com" ]
noreply@github.com
b865fa83e9b8e72b08e144110aa75b200cf807d4
7ec04fc867d0a48fffc05c65bff9217cfe211fe7
/HW/统计字符串/countHotal.py
31f95d4500041b28fa209a1a7c4dae778ad024c5
[]
no_license
Cherry93/pythonPractic
3b9d1f99803503073bbb2f3a58009665338bd278
2889183af6c9a01ab47895b23e2d6ce8c288fd4d
refs/heads/master
2021-08-31T16:41:56.655989
2017-12-22T03:53:18
2017-12-22T03:53:18
115,008,198
0
0
null
null
null
null
UTF-8
Python
false
false
520
py
''' 统计加州旅馆中所有单词出现的次数,并降序打印 ''' import collections file = input("Enter a filename:") with open(file, 'r') as fpr: content = fpr.read() content = content.replace("\n", '') content1 = content.split() print(content1) print(content1[0].lower()) print(len(content1)) list =[] for i in range(0,len(content1)): list.append(content1[i].lower()) print(list) print("\n各单词出现的个数:\n%s"%collections.Counter(list)) #content2 = content1.lower() #print(content1)
[ "358544104@qq.com" ]
358544104@qq.com
071cdfa73d8b8928b72c1e5dd2ba4a8ba6f7578c
b7850a5605eea1ab876140e2ab9eea9b5e3b6305
/ControlFlujo/for.py
80c2c78706b7c7dc8c074f7ee4798cbe2b99f7a3
[]
no_license
lagarridom/PythonSemestral19_1
db37159983f842a2310f676e167f877fe93c6706
aad25b1a84822f52796f5faed251a62037a04c3a
refs/heads/master
2020-03-27T06:34:45.149558
2018-10-17T02:37:05
2018-10-17T02:37:05
146,116,479
1
1
null
null
null
null
UTF-8
Python
false
false
589
py
#10.- """ for ITERADOR in ITERABLE: instrucciones """ for i in [10, 20, 30, 40, 50]: print(i) nombres = ["luis", "pato", "gabriel"] for nom in nombres: print("El es"+nom) for elemento in ("cadena", 3, 3.4, True): print(type(elemento)) diccionario = {"lunes":"pollo","martes":"pescado","miercoles":"carne"} for llave in diccionario: print("Los %s me gusta comer %s"%(llave,diccionario[llave])) #print("Los {} me gusta comer {}".format(llave,diccionario[llave])) lista = [("Jorge",10),("Gueva",9),("Ana",10)] for nombre,calif in lista: print("%s obtuvo %s"%(nombre,calif))
[ "noreply@github.com" ]
noreply@github.com
96c1f73d17d18f7906615ca48bc2e2d25d8b7259
caa06eca3eef2549d5088f6487201f734b35822e
/multi_ie/EE/model/multi_pointer_net.py
0d2334ed6d7a09e87757e36528cedd3c228713c5
[]
no_license
kelvincjr/shared
f947353d13e27530ba44ea664e27de51db71a5b6
4bc4a12b0ab44c6847a67cbd7639ce3c025f38f8
refs/heads/master
2023-06-23T19:38:14.801083
2022-05-17T09:45:22
2022-05-17T09:45:22
141,774,490
6
1
null
2023-06-12T21:30:07
2018-07-21T02:22:34
Python
UTF-8
Python
false
false
7,694
py
# _*_ coding:utf-8 _*_ import warnings import numpy as np import torch import torch.nn as nn from transformers import BertModel from transformers import BertPreTrainedModel from .layernorm import ConditionalLayerNorm #from utils.data_util import batch_gather warnings.filterwarnings("ignore") def batch_gather(data: torch.Tensor, index: torch.Tensor): length = index.shape[0] t_index = index.cpu().numpy() t_data = data.cpu().data.numpy() result = [] for i in range(length): result.append(t_data[i, t_index[i], :]) return torch.from_numpy(np.array(result)).to(data.device) class ERENet(nn.Module): """ ERENet : entity relation jointed extraction """ def __init__(self, encoder, classes_num): super().__init__() self.classes_num = classes_num # BERT model self.bert = encoder config = encoder.config self.token_entity_emb = nn.Embedding(num_embeddings=2, embedding_dim=config.hidden_size, padding_idx=0) # self.encoder_layer = TransformerEncoderLayer(config.hidden_size, nhead=4) # self.transformer_encoder = TransformerEncoder(self.encoder_layer, num_layers=1) self.LayerNorm = ConditionalLayerNorm(config.hidden_size, eps=config.layer_norm_eps) # pointer net work self.po_dense = nn.Linear(config.hidden_size, self.classes_num * 2) self.subject_dense = nn.Linear(config.hidden_size, 2) self.loss_fct = nn.BCEWithLogitsLoss(reduction='none') #self.init_weights() def forward(self, q_ids=None, passage_ids=None, segment_ids=None, attention_mask=None, subject_ids=None, subject_labels=None, object_labels=None, eval_file=None, is_eval=False): mask = (passage_ids != 0).float() bert_encoder = self.bert(passage_ids, token_type_ids=segment_ids, attention_mask=mask)[0] if not is_eval: # subject_encoder = self.token_entity_emb(token_type_ids) # context_encoder = bert_encoder + subject_encoder sub_start_encoder = batch_gather(bert_encoder, subject_ids[:, 0]) sub_end_encoder = batch_gather(bert_encoder, subject_ids[:, 1]) subject = torch.cat([sub_start_encoder, sub_end_encoder], 1) context_encoder = self.LayerNorm(bert_encoder, subject) sub_preds = self.subject_dense(bert_encoder) po_preds = self.po_dense(context_encoder).reshape(passage_ids.size(0), -1, self.classes_num, 2) subject_loss = self.loss_fct(sub_preds, subject_labels) # subject_loss = F.binary_cross_entropy(F.sigmoid(sub_preds) ** 2, subject_labels, reduction='none') subject_loss = subject_loss.mean(2) subject_loss = torch.sum(subject_loss * mask.float()) / torch.sum(mask.float()) po_loss = self.loss_fct(po_preds, object_labels) # po_loss = F.binary_cross_entropy(F.sigmoid(po_preds) ** 4, object_labels, reduction='none') po_loss = torch.sum(po_loss.mean(3), 2) po_loss = torch.sum(po_loss * mask.float()) / torch.sum(mask.float()) loss = subject_loss + po_loss return loss else: subject_preds = nn.Sigmoid()(self.subject_dense(bert_encoder)) answer_list = list() for qid, sub_pred in zip(q_ids.cpu().numpy(), subject_preds.cpu().numpy()): context = eval_file[qid].bert_tokens start = np.where(sub_pred[:, 0] > 0.6)[0] end = np.where(sub_pred[:, 1] > 0.5)[0] subjects = [] for i in start: j = end[end >= i] if i == 0 or i > len(context) - 2: continue if len(j) > 0: j = j[0] if j > len(context) - 2: continue subjects.append((i, j)) answer_list.append(subjects) qid_ids, bert_encoders, pass_ids, subject_ids, token_type_ids = [], [], [], [], [] for i, subjects in enumerate(answer_list): if subjects: qid = q_ids[i].unsqueeze(0).expand(len(subjects)) pass_tensor = passage_ids[i, :].unsqueeze(0).expand(len(subjects), passage_ids.size(1)) new_bert_encoder = bert_encoder[i, :, :].unsqueeze(0).expand(len(subjects), bert_encoder.size(1), bert_encoder.size(2)) token_type_id = torch.zeros((len(subjects), passage_ids.size(1)), dtype=torch.long) for index, (start, end) in enumerate(subjects): token_type_id[index, start:end + 1] = 1 qid_ids.append(qid) pass_ids.append(pass_tensor) subject_ids.append(torch.tensor(subjects, dtype=torch.long)) bert_encoders.append(new_bert_encoder) token_type_ids.append(token_type_id) if len(qid_ids) == 0: subject_ids = torch.zeros(1, 2).long().to(bert_encoder.device) qid_tensor = torch.tensor([-1], dtype=torch.long).to(bert_encoder.device) po_tensor = torch.zeros(1, bert_encoder.size(1)).long().to(bert_encoder.device) return qid_tensor, subject_ids, po_tensor qids = torch.cat(qid_ids).to(bert_encoder.device) pass_ids = torch.cat(pass_ids).to(bert_encoder.device) bert_encoders = torch.cat(bert_encoders).to(bert_encoder.device) # token_type_ids = torch.cat(token_type_ids).to(bert_encoder.device) subject_ids = torch.cat(subject_ids).to(bert_encoder.device) flag = False split_heads = 1024 bert_encoders_ = torch.split(bert_encoders, split_heads, dim=0) pass_ids_ = torch.split(pass_ids, split_heads, dim=0) # token_type_ids_ = torch.split(token_type_ids, split_heads, dim=0) subject_encoder_ = torch.split(subject_ids, split_heads, dim=0) po_preds = list() for i in range(len(bert_encoders_)): bert_encoders = bert_encoders_[i] # token_type_ids = token_type_ids_[i] pass_ids = pass_ids_[i] subject_encoder = subject_encoder_[i] if bert_encoders.size(0) == 1: flag = True # print('flag = True**********') bert_encoders = bert_encoders.expand(2, bert_encoders.size(1), bert_encoders.size(2)) subject_encoder = subject_encoder.expand(2, subject_encoder.size(1)) # pass_ids = pass_ids.expand(2, pass_ids.size(1)) sub_start_encoder = batch_gather(bert_encoders, subject_encoder[:, 0]) sub_end_encoder = batch_gather(bert_encoders, subject_encoder[:, 1]) subject = torch.cat([sub_start_encoder, sub_end_encoder], 1) context_encoder = self.LayerNorm(bert_encoders, subject) po_pred = self.po_dense(context_encoder).reshape(subject_encoder.size(0), -1, self.classes_num, 2) if flag: po_pred = po_pred[1, :, :, :].unsqueeze(0) po_preds.append(po_pred) po_tensor = torch.cat(po_preds).to(qids.device) po_tensor = nn.Sigmoid()(po_tensor) return qids, subject_ids, po_tensor
[ "deco_2004@163.com" ]
deco_2004@163.com
e33f5e747a3394df1d4ab13d0f881353585a66d3
fb9b320109ba55fc68fab793ce7a77058dc8b682
/semi-supervised-learning/visual/score_visual_lfs.py
044956da5eebf4ddcac680083ba6c027aec29002
[]
no_license
NLP4H/MSBC
c3b03806666584a4fa1cc7328ba9d45f061d2a77
60b0b89496eb28707d323b595af7a411dbd84768
refs/heads/master
2022-10-09T19:02:48.998958
2020-06-02T07:15:11
2020-06-02T07:15:11
268,724,086
2
0
null
null
null
null
UTF-8
Python
false
false
14,366
py
""" Snorkel labelling functions for visual score Note: This visual subscore is not based on Neurostatus defns but based on heureustic information provided by Zhen based on advice given from MS clinicians """ import os import pandas as pd import re import numpy as np from nltk.tokenize import word_tokenize, sent_tokenize import snorkel from snorkel.labeling import labeling_function def predict_visual_acuity(note): """ Rules 1: 20/20 – 20/30 2: 20/30 – 20/60 3: 20/60 – 20/100 4: 20/100 – 20/200 Input: note Returns: raw visual acuity score """ score = -1 # Pattern p = re.compile(r" 20\/\d{2,3}") p2 = re.compile(r"visual acuity", re.IGNORECASE) if len(re.findall(p, note)) > 0: # List of possible visual acuities in each note visual_acuities = [] for acuity in re.findall(p, note): visual_acuities.append(int(acuity[4:])) # Take the worst disability worst_eye = max(visual_acuities) best_eye = min(visual_acuities) # vision improvement -> remove the worst one sentences = sent_tokenize(note) for sent in sentences: # in each sentence, look for visual aquity number and "vision improved" token if len(visual_acuities) > 1 and len(re.findall(r"(?:Vision|vision)", sent)) > 0 and len(re.findall(r"improv", sent)) > 0: # If originally is finger counting, than no use to remove if len(re.findall(r"finger counting vision", note)) > 0: break else: visual_acuities.remove(max(visual_acuities)) worst_eye = max(visual_acuities) break if len(visual_acuities) > 1 and len(re.findall(p, sent)) > 0 and len(re.findall(r"improv", sent)) > 0: visual_acuities.remove(max(visual_acuities)) worst_eye = max(visual_acuities) break # Vision recover if len(re.findall(r"(?:Vision|vision) recover", sent)) > 0: if len(re.findall(p, sent)) > 0: visual_acuities = [] for acuity in re.findall(p, sent): visual_acuities.append(int(acuity[4:])) worst_eye = max(visual_acuities) best_eye = min(visual_acuities) break else: score = 0 # print("Visual Acuity: ", score) return score # print("worst:", worst_eye) # print("best:", best_eye) # 20/20 normal if worst_eye == 20: score = 0 # print("Visual Acuity: ", score) return score # 1: 20/20 – 20/30 elif worst_eye > 20 and worst_eye <= 30: score = 1 # print("Visual Acuity: ", score) return score # 2: 20/30 – 20/60 elif worst_eye > 30 and worst_eye <= 60: score = 2 # print("Visual Acuity: ", score) return score # 3: 20/60 – 20/100 elif worst_eye > 60 and worst_eye <= 100: score = 3 # print("Visual Acuity: ", score) return score # 4: 20/100 – 20/200 elif (worst_eye > 100 and worst_eye <= 200) or \ (worst_eye != best_eye and worst_eye > 60 and worst_eye <= 100 and best_eye > 60 and best_eye <= 100): score = 4 # print("Visual Acuity: ", score) return score # 5: > 200 elif (worst_eye > 200) or \ (worst_eye != best_eye and worst_eye > 100 and worst_eye <= 200 and best_eye > 60 and best_eye <= 200): score = 5 # print("Visual Acuity: ", score) return score # 6: worst eye > 200, best eye >= 60 elif (worst_eye > 200): score = 6 # print("Visual Acuity: ", score) return score # "Visual acuity" is detected elif len(re.findall(p2, note)) > 0: sentences = sent_tokenize(note) for sent in sentences: if len(re.findall(p2, sent)) > 0 and len(re.findall(r"normal|Normal", sent)) > 0: score = 0 # print("Visual Acuity: ", score) return score # print("Visual Acuity: ", score) return score def predict_pallor(note): """ Check whether there's disc pallor Input: note Returns: score for disc pallor (maximum 1) """ # Patterns p = re.compile(r" disk | disc |fundoscopy| fundi | fundus|optic nerve", re.IGNORECASE) p_neg = re.compile(r" no | not |normal|unremarkable|crisp", re.IGNORECASE) p_abnormal = re.compile(r"pallor|pale", re.IGNORECASE) # Predictions score = -1 sentences = sent_tokenize(note) for sent in sentences: if len(re.findall(r"optic atrophy", sent)) > 0: score = 1 break if len(re.findall(r"temporal pallor|significant pallor|bilateral optic disc pallor", sent)) > 0: score = 1 break if len(re.findall(p, sent)) > 0: if len(re.findall(p_neg, sent)) > 0: score = 0 break elif len(re.findall(p_abnormal, sent)) > 0: score = 1 break # print("Pallor:", score) return score def predict_scotoma(note): """ Check scotoma 0: normal 1: small / no mention of size 2: large Input: note Returns: score for scotoma """ # Patterns p = re.compile(r"scotoma", re.IGNORECASE) p_neg = re.compile(r" no | deni|not have|not had", re.IGNORECASE) # Predictions score = -1 sentences = sent_tokenize(note) for sent in sentences: if len(re.findall(p, sent)) > 0: # print(sent) # Negation if len(re.findall(p_neg, sent)) > 0: score = 0 break # Large elif len(re.findall(r"large|Large", sent)) > 0: score = 2 break else: score = 1 break # print("Scotoma: ", score) return score def predict_visual_fields(note): """ Outputs: 0: if no change in visual field 1: if visual field got worst """ p = re.compile(r"visual field", re.IGNORECASE) p_neg = re.compile(r"full|intact|normal") # p2 = re.compile(r"hemianopsia", re.IGNORECASE) score = -1 sentences = sent_tokenize(note) for sent in sentences: if len(re.findall(p, sent)) > 0: if len(re.findall(p_neg, sent)) > 0: score = 0 elif len(re.findall(r"restrict", sent)) > 0: score = 1 # print("Visual Fields: ", score) return score def general_rule(note): """ Zhen's heurestics (developed through meetings with MS clinicians who label) Apply general rules where there's no specific description in the notes 1. Finger Counting 2. Light Perception """ # Normal # Some level of blindness # finger counting p1 = re.compile(r"count finger acuity|remains blind|left blind|right blind", re.IGNORECASE) score = -1 sentences = sent_tokenize(note) # TODO: Black and white|shapes and shadows for sent in sentences: # Normal if len(re.findall(r"no visual symptom", sent)) > 0: # print("No visual symptons") score = 0 break if len(re.findall(r"neurological exam", sent)) > 0 and len(re.findall(r"normal", sent)) > 0: # print("Neurological exam normal") score = 0 break if len(re.findall(r"otherwise|Otherwise", sent)) > 0 and len(re.findall(r"normal", sent)) > 0 and len(re.findall(r"visual|vision", sent)) == 0: score = 0 break if len(re.findall(r"EDSS", sent)) > 0 and len(re.findall(r"based on sensory", sent)) > 0: score = 0 break # Abnormal if len(re.findall(p1, sent)) > 0: # print("Blind/Finger counting") score = 6 break elif len(re.findall(r"finger counting", sent)) > 0 and len(re.findall(r"foot", sent)) > 0: # print("Finger counting 1 ft") score = 5 break elif len(re.findall(r"finger counting", sent)) > 0 and len(re.findall(r"2 feet|two feet|3 feet|three feet", sent)) > 0: # print("Finger counting 2/3 ft") score = 4 break elif len(re.findall(r"finger counting", sent)) > 0 and len(re.findall(r"light perception", sent)) > 0: # print("Finger counting & light perception") score = 6 break elif len(re.findall(r"EDSS", sent)) > 0 and len(re.findall(r"\s4", sent)) > 0 and len(re.findall(r"vision alone", sent)) > 0: # print("EDSS 4 related to vision") score = 6 break elif len(re.findall(r"EDSS", sent)) > 0 and len(re.findall(r"\s3", sent)) > 0 and len(re.findall(r"vision|visual sign", sent)) > 0: # print("EDSS 3 related to vision") score = 4 break elif len(re.findall(r"EDSS", sent)) > 0 and len(re.findall(r"\s2", sent)) > 0 and len(re.findall(r"vision|visual sign", sent)) > 0: score = 2 break elif len(re.findall(r"EDSS", sent)) > 0 and len(re.findall(r"\s4", sent)) > 0 and len(re.findall(r"loss of vision", sent)) > 0: # print("EDSS 4 related to vision") score = 4 break phrases = sent.split(",") for phrase in phrases: if len(re.findall(r"vision|visual", phrase)) > 0 and len(re.findall(r"significant", phrase)) > 0 and len(re.findall(r"impair", phrase)) > 0: if len(re.findall(r"improv", note)) > 0: break score = 6 break # print("General Rule: ", score) return score def select_neuro_exam(note): """ Function used for Zhen's heurestics """ p = re.compile(r"NEUROLOGICAL EXAMINATION:|EXAMINATION:|NEUROLOGICAL|(?:Neurological|neurological|neurologic|Neurologic) examination") p1 = re.compile(r"Cranial|Visual|Vision|On examination day") p2 = re.compile(r"examination|exam", re.IGNORECASE) sentences = sent_tokenize(note) start_index = 0 if len(re.findall(p, note)) > 0: for j in range(len(sentences)): if len(re.findall(p, sentences[j])) > 0: # start index = first sentence to mention neurological exam start_index = j else: for j in range(len(sentences)): if len(re.findall(p1, sentences[j])) > 0: start_index = j break elif len(re.findall(p2, sentences[j])) > 0: start_index = j break selected_note = " ".join([sentences[j] for j in range(start_index, len(sentences))]) return selected_note @labeling_function() def LF_visual_original(df_row): """ Visual subscore prediction based on Zhen's heurestics (developed through meeting with MS clinicians) Visual subscore is determined from the highest potential visual subscore from general_rule, or outputs from predict_visual_acuity, predict_pallor, predict_scotoma and predict_visual_fields This doesn't match with the neurostatus definitions, but seems to be a heurestics that's applied when labelling the function This apparently gives higher accuracy than just following neurostatus defns """ note = df_row.text if "edss_19" in np.asarray(df_row.index): edss_categorical = df_row.edss_19 else: edss_categorical = -1 # Unknown by default score = -1 selected_note = select_neuro_exam(note) # EDSS = 0 all scores 0 if edss_categorical == 0: score = 0 else: score = max(general_rule(selected_note), predict_visual_acuity(selected_note), predict_pallor(selected_note), predict_scotoma(selected_note), predict_visual_fields(selected_note)) return score def get_visual_lfs(): # Uncomment to test just new LFs return [LF_visual_original] # RULES # visual subscore depends on visual acuity, visual fields, scotoma, disc pallor # Visual acuity # 1: 20/20 – 20/30 # 2: 20/30 – 20/60 # 3: 20/60 – 20/100 # 4: 20/100 – 20/200 # disk pallor: # 0: none # 1: present # scotom, # 0: normal # 1: small / no mention of size # 2: large # visual fields # 0: healthy # 1: decline / restricted # score 0 # disc pallor: N/A = 0 # scotoma: N/A = 0 # visual field: N/A # visual acuity: normal = 1 # score 1 # if either one or all (and/or) # disc pallor: true = 1 # scotoma: small = 1 # visual field: N/A = 0 # visual acuity: 20/30 (0.67) - 20/20 (1.0) of worse eye = 2 # score 2: # disc pallor: N/A = 0 # scotoma: N/A = 0 # visual field: N/A = 0 # visual acuity: 20/30(0.67) - 20/59(0.34) of worse eye with maximal visual acuity (corrected) = 2 # score 3: # disc pallor: N/A = 0 # scotoma: large = 2 # visual field: moderate decrease = 1 # visual acuity: 20/60(0.33) - 20/99(0.21) of worse eye with maximal visual acuity (corrected) = 3 # score 4: # disc pallor: N/A = 0 # scotoma: N/A # visual field: decrease in worse eye # visual acuity: # 20/100(0.2) - 20/200(0.1) of worse eye with maximal visual acuity (corrected) = 4 # grade 3 # < 20/60 (0.33) for better eye = 1, 2 # score 5: # disc pallor: N/A = 0 # scotoma: N/A = 0 # visual field: N/A = 0 # visual acuity: # < 20/200 (0.1) for worse eye with maximal visual acuity = 1,2,3,4 # grade 4 # < 20/60 (0.33) for better eye with maximal visual acuity = 1,2 # score 6: # disc pallor: N/A = 0 # scotoma: N/A = 0 # visual field: N/A =0 # visual acuity: # grade 5 # < 20/60 (0.33) for better eye with maximal visual acuity = 1,2
[ "michalmalyska@Michals-MacBook-Air.local" ]
michalmalyska@Michals-MacBook-Air.local
1e6eba4819d445e4bc1753cf73177ddf6931adac
ae09e15b3f4ac2c6f5a0d3f7a0d73c6def37ab2b
/joinquant/get_data.py
eb55a21d0de07f52edf198b705d82a416b5a4de2
[]
no_license
fswzb/sensequant
db81053685a985fe50e7082b6f2e65e2e6540212
b5fdac13f5caf2d00b99506c1c25389b406d9b17
refs/heads/master
2020-08-01T15:53:35.102784
2016-08-01T03:56:49
2016-08-01T03:56:49
null
0
0
null
null
null
null
UTF-8
Python
false
false
3,713
py
import pandas as pd import numpy as np import itertools from src import configure import sys # Add the Test Folder path to the sys.path list sys.path.append('/home/lcc/sensequant/code') # Now you can import your module from src.preprocess import in_day_unit class get_data(): def __init__(self, fname): self.fname = configure.tech_hdf_file def history(self, start_date=None, end_date=None, count=None, filed='avg', stock_list=None): ''' the resulting data frame is in a unit of day start_date OR count ''' def read_data(fname): df_all = pd.DataFrame() for stock in stock_list: df = pd.read_hdf(fname, stock) df = df.drop('index') df = in_day_unit(df) df_all = df_all.append(df) return df_all def open_(df): return df.loc[0, 'open'] def close_(df): return df.iloc[-1]['close'] def low(df): if count: return np.sort(df.low.values)[:count] def high(df): return np.sort(df.high.values)[-count:] def avg(df): return np.average(df[:count].close.values) def pre_close(df): return scalify(df[df.date==np.sort(df.date)[-count]].close.values) def integrate_into_df(series, colname): return pd.DataFrame(series, columns=[colname]).reset_index() if start_date and count: raise ValueError('CAN set count or start_date!') df_all = read_data(self.fname, stock_list) df_all = df_all[df_all.date > pd.to_datetime(start_date)] if start_date else df_all groups = df_all.groupby('stock_id') if filed == 'open': result = groups.apply(open_) elif filed == 'low': result = groups.apply(close_) elif filed == 'high': result = groups.apply(high) elif filed == 'avg': result = groups.apply(ave) elif filed == 'pre_close': result = groups.apply(pre_close) else: raise ValueError('No such filed') return integrate_into_df(result, filed) def get_fundamentals(self, start_date, end_date, stock_id, colname_list): def has_same_element(l1, l2): ''' check whether two list have mutual elements ''' for e in l1: if e in l2: return True return False col4check = json.loads(open(configure.colnames_in_each_fundamental_df).read()) msk = (df.date>=start_date)&(df.date<=end_date)&(df.stock_id==stock_id) result = pd.DataFrame() for df_name, columns in col4check.items(): if has_same_element(stock_id, colname_list): df = pd.read_hdf(configure.fundamental_hdf_file, df_name, columns=columns_list+'stock_id'+'date', where=['date>=pd.to_datetime(%s)' % start_date, 'date<=pd.to_datetime(%s)' % end_date, 'stock_id==(%s)'% stock_id]) if result.empty: result = df else: result = pd.merge(info, share, on=['stock_id', 'date'], how='outer') return result def get_fundamental_items(self): col4check = json.loads(open(configure.colnames_in_each_fundamental_df).read()) for k,v in col4check.items(): print ('%s: %s'(% k, % v))
[ "rylanlzc@gmail.com" ]
rylanlzc@gmail.com
090808a81f4df895399f7bf0cacf2ebba9dc778e
dac7095e7b5ad4dae993871c1ae45cbb7a5ce5f7
/Character/14.Yashiro/Re/Yashiro_C.py
2c0cd5afc2266f92fa0884ed0406cc049d41656d
[]
no_license
Lastation/RenewalAniChaos
d12a8423f4b83cb019495c59ed059451e67e0483
c3edb29af58925de55c11110ccaf927d2b5d1b39
refs/heads/master
2023-08-24T11:28:35.614844
2023-08-22T21:23:14
2023-08-22T21:23:14
246,617,812
1
0
null
null
null
null
UTF-8
Python
false
false
2,242
py
import variable as v; import func.trig as trg; function main(playerID) { if (v.P_WaitMain[playerID] == 0) { if (v.P_CountMain[playerID] == 0) { KillUnitAt(All, " Creep. Dunkelheit", "Anywhere", playerID); if (v.P_LoopMain[playerID] < 2) { var d = 0; var n = 8; var r = 75 + 75 * v.P_LoopMain[playerID]; trg.Shape_Circle(playerID, 1, " Creep. Dunkelheit", d, n, r); trg.Shape_Circle(playerID, 1, "Kakaru (Twilight)", d, n, r); MoveLocation(v.P_LocationID[playerID], v.P_UnitID[playerID], playerID, "Anywhere"); MoveUnit(All, " Creep. Dunkelheit", playerID, "[Skill]Unit_Wait_ALL", v.P_LocationID[playerID]); Order(" Creep. Dunkelheit", playerID, "Anywhere", Attack, v.P_LocationID[playerID]); KillUnitAt(All, "Kakaru (Twilight)", "Anywhere", playerID); } else if (v.P_LoopMain[playerID] == 3) { var d = 0; var n = 8; var r = 150; trg.Shape_Circle(playerID, 1, "40 + 1n Ghost", d, n, r); MoveLocation(v.P_LocationID[playerID], v.P_UnitID[playerID], playerID, "Anywhere"); MoveUnit(All, "40 + 1n Ghost", playerID, "[Skill]Unit_Wait_ALL", v.P_LocationID[playerID]); Order("40 + 1n Ghost", playerID, "Anywhere", Attack, v.P_LocationID[playerID]); } else if (v.P_LoopMain[playerID] == 5) { KillUnitAt(All, "40 + 1n Ghost", "Anywhere", playerID); for (var i = 0; i < 3; i++) { var d = 0; var n = 8; var r = 50 + 50 * i; trg.Shape_Circle(playerID, 1, "40 + 1n Zergling", d, n, r); } KillUnitAt(All, "40 + 1n Zergling", "Anywhere", playerID); } trg.Main_Wait(160); v.P_LoopMain[playerID] += 1; if (v.P_LoopMain[playerID] == 6) { v.P_CountMain[playerID] += 1; v.P_LoopMain[playerID] = 0; } } else if (v.P_CountMain[playerID] == 1) { trg.SkillEnd(); } } }
[ "aaiiiho@gmail.com" ]
aaiiiho@gmail.com
7d3e8a94ef63a6dd931ce66016c56a243fb7a2e9
7a402c6bb3887af56ff8609453ef926fa41291a5
/LightGBMwithSImpleFeatures.py
ebcffd5eb8005c9fe27613e44bc60d7577bc6206
[]
no_license
lizihaoleo/home-credit-default-risk
cdcfe2ee7768b553205f3121a946535122b8626b
1a9f8d3ab107f8b3ed59634db71382f4965ecb0b
refs/heads/master
2021-10-18T05:27:36.397640
2019-02-14T03:54:31
2019-02-14T03:54:31
170,621,157
0
0
null
null
null
null
UTF-8
Python
false
false
17,019
py
# HOME CREDIT DEFAULT RISK COMPETITION # Most features are created by applying min, max, mean, sum and var functions to grouped tables. # Little feature selection is done and overfitting might be a problem since many features are related. # The following key ideas were used: # - Divide or subtract important features to get rates (like annuity and income) # - In Bureau Data: create specific features for Active credits and Closed credits # - In Previous Applications: create specific features for Approved and Refused applications # - Modularity: one function for each table (except bureau_balance and application_test) # - One-hot encoding for categorical features # All tables are joined with the application DF using the SK_ID_CURR key (except bureau_balance). # You can use LightGBM with KFold or Stratified KFold. Please upvote if you find usefull, thanks! # Update 16/06/2018: # - Added Payment Rate feature # - Removed index from features # - Set early stopping to 200 rounds # - Use standard KFold CV (not stratified) # Public LB increased to 0.792 import numpy as np import pandas as pd import gc import time from contextlib import contextmanager from lightgbm import LGBMClassifier from sklearn.metrics import roc_auc_score, roc_curve from sklearn.model_selection import KFold, StratifiedKFold import matplotlib.pyplot as plt import seaborn as sns import warnings warnings.simplefilter(action='ignore', category=FutureWarning) @contextmanager def timer(title): t0 = time.time() yield print("{} - done in {:.0f}s".format(title, time.time() - t0)) # One-hot encoding for categorical columns with get_dummies def one_hot_encoder(df, nan_as_category = True): original_columns = list(df.columns) categorical_columns = [col for col in df.columns if df[col].dtype == 'object'] df = pd.get_dummies(df, columns= categorical_columns, dummy_na= nan_as_category) new_columns = [c for c in df.columns if c not in original_columns] return df, new_columns # Preprocess application_train.csv and application_test.csv def application_train_test(num_rows = None, nan_as_category = False): # Read data and merge df = pd.read_csv('./application_train.csv', nrows= num_rows) test_df = pd.read_csv('./application_test.csv', nrows= num_rows) print("Train samples: {}, test samples: {}".format(len(df), len(test_df))) df = df.append(test_df).reset_index() # Optional: Remove 4 applications with XNA CODE_GENDER (train set) df = df[df['CODE_GENDER'] != 'XNA'] # Categorical features with Binary encode (0 or 1; two categories) for bin_feature in ['CODE_GENDER', 'FLAG_OWN_CAR', 'FLAG_OWN_REALTY']: df[bin_feature], uniques = pd.factorize(df[bin_feature]) # Categorical features with One-Hot encode df, cat_cols = one_hot_encoder(df, nan_as_category) # NaN values for DAYS_EMPLOYED: 365.243 -> nan df['DAYS_EMPLOYED'].replace(365243, np.nan, inplace= True) # Some simple new features (percentages) df['DAYS_EMPLOYED_PERC'] = df['DAYS_EMPLOYED'] / df['DAYS_BIRTH'] df['INCOME_CREDIT_PERC'] = df['AMT_INCOME_TOTAL'] / df['AMT_CREDIT'] df['INCOME_PER_PERSON'] = df['AMT_INCOME_TOTAL'] / df['CNT_FAM_MEMBERS'] df['ANNUITY_INCOME_PERC'] = df['AMT_ANNUITY'] / df['AMT_INCOME_TOTAL'] df['PAYMENT_RATE'] = df['AMT_ANNUITY'] / df['AMT_CREDIT'] del test_df gc.collect() return df # Preprocess bureau.csv and bureau_balance.csv def bureau_and_balance(num_rows = None, nan_as_category = True): bureau = pd.read_csv('./bureau.csv', nrows = num_rows) bb = pd.read_csv('./bureau_balance.csv', nrows = num_rows) bb, bb_cat = one_hot_encoder(bb, nan_as_category) bureau, bureau_cat = one_hot_encoder(bureau, nan_as_category) # Bureau balance: Perform aggregations and merge with bureau.csv bb_aggregations = {'MONTHS_BALANCE': ['min', 'max', 'size']} for col in bb_cat: bb_aggregations[col] = ['mean'] bb_agg = bb.groupby('SK_ID_BUREAU').agg(bb_aggregations) bb_agg.columns = pd.Index([e[0] + "_" + e[1].upper() for e in bb_agg.columns.tolist()]) bureau = bureau.join(bb_agg, how='left', on='SK_ID_BUREAU') bureau.drop(['SK_ID_BUREAU'], axis=1, inplace= True) del bb, bb_agg gc.collect() # Bureau and bureau_balance numeric features num_aggregations = { 'DAYS_CREDIT': ['min', 'max', 'mean', 'var'], 'DAYS_CREDIT_ENDDATE': ['min', 'max', 'mean'], 'DAYS_CREDIT_UPDATE': ['mean'], 'CREDIT_DAY_OVERDUE': ['max', 'mean'], 'AMT_CREDIT_MAX_OVERDUE': ['mean'], 'AMT_CREDIT_SUM': ['max', 'mean', 'sum'], 'AMT_CREDIT_SUM_DEBT': ['max', 'mean', 'sum'], 'AMT_CREDIT_SUM_OVERDUE': ['mean'], 'AMT_CREDIT_SUM_LIMIT': ['mean', 'sum'], 'AMT_ANNUITY': ['max', 'mean'], 'CNT_CREDIT_PROLONG': ['sum'], 'MONTHS_BALANCE_MIN': ['min'], 'MONTHS_BALANCE_MAX': ['max'], 'MONTHS_BALANCE_SIZE': ['mean', 'sum'] } # Bureau and bureau_balance categorical features cat_aggregations = {} for cat in bureau_cat: cat_aggregations[cat] = ['mean'] for cat in bb_cat: cat_aggregations[cat + "_MEAN"] = ['mean'] bureau_agg = bureau.groupby('SK_ID_CURR').agg({**num_aggregations, **cat_aggregations}) bureau_agg.columns = pd.Index(['BURO_' + e[0] + "_" + e[1].upper() for e in bureau_agg.columns.tolist()]) # Bureau: Active credits - using only numerical aggregations active = bureau[bureau['CREDIT_ACTIVE_Active'] == 1] active_agg = active.groupby('SK_ID_CURR').agg(num_aggregations) active_agg.columns = pd.Index(['ACTIVE_' + e[0] + "_" + e[1].upper() for e in active_agg.columns.tolist()]) bureau_agg = bureau_agg.join(active_agg, how='left', on='SK_ID_CURR') del active, active_agg gc.collect() # Bureau: Closed credits - using only numerical aggregations closed = bureau[bureau['CREDIT_ACTIVE_Closed'] == 1] closed_agg = closed.groupby('SK_ID_CURR').agg(num_aggregations) closed_agg.columns = pd.Index(['CLOSED_' + e[0] + "_" + e[1].upper() for e in closed_agg.columns.tolist()]) bureau_agg = bureau_agg.join(closed_agg, how='left', on='SK_ID_CURR') del closed, closed_agg, bureau gc.collect() return bureau_agg # Preprocess previous_applications.csv def previous_applications(num_rows = None, nan_as_category = True): prev = pd.read_csv('./previous_application.csv', nrows = num_rows) prev, cat_cols = one_hot_encoder(prev, nan_as_category= True) # Days 365.243 values -> nan prev['DAYS_FIRST_DRAWING'].replace(365243, np.nan, inplace= True) prev['DAYS_FIRST_DUE'].replace(365243, np.nan, inplace= True) prev['DAYS_LAST_DUE_1ST_VERSION'].replace(365243, np.nan, inplace= True) prev['DAYS_LAST_DUE'].replace(365243, np.nan, inplace= True) prev['DAYS_TERMINATION'].replace(365243, np.nan, inplace= True) # Add feature: value ask / value received percentage prev['APP_CREDIT_PERC'] = prev['AMT_APPLICATION'] / prev['AMT_CREDIT'] # Previous applications numeric features num_aggregations = { 'AMT_ANNUITY': ['min', 'max', 'mean'], 'AMT_APPLICATION': ['min', 'max', 'mean'], 'AMT_CREDIT': ['min', 'max', 'mean'], 'APP_CREDIT_PERC': ['min', 'max', 'mean', 'var'], 'AMT_DOWN_PAYMENT': ['min', 'max', 'mean'], 'AMT_GOODS_PRICE': ['min', 'max', 'mean'], 'HOUR_APPR_PROCESS_START': ['min', 'max', 'mean'], 'RATE_DOWN_PAYMENT': ['min', 'max', 'mean'], 'DAYS_DECISION': ['min', 'max', 'mean'], 'CNT_PAYMENT': ['mean', 'sum'], } # Previous applications categorical features cat_aggregations = {} for cat in cat_cols: cat_aggregations[cat] = ['mean'] prev_agg = prev.groupby('SK_ID_CURR').agg({**num_aggregations, **cat_aggregations}) prev_agg.columns = pd.Index(['PREV_' + e[0] + "_" + e[1].upper() for e in prev_agg.columns.tolist()]) # Previous Applications: Approved Applications - only numerical features approved = prev[prev['NAME_CONTRACT_STATUS_Approved'] == 1] approved_agg = approved.groupby('SK_ID_CURR').agg(num_aggregations) approved_agg.columns = pd.Index(['APPROVED_' + e[0] + "_" + e[1].upper() for e in approved_agg.columns.tolist()]) prev_agg = prev_agg.join(approved_agg, how='left', on='SK_ID_CURR') # Previous Applications: Refused Applications - only numerical features refused = prev[prev['NAME_CONTRACT_STATUS_Refused'] == 1] refused_agg = refused.groupby('SK_ID_CURR').agg(num_aggregations) refused_agg.columns = pd.Index(['REFUSED_' + e[0] + "_" + e[1].upper() for e in refused_agg.columns.tolist()]) prev_agg = prev_agg.join(refused_agg, how='left', on='SK_ID_CURR') del refused, refused_agg, approved, approved_agg, prev gc.collect() return prev_agg # Preprocess POS_CASH_balance.csv def pos_cash(num_rows = None, nan_as_category = True): pos = pd.read_csv('./POS_CASH_balance.csv', nrows = num_rows) pos, cat_cols = one_hot_encoder(pos, nan_as_category= True) # Features aggregations = { 'MONTHS_BALANCE': ['max', 'mean', 'size'], 'SK_DPD': ['max', 'mean'], 'SK_DPD_DEF': ['max', 'mean'] } for cat in cat_cols: aggregations[cat] = ['mean'] pos_agg = pos.groupby('SK_ID_CURR').agg(aggregations) pos_agg.columns = pd.Index(['POS_' + e[0] + "_" + e[1].upper() for e in pos_agg.columns.tolist()]) # Count pos cash accounts pos_agg['POS_COUNT'] = pos.groupby('SK_ID_CURR').size() del pos gc.collect() return pos_agg # Preprocess installments_payments.csv def installments_payments(num_rows = None, nan_as_category = True): ins = pd.read_csv('./installments_payments.csv', nrows = num_rows) ins, cat_cols = one_hot_encoder(ins, nan_as_category= True) # Percentage and difference paid in each installment (amount paid and installment value) ins['PAYMENT_PERC'] = ins['AMT_PAYMENT'] / ins['AMT_INSTALMENT'] ins['PAYMENT_DIFF'] = ins['AMT_INSTALMENT'] - ins['AMT_PAYMENT'] # Days past due and days before due (no negative values) ins['DPD'] = ins['DAYS_ENTRY_PAYMENT'] - ins['DAYS_INSTALMENT'] ins['DBD'] = ins['DAYS_INSTALMENT'] - ins['DAYS_ENTRY_PAYMENT'] ins['DPD'] = ins['DPD'].apply(lambda x: x if x > 0 else 0) ins['DBD'] = ins['DBD'].apply(lambda x: x if x > 0 else 0) # Features: Perform aggregations aggregations = { 'NUM_INSTALMENT_VERSION': ['nunique'], 'DPD': ['max', 'mean', 'sum'], 'DBD': ['max', 'mean', 'sum'], 'PAYMENT_PERC': ['max', 'mean', 'sum', 'var'], 'PAYMENT_DIFF': ['max', 'mean', 'sum', 'var'], 'AMT_INSTALMENT': ['max', 'mean', 'sum'], 'AMT_PAYMENT': ['min', 'max', 'mean', 'sum'], 'DAYS_ENTRY_PAYMENT': ['max', 'mean', 'sum'] } for cat in cat_cols: aggregations[cat] = ['mean'] ins_agg = ins.groupby('SK_ID_CURR').agg(aggregations) ins_agg.columns = pd.Index(['INSTAL_' + e[0] + "_" + e[1].upper() for e in ins_agg.columns.tolist()]) # Count installments accounts ins_agg['INSTAL_COUNT'] = ins.groupby('SK_ID_CURR').size() del ins gc.collect() return ins_agg # Preprocess credit_card_balance.csv def credit_card_balance(num_rows = None, nan_as_category = True): cc = pd.read_csv('./credit_card_balance.csv', nrows = num_rows) cc, cat_cols = one_hot_encoder(cc, nan_as_category= True) # General aggregations cc.drop(['SK_ID_PREV'], axis= 1, inplace = True) cc_agg = cc.groupby('SK_ID_CURR').agg(['min', 'max', 'mean', 'sum', 'var']) cc_agg.columns = pd.Index(['CC_' + e[0] + "_" + e[1].upper() for e in cc_agg.columns.tolist()]) # Count credit card lines cc_agg['CC_COUNT'] = cc.groupby('SK_ID_CURR').size() del cc gc.collect() return cc_agg # LightGBM GBDT with KFold or Stratified KFold # Parameters from Tilii kernel: https://www.kaggle.com/tilii7/olivier-lightgbm-parameters-by-bayesian-opt/code def kfold_lightgbm(df, num_folds, stratified = False, debug= False): # Divide in training/validation and test data train_df = df[df['TARGET'].notnull()] test_df = df[df['TARGET'].isnull()] print("Starting LightGBM. Train shape: {}, test shape: {}".format(train_df.shape, test_df.shape)) del df gc.collect() # Cross validation model if stratified: folds = StratifiedKFold(n_splits= num_folds, shuffle=True, random_state=1001) else: folds = KFold(n_splits= num_folds, shuffle=True, random_state=1001) # Create arrays and dataframes to store results oof_preds = np.zeros(train_df.shape[0]) sub_preds = np.zeros(test_df.shape[0]) feature_importance_df = pd.DataFrame() feats = [f for f in train_df.columns if f not in ['TARGET','SK_ID_CURR','SK_ID_BUREAU','SK_ID_PREV','index']] for n_fold, (train_idx, valid_idx) in enumerate(folds.split(train_df[feats], train_df['TARGET'])): train_x, train_y = train_df[feats].iloc[train_idx], train_df['TARGET'].iloc[train_idx] valid_x, valid_y = train_df[feats].iloc[valid_idx], train_df['TARGET'].iloc[valid_idx] # LightGBM parameters found by Bayesian optimization clf = LGBMClassifier( nthread=4, n_estimators=10000, learning_rate=0.02, num_leaves=34, colsample_bytree=0.9497036, subsample=0.8715623, max_depth=8, reg_alpha=0.041545473, reg_lambda=0.0735294, min_split_gain=0.0222415, min_child_weight=39.3259775, silent=-1, verbose=-1, ) clf.fit(train_x, train_y, eval_set=[(train_x, train_y), (valid_x, valid_y)], eval_metric= 'auc', verbose= 100, early_stopping_rounds= 200) oof_preds[valid_idx] = clf.predict_proba(valid_x, num_iteration=clf.best_iteration_)[:, 1] sub_preds += clf.predict_proba(test_df[feats], num_iteration=clf.best_iteration_)[:, 1] / folds.n_splits fold_importance_df = pd.DataFrame() fold_importance_df["feature"] = feats fold_importance_df["importance"] = clf.feature_importances_ fold_importance_df["fold"] = n_fold + 1 feature_importance_df = pd.concat([feature_importance_df, fold_importance_df], axis=0) print('Fold %2d AUC : %.6f' % (n_fold + 1, roc_auc_score(valid_y, oof_preds[valid_idx]))) del clf, train_x, train_y, valid_x, valid_y gc.collect() print('Full AUC score %.6f' % roc_auc_score(train_df['TARGET'], oof_preds)) # Write submission file and plot feature importance if not debug: test_df['TARGET'] = sub_preds test_df[['SK_ID_CURR', 'TARGET']].to_csv(submission_file_name, index= False) display_importances(feature_importance_df) return feature_importance_df # Display/plot feature importance def display_importances(feature_importance_df_): cols = feature_importance_df_[["feature", "importance"]].groupby("feature").mean().sort_values(by="importance", ascending=False)[:40].index best_features = feature_importance_df_.loc[feature_importance_df_.feature.isin(cols)] plt.figure(figsize=(8, 10)) sns.barplot(x="importance", y="feature", data=best_features.sort_values(by="importance", ascending=False)) plt.title('LightGBM Features (avg over folds)') plt.tight_layout() plt.savefig('lgbm_importances01.png') def main(debug = False): num_rows = 10000 if debug else None df = application_train_test(num_rows) with timer("Process bureau and bureau_balance"): bureau = bureau_and_balance(num_rows) print("Bureau df shape:", bureau.shape) df = df.join(bureau, how='left', on='SK_ID_CURR') del bureau gc.collect() with timer("Process previous_applications"): prev = previous_applications(num_rows) print("Previous applications df shape:", prev.shape) df = df.join(prev, how='left', on='SK_ID_CURR') del prev gc.collect() with timer("Process POS-CASH balance"): pos = pos_cash(num_rows) print("Pos-cash balance df shape:", pos.shape) df = df.join(pos, how='left', on='SK_ID_CURR') del pos gc.collect() with timer("Process installments payments"): ins = installments_payments(num_rows) print("Installments payments df shape:", ins.shape) df = df.join(ins, how='left', on='SK_ID_CURR') del ins gc.collect() with timer("Process credit card balance"): cc = credit_card_balance(num_rows) print("Credit card balance df shape:", cc.shape) df = df.join(cc, how='left', on='SK_ID_CURR') del cc gc.collect() with timer("Run LightGBM with kfold"): feat_importance = kfold_lightgbm(df, num_folds= 5, stratified= False, debug= debug) if __name__ == "__main__": submission_file_name = "submission_kernel02.csv" with timer("Full model run"): main()
[ "lizihaoleo@gmail.com" ]
lizihaoleo@gmail.com
357383ee731f6f3e548ae395bba056b0b0a71aeb
9acbda554a136235ef2f605281ec39e6c6d8cd63
/billman/settings.py
7b47311e624aa21bf2d35d7e64239afeefdaf76a
[]
no_license
arpheno/billman
23dfa9d6b05ce022ec8b2165c6e5bdf2dad51d91
11b6d9386d4e450f75b9c809a48626b2379a39eb
refs/heads/master
2020-04-10T09:22:55.242570
2014-06-07T17:37:51
2014-06-07T17:37:51
null
0
0
null
null
null
null
UTF-8
Python
false
false
2,086
py
""" Django settings for billman project. For more information on this file, see https://docs.djangoproject.com/en/1.6/topics/settings/ For the full list of settings and their values, see https://docs.djangoproject.com/en/1.6/ref/settings/ """ # Build paths inside the project like this: os.path.join(BASE_DIR, ...) import os BASE_DIR = os.path.dirname(os.path.dirname(__file__)) # Quick-start development settings - unsuitable for production # See https://docs.djangoproject.com/en/1.6/howto/deployment/checklist/ # SECURITY WARNING: keep the secret key used in production secret! SECRET_KEY = 'abwsz7w&+fzjjwjn%uql(*f=s^fy_$d1d#oc$9)q_v@weh9fp#' # SECURITY WARNING: don't run with debug turned on in production! DEBUG = True TEMPLATE_DEBUG = True ALLOWED_HOSTS = [] # Application definition INSTALLED_APPS = ( 'django.contrib.admin', 'django.contrib.auth', 'django.contrib.contenttypes', 'django.contrib.sessions', 'django.contrib.messages', 'django.contrib.staticfiles', ) TEMPLATE_DIRS = (os.path.join(BASE_DIR, 'templates'),) STATICFILES_DIRS = (os.path.join(BASE_DIR, 'static'),) MIDDLEWARE_CLASSES = ( 'django.contrib.sessions.middleware.SessionMiddleware', 'django.middleware.common.CommonMiddleware', 'django.middleware.csrf.CsrfViewMiddleware', 'django.contrib.auth.middleware.AuthenticationMiddleware', 'django.contrib.messages.middleware.MessageMiddleware', 'django.middleware.clickjacking.XFrameOptionsMiddleware', ) ROOT_URLCONF = 'billman.urls' WSGI_APPLICATION = 'billman.wsgi.application' # Database # https://docs.djangoproject.com/en/1.6/ref/settings/#databases DATABASES = { 'default': { 'ENGINE': 'django.db.backends.sqlite3', 'NAME': os.path.join(BASE_DIR, 'db.sqlite3'), } } # Internationalization # https://docs.djangoproject.com/en/1.6/topics/i18n/ LANGUAGE_CODE = 'en-us' TIME_ZONE = 'UTC' USE_I18N = True USE_L10N = True USE_TZ = True # Static files (CSS, JavaScript, Images) # https://docs.djangoproject.com/en/1.6/howto/static-files/ STATIC_URL = '/static/'
[ "ciupakm@gmail.com" ]
ciupakm@gmail.com
f107a42e17a213bb257e6dc9bee18367a2d43d35
c7a867c33675d48c9bcc73c70c27cac085661ebb
/extractor/POSMmanagement/process.py
a6b472f4c094cbd47cf5cd7e2c6fc14894009bd4
[]
no_license
v55448330/posm
3e4cbcb22f5eae17c956eb02346a8fc5a932966c
a53c15337301a769ac3b9bde54ab845ac0fe5211
refs/heads/master
2020-05-29T11:05:40.229015
2015-03-29T08:21:22
2015-03-29T08:21:22
47,541,652
1
0
null
2015-12-07T09:15:42
2015-12-07T09:15:42
null
UTF-8
Python
false
false
4,559
py
# -*- coding: utf-8 -*- import logging LOG = logging.getLogger(__file__) import subprocess import psycopg2 import sys import os from .utils import proc_exec class ProcessManagement(): def __init__(self, settings, verbose=False): self.verbose = verbose self.settings = settings.get_settings() self.db_params = settings.db_params def processAdminLevels(self, settings_file): command = [ 'python', 'extract.py', '--settings', settings_file, '--problems_as_geojson' ] LOG.debug('Command: %s', ' '.join(command)) proc = subprocess.Popen( command, stdout=subprocess.PIPE, stderr=subprocess.PIPE, shell=False ) # execute the process ... .wait() admin_level_data_path = os.path.join( self.settings.get('sources').get('data_directory'), '{}.pbf'.format( self.settings.get('sources').get('admin_levels_file') ) ) LOG.info( 'Processing admin levels %s', admin_level_data_path ) msg = proc_exec(proc, self.verbose) if proc.returncode != 0: LOG.error('Admin level processing has not exited cleanly!') LOG.error(msg) sys.exit(99) def deconstructGeometry(self): conn = psycopg2.connect(**self.db_params) cur = conn.cursor() try: cur.execute("set search_path = \"$user\", 'public', 'topology';") LOG.info('Deconstructing geometry...') cur.execute('select deconstruct_geometry();') conn.commit() except psycopg2.ProgrammingError, e: LOG.error('Unhandeld error: (%s) %s', e.pgcode, e.pgerror) raise e cur.close() conn.close() def createBaseTopology(self): conn = psycopg2.connect(**self.db_params) cur = conn.cursor() try: cur.execute("set search_path = \"$user\", 'public', 'topology';") LOG.info('Initializing topology...') cur.execute('select init_base_topology();') except psycopg2.ProgrammingError, e: LOG.error('Unhandeld error: (%s) %s', e.pgcode, e.pgerror) raise e cur.execute('SELECT osm_id FROM all_geom order by osm_id asc') osm_ids = cur.fetchall() cur.execute('SELECT count(osm_id) FROM all_geom') total = cur.fetchone()[0] try: for idx, osm_id in enumerate(osm_ids): LOG.debug( 'Creating topology for %s ... (%s/%s)', osm_id[0], idx+1, total ) cur.execute( "set search_path = \"$user\", 'public', 'topology';" ) cur.execute('select create_base_topology_for_id(%s);', osm_id) conn.commit() except psycopg2.ProgrammingError, e: LOG.error('Unhandeld error: (%s) %s', e.pgcode, e.pgerror) raise e cur.close() conn.close() def simplifyAdminLevels(self, tolerance=0.001): conn = psycopg2.connect(**self.db_params) cur = conn.cursor() try: cur.execute("set search_path = \"$user\", 'public', 'topology';") LOG.info('Simplifying admin_levels ...') cur.execute('select simplify_dissolve(%s);', (tolerance,)) conn.commit() except psycopg2.ProgrammingError, e: LOG.error('Unhandeld error: (%s) %s', e.pgcode, e.pgerror) raise e cur.close() conn.close() def convertToGeoJson(self, settings_file, *args): if len(args) > 0: command = [ 'python', 'generate_geojson.py', '--rm', '--settings', settings_file ] command += [arg for arg in args] else: command = [ 'python', 'generate_geojson.py', '--rm', '--all', '--settings', settings_file ] LOG.debug('Command: %s', ' '.join(command)) proc = subprocess.Popen( command, stdout=subprocess.PIPE, stderr=subprocess.PIPE, shell=False ) # execute the process ... .wait() LOG.info('Converting to geojson ... exported_geojson.zip') msg = proc_exec(proc, self.verbose) if proc.returncode != 0: LOG.error('Converting to geojson has not exited cleanly!') LOG.error(msg) sys.exit(99)
[ "dodobas@candela-it.com" ]
dodobas@candela-it.com
20f38bed70c1832fcabd62c4095639a2b179b570
2b63c99eaa0c4866e5ea956ab0c5207b01ce6c0f
/bin/easy_install-3.2
2753b93978d3263f58bff7abb72623f52d4505d4
[]
no_license
raspicamproject/raspi-python
6f55b7c989e6f98d16a06f2f3b245f296af5d2a4
42380d47fb7974b63f21f56bb1093c246074ed76
refs/heads/master
2021-01-15T11:25:38.106055
2015-03-19T16:05:43
2015-03-19T16:05:43
null
0
0
null
null
null
null
UTF-8
Python
false
false
346
2
#!/home/pi/raspi-cam/bin/python3 # EASY-INSTALL-ENTRY-SCRIPT: 'distribute==0.6.24','console_scripts','easy_install-3.2' __requires__ = 'distribute==0.6.24' import sys from pkg_resources import load_entry_point if __name__ == '__main__': sys.exit( load_entry_point('distribute==0.6.24', 'console_scripts', 'easy_install-3.2')() )
[ "lebras.simon@gmail.com" ]
lebras.simon@gmail.com
b0dd12f359d47826a34ff45e3eec933fcad07418
b9bd74516dc0f24d336f35b14780a171030ea529
/18.随机森林算法.py
ba4d7db1198edb81bfe134670e0a5c5bf50c3801
[]
no_license
DataIsStrength/scikit-learn
814326c03743cacb62b68bb5cfda8e4e75b2fe09
2d581032404f67ad628da75a64c1cba514b77bf7
refs/heads/main
2023-08-29T07:47:28.600815
2021-10-30T14:29:26
2021-10-30T14:29:26
422,897,387
0
0
null
null
null
null
UTF-8
Python
false
false
2,643
py
# -*- coding: utf-8 -*- """ 随机森林算法 """ ''' 1.随机森林: 随机森林是包含多个决策树的分类器,其输出的类别是由个别树输出的类别的众数而定。 例如,训练了5个树,其中4个树结果是True,1个数结果False,那么最终结果就是True。 2.Bootstrap抽样: 即随机有放回抽样,是一种重抽样的方法,为了形成随机森林的多个决策树, 要采用Bootstrap抽样,具体过程如下: ①抽取样本:在N个样本中Bootstrap抽取N个,形成一个树的训练数据集。 ②选择特征:如果一共有M个特征,则选择m个来训练决策树,m<<M,这样的好处是可以降维。 ''' import pandas as pd #导入字典向量化类 from sklearn.feature_extraction import DictVectorizer #导入数据集划分函数 from sklearn.model_selection import train_test_split #导入随机森林预估器类 from sklearn.ensemble import RandomForestClassifier #导入网格搜索函数 from sklearn.model_selection import GridSearchCV #从网站下载数据 data=pd.read_csv('titanic.csv') #选择特征值 x=data[['pclass','age','sex']].copy() #选择目标值 y=data['survived'].copy() #缺失值处理,inplace设置为True表示对原始数据进行修改 #如果inplace设置为False,则修改后需要赋值给一个新的变量,而原数据不变 x['age'].fillna(x['age'].mean(),inplace=True) #特征工程 #将x转换成字典数据x.to_dict,设置orient参数可以调整格式,一般常用records x=x.to_dict(orient='records') #实例化字典向量化类 transform=DictVectorizer(sparse=False) #调用fit_transform x=transform.fit_transform(x) print(transform.get_feature_names()) print(x) #划分数据集,设置测试集占比30% x_train,x_test,y_train,y_test=train_test_split(x,y,test_size=0.3) #实例化随机森林预估器类 estimator=RandomForestClassifier() #设置备选超参数,n_estimators是决策树的数量,max_depth是单个树的最大深度 param={'n_estimators':[120,200,300,500,800,1200],'max_depth':[5,8,15,25,30]} #对模型进行2折交叉验证的网格搜索 estimator=GridSearchCV(estimator,param_grid=param,cv=2) #训练模型 estimator.fit(x_train,y_train) #验证和评估模型 print('预测的准确率为:',estimator.score(x_test,y_test)) ''' 随机森林算法总结: 1.在当前所有算法中,具有极好的准确率。 2.能够有效地运行在大数据集上,处理具有高维特征的输入样本,而且不需要降维。 3.能够评估各个特征在分类问题上的重要性。 '''
[ "noreply@github.com" ]
noreply@github.com
ed42e42af0566b7f01fc67a855eb8e7c557d20a2
588de5142e35002b618c5777751f141c075be8fd
/hourofci/hourofci_widgets.py
2b89a07e5c04fda80ff9b97562caa96a587a59f7
[]
no_license
IrisZhiweiYan/test_hourofci
5464035164f4657351c70bc81dfd52694ccd9096
a4371f88fcd091ae939655da5087b1010a7795f9
refs/heads/master
2021-07-14T01:22:27.932172
2020-11-09T05:40:27
2020-11-09T05:40:27
219,059,799
0
0
null
null
null
null
UTF-8
Python
false
false
835
py
from hourofci import * from .buttons import button # import buttons # CHANGE in v2: pass the answer catalog in the notebook to the widget function def IntSlider(question, hash_answer_catalog): # start_time = time.time() int_range = widgets.IntSlider() display(int_range) value = 10 # Iris: where to get change? def on_value_change(change): # CHANGE: append -> replace (only keep the last answers between two submissions) Answer_Dict[question]=[change["new"]] int_range.observe(on_value_change, names='value') # Button Evaluator with arguments (desired_answer, frmt) | Fmrt is the format to evaluate like single item, list, dict, etc # CHANGE in v2: pass the answer catalog to the submit button function to valid button(question, hash_answer_catalog)
[ "irisgogo.yan@gmail.com" ]
irisgogo.yan@gmail.com
9f8c5b37b12ed48e2504be0c4a85627772bf19c6
f188c8791f10b481c59b064e5632f2b8e6468a15
/pyhrv/tools.py
cbfc8d21cab829d8db98694990d6728b1c3940bf
[ "BSD-3-Clause" ]
permissive
RuWander/pyhrv
a440b0f0e7bb199718fefcbc6f56a509914f4585
61b233e907a0223e2a2bf1a5962ee2e9ae8ed36c
refs/heads/master
2021-03-27T05:56:17.351755
2020-03-17T13:28:39
2020-03-17T13:28:39
247,793,858
0
0
BSD-3-Clause
2020-03-16T18:55:30
2020-03-16T18:55:29
null
UTF-8
Python
false
false
45,737
py
#!/usr/bin/env python -W ignore::FutureWarning # -*- coding: utf-8 -*- """ pyHRV - Heart Rate Variability Toolbox - Tools ---------------------------------------------- This module provides support tools for HRV analysis such as the computation of HRV relevant data series (NNI, NNI differences Heart Rate) and Notes ----- .. This module is part of the master thesis "Development of an Open-Source Python Toolbox for Heart Rate Variability (HRV)". .. This module is a contribution to the open-source biosignal processing toolbox 'BioSppy': https://github.com/PIA-Group/BioSPPy Author ------ .. Pedro Gomes, pgomes92@gmail.com Thesis Supervisors ------------------ .. Hugo Silva, PhD, Instituto de Telecomunicacoes, PLUX wireless biosignals S.A. .. Prof. Dr. Petra Margaritoff, University of Applied Sciences Hamburg Docs ---- .. You can find the documentation for this module here: https://pyhrv.readthedocs.io/en/latest/_pages/api/tools.html Last Update ----------- 12-11-2019 :copyright: (c) 2018 by Pedro Gomes :license: BSD 3-clause, see LICENSE for more details. """ # Compatibility from __future__ import absolute_import, division # Imports import os import sys import warnings import json import numpy as np import matplotlib as mpl import matplotlib.pyplot as plt import datetime as dt from matplotlib.projections import register_projection # BioSPPy imports import biosppy # Local imports import pyhrv import pyhrv.time_domain import pyhrv.frequency_domain import pyhrv.nonlinear # Turn off toolbox triggered warnings warnings.filterwarnings('ignore', category=FutureWarning) warnings.filterwarnings('ignore', category=RuntimeWarning) def nn_intervals(rpeaks=None): """Computes the NN intervals [ms] between successive R-peaks. Docs: https://pyhrv.readthedocs.io/en/latest/_pages/api/tools.html#nn-intervals-nn-intervals Parameter --------- rpeaks : array R-peak times in [ms] or [s] Returns ------- nni : array NN intervals in [ms] Raises ------ TypeError If no data provided for 'rpeaks' TypeError If data format is not list or numpy array TypeError If 'rpeaks' array contains non-integer or non-float value Notes ----- .. You can find the documentation for this function here: https://pyhrv.readthedocs.io/en/latest/_pages/api/tools.html#nn-intervals-nn-intervals """ # Check input signal if rpeaks is None: raise TypeError("No data for R-peak locations provided. Please specify input data.") elif type(rpeaks) is not list and not np.ndarray: raise TypeError("List, tuple or numpy array expected, received %s" % type(rpeaks)) # if all(isinstance(n, int) for n in rpeaks) is False or all(isinstance(n, float) for n in rpeaks) is False: # raise TypeError("Incompatible data type in list or numpy array detected (only int or float allowed).") # Confirm numpy arrays & compute NN intervals rpeaks = np.asarray(rpeaks) nn_int = np.zeros(rpeaks.size - 1) for i in range(nn_int.size): nn_int[i] = rpeaks[i + 1] - rpeaks[i] return pyhrv.utils.nn_format(nn_int) def nni_diff(nni=None): """Computes the series of differences between successive NN intervals [ms]. Docs: https://pyhrv.readthedocs.io/en/latest/_pages/api/tools.html#nn-interval-differences-nn-diff Parameters ---------- nni : array NN intervals in [ms] or [s]. Returns ------- nni_diff_ : numpy array Difference between successive NN intervals in [ms]. Raises ------ TypeError If no data provided for 'rpeaks'. TypeError If no list or numpy array is provided. TypeError If NN interval array contains non-integer or non-float value. Notes .. You can find the documentation for this module here: https://pyhrv.readthedocs.io/en/latest/_pages/api/tools.html#nn-interval-differences-nn-diff """ # Check input signal if nni is None: raise TypeError("No data for R-peak locations provided. Please specify input data.") elif type(nni) is not list and type(nni) is not np.ndarray: raise TypeError("List or numpy array expected, received %s" % type(nni)) elif all(isinstance(x, int) for x in nni) and all(isinstance(x, float) for x in nni): raise TypeError("'nni' data contains non-int or non-float data.") else: nn = pyhrv.utils.nn_format(nni) # Confirm numpy arrays & compute NN interval differences nn_diff_ = np.zeros(nn.size - 1) for i in range(nn.size - 1): nn_diff_[i] = abs(nn[i + 1] - nn[i]) return np.asarray(nn_diff_) def plot_ecg(signal=None, t=None, sampling_rate=1000., interval=None, rpeaks=True, figsize=None, title=None, show=True): """Plots ECG lead-I like signal on a medical grade ECG paper-like figure layout. Docs: https://pyhrv.readthedocs.io/en/latest/_pages/api/tools.html#plot-ecg-plot-ecg Parameters ---------- signal : array ECG lead-I like signal (filtered or unfiltered) t : array, optional Time vector for the ECG lead-I like signal (default: None) sampling_rate : int, float, optional Sampling rate of the acquired signal in [Hz] (default: 1000Hz) interval : array, 2-element, optional Visualization interval of the ECG lead-I like signal plot (default: None: [0s, 10s] rpeaks : bool, optional If True, marks R-peaks in ECG lead-I like signal (default: True) figsize : array, optional Matplotlib figure size (width, height) (default: None: (12, 4)) title : str, optional Plot figure title (default: None). show : bool, optional If True, shows the ECG plot figure(default: True) Returns ------- fig_ecg : matplotlib figure object Matplotlib figure of ECG plot Raises ------ TypeError If no ECG data provided. Notes ---- .. The 'rpeaks' parameter will have no effect if there are more then 50 r-epaks within the visualization interval. In this case, no markers will be set to avoid overloading the plot .. You can find the documentation for this function here: https://pyhrv.readthedocs.io/en/latest/_pages/api/tools.html#plot-ecg-plot-ecg """ # Check input data if signal is None: raise TypeError("No ECG data provided. Please specify input data.") else: # Confirm numpy signal = np.asarray(signal) # Compute time vector if t is None: t = pyhrv.utils.time_vector(signal, sampling_rate=sampling_rate) # Configure interval of visualized signal if interval is 'complete': interval = [0, t[-1]] else: interval = pyhrv.utils.check_interval(interval, limits=[0, t[-1]], default=[0, 10]) # Prepare figure if figsize is None: figsize = (12, 4) fig_ecg = plt.figure(figsize=figsize) ax = fig_ecg.add_subplot(111) # Configure axis according to according to BITalino ECG sensor ranges if signal.max() > 1.5: y_min = int(signal.min() - (signal.max() - signal.min()) * 0.2) y_max = int(signal.max() + (signal.max() - signal.min()) * 0.2) unit = '-' y_minor = np.linspace(y_min, y_max, 12) y_major = np.linspace(y_min, y_max, 4) elif signal.max() < 1.0: y_min, y_max = -1., 1., unit = 'mV' y_minor = np.arange(-0.9, y_min, 0.1) y_major = np.arange(-1.0, y_max + 0.5, 0.5) else: y_min, y_max = -1.5, 1.5, unit = 'mV' y_minor = np.arange(-1.4, y_min, 0.1) y_major = np.arange(y_min, y_max + 0.5, 0.5) ax.axis([interval[0], interval[1], y_min, y_max]) ax.set_xlabel('Time [$s$]') ax.set_ylabel('ECG [$%s$]' % unit) # Set ticks as ECG paper (box height ~= 0.1mV; width ~= 0.1s when using default values) n = int(interval[1] / 10) try: ax.set_xticks(np.arange(0.0, interval[1] + 0.1, float(n)/5), minor=True) ax.xaxis.grid(which='minor', color='salmon', lw=0.3) ax.set_xticks(np.arange(0, interval[1] + 0.1, n)) ax.xaxis.grid(which='major', color='r', lw=0.7) ax.set_yticks(y_minor, minor=True) ax.yaxis.grid(which='minor', color='salmon', lw=0.3) ax.set_yticks(y_major) ax.yaxis.grid(which='major', color='r', lw=0.7) except: ax.grid(False) # Add legend unit = '' if unit == '-' else unit text_ = 'Division (x): %is\nDivision (y): %.1f%s' % (n, (np.abs(y_major[1] - y_major[0])), unit) ax.text(0.88, 0.85, text_, transform=ax.transAxes, fontsize=9, bbox=dict(boxstyle='round', facecolor='white', alpha=0.9)) # Plot ECG lead-I like signal ax.plot(t, signal, 'r') fig_ecg.tight_layout() # Plot r-peaks rps = biosppy.signals.ecg.ecg(signal=signal, sampling_rate=sampling_rate, show=False)[2] p = [float(signal[x]) for x in rps] r = t[rps] if rpeaks: ax.plot(r, p, 'g*', alpha=0.7) # Add title if title is not None: ax.set_title('ECG lead-I like signal - %s' % str(title)) else: ax.set_title('ECG lead-I like signal') # Show plot if show: plt.show() # Output args = (fig_ecg, ) names = ('ecg_plot', ) return biosppy.utils.ReturnTuple(args, names) def tachogram(nni=None, signal=None, rpeaks=None, sampling_rate=1000., hr=True, interval=None, title=None, figsize=None, show=True): """Plots Tachogram (NNI & HR) of an ECG lead-I like signal, NNI or R-peak series. Docs: https://pyhrv.readthedocs.io/en/latest/_pages/api/tools.html#tachogram-tachogram Parameters ---------- nni : array NN intervals in [ms] or [s]. rpeaks : array R-peak times in [ms] or [s]. signal : array, optional ECG lead-I like signal. sampling_rate : int, float Sampling rate of the acquired signal in [Hz]. hr : bool, optional If True, plots series of heart rate data in [bpm] (default: True). interval : list, optional Sets visualization interval of the signal (default: [0, 10]). title : str, optional Plot figure title (default: None). figsize : array, optional Matplotlib figure size (width, height) (default: (12, 4)). show : bool, optional If True, shows plot figure (default: True). Returns ------- fig : matplotlib.pyplot figure Tachogram figure & graph Raises ------ TypeError If no input data for 'nni', 'rpeaks' or 'signal' is provided Notes ----- .. NN intervals are derived from the ECG lead-I like signal if 'signal' is provided. .. If both 'nni' and 'rpeaks' are provided, 'rpeaks' will be chosen over the 'nn' and the 'nni' data will be computed from the 'rpeaks'. .. If both 'nni' and 'signal' are provided, 'nni' will be chosen over 'signal'. .. If both 'rpeaks' and 'signal' are provided, 'rpeaks' will be chosen over 'signal'. """ # Check input if signal is not None: rpeaks = biosppy.signals.ecg.ecg(signal=signal, sampling_rate=sampling_rate, show=False)[2] elif nni is None and rpeaks is None: raise TypeError('No input data provided. Please specify input data.') # Get NNI series nni = pyhrv.utils.check_input(nni, rpeaks) # Time vector back to ms t = np.cumsum(nni) / 1000. # Configure interval of visualized signal if interval is 'complete': interval = [0, t[-1]] else: interval = pyhrv.utils.check_interval(interval, limits=[0, t[-1]], default=[0, 10]) # Prepare figure if figsize is None: figsize = (12, 4) fig = plt.figure(figsize=figsize) ax = fig.add_subplot(111) # X-Axis configuration # Set x-axis format to seconds if the duration of the signal <= 60s if interval[1] <= 60: ax.set_xlabel('Time [s]') # Set x-axis format to MM:SS if the duration of the signal > 60s and <= 1h elif 60 < interval[1] <= 3600: ax.set_xlabel('Time [MM:SS]') formatter = mpl.ticker.FuncFormatter(lambda ms, x: str(dt.timedelta(seconds=ms))[2:]) ax.xaxis.set_major_formatter(formatter) # Set x-axis format to HH:MM:SS if the duration of the signal > 1h else: ax.set_xlabel('Time [HH:MM:SS]') formatter = mpl.ticker.FuncFormatter(lambda ms, x: str(dt.timedelta(seconds=ms))) ax.xaxis.set_major_formatter(formatter) try: n = int(interval[1] / 10) ax.set_xticks(np.arange(0, interval[1] + n, n)) except Exception as e: ax.grid(False) # Y-Axis configuration (min, max set to maximum of the visualization interval) ax.set_ylabel('NN Interval [$ms$]') nn_min = np.min(nni[np.argwhere(np.logical_and(interval[0] <= t, t <= interval[1]))]) nn_max = np.max(nni[np.argwhere(np.logical_and(interval[0] <= t, t <= interval[1]))]) ax.axis([interval[0], interval[1], nn_min * 0.9, nn_max * 1.1]) # Plot 'x' markers only if less than 50 rpeaks are within the given data, otherwise don't add them if np.argwhere(t < interval[1]).size < 50: l1 = ax.plot(t, nni, color='g', label='NN Intervals', marker='x', linestyle='--', linewidth=0.8) ax.vlines(t, 200, 3000, linestyles='--', linewidth=0.5, alpha=0.7, colors='lightskyblue') else: l1 = ax.plot(t, nni, color='g', label='NN Intervals', linestyle='--', linewidth=0.8) lns = [] # Plot heart rate signal if hr: ax2 = ax.twinx() bpm_values = heart_rate(nni) hr_min = heart_rate(nn_max) hr_max = heart_rate(nn_min) ax2.set_ylabel('Heart Rate [$1/min$]', rotation=270, labelpad=15) ax2.axis([interval[0], interval[1], hr_min * 0.9, hr_max * 1.1]) # Plot 'x' markers only if less than 50 rpeaks are within the given data, otherwise don't add them if np.argwhere(t < interval[1]).size < 50: l2 = ax2.plot(t, bpm_values, color='red', label='Heart Rate', marker='x', linestyle='--', linewidth=0.8) else: l2 = ax2.plot(t, bpm_values, color='red', label='Heart Rate', linestyle='--', linewidth=0.8) lns = l1 + l2 labs = [l.get_label() for l in lns] ax.legend(lns, labs, loc=1) else: ax.legend(loc=1) # Add title if title is not None: ax.set_title('Tachogram - %s' % str(title)) else: ax.set_title('Tachogram') # Show plot if show: plt.show() # Output args = (fig, ) names = ('tachogram_plot', ) return biosppy.utils.ReturnTuple(args, names) def heart_rate(nni=None, rpeaks=None): """Computes a series of Heart Rate values in [bpm] from a series of NN intervals or R-peaks in [ms] or [s] or the HR from a single NNI. Docs: https://pyhrv.readthedocs.io/en/latest/_pages/api/tools.html#heart-rate-heart-rate Parameters ---------- nni : int, float, array NN intervals in [ms] or [s]. rpeaks : int, float, array R-peak times in [ms] or [s]. Returns ------- bpm : list, numpy array, float Heart rate computation [bpm]. Float value if 1 NN interval has been provided Float array if series of NN intervals or R-peaks are provided. Raises ------ TypeError If no input data for 'rpeaks' or 'nn_intervals provided. TypeError If provided NN data is not provided in float, int, list or numpy array format. Notes ----- .. You can find the documentation for this module here: https://pyhrv.readthedocs.io/en/latest/_pages/api/tools.html#heart-rate-heart-rate """ # Check input if nni is None and rpeaks is not None: # Compute NN intervals if rpeaks array is given; only 1 interval if 2 r-peaks provided nni = nn_intervals(rpeaks) if len(rpeaks) > 2 else int(np.abs(rpeaks[1] - rpeaks[0])) elif nni is not None: # Use given NN intervals & confirm numpy if series of NN intervals is provided if type(nni) is list or type(nni) is np.ndarray: nni = pyhrv.utils.nn_format(nni) if len(nni) > 1 else nni[0] elif type(nni) is int or float: nni = int(nni) if nni > 10 else int(nni) / 1000 else: raise TypeError("No data for R-peak locations or NN intervals provided. Please specify input data.") # Compute heart rate data if type(nni) is int: return 60000. / float(nni) elif type(nni) is np.ndarray: return np.asarray([60000. / float(x) for x in nni]) else: raise TypeError("Invalid data type. Please provide data in int, float, list or numpy array format.") def heart_rate_heatplot(nni=None, rpeaks=None, signal=None, sampling_rate=1000., age=18, gender='male', interval=None, figsize=None, show=True): """Graphical visualization & classification of HR performance based on normal HR ranges by age and gender. Docs: https://pyhrv.readthedocs.io/en/latest/_pages/api/tools.html#heart-rate-heatplot-hr-heatplot Parameters ---------- nni : array NN intervals in [ms] or [s]. rpeaks : array R-peak times in [ms] or [s]. signal : array, optional ECG lead-I like signal. sampling_rate : int, float, optional Sampling rate of the acquired signal in [Hz]. age : int, float Age of the subject (default: 18). gender : str Gender of the subject ('m', 'male', 'f', 'female'; default: 'male'). interval : list, optional Sets visualization interval of the signal (default: [0, 10]). figsize : array, optional Matplotlib figure size (width, height) (default: (12, 4)). show : bool, optional If True, shows plot figure (default: True). Returns ------- hr_heatplot : biosppy.utils.ReturnTuple object Raises ------ TypeError If no input data for 'nni', 'rpeaks' or 'signal' is provided Notes ----- .. If both 'nni' and 'rpeaks' are provided, 'rpeaks' will be chosen over the 'nn' and the 'nni' data will be computed from the 'rpeaks' .. Modify the 'hr_heatplot.json' file to write own database values """ # Helper function def _get_classification(val, data): for key in data.keys(): if data[key][0] <= int(val) <= data[key][1]: return key # Check input if signal is not None: rpeaks = biosppy.signals.ecg.ecg(signal=signal, sampling_rate=sampling_rate, show=False)[2] elif nni is None and rpeaks is None: raise TypeError('No input data provided. Please specify input data.') # Get NNI series nn = pyhrv.utils.check_input(nni, rpeaks) # Compute HR data and hr_data = heart_rate(nn) t = np.cumsum(nn) / 1000 interval = pyhrv.utils.check_interval(interval, limits=[0, t[-1]], default=[0, t[-1]]) # Prepare figure if figsize is None: figsize = (12, 5) fig, (ax, ax1, ax2) = plt.subplots(3, 1, figsize=figsize, gridspec_kw={'height_ratios': [12, 1, 1]}) ax1.axis("off") fig.suptitle("Heart Rate Heat Plot (%s, %s)" % (gender, age)) # X-Axis configuration # Set x-axis format to seconds if the duration of the signal <= 60s if interval[1] <= 60: ax.set_xlabel('Time [s]') # Set x-axis format to MM:SS if the duration of the signal > 60s and <= 1h elif 60 < interval[1] <= 3600: ax.set_xlabel('Time [MM:SS]') formatter = mpl.ticker.FuncFormatter(lambda ms, x: str(dt.timedelta(seconds=ms))[2:]) ax.xaxis.set_major_formatter(formatter) # Set x-axis format to HH:MM:SS if the duration of the signal > 1h else: ax.set_xlabel('Time [HH:MM:SS]') formatter = mpl.ticker.FuncFormatter(lambda ms, x: str(dt.timedelta(seconds=ms))) ax.xaxis.set_major_formatter(formatter) # Set gender if gender not in ["male", "m", "female", "f"]: raise ValueError("Unknown gender '%s' for this database." % gender) else: if gender == 'm': gender = 'male' elif gender == 'f': gender = 'female' # Load comparison data from database database = json.load(open(os.path.join(os.path.split(__file__)[0], './files/hr_heatplot.json'))) # Get database values if age > 17: for key in database["ages"].keys(): if database["ages"][key][0] - 1 < age < database["ages"][key][1] + 1: _age = database["ages"][key][0] color_map = database["colors"] data = database[gender][str(_age)] order = database["order"] # Plot with information based on reference database: # Create classifier counter (preparation for steps after the plot) classifier_counter = {} for key in data.keys(): classifier_counter[key] = 0 # Add threshold lines based on the comparison data for threshold in data.keys(): ax.hlines(data[threshold][0], 0, t[-1], linewidth=0.4, alpha=1, color=color_map[threshold]) ax.plot(t, hr_data, 'k--', linewidth=0.5) # Add colorized HR markers old_classifier = _get_classification(hr_data[0], data) start_index = 0 end_index = 0 for hr_val in hr_data: classifier_counter[old_classifier] += 1 current_classifier = _get_classification(hr_val, data) if current_classifier != old_classifier: ax.plot(t[start_index:end_index], hr_data[start_index:end_index], 'o', markerfacecolor=color_map[old_classifier], markeredgecolor=color_map[old_classifier]) start_index = end_index old_classifier = current_classifier end_index += 1 # Compute distribution of HR values in % percentages = {} _left = 0 legend = [] for i in range(7): classifier = str(order[str(i)][0]) percentages[classifier] = float(classifier_counter[classifier]) / hr_data.size * 100 ax2.barh("", percentages[classifier], left=_left, color=color_map[classifier]) _left += percentages[classifier] legend.append(mpl.patches.Patch(label="%s\n(%.2f%s)" % (order[str(i)][1], percentages[classifier], "$\%$"), fc=color_map[classifier])) ax.legend(handles=legend, loc=8, ncol=7) elif age <= 0: raise ValueError("Age cannot be <= 0.") else: warnings.warn("No reference data for age %i available." % age) ax.plot(t, hr_data, 'k--', linewidth=0.5) ax2.plot("", 0) # Set axis limits ax.axis([interval[0], interval[1], hr_data.min() * 0.7, hr_data.max() * 1.1]) ax.set_ylabel('Heart Rate [$1/min$]') ax2.set_xlim([0, 100]) ax2.set_xlabel("Distribution of HR over the HR classifiers [$\%$]") # Show plot if show: plt.show() # Output return biosppy.utils.ReturnTuple((fig, ), ('hr_heatplot', )) def time_varying(nni=None, rpeaks=None, parameter='sdnn', window='n20', interpolation=None, show=True, mode='normal'): """Computes time varying plot of a pyHRV parameter at every NNI of the input NNI (or rpeak) series using a moving time window or a moving NNI window. Parameters ---------- nni : array NN-Intervals in [ms] or [s] rpeaks : array R-peak locations in [ms] or [s] parameter : string pyHRV parameter key for which the time varying computation is to be plotted (check the hrv_keys.json file for a full list of available keys) window : string Time varying window configuration using the following syntax: 'tX' for using a moving time window, with X being the window interval before and after the current NNI Example: t20 generates a time window of 20s before and 20s after each NNI for the computation of th pyHRV parameter OR 'nX' for using a moving NNI window, with X being the number of NNI included before and after the current NNI Example: n20 generates a window which includes 20 NNI before and 20 NNI after the current NNI interpolation : int (optional) Frequency at which the computed parameter signal is be resampled and interpolated (for example to create a parameter signal with the same sampling frequency of the original ECG signal) show : bool, optional If true, show time varying plot (default: True) mode : Returns ------- """ # Check input series nn = pyhrv.utils.check_input(nni, rpeaks) # Check if parameter is on the list of invalid parameters (computational time of these parameters are too long or # the parameters are input parameters for PSD functions invalid_parameters = ['plot', 'tinn_m', 'tinn_n', 'fft_nfft', 'fft_window', 'fft_resampling_frequency', 'fft_interpolation', 'ar_nfft', 'ar_order', 'lomb_nfft', 'lomb_ma'] # Check selected parameter if parameter is None: raise TypeError("No parameter set for 'parameter'") elif parameter in invalid_parameters: raise ValueError("Parameter '%s' is not supported by this function. Please select another one." % parameter) elif parameter not in pyhrv.utils.load_hrv_keys_json().keys(): raise ValueError("Unknown parameter '%s' (not a pyHRV parameter)." % parameter) # Check window and decode window configuration if window[0] != 't' and window[0] != 'n': raise ValueError("Invalid mode '%s'. Please select 't' for a time window or 'n' for a NNI window." % window[0]) elif int(window[1:]) <= 0: raise ValueError("'window' cannot be <= 0.") else: window_mode = window[0] window_size = int(window[1:]) # Internal helper function def _compute_parameter(array, func): try: # Try to pass the show and mode argument to to suppress PSD plots val = eval(func + '(nni=array, mode=\'dev\')[0][\'%s\']' % parameter) except TypeError as e: if 'mode' in str(e): try: # If functions has now mode feature but 'mode' argument, but a plotting feature val = eval(func + '(nni=array, plot=False)[\'%s\']' % parameter) except TypeError as a: try: val = eval(func + '(nni=array, show=False)[\'%s\']' % parameter) except TypeError as ae: if 'plot' in str(ae): # If functions has now plotting feature try regular function val = eval(func + '(nni=array)[\'%s\']' % parameter) else: val = eval(func + '(nni=array)[\'%s\']' % parameter) return val # Vars parameter_values = np.asarray([]) # Get hrv_keys & the respective function hrv_keys = pyhrv.utils.load_hrv_keys_json() parameter_func = hrv_keys[parameter][-1] parameter_label = hrv_keys[parameter][1] parameter_unit = hrv_keys[parameter][2] # Beat window computation if window_mode == 'n': for i, _ in enumerate(nni): if i == 0: continue # Incomplete initial window elif i <= (window_size - 1): vals = nn[:(i + window_size + 1)] parameter_values = np.append(parameter_values, _compute_parameter(vals, parameter_func)) # Complete Window elif i < (nni.size - window_size): vals = nn[i - window_size: i + window_size + 1] parameter_values = np.append(parameter_values, _compute_parameter(vals, parameter_func)) # Incomplete ending window else: vals = nn[i - window_size:i] parameter_values = np.append(parameter_values, _compute_parameter(vals, parameter_func)) # Time window computation elif window_mode == 't': t = np.cumsum(nn) / 1000 for i, _t in enumerate(t): if i == 0: continue # Incomplete initial window elif _t <= window_size: # t_vals = np.where((t <= _t) & (t <== (_t + window_size))) indices = np.where(t <= (_t + window_size))[0] parameter_values = np.append(parameter_values, _compute_parameter(nn[indices], parameter_func)) # Complete Window elif _t < t[-1] - window_size: indices = np.where(((_t - window_size) <= t) & (t <= (_t + window_size)))[0] parameter_values = np.append(parameter_values, _compute_parameter(nn[indices], parameter_func)) # Incomplete end window else: indices = np.where(((_t - window_size) <= t) & (t <= t[-1]))[0] parameter_values = np.append(parameter_values, _compute_parameter(nn[indices], parameter_func)) # Interpolation (optional) and time vector if interpolation is not None: t = np.cumsum(nn) f_interpol = sp.interpolate.interp1d(t, parameter_values, 'cubic') t = np.arange(t[0], t[-1], 1000. / interpolation) parameter_values = f_interpol(t) t /= 1000. else: t = np.cumsum(nn) / 1000 # Define start and end intervals if window_mode == 'n': indices = np.arange(0, len(nn)) start_interval = np.where(indices < window_size + 1)[0] valid_interval = np.where((indices >= (window_size + 1)) & (indices <= (indices[-1] - window_size)))[0] end_interval = np.where(indices > (indices[-1] - window_size))[0][:-1] elif window_mode == 't': start_interval = np.where(t < window_size)[0] valid_interval = np.where((t >= window_size) & (t <= t[-1] - window_size))[0] end_interval = np.where(t > t[-1] - window_size)[0][:-1] y_min, y_max = 0, parameter_values.max() * 1.2 # Figure fig = plt.figure(figsize=(12, 4)) ax = fig.add_subplot(111) _win_mode = "NNI Window: %i Intervals" % window_size if window_mode == 'n' else "Time Window: %is" % window_size fig.suptitle('Time Varying - %s Evolution' % parameter_label) ax.set_title('(%s)' % _win_mode, size=10) ax.set_ylabel('%s [$%s$]' % (parameter.upper(), parameter_unit)) ax.set_xlim([0, t[-1]]) ax.set_ylim([y_min, y_max]) # Plot start values (except the very first NNI) ax.plot(t[1:window_size + 1], parameter_values[1:window_size + 1], 'r--') # Plot valid values ax.plot(t[valid_interval], parameter_values[valid_interval], 'g') # Plot final values ax.plot(t[end_interval], parameter_values[end_interval], 'r--') # X-Axis configuration # Set x-axis format to seconds if the duration of the signal <= 60s if t[-1] <= 60: ax.set_xlabel('Time [s]') # Set x-axis format to MM:SS if the duration of the signal > 60s and <= 1h elif 60 < t[-1] <= 3600: ax.set_xlabel('Time [MM:SS]') formatter = mpl.ticker.FuncFormatter(lambda ms, x: str(dt.timedelta(seconds=ms))[2:]) ax.xaxis.set_major_formatter(formatter) # Set x-axis format to HH:MM:SS if the duration of the signal > 1h else: ax.set_xlabel('Time [HH:MM:SS]') formatter = mpl.ticker.FuncFormatter(lambda ms, x: str(dt.timedelta(seconds=ms))) ax.xaxis.set_major_formatter(formatter) # Window areas legends = [] ax.vlines(t[window_size], y_min, y_max, color='r') ax.fill_between([0, t[window_size]], [y_max, y_max], facecolor='r', alpha=0.3) ax.vlines(t[parameter_values.size - window_size - 1], y_min, y_max, color='r') ax.fill_between([t[parameter_values.size - window_size - 1], t[-1]], [y_max, y_max], facecolor='r', alpha=0.3) legends.append(mpl.patches.Patch(fc='g', label='Complete Window')) legends.append(mpl.patches.Patch(fc='r', label='Incomplete Window', alpha=0.3)) # Recommended minimum window size # TODO in future versions: add available recommended minimum durations to the HRV keys json file parameter_minimum = 50 if t[window_size] < parameter_minimum: ax.vlines(parameter_minimum, y_min, y_max, color='orange') ax.fill_between([t[window_size], parameter_minimum], [y_max, y_max], color='orange', alpha=0.3) legends.append(mpl.patches.Patch(fc='orange', label='Recommended Minimum Window Size (%is)' % parameter_minimum, alpha=0.3)) ax.legend(handles=legends, loc=8, framealpha=1., ncol=3) # Add overall value val = _compute_parameter(nn, parameter_func) ax.hlines(val, 0, t[-1], linestyles='--', linewidth=0.7) ax.text(1, val + 1, 'Overall') # Check mode if mode not in ['normal', 'dev', 'devplot']: warnings.warn("Unknown mode '%s'. Will proceed with 'normal' mode." % mode, stacklevel=2) mode = 'normal' if mode == 'normal': if show: plt.show() # Output args = (fig,) names = ("time_varying_%s" % parameter,) return biosppy.utils.ReturnTuple(args, names) elif mode == 'dev': return t, parameter_values, parameter elif mode == 'devplot': if mode == 'normal': if show: plt.show() # Output args = (fig, ) names = ("time_varying_%s" % parameter, ) return biosppy.utils.ReturnTuple(args, names), t, parameter_values, parameter def radar_chart(nni=None, rpeaks=None, comparison_nni=None, comparison_rpeaks=None, parameters=None, reference_label='Reference', comparison_label='Comparison', show=True, legend=True): """Plots a radar chart of HRV parameters to visualize the evolution the parameters computed from a NNI series (e.g. extracted from an ECG recording while doing sports) compared to a reference/baseline NNI series ( e.g. extracted from an ECG recording while at rest). The radarchart normalizes the values of the reference NNI series with the values extracted from the baseline NNI series being used as the 100% reference values. Example: Reference NNI series: SDNN = 100ms → 100% Comparison NNI series: SDNN = 150ms → 150% The radar chart is not limited by the number of HRV parameters to be included in the chart; it dynamically adjusts itself to the number of compared parameters. Docs: https://pyhrv.readthedocs.io/en/latest/_pages/api/tools.html#radar-chart-radar-chart Parameters ---------- nni : array Baseline or reference NNI series in [ms] or [s] (default: None) rpeaks : array Baseline or referene R-peak series in [ms] or [s] (default: None) comparison_nni : array Comparison NNI series in [ms] or [s] (default: None) comparison_rpeaks : array Comparison R-peak series in [ms] or [s] (default: None) parameters : list List of pyHRV parameters (see keys of the hrv_keys.json file for a full list of available parameters). The list must contain more than 1 pyHRV parameters (default: None) reference_label : str, optional Plot label of the reference input data (e.g. 'ECG while at rest'; default: 'Reference') comparison_label : str, optional Plot label of the comparison input data (e.g. 'ECG while running'; default: 'Comparison') show : bool, optional If True, shows plot figure (default: True). legend : bool, optional If true, add a legend with the computed results to the plot (default: True) Returns (biosppy.utils.ReturnTuple Object) ------------------------------------------ [key : format] Description. reference_results : dict Results of the computed HRV parameters of the reference NNI series Keys: parameters listed in the input parameter 'parameters' comparison results : dict Results of the computed HRV parameters of the comparison NNI series Keys: parameters listed in the input parameter 'parameters' radar_plot : matplotlib figure Figure of the generated radar plot Raises ------ TypeError If an error occurred during the computation of a parameter TypeError If no input data is provided for the baseline/reference NNI or R-peak series TypeError If no input data is provided for the comparison NNI or R-peak series TypeError If no selection of pyHRV parameters is provided ValueError If less than 2 pyHRV parameters were provided Notes ----- .. If both 'nni' and 'rpeaks' are provided, 'rpeaks' will be chosen over the 'nn' and the 'nni' data will be computed from the 'rpeaks' .. If both 'comparison_nni' and 'comparison_rpeaks' are provided, 'comparison_rpeaks' will be chosen over the the 'comparison_nni' and the nni data will be computed from the 'comparison_rpeaks' """ # Helper function & variables para_func = pyhrv.utils.load_hrv_keys_json() unknown_parameters, ref_params, comp_params = [], {}, {} def _compute_parameter(nni_series, parameter): # Get function name for the requested parameter func = para_func[parameter][-1] try: # Try to pass the show and mode argument to to suppress PSD plots index = 0 if parameter.endswith('_vlf'): parameter = parameter.replace('_vlf', '') elif parameter.endswith('_lf'): index = 1 parameter = parameter.replace('_lf', '') elif parameter.endswith('_hf'): index = 2 parameter = parameter.replace('_hf', '') val = eval(func + '(nni=nni_series, mode=\'dev\')[0][\'%s\']' % (parameter)) val = val[index] except TypeError as e: if 'mode' in str(e): try: # If functions has now mode feature but 'mode' argument, but a plotting feature val = eval(func + '(nni=nni_series, plot=False)[\'%s\']' % parameter) except TypeError as a: if 'plot' in str(a): # If functions has now plotting feature try regular function val = eval(func + '(nni=nni_series)[\'%s\']' % parameter) else: raise TypeError(e) return val # Check input data if nni is None and rpeaks is None: raise TypeError("No input data provided for baseline or reference NNI. Please specify the reference NNI series.") else: nn = pyhrv.utils.check_input(nni, rpeaks) if comparison_nni is not None and comparison_rpeaks is not None: raise TypeError("No input data provided for comparison NNI. Please specify the comarison NNI series.") else: comp_nn = pyhrv.utils.check_input(comparison_nni, comparison_rpeaks) if parameters is None: raise TypeError("No input list of parameters provided for 'parameters'. Please specify a list of the parameters" "to be computed and compared.") elif len(parameters) < 2: raise ValueError("Not enough parameters selected for a radar chart. Please specify at least 2 HRV parameters " "listed in the 'hrv_keys.json' file.") # Check for parameter that require a minimum duration to be computed & remove them if the criteria is not met if nn.sum() / 1000. <= 600 or comp_nn.sum() / 1000. <= 600: for p in ['sdann', 'sdnn_index']: if p in parameters: parameters.remove(p) warnings.warn("Input NNI series are too short for the computation of the '%s' parameter. This " "parameter has been removed from the parameter list." % p, stacklevel=2) # Register projection of custom RadarAxes class register_projection(pyhrv.utils.pyHRVRadarAxes) # Check if the provided input parameter exists in pyHRV (hrv_keys.json) & compute available parameters for p in parameters: p = p.lower() if p not in para_func.keys(): # Save unknown parameters unknown_parameters.append(p) else: # Compute available parameters ref_params[p] = _compute_parameter(nn, p) comp_params[p] = _compute_parameter(comp_nn, p) # Check if any parameters could not be computed (returned as None or Nan) and remove them # (avoids visualization artifacts) if np.isnan(ref_params[p]) or np.isnan(comp_params[p]): ref_params.pop(p) comp_params.pop(p) warnings.warn("The parameter '%s' could not be computed and has been removed from the parameter list." % p) # Raise warning pointing out unknown parameters if unknown_parameters != []: warnings.warn("Unknown parameters '%s' will not be computed." % unknown_parameters, stacklevel=2) # Prepare plot colors = ['lightskyblue', 'salmon'] if legend: fig, (ax_l, ax) = plt.subplots(1, 2, figsize=(12, 6), subplot_kw=dict(projection='radar')) else: fig, ax = plt.subplots(1, 1, figsize=(8, 8), subplot_kw={'projection': 'radar'}) theta = np.linspace(0, 2 * np.pi, len(ref_params.keys()), endpoint=False) ax.theta = theta # Prepare plot data ax.set_varlabels([para_func[s][1].replace(' ', '\n') for s in ref_params.keys()]) ref_vals = [100 for x in ref_params.keys()] com_vals = [comp_params[p] / ref_params[p] * 100 for p in ref_params.keys()] # Plot data for i, vals in enumerate([ref_vals, com_vals]): ax.plot(theta, vals, color=colors[i]) ax.fill(theta, vals, color=colors[i], alpha=0.3) title = "HRV Parameter Radar Chart\nReference NNI Series (%s) vs. Comparison NNI Series (%s)\n" % (colors[0], colors[1]) \ + r"(Chart values in $\%$, Reference NNI parameters $\hat=$100$\%$)" # Add legend to second empty plot if legend: ax_l.set_title(title, horizontalalignment='center') legend = [] # Helper function def _add_legend(label, fc="white"): return legend.append(mpl.patches.Patch(fc=fc, label="\n" + label)) # Add list of computed parameters _add_legend(reference_label, colors[0]) for p in ref_params.keys(): _add_legend("%s:" % para_func[p][1]) # Add list of comparison parameters _add_legend(comparison_label, colors[1]) for p in ref_params.keys(): u = para_func[p][2] if para_func[p][2] != "-" else "" _add_legend("%.2f%s vs. %.2f%s" % (ref_params[p], u, comp_params[p], u)) # Add relative differences _add_legend("") for i, _ in enumerate(ref_params.keys()): val = com_vals[i] - 100 _add_legend("+%.2f%s" % (val, r"$\%$") if val > 0 else "%.2f%s" % (val, r"$\%$")) ax_l.legend(handles=legend, ncol=3, frameon=False, loc=7) ax_l.axis('off') else: ax.set_title(title, horizontalalignment='center') # Show plot if show: plt.show() # Output args = (ref_params, comp_params, fig, ) names = ('reference_results', 'comparison_results', 'radar_plot', ) return biosppy.utils.ReturnTuple(args, names) def hrv_export(results=None, path=None, efile=None, comment=None, plots=False): """ Exports HRV results into a JSON file. Docs: https://pyhrv.readthedocs.io/en/latest/_pages/api/tools.html#hrv-export-hrv-export Parameters ---------- results : dict, biosppy.utils.ReturnTuple object Results of the HRV analysis path : str Absolute path of the output directory efile : str, optional Output file name comment : str, optional Optional comment plots : bool, optional If True, save figures of the results in .png format Returns ------- efile : str Absolute path of the output export file (may vary from the input data) Raises ------ TypeError No input data provided TypeError Unsupported data format provided (other than dict, or biosppy.utils.ReturnTuple object.) TypeError If no file or directory path provided Notes ----- .. If 'path' is a file handler, 'efile' will be ignored. .. Creates file with automatic name generation if only an output path is provided. .. Output file name may vary from input file name due changes made to avoid overwrting existing files (your results are important after all!). .. Existing files will not be overwritten, instead the new file will consist of the given file name with an (incremented) identifier (e.g. '_1') that will be added at the end of the provided file name. .. You can find the documentation for this function here: https://pyhrv.readthedocs.io/en/latest/_pages/api/tools.html#hrv-export-hrv-export """ # Check input (if available & biosppy.utils.ReturnTuple object) if results is None: raise TypeError("No results data provided. Please specify input data.") elif results is not type(dict()) and isinstance(results, biosppy.utils.ReturnTuple) is False: raise TypeError("Unsupported data format: %s. " "Please provide input data as Python dictionary or biosppy.utils.ReturnTuple object." % type(results)) if path is None: raise TypeError("No file name or directory provided. Please specify at least an output directory.") elif type(path) is str: if efile is None: # Generate automatic file name efile = 'hrv_export' + dt.datetime.now().strftime('_%Y-%m-%d_%H-%M-%S') + '.json' path += efile else: # Check if file name has an '.json' extension _, fformat = os.path.splitext(efile) if fformat != 'json': path = path + efile + '.json' else: path = path + efile elif type(path) is file: path_ = path.name path.close() path = path_ efile, _ = pyhrv.utils.check_fname(path, 'json', efile) # Get HRV parameters params = json.load(open(os.path.join(os.path.split(__file__)[0], './files/hrv_keys.json'), 'r')) # Save plot figures if plots: for key in results.keys(): if isinstance(results[key], plt.Figure) and key in params.keys(): results[key].savefig(os.path.splitext(efile)[0] + '_' + str(key), dpi=300) # Prepare output dictionary output = {'Name': efile, 'Comment': str(comment)} for key in results.keys(): if isinstance(results[key], biosppy.utils.ReturnTuple): output[key] = dict(results[key]) elif isinstance(results[key], tuple): output[key] = list(results[key]) elif isinstance(results[key], str): output[key] = results[key] elif isinstance(results[key], range): output[key] = list(results[key]) elif results[key] is None: output[key] = 'n/a' elif 'plot' not in str(key) and 'histogram' not in str(key): output[key] = float(results[key]) if str(results[key]) != 'nan' else 'n/a' json.encoder.FLOAT_REPR = lambda o: format(o, 'f') with open(efile, 'w+') as f: json.dump(output, f, sort_keys=True, indent=4, separators=(',', ': ')) return str(efile) def hrv_import(hrv_file=None): """Imports HRV results stored in JSON files generated with the 'hrv_export()' function. Docs: https://pyhrv.readthedocs.io/en/latest/_pages/api/tools.html#hrv-import-hrv-import Parameters ---------- hrv_file : file object, str File handler or absolute string path of the HRV JSON file Returns ------- output : biosppy.utils.ReturnTuple object All imported results. Raises ------ TypeError No input data provided. Notes ----- .. You can find the documentation for this function here: https://pyhrv.readthedocs.io/en/latest/_pages/api/tools.html#hrv-import-hrv-import """ # Check input data and load JSON file content if hrv_file is None: raise TypeError("No input data provided. Please specify input data.") elif type(hrv_file) is str: data = json.load(open(hrv_file, 'r')) elif isinstance(hrv_file, file): data = json.load(hrv_file) results = dict() for key in data.keys(): results[str(key)] = data[key] if type(data[key]) is not str else str(data[key]) # Create biosppy.utils.ReturnTuple object from imported data return biosppy.utils.ReturnTuple(results.values(), results.keys()) if __name__ == "__main__": """ Example Script - HRV Tools """ import pyhrv from biosppy.signals.ecg import ecg # Load a Sample Signal nni = pyhrv.utils.load_sample_nni() # Load OpenSignals (r)evolution ECG sample file signal = np.loadtxt('./files/SampleECG.txt')[:, -1] # Filter data & get r-peak locations [ms] signal, rpeaks = ecg(signal, show=False)[1:3] # Plot ECG for the interval of 0s and 22s plot_ecg(signal, interval=[0, 22]) # Plot Tachogram for the interval of 0s and 22s tachogram(nni, interval=[0, 22]) # Heart Rate Heatplot to highlight HR performance compared to a sports database heart_rate_heatplot(nni, gender='male', age=28) # Time Varying is designed to show the evolution of HRV parameters over time using a moving window # Define a moving window of 3 NNIs before and after the current NNI using the NNI window indicator 'n' time_varying(nni, parameter='sdnn', window='n3') # Define a moving window of 3 seconds before and after the current NNI using the time window indicator 't' time_varying(nni, parameter='sdnn', window='t3') # Radar charts are created dynamically, depending on the number of parameters used as input # For this example, let's split he test NNI series into two segments & select a list of 6 parameters ref_nni = nni[:100] comp_nni = nni[100:200] params = ['nni_mean', 'nni_max', 'sdnn', 'rmssd', 'sdsd', 'nn50', 'nn20'] radar_chart(ref_nni, comparison_nni=comp_nni, parameters=params) # Now with only 3 parameters params = ['nni_mean', 'sdnn', 'rmssd'] radar_chart(ref_nni, comparison_nni=comp_nni, parameters=params) # Export and import HRV results into and from JSON files: # First, compute hrv parameters results = pyhrv.hrv(nni, show=False) hrv_export(results, path='./files/', efile='SampleExport') hrv_import('./files/SampleExport.json')
[ "pgomes92@gmail.com" ]
pgomes92@gmail.com
215ae356baf15b509dbf0205fdc664d254fcde92
09fbe030de6322ab7d141276fc1f8019cc1604e5
/migrations/versions/1da835fbb866_season_model.py
9bdfea12529d693c6a23bdc367ef38e0eb5d0bd2
[]
no_license
cpkm/darts-site
0f2dece8c563adfa67f85031d0a99ac947e1508a
1be939ffe8ba2e58e093adac81993aa19079d1e1
refs/heads/master
2022-12-10T00:30:09.173821
2021-11-24T01:43:03
2021-11-24T01:43:03
158,226,996
0
2
null
2022-12-08T01:22:29
2018-11-19T13:24:05
Python
UTF-8
Python
false
false
1,492
py
"""season model Revision ID: 1da835fbb866 Revises: 8689d43c428c Create Date: 2018-11-27 16:58:53.743252 """ from alembic import op import sqlalchemy as sa # revision identifiers, used by Alembic. revision = '1da835fbb866' down_revision = '8689d43c428c' branch_labels = None depends_on = None def upgrade(): # ### commands auto generated by Alembic - please adjust! ### op.create_table('season', sa.Column('id', sa.Integer(), nullable=False), sa.Column('season_name', sa.String(length=64), nullable=True), sa.Column('start_date', sa.Date(), nullable=True), sa.Column('end_date', sa.Date(), nullable=True), sa.PrimaryKeyConstraint('id', name=op.f('pk_season')) ) with op.batch_alter_table('season', schema=None) as batch_op: batch_op.create_index(batch_op.f('ix_season_end_date'), ['end_date'], unique=False) batch_op.create_index(batch_op.f('ix_season_season_name'), ['season_name'], unique=True) batch_op.create_index(batch_op.f('ix_season_start_date'), ['start_date'], unique=False) # ### end Alembic commands ### def downgrade(): # ### commands auto generated by Alembic - please adjust! ### with op.batch_alter_table('season', schema=None) as batch_op: batch_op.drop_index(batch_op.f('ix_season_start_date')) batch_op.drop_index(batch_op.f('ix_season_season_name')) batch_op.drop_index(batch_op.f('ix_season_end_date')) op.drop_table('season') # ### end Alembic commands ###
[ "20861192+cpkm@users.noreply.github.com" ]
20861192+cpkm@users.noreply.github.com
9435b62b274dd42e74992cca64e569aa33c081d9
00adb3ceec4e37f8384f575d2711a27ca94327bb
/solutions/836_Rectangle_Overlap/solution_arsho.py
9cc8700ed4892cd4bd78dee1949d0945fb8ad33a
[]
no_license
arsho/leetcode
8c7d40de5a2579f06c58474540d704aaae41d81a
55d52f7b150968ce348782ca30e573ae1cf1bd53
refs/heads/master
2023-01-23T20:46:00.859736
2023-01-22T15:55:18
2023-01-22T15:55:18
123,886,220
6
10
null
null
null
null
UTF-8
Python
false
false
476
py
""" Title : 836. Rectangle Overlap Category : Math URL : https://leetcode.com/problems/rectangle-overlap/ Author : arsho Created : 23 March 2021 """ from typing import List class Solution: def isRectangleOverlap(self, rect1: List[int], rect2: List[int]) -> bool: x_distance = min(rect1[2], rect2[2]) - max(rect1[0], rect2[0]) y_distance = min(rect1[3], rect2[3]) - max(rect1[1], rect2[1]) return x_distance > 0 and y_distance > 0
[ "shahariarrabby@gmail.com" ]
shahariarrabby@gmail.com
44c995550d05f889cc581a0508223c1b95b5eb2d
f3075dd8f88e2d1d3c6ee350098d7bd42238b92c
/naver_map/naver_gigye.py
1441572bc7711d0e83c1a8ca27bc707c1bafb40c
[]
no_license
jjjjooonno/kmong
8257a208918b947569d8868605051c7c801f5fa6
6a38f5fa4ff031602c70c6ed925fa3abdb6af38d
refs/heads/master
2020-03-28T22:19:48.693145
2018-09-18T03:35:39
2018-09-18T03:35:39
149,224,447
0
0
null
null
null
null
UTF-8
Python
false
false
1,420
py
from selenium import webdriver from bs4 import BeautifulSoup from pandas import * import time import re dt = read_excel('행정구역코드(법정동코드)_전체본.xls',0) dt_dong = dt['법정동명'] dt_dong_new = [] for i in dt_dong: if i[-1] == '동': dt_dong_new.append(i) query = [] for i in dt_dong_new: query.append(i+' 제조업 > 기계, 장비제조') names = [] tels = [] addrss = [] dr = webdriver.Chrome('/Users/joono/chromedriver') dr.get('https://map.naver.com/') dr.find_element_by_xpath('//*[@id="search-input"]').send_keys(query[0]) dr.find_element_by_xpath('//*[@id="header"]/div[1]/fieldset/button').click() time.sleep(1) drt = dr.page_source soup = BeautifulSoup(drt,'html.parser') num = soup.find('span',attrs = {'class':'n'}).text[] whole = soup.find_all('dl',attrs={'class':'lsnx_det'}) for i in whole: print(str(i)) i = str(i) if 'href=\"#\">' in i: name = i.split('href="#">')[1].split('</a>')[0] names.append(name.strip()) else: names.append('없음') if 'class=\"addr\">' in i: addr = i.split('class="addr">')[1].split('<a')[0] addrss.append(addr.strip()) else: addrss.append('없음') if 'class=\"tel\">' in i: tel1 = i.split('class="tel">')[1].split('</dd>')[0] tels.append(tel1.strip()) else: tels.append('없음') print(names) print(tels) print(addrss)
[ "jjjjooonno@gmail.com" ]
jjjjooonno@gmail.com
ca7766b7e9d85b36afe08c3d90244d6b446ff5e9
99d3d6132908841edcbc9deec3f590b0e9eba5b6
/venvpip/bin/macho_standalone
14da32f6ac9d00114ceb371da83895ecd90dbe2a
[]
no_license
nataliawcislo/cutvideo
7979066b6ee146d78308832e137dbe0028ccca11
66401b15bff3a7e6a01d8eb2d7e308b8bd04d302
refs/heads/main
2023-07-28T07:05:06.709878
2021-09-14T13:37:08
2021-09-14T13:37:08
391,267,249
0
0
null
null
null
null
UTF-8
Python
false
false
266
#!/Users/natalka/PycharmProjects/cutvideo/venvpip/bin/python # -*- coding: utf-8 -*- import re import sys from macholib.macho_standalone import main if __name__ == '__main__': sys.argv[0] = re.sub(r'(-script\.pyw|\.exe)?$', '', sys.argv[0]) sys.exit(main())
[ "wcislonatalia1998@gmail.com" ]
wcislonatalia1998@gmail.com
bcfcfd42d82934ef66bd39ecc5139583c6a927df
f62ff90d7850af458d8f12386fc9ee9134dbe7c1
/Plots/Showplots/Model_3/Current_Voltage_Curves.py
2d9023dab4df536df56c4202551adad30523eb73
[]
no_license
AlexSchmid22191/EIS_R_Sim
51b431f078cb455fc38637c192436c0523449565
851b061e60811e1e58a5b2fd4e393e529c3f86ac
refs/heads/master
2023-06-27T17:40:59.177270
2021-07-22T11:50:27
2021-07-22T11:50:27
380,768,174
0
0
null
null
null
null
UTF-8
Python
false
false
2,606
py
from matplotlib.pyplot import subplots, show from matplotlib.style import use from numpy import load, log10 use('../Show.mplstyle') data = load('../../../Currents_Resistances_Model_3/Current_Data_Model_3.npy') fig_hi, ax_hi = subplots(nrows=2, figsize=(6, 8)) fig_me, ax_me = subplots(nrows=2, figsize=(6, 8)) fig_lo, ax_lo = subplots(nrows=2, figsize=(6, 8)) # High oxygen partial pressures for i in (1400, 1500, 1600, 1700, 1800): ax_hi[0].plot(data['overpotential'][1::25, i], abs(data['current'][1::25, i]), linestyle='-', label='$10^{%d}$ bar' % log10(data['pressure'][1, i])) ax_hi[1].plot(data['overpotential'][0::25, i], data['current'][0::25, i], linestyle='-', label='$10^{%d}$ bar' % log10(data['pressure'][1, i])) # Medium oxygen partial pressures for i in (1000, 1100, 1200, 1300): ax_me[0].plot(data['overpotential'][1::25, i], abs(data['current'][1::25, i]), linestyle='-', label='$10^{%d}$ bar' % log10(data['pressure'][1, i])) ax_me[1].plot(data['overpotential'][0::25, i], data['current'][0::25, i], linestyle='-', label='$10^{%d}$ bar' % log10(data['pressure'][1, i])) # Low oxygen partial pressures for i in (500, 600, 700, 800, 900): ax_lo[0].plot(data['overpotential'][1::25, i], abs(data['current'][1::25, i]), linestyle='-', label='$10^{%d}$ bar' % log10(data['pressure'][1, i])) ax_lo[1].plot(data['overpotential'][0::25, i], data['current'][0::25, i], linestyle='-', label='$10^{%d}$ bar' % log10(data['pressure'][1, i])) ax_hi[0].set_yscale('log') ax_me[0].set_yscale('log') ax_lo[0].set_yscale('log') ax_hi[1].set_yscale('symlog', linthreshy=1e-1) ax_me[1].set_yscale('symlog', linthreshy=1e-4) ax_lo[1].set_yscale('symlog', linthreshy=1e-9) # ax_hi[0].set_ylim(1e-3, 1e5) # ax_hi[1].set_ylim(-1e5, 1e0) # ax_me[0].set_ylim(1e-6, 1e0) # ax_me[1].set_ylim(-1e0, 1e0) # ax_lo[0].set_ylim(1e-10, 1e0) # ax_lo[1].set_ylim(-1e-4, 1e1) for ax in (ax_hi[0], ax_hi[1], ax_me[0], ax_me[1], ax_lo[0], ax_lo[1]): ax.set_ylabel('Absolute current density (A/m²)') ax.set_xlabel('Overpotential (V)') ax.legend() # fig_hi.tight_layout() # fig_hi.savefig('Plots/Current_Voltage_Curves_Hi.pdf') # fig_hi.savefig('Plots/Current_Voltage_Curves_Hi.png') # # fig_me.tight_layout() # fig_me.savefig('Plots/Current_Voltage_Curves_Me.pdf') # fig_me.savefig('Plots/Current_Voltage_Curves_Me.png') # # fig_lo.tight_layout() # fig_lo.savefig('Plots/Current_Voltage_Curves_Lo.pdf') # fig_lo.savefig('Plots/Current_Voltage_Curves_Lo.png') show()
[ "Alex.Schmid91@gmail.com" ]
Alex.Schmid91@gmail.com
f6eee75f1210f4f5852d81892eb4e17c12279a21
f7dc26779dd5491480b3e88fa1ab38eee99eacad
/third-party-synthetic/third-party-tester/tests/testdefinitions/ping_test.py
d2e1480b1f6af85bfd054fced4bfae7b7be55a4f
[ "BSD-3-Clause", "Apache-2.0" ]
permissive
Dynatrace/dynatrace-api
13e068a17c08f0e7609438c043066ff4dc6b9357
5d85759bbead72dc708581979987c41912393410
refs/heads/master
2023-08-11T13:46:46.051080
2023-07-24T13:33:50
2023-07-24T13:33:50
45,841,623
99
87
BSD-3-Clause
2023-07-08T22:51:18
2015-11-09T14:04:12
Python
UTF-8
Python
false
false
2,810
py
import os from tests import Test, TestStep, TestArgument from datetime import timedelta import pings class PingTest(Test): """Ping test class. It measures the time needed for a single ICMP probe to check server availability. """ TEST_NAME = 'ping' TEST_HELP_DESCRIPTION = "Test if given host responds to ping message and measures response time" HOSTNAME_FLAG = 'hostname' TEST_ARGUMENTS = [ TestArgument( flag_names=['--' + HOSTNAME_FLAG], flag_args={ 'required': True, 'nargs': 1, 'metavar': HOSTNAME_FLAG, 'help': 'Address of a host to test' } ), ] def __init__(self, args): """Create a PingTest class instance. Extends the base class __init__() method. Args: args: Command line arguments in dict form """ super().__init__() self.hostname = args[self.HOSTNAME_FLAG][0] self.dynatrace_test_name = 'ICMP ping test for {hostname}'.format(hostname=self.hostname) self.steps.append(PingTest.PingStep(self.hostname)) class PingStep(TestStep): """ICMP ping test class.""" def __init__(self, hostname): """Create PingStep class instance. Args: hostname: IP or hostname of the host to ping """ test_step_name = 'ICMP ping test for {hostname}'.format(hostname=hostname) super().__init__(test_step_name) # Check if running as root at posix systems if os.name != "nt" and os.geteuid() != 0: self.logger.error( 'Operation not permitted - Note that ICMP messages ' 'can only be sent from processes running as root.' ) exit(1) self.pinger = pings.Ping() self.hostname = hostname def __call__(self): """Execute the test step. Overrides the base class implementation. """ self.logger.info("Sending ICMP probe to {}".format(self.hostname)) self.set_started() ping_response = self.pinger.ping(self.hostname) # Check if ICMP message was successfully received if ping_response.ret_code != pings.consts.SUCCESS: self.logger.error("ICMP probing failed") # Fail test by returning without calling self.set_passed() return # Only one ICMP probe is sent, so min time is the same as max and avg times self.duration = timedelta(milliseconds=ping_response.min_rtt) self.set_passed() self.logger.info("{} responded successfully".format(self.hostname))
[ "pawel.nalezyty@dynatrace.com" ]
pawel.nalezyty@dynatrace.com
1b2d9640e14054aa031bb6464300d7d9f16da3d7
14da21d280a8508de934a68508612af36b42c43c
/mysite/text/migrations/backup/0017_auto_20200714_1701.py
435eeb64ccf5c4e3f0f3a106b2040fe4d5469db4
[]
no_license
Shoupinglianginnolux/textmining
dd8bc2e2d5c8ea576fffb8c197f6bcffa64272e1
c811589c5d675f65a6211aec1df5d408ca1bd98c
refs/heads/main
2023-08-07T12:52:05.709080
2021-10-01T07:07:20
2021-10-01T07:07:20
412,362,431
0
0
null
null
null
null
UTF-8
Python
false
false
1,534
py
# Generated by Django 3.0.4 on 2020-07-14 17:01 from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('text', '0016_auto_20200713_1111'), ] operations = [ migrations.CreateModel( name='TMPSRQ', fields=[ ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), ('SRNumber', models.CharField(max_length=12, verbose_name='SRNumber')), ('SRType', models.CharField(max_length=5, verbose_name='SRType')), ('CreatedDate', models.DateTimeField(blank=True, null=True)), ('Model', models.CharField(max_length=15, verbose_name='Model')), ('SerialNumber', models.CharField(blank=True, max_length=25, null=True)), ('ErrorCode', models.CharField(max_length=30)), ('InternalNotes', models.CharField(max_length=300)), ('PredictErrorCode', models.CharField(blank=True, max_length=10, null=True)), ('ReviseErrorCode', models.CharField(blank=True, max_length=10, null=True)), ('Train', models.BooleanField(blank=True, default=True, null=True)), ('UploadDate', models.DateField(auto_now=True, null=True)), ], ), migrations.AddField( model_name='srqs', name='ReviseErrorCode', field=models.CharField(blank=True, max_length=10, null=True), ), ]
[ "shouping.liang@innolux.com" ]
shouping.liang@innolux.com
098f5bd193fa7c2e62926d0e9b70554529503192
1d218b6af654a77aa8a3592e40f93fbc4f584319
/cirq-core/cirq/sim/act_on_state_vector_args.py
07f86b370d922667fa62ff5ea2779f0a32ba82d1
[ "Apache-2.0" ]
permissive
eendebakpt/Cirq
dcf2d15781aea201760b37140ab1505ef570ee1c
f8ffbea1b68b483f8dc716781f2c514a02aa765e
refs/heads/master
2023-08-30T10:30:41.653293
2021-09-23T19:01:32
2021-09-23T19:01:32
409,719,785
0
0
Apache-2.0
2021-09-23T19:29:17
2021-09-23T19:29:16
null
UTF-8
Python
false
false
13,551
py
# Copyright 2018 The Cirq Developers # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # https://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Objects and methods for acting efficiently on a state vector.""" from typing import Any, Tuple, TYPE_CHECKING, Union, Dict, List, Sequence, Iterable import numpy as np from cirq import linalg, protocols, sim from cirq._compat import deprecated_parameter from cirq.sim.act_on_args import ActOnArgs, strat_act_on_from_apply_decompose from cirq.linalg import transformations if TYPE_CHECKING: import cirq def _rewrite_deprecated_args(args, kwargs): if len(args) > 3: kwargs['axes'] = args[3] if len(args) > 4: kwargs['prng'] = args[4] if len(args) > 5: kwargs['log_of_measurement_results'] = args[5] if len(args) > 6: kwargs['qubits'] = args[6] return args[:3], kwargs class ActOnStateVectorArgs(ActOnArgs): """State and context for an operation acting on a state vector. There are two common ways to act on this object: 1. Directly edit the `target_tensor` property, which is storing the state vector of the quantum system as a numpy array with one axis per qudit. 2. Overwrite the `available_buffer` property with the new state vector, and then pass `available_buffer` into `swap_target_tensor_for`. """ @deprecated_parameter( deadline='v0.13', fix='No longer needed. `protocols.act_on` infers axes.', parameter_desc='axes', match=lambda args, kwargs: 'axes' in kwargs or ('prng' in kwargs and len(args) == 4) or (len(args) > 4 and isinstance(args[4], np.random.RandomState)), rewrite=_rewrite_deprecated_args, ) def __init__( self, target_tensor: np.ndarray, available_buffer: np.ndarray, prng: np.random.RandomState, log_of_measurement_results: Dict[str, Any], qubits: Sequence['cirq.Qid'] = None, axes: Iterable[int] = None, ): """Inits ActOnStateVectorArgs. Args: target_tensor: The state vector to act on, stored as a numpy array with one dimension for each qubit in the system. Operations are expected to perform inplace edits of this object. available_buffer: A workspace with the same shape and dtype as `target_tensor`. Used by operations that cannot be applied to `target_tensor` inline, in order to avoid unnecessary allocations. Passing `available_buffer` into `swap_target_tensor_for` will swap it for `target_tensor`. qubits: Determines the canonical ordering of the qubits. This is often used in specifying the initial state, i.e. the ordering of the computational basis states. prng: The pseudo random number generator to use for probabilistic effects. log_of_measurement_results: A mutable object that measurements are being recorded into. axes: The indices of axes corresponding to the qubits that the operation is supposed to act upon. """ super().__init__(prng, qubits, axes, log_of_measurement_results) self.target_tensor = target_tensor self.available_buffer = available_buffer def swap_target_tensor_for(self, new_target_tensor: np.ndarray): """Gives a new state vector for the system. Typically, the new state vector should be `args.available_buffer` where `args` is this `cirq.ActOnStateVectorArgs` instance. Args: new_target_tensor: The new system state. Must have the same shape and dtype as the old system state. """ if new_target_tensor is self.available_buffer: self.available_buffer = self.target_tensor self.target_tensor = new_target_tensor # TODO(#3388) Add documentation for Args. # pylint: disable=missing-param-doc def subspace_index( self, axes: Sequence[int], little_endian_bits_int: int = 0, *, big_endian_bits_int: int = 0 ) -> Tuple[Union[slice, int, 'ellipsis'], ...]: """An index for the subspace where the target axes equal a value. Args: little_endian_bits_int: The desired value of the qubits at the targeted `axes`, packed into an integer. The least significant bit of the integer is the desired bit for the first axis, and so forth in increasing order. Can't be specified at the same time as `big_endian_bits_int`. When operating on qudits instead of qubits, the same basic logic applies but in a different basis. For example, if the target axes have dimension [a:2, b:3, c:2] then the integer 10 decomposes into [a=0, b=2, c=1] via 7 = 1*(3*2) + 2*(2) + 0. big_endian_bits_int: The desired value of the qubits at the targeted `axes`, packed into an integer. The most significant bit of the integer is the desired bit for the first axis, and so forth in decreasing order. Can't be specified at the same time as `little_endian_bits_int`. When operating on qudits instead of qubits, the same basic logic applies but in a different basis. For example, if the target axes have dimension [a:2, b:3, c:2] then the integer 10 decomposes into [a=1, b=2, c=0] via 7 = 1*(3*2) + 2*(2) + 0. Returns: A value that can be used to index into `target_tensor` and `available_buffer`, and manipulate only the part of Hilbert space corresponding to a given bit assignment. Example: If `target_tensor` is a 4 qubit tensor and `axes` is `[1, 3]` and then this method will return the following when given `little_endian_bits=0b01`: `(slice(None), 0, slice(None), 1, Ellipsis)` Therefore the following two lines would be equivalent: args.target_tensor[args.subspace_index(0b01)] += 1 args.target_tensor[:, 0, :, 1] += 1 """ return linalg.slice_for_qubits_equal_to( axes, little_endian_qureg_value=little_endian_bits_int, big_endian_qureg_value=big_endian_bits_int, qid_shape=self.target_tensor.shape, ) # pylint: enable=missing-param-doc def _act_on_fallback_( self, action: Union['cirq.Operation', 'cirq.Gate'], qubits: Sequence['cirq.Qid'], allow_decompose: bool = True, ) -> bool: strats = [ _strat_act_on_state_vector_from_apply_unitary, _strat_act_on_state_vector_from_mixture, _strat_act_on_state_vector_from_channel, ] if allow_decompose: strats.append(strat_act_on_from_apply_decompose) # Try each strategy, stopping if one works. for strat in strats: result = strat(action, self, qubits) if result is False: break # coverage: ignore if result is True: return True assert result is NotImplemented, str(result) raise TypeError( "Can't simulate operations that don't implement " "SupportsUnitary, SupportsConsistentApplyUnitary, " "SupportsMixture or is a measurement: {!r}".format(action) ) def _perform_measurement(self, qubits: Sequence['cirq.Qid']) -> List[int]: """Delegates the call to measure the state vector.""" bits, _ = sim.measure_state_vector( self.target_tensor, self.get_axes(qubits), out=self.target_tensor, qid_shape=self.target_tensor.shape, seed=self.prng, ) return bits def _on_copy(self, target: 'ActOnStateVectorArgs'): target.target_tensor = self.target_tensor.copy() target.available_buffer = self.available_buffer.copy() def _on_kronecker_product(self, other: 'ActOnStateVectorArgs', target: 'ActOnStateVectorArgs'): target_tensor = transformations.state_vector_kronecker_product( self.target_tensor, other.target_tensor ) target.target_tensor = target_tensor target.available_buffer = np.empty_like(target_tensor) def _on_factor( self, qubits: Sequence['cirq.Qid'], extracted: 'ActOnStateVectorArgs', remainder: 'ActOnStateVectorArgs', validate=True, atol=1e-07, ): axes = self.get_axes(qubits) extracted_tensor, remainder_tensor = transformations.factor_state_vector( self.target_tensor, axes, validate=validate, atol=atol ) extracted.target_tensor = extracted_tensor extracted.available_buffer = np.empty_like(extracted_tensor) remainder.target_tensor = remainder_tensor remainder.available_buffer = np.empty_like(remainder_tensor) def _on_transpose_to_qubit_order( self, qubits: Sequence['cirq.Qid'], target: 'ActOnStateVectorArgs' ): axes = self.get_axes(qubits) new_tensor = transformations.transpose_state_vector_to_axis_order(self.target_tensor, axes) target.target_tensor = new_tensor target.available_buffer = np.empty_like(new_tensor) def sample( self, qubits: Sequence['cirq.Qid'], repetitions: int = 1, seed: 'cirq.RANDOM_STATE_OR_SEED_LIKE' = None, ) -> np.ndarray: indices = [self.qubit_map[q] for q in qubits] return sim.sample_state_vector( self.target_tensor, indices, qid_shape=tuple(q.dimension for q in self.qubits), repetitions=repetitions, seed=seed, ) def _strat_act_on_state_vector_from_apply_unitary( unitary_value: Any, args: 'cirq.ActOnStateVectorArgs', qubits: Sequence['cirq.Qid'], ) -> bool: new_target_tensor = protocols.apply_unitary( unitary_value, protocols.ApplyUnitaryArgs( target_tensor=args.target_tensor, available_buffer=args.available_buffer, axes=args.get_axes(qubits), ), allow_decompose=False, default=NotImplemented, ) if new_target_tensor is NotImplemented: return NotImplemented args.swap_target_tensor_for(new_target_tensor) return True def _strat_act_on_state_vector_from_mixture( action: Any, args: 'cirq.ActOnStateVectorArgs', qubits: Sequence['cirq.Qid'] ) -> bool: mixture = protocols.mixture(action, default=None) if mixture is None: return NotImplemented probabilities, unitaries = zip(*mixture) index = args.prng.choice(range(len(unitaries)), p=probabilities) shape = protocols.qid_shape(action) * 2 unitary = unitaries[index].astype(args.target_tensor.dtype).reshape(shape) linalg.targeted_left_multiply( unitary, args.target_tensor, args.get_axes(qubits), out=args.available_buffer ) args.swap_target_tensor_for(args.available_buffer) if protocols.is_measurement(action): key = protocols.measurement_key_name(action) args.log_of_measurement_results[key] = [index] return True def _strat_act_on_state_vector_from_channel( action: Any, args: 'cirq.ActOnStateVectorArgs', qubits: Sequence['cirq.Qid'] ) -> bool: kraus_operators = protocols.kraus(action, default=None) if kraus_operators is None: return NotImplemented def prepare_into_buffer(k: int): linalg.targeted_left_multiply( left_matrix=kraus_tensors[k], right_target=args.target_tensor, target_axes=args.get_axes(qubits), out=args.available_buffer, ) shape = protocols.qid_shape(action) kraus_tensors = [e.reshape(shape * 2).astype(args.target_tensor.dtype) for e in kraus_operators] p = args.prng.random() weight = None fallback_weight = 0 fallback_weight_index = 0 for index in range(len(kraus_tensors)): prepare_into_buffer(index) weight = np.linalg.norm(args.available_buffer) ** 2 if weight > fallback_weight: fallback_weight_index = index fallback_weight = weight p -= weight if p < 0: break assert weight is not None, "No Kraus operators" if p >= 0 or weight == 0: # Floating point error resulted in a malformed sample. # Fall back to the most likely case. prepare_into_buffer(fallback_weight_index) weight = fallback_weight index = fallback_weight_index args.available_buffer /= np.sqrt(weight) args.swap_target_tensor_for(args.available_buffer) if protocols.is_measurement(action): key = protocols.measurement_key_name(action) args.log_of_measurement_results[key] = [index] return True
[ "noreply@github.com" ]
noreply@github.com
65d5f34158a47d8bcadc80486e61ebf04640a721
2e2ecaaace5bcc70ccfad103ddb3692db82ec655
/django/background_scripts/microsoft/get_utilization.py
0c87906410b428638748f0f7ee6e6701ca43fa35
[]
no_license
RobertRajcool/Angular4-Django
32f3d24c4dbbc42abe671fd65fe98637834a388b
4424e25f6d9f982d682d13ae0e620dd0a01348a7
refs/heads/master
2021-07-10T17:53:36.222339
2017-10-10T10:17:56
2017-10-10T10:17:56
104,564,564
0
3
null
null
null
null
UTF-8
Python
false
false
15,536
py
from decimal import Decimal from . import microsoft_base from redington import settings from cloudtemplates.models import CloudRates from billing.models import CloudServiceConsumptions from customers.models import Customers, CloudAccounts from products.models import Products, VendorDetails from django.db.models import ObjectDoesNotExist, Q from django.core.exceptions import MultipleObjectsReturned from cloudapp.defaults import AppDefaults import datetime from datetime import timedelta, tzinfo from django.utils import timezone from cloudapp.generics.caculator import calculate_azure_partner_cost import pprint import requests import subprocess import os.path import json import functools import uuid import sys import pytz class UtilizationRecords(microsoft_base.MicrosoftBase): def __init__(self, tenantId, subscriptionId, startDate, endDate): super(UtilizationRecords, self).__init__() self.tenantId = tenantId self.subscriptionId = subscriptionId self.startDate = startDate self.endDate = endDate self.consolidated_rates = {} self.grouped_records = {} self.grouped_calculations = {} self.ignored_rate_names = [ 'Data Transfer In (GB)', ] self.consolidated_rate_names = [ 'Data Transfer Out (GB)' ] # Main method to get the utilizations def getUtilization(self): access_headers = self.getAccessHeaders() url = 'https://api.partnercenter.microsoft.com/v1/customers/' \ '{}/subscriptions/{}/utilizations/azure?' \ 'start_time={}&end_time={}&granularity=Daily&show_details=True'. \ format(self.tenantId, self.subscriptionId, self.startDate, self.endDate) utilization_records_out = requests.get(url, headers=access_headers) utilization_records_out.encoding = 'utf-8-sig' utilization_records = utilization_records_out.text self.process_records(utilization_records, self.grouped_records, self.grouped_calculations, self.consolidated_rates) if len(self.consolidated_rates) > 0: """ Querying vendor & customer """ vendor = VendorDetails.objects.filter(vendor_name=AppDefaults.cloud_vendor_codes(return_as='name', query_str='MS')).first() account_type = AppDefaults.cloud_vendor_codes(return_as='code', query_str=vendor.vendor_name) cloud_accounts = CloudAccounts.objects.filter(details__tenant_id=self.tenantId.upper(), type=account_type ) """ Try for lowercase """ if not cloud_accounts.exists(): cloud_accounts = CloudAccounts.objects.filter(details__tenant_id=self.tenantId.lower(), type=account_type ) customer = None if cloud_accounts.exists(): cloud_account = cloud_accounts.first() customer = cloud_account.customer customer_cloud_acc_details = cloud_account.details standard_discount = 10 if 'standard_discount' in customer_cloud_acc_details \ and customer_cloud_acc_details['standard_discount'] is not None \ and customer_cloud_acc_details['standard_discount'] != '': standard_discount = float(customer_cloud_acc_details['standard_discount']) for name, entries in self.consolidated_rates.items(): by_region = {} totals = 0 for entry in entries: name_with_location = str.format('{}|{}', name, entry[6]) region_entry = by_region.setdefault(name_with_location, []) region_entry.append(entry) totals = totals + entry[7] for item in by_region: split_values = item.split('|') if split_values: product_name = split_values[0] location = split_values[1] daily_records = by_region[item] for rec in daily_records: start_date = self.str_to_datetime(rec[0]) date_of_recording = None if start_date.month == 1: if start_date.day >= 22: date_of_recording = datetime.datetime(start_date.year, start_date.month, 22, 0, 0, 0, tzinfo=pytz.UTC) else: date_of_recording = datetime.datetime(start_date.year - 1, 12, 22, 0, 0, 0, tzinfo=pytz.UTC) else: if start_date.day >= 22: date_of_recording = datetime.datetime(start_date.year, start_date.month, 22, 0, 0, 0, tzinfo=pytz.UTC) else: date_of_recording = datetime.datetime(start_date.year, start_date.month - 1, 22, 0, 0, 0, tzinfo=pytz.UTC) # Check if there isa record on the 22nd (as we store all storage only on the 22nd consumption = CloudServiceConsumptions.objects.filter( linked_account_id=self.tenantId, subscription_id=self.subscriptionId, item_description=product_name, region=location, usage_start_date=date_of_recording ) cloud_rate = CloudRates.objects.get(uuid=rec[2]) if consumption.exists(): consumption = consumption[0] consumption.usage_quantity = consumption.usage_quantity + Decimal(rec[7]) if consumption.usage_quantity > 5: cost = calculate_azure_partner_cost( (float(consumption.usage_quantity) - 5) * float(cloud_rate.rate), standard_discount) consumption.unblended_cost = Decimal(cost) consumption.save() else: consumption = CloudServiceConsumptions() consumption.customer = customer consumption.vendor = vendor consumption.record_id = cloud_rate.uuid consumption.usage_start_date = date_of_recording end_date = date_of_recording + timedelta(days=1) consumption.usage_end_date = end_date consumption.payer_account_id = self.csp_domain consumption.linked_account_id = self.tenantId consumption.pricing_plan_id = '' consumption.product_name = rec[4] consumption.usage_type = rec[5] consumption.item_description = rec[3] consumption.usage_quantity = rec[7] consumption.region = location if location else 'N/A' consumption.rate_id = cloud_rate.id consumption.subscription_id = self.subscriptionId consumption.unblended_cost = 0 # Always 0 when we start consumption.save() # pprint.pprint(by_region) pprint.pprint(self.grouped_records) pprint.pprint(self.grouped_calculations) total = functools.reduce(lambda x, y: x + y, self.grouped_calculations.values()) pprint.pprint(total) def str_to_datetime(self, dt_string): """ Converts date string into UTC datetime object """ return datetime.datetime.strptime(dt_string, "%Y-%m-%d").replace( tzinfo=timezone.utc) if dt_string is not None else datetime.datetime.utcnow() # Recursive Block to keep returning records till we dont have any more continuation records...SPIN SPIN SPIN def process_records(self, utilization_records, grouped_records, grouped_calculations, consolidated_rates): out_file = open('/tmp/{}.json'.format(self.subscriptionId), 'w') out_file.write(utilization_records) out_file.close() if os.path.exists('/tmp/{}.json'.format(self.subscriptionId)): proc = subprocess.Popen( ["jq", "-c", '.items[] | [(.usageStartTime | sub("(?<before>.*)[-+]\\\\d{2}:\\\\d{2}"; .before ) | ' 'strptime("%Y-%m-%dT%H:%M:%S") | strftime("%Y-%m-%d")), ' '(.usageEndTime | sub("(?<before>.*)[-+]\\\\d{2}:\\\\d{2}"; .before ) | ' 'strptime("%Y-%m-%dT%H:%M:%S") | strftime("%Y-%m-%d")), ' '.resource.id, .resource.name, .resource.category, .resource.subcategory, .resource.region, .quantity]' ], stdout=subprocess.PIPE, stdin=open('/tmp/{}.json'.format(self.subscriptionId))) """ Querying vendor & customer """ vendor = VendorDetails.objects.filter(vendor_name=AppDefaults.cloud_vendor_codes(return_as='name', query_str='MS')).first() account_type = AppDefaults.cloud_vendor_codes(return_as='code', query_str=vendor.vendor_name) cloud_accounts = CloudAccounts.objects.filter(details__tenant_id=self.tenantId.upper(), type=account_type ) """ Try for lowercase """ if not cloud_accounts.exists(): cloud_accounts = CloudAccounts.objects.filter(details__tenant_id=self.tenantId.lower(), type=account_type ) customer = None if cloud_accounts.exists(): cloud_account = cloud_accounts.first() customer = cloud_account.customer customer_cloud_acc_details = cloud_account.details standard_discount = 10 if 'standard_discount' in customer_cloud_acc_details \ and customer_cloud_acc_details['standard_discount'] is not None \ and customer_cloud_acc_details['standard_discount'] != '': standard_discount = float(customer_cloud_acc_details['standard_discount']) else: sys.exit( '\033[0;37;41mSeems there is no customer for tenant id: %s. Terminating ...\033[0m' % self.tenantId) for line in proc.stdout.readlines(): line = json.loads(line.decode()) utilization_start_date = self.str_to_datetime(line[0]) utilization_end_date = self.str_to_datetime(line[1]) resource_uuid = line[2] name = line[3] category = line[4] subcategory = line[5] location = line[6] quantity = line[7] if name in self.ignored_rate_names: continue if name in self.consolidated_rate_names: consolidated_rate_name_value = consolidated_rates.setdefault(name, []) consolidated_rate_name_value.append(line) continue try: cloud_rate = CloudRates.objects.get(uuid=resource_uuid) full_name = str.format('{}|{}|{}|{}', category, subcategory, name, location) current_util = grouped_records.setdefault(full_name, 0) grouped_records[full_name] = current_util + quantity current_prices = grouped_calculations.setdefault(full_name, 0) grouped_calculations[full_name] = current_prices + (quantity * float(cloud_rate.rate)) # Store in the DB consumption = CloudServiceConsumptions() consumption.customer = customer consumption.vendor = vendor consumption.record_id = cloud_rate.uuid consumption.usage_start_date = utilization_start_date consumption.usage_end_date = utilization_end_date consumption.payer_account_id = self.csp_domain consumption.linked_account_id = self.tenantId consumption.pricing_plan_id = '' consumption.product_name = category consumption.usage_type = subcategory consumption.item_description = name consumption.usage_quantity = quantity consumption.region = location if location else 'N/A' consumption.rate_id = cloud_rate.id consumption.subscription_id = self.subscriptionId consumption.unblended_cost = calculate_azure_partner_cost(quantity * float(cloud_rate.rate), standard_discount) consumption.save() except ObjectDoesNotExist: print( "could not find for %s %s %s %s %s" % ( category, subcategory, name, location, utilization_start_date)) # Delete the file os.remove('/tmp/{}.json'.format(self.subscriptionId)) # Check if there are further entries json_output = json.loads(utilization_records) if 'next' in json_output['links']: url = 'https://api.partnercenter.microsoft.com/v1/' + json_output['links']['next']['uri'] continuation_header = {json_output['links']['next']['headers'][0]['key']: json_output['links']['next']['headers'][0]['value']} access_headers = self.getAccessHeaders() access_headers.update(continuation_header) utilization_records_out = requests.get(url, headers=access_headers) utilization_records_out.encoding = 'utf-8-sig' utilization_records = utilization_records_out.text self.process_records(utilization_records, grouped_records, grouped_calculations, consolidated_rates)
[ "doss.cclawrance226@gmail.com" ]
doss.cclawrance226@gmail.com
0316e5d65124316a03b71cca0c7a19fb8a058d61
eaf2c1331eeef64d276d7d77127191a0b7ac6908
/home/models.py
a8f4e4b4114e33c3a14700d090085fb9619d7358
[]
no_license
anubhavsrwn/Basic-Django-App
9378d9244bc34437498c7792646d0fe933e5c40a
de6ccdb378927b35c6c7b916c80290f4d573ea69
refs/heads/master
2023-06-26T21:39:20.879087
2021-07-20T08:33:40
2021-07-20T08:33:40
387,726,678
1
0
null
null
null
null
UTF-8
Python
false
false
281
py
from django.db import models # Create your models here. class Contact(models.Model): name = models.CharField(max_length=122) email = models.CharField(max_length=122) phone = models.CharField(max_length=12) desc = models.TextField() date = models.DateField()
[ "anubhav.s@technovert.net" ]
anubhav.s@technovert.net
66768fa558987437e75dd0b92ce3c1a561c04efb
b9650416b09de5ca8f4063a066e56586060f1cf0
/Tutorial29/tutorial29.py
27ebdb2a5c415b4b27e6a0505715c2550d1b34f9
[]
no_license
giova0/cursopython
e0ef2f4edacab706ed896df5b8b22fdc3e9f1b23
19790bb6f82dcae80d10df9f5d84df85ae4fea54
refs/heads/master
2023-02-21T10:44:51.597349
2021-01-26T00:35:02
2021-01-26T00:35:02
null
0
0
null
null
null
null
UTF-8
Python
false
false
85
py
#import modulo from modulo import * suma(8,9) resta(4,5) multiplicacion(6,7) input()
[ "jesusjppr@gmail.com" ]
jesusjppr@gmail.com
e469b1c0df202afbdf63411ba8abdbd4527e1190
493f99b210303d019f62195ae8dde9d02ee1b81f
/indy_node/test/api/test_rich_schema_objects_reply.py
56dbda7791bdd23e1fda9ccddcde4b9f7330077d
[ "Apache-2.0" ]
permissive
darklordz-217/indy-node
745baa357fe739bac20433cb2daa0f7c5a2f2caf
4d2f6a9dc0ff136117f8766a4f2cf70b239404e0
refs/heads/master
2022-10-06T17:01:12.414734
2020-06-11T08:49:17
2020-06-11T08:49:17
271,472,931
2
0
Apache-2.0
2020-06-11T09:04:23
2020-06-11T06:48:21
Python
UTF-8
Python
false
false
2,499
py
import json import pytest from indy_common.constants import JSON_LD_CONTEXT, RS_CONTEXT_TYPE_VALUE, RICH_SCHEMA, RICH_SCHEMA_ENCODING, \ RICH_SCHEMA_MAPPING, RICH_SCHEMA_CRED_DEF, RS_CRED_DEF_TYPE_VALUE, RS_MAPPING_TYPE_VALUE, \ RS_ENCODING_TYPE_VALUE, RS_SCHEMA_TYPE_VALUE, RICH_SCHEMA_PRES_DEF, RS_PRES_DEF_TYPE_VALUE from indy_node.test.api.helper import validate_write_reply, validate_rich_schema_txn, sdk_build_rich_schema_request from indy_node.test.rich_schema.templates import RICH_SCHEMA_EX1, W3C_BASE_CONTEXT, RICH_SCHEMA_ENCODING_EX1, \ RICH_SCHEMA_MAPPING_EX1, RICH_SCHEMA_CRED_DEF_EX1, RICH_SCHEMA_PRES_DEF_EX1 from plenum.common.util import randomString from plenum.test.helper import sdk_get_reply, sdk_sign_and_submit_req # The order of creation is essential as some rich schema object reference others by ID # Encoding's id must be equal to the one used in RICH_SCHEMA_MAPPING_EX1 @pytest.mark.parametrize('txn_type, rs_type, content, rs_id', [(JSON_LD_CONTEXT, RS_CONTEXT_TYPE_VALUE, W3C_BASE_CONTEXT, randomString()), (RICH_SCHEMA, RS_SCHEMA_TYPE_VALUE, RICH_SCHEMA_EX1, RICH_SCHEMA_EX1['@id']), (RICH_SCHEMA_ENCODING, RS_ENCODING_TYPE_VALUE, RICH_SCHEMA_ENCODING_EX1, "did:sov:1x9F8ZmxuvDqRiqqY29x6dx9oU4qwFTkPbDpWtwGbdUsrCD"), (RICH_SCHEMA_MAPPING, RS_MAPPING_TYPE_VALUE, RICH_SCHEMA_MAPPING_EX1, RICH_SCHEMA_MAPPING_EX1['@id']), (RICH_SCHEMA_CRED_DEF, RS_CRED_DEF_TYPE_VALUE, RICH_SCHEMA_CRED_DEF_EX1, randomString()), (RICH_SCHEMA_PRES_DEF, RS_PRES_DEF_TYPE_VALUE, RICH_SCHEMA_PRES_DEF_EX1, RICH_SCHEMA_PRES_DEF_EX1['@id'])]) def test_rich_schema_object_reply_is_valid(looper, sdk_pool_handle, sdk_wallet_steward, txn_type, rs_type, content, rs_id): request = sdk_build_rich_schema_request(looper, sdk_wallet_steward, txn_type=txn_type, rs_id=rs_id, rs_name=randomString(), rs_version='1.0', rs_type=rs_type, rs_content=json.dumps(content)) reply = sdk_get_reply(looper, sdk_sign_and_submit_req(sdk_pool_handle, sdk_wallet_steward, request))[1] validate_write_reply(reply) validate_rich_schema_txn(reply['result']['txn'], txn_type)
[ "alexander.sherbakov@dsr-corporation.com" ]
alexander.sherbakov@dsr-corporation.com
c3589908c3d02252488818d8c7ea24447f365be5
af0d9efc37cc79b170cafcee1a5044588167761c
/clean.py
0dc4360098b244e5377974c16aec64899d8bb40d
[]
no_license
AcmeCleanPower/HRSLToolkit
41ba3a817bc5dfb59bdb1bca32def61effbc3902
4b5144c775b6bb7292dfec132eabc839cc9c7374
refs/heads/master
2021-01-18T20:04:18.656066
2017-04-07T20:43:14
2017-04-07T20:43:14
86,934,250
0
1
null
null
null
null
UTF-8
Python
false
false
572
py
import numpy as np import scipy as sp import matplotlib.pyplot as plt from skimage import morphology from skimage.external import tifffile as tif # borrowed from tiff_tools.py def read_array_from_tiff(fin, band=1): tiff = gdal.Open(fin) return np.array(tiff.GetRasterBand(band).ReadAsArray()) def med_filter(mr, n=8): med_denoise = sp.ndimage.median_filter(mr, n) return med_denoise def gauss_filter(mr, n=8): gauss_denoise = sp.ndimage.gaussian_filter(mr, n) return gauss_denoise # def tifsave(denoised, name='denoised.tif'): # tif.imsave(name, denoised)
[ "stephen.s.c.chan@gmail.com" ]
stephen.s.c.chan@gmail.com
61dea01dd75ab8c90bcab212f1a036edbc2ab457
2bebcab2997a08695378d798ffb196d49ad1b846
/orders/migrations/0010_auto_20200826_1214.py
e324a93b00264f69ef948c91908114bec10504d1
[]
no_license
lit-lucy/Pizza-orders
739049825ac9af266d23b2d2a2af53236b1c728b
46b5c5aa2bb5e3d5f7c5901f9c06ccec12c8c0d0
refs/heads/master
2023-05-31T19:10:42.206173
2020-09-04T09:42:49
2020-09-04T09:42:49
272,914,438
0
0
null
2021-06-10T23:03:50
2020-06-17T08:03:56
Python
UTF-8
Python
false
false
432
py
# Generated by Django 3.0.7 on 2020-08-26 12:14 from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('orders', '0009_auto_20200826_1027'), ] operations = [ migrations.AlterField( model_name='order', name='delivery_type', field=models.IntegerField(choices=[(1, 'Pick up in a restaurant')], default=1), ), ]
[ "love@MacBook-Pro.local" ]
love@MacBook-Pro.local
8fcbe2dc5cae2f366d06acf5e7f587d9893d8b85
7e181f4925d24c95924920647a8d007f6a609821
/venv/bin/django-admin.py
7aded3eddca9f1de5217af11ebfb2746615b12e1
[]
no_license
Tamim101/portfolio-update
af76cd1db34d4c57658f39b032c40689ad5afe8c
bf52a72eb45c485cad578ef2a81536d8827899dc
refs/heads/master
2023-04-23T03:47:14.815150
2021-05-02T04:29:11
2021-05-02T04:29:11
363,565,323
1
0
null
null
null
null
UTF-8
Python
false
false
726
py
#!/Users/mujahid/Documents/Django/Personal_Portfolio_/My_Personal_Portfolio/venv/bin/python # When the django-admin.py deprecation ends, remove this script. import warnings from django.core import management try: from django.utils.deprecation import RemovedInDjango40Warning except ImportError: raise ImportError( 'django-admin.py was deprecated in Django 3.1 and removed in Django ' '4.0. Please manually remove this script from your virtual environment ' 'and use django-admin instead.' ) if __name__ == "__main__": warnings.warn( 'django-admin.py is deprecated in favor of django-admin.', RemovedInDjango40Warning, ) management.execute_from_command_line()
[ "tamimkhan7133@gmail.com" ]
tamimkhan7133@gmail.com
d6e158a754d97f4f5e0dedfbf9ad93d4b43e0abe
ec28e7f3290069451ec8889efa4e22b5930979c0
/strategery/engine.py
c9ee8a83a5a9f31f423dbb23634a73f29cba5401
[ "MIT" ]
permissive
rcgale/strategery
e226c0deb53a8ff35aa32a25b963807ffc98859f
d1608ea59587d7e49db0bdf788e3243d4d42081a
refs/heads/master
2021-06-23T16:06:45.218568
2021-03-06T01:42:15
2021-03-06T01:42:15
206,894,562
0
0
null
null
null
null
UTF-8
Python
false
false
3,487
py
import inspect import sys import time from functools import lru_cache from strategery.exceptions import TaskError, StrategyError from strategery.logging import BypassLogger from strategery.strategy import get_strategy from strategery.tasks import Task, get_key logger = None def execute(*args, targets, input=None, preprocessed=None): resolved_logger = logger or BypassLogger() input = __renaming_preprocessed_to_input(preprocessed, input) if type(input) is list or type(input) is tuple: # Convert lists/tuples to type-indexed dictionary input = {type(p): p for p in input} queue = get_strategy(tuple(targets), preprocessed_keys=tuple(input.keys())) print('Processing strategy:', file=resolved_logger) for n, stage in enumerate(queue): print('Phase {}: {}'.format(n, [t.name() for t in stage]), file=resolved_logger) print("\n", file=resolved_logger) # Populate with preprocessed processed = input for stage in queue: for task in stage: if task not in processed: try: ts = time.time() __assert_task_type(task) dependencies = __resolve_task_dependencies(task, processed) processed[task] = task(*dependencies) te = time.time() print('[%2.2f sec] Processed: %r ' % (te - ts, task.name()), file=resolved_logger) except Exception as e: raise TaskError('Stategery failed at task {t}, found at approximately "{f}".\n\nInner error:\n{et}: {e}'.format( t=task.name(), et=type(e).__name__, e=e, f=task.code_file_colon_line(), )) return tuple([processed[get_key(t)] for t in targets]) def __renaming_preprocessed_to_input(preprocessed, input): if preprocessed: __warn_once( 'strategery warning: the argument `preprocessed` has been renamed to `input` ' 'and will be removed in a future version.', ) if input and preprocessed: raise Exception('Cannot specify both `input` and `preprocessed') return input or preprocessed or {} @lru_cache(1) def __warn_once(message): print(message, file=sys.stderr) def __assert_task_type(task): if not inspect.isfunction(task) and not inspect.isclass(task) and not hasattr(type(task), '__call__'): raise Exception("Task cannot be processed, '{t}' is not a function or a class.".format(t=task.name)) def __resolve_task_dependencies(task: Task, processed): if len(task.parameters) != len(task.dependencies): raise StrategyError('Stategery task {t} expects parameters {p}, @fed_by decorator only accounts for {d}'.format( t=task.name(), p=[k for k in task.signature.parameters.keys()], d=[d.name() for d in task.dependencies] )) values = [] for parameter, dependency in zip(task.parameters.values(), task.dependencies): if dependency in processed: values.append(processed[dependency]) elif parameter.default != inspect._empty: values.append(parameter.default) else: raise StrategyError('Strategery task {t} could not resolve parameter {p}.'.format( t=task.name(), p=parameter.name )) return values
[ "galer@ohsu.edu" ]
galer@ohsu.edu
3d037302f28cf01673df247c203d6585ef894a16
b7a03c36c8f7d95d7cac3628f7c9d4834f33b2b9
/week6_EDA_streamlit/day1_streamlit/streamlit_class/utils/dataframes.py
8761c1b955c376c140a4d4f4ee5bc17d670bef0f
[ "MIT" ]
permissive
Jasc94/thebridge
dcca51d2d9aef7044f87795025678cb4340fb109
e194e7f5948dc52bc62fc06bd4ae58a031d3aa7a
refs/heads/main
2023-07-16T04:12:37.673222
2021-08-16T16:00:04
2021-08-16T16:00:04
361,670,388
0
1
null
null
null
null
UTF-8
Python
false
false
1,190
py
import streamlit as st import pandas as pd def get_data_from_df(df): selected_values = df.iloc[:10,:].values return str(selected_values) @st.cache(suppress_st_warning=True) def load_csv_df(uploaded_file): df = None if uploaded_file != None: #uploaded_file.seek(0) df = pd.read_csv(uploaded_file, nrows=200) # Cargame las primeras 200 filas #st.write("csv Readed¡") st.balloons() # Muestra unos globos cuando cargamos el archivo exitosamente return df @st.cache(suppress_st_warning=True) def load_normal_csv(uploaded_file): df = None if uploaded_file != None: #uploaded_file.seek(0) df = pd.read_csv(uploaded_file, nrows=200) # Cargame las primeras 200 filas #st.write("csv Readed¡") return df # Para cargar los dataframes con fin de utilizarlo como un mapa @st.cache(suppress_st_warning=True) def load_csv_for_map(csv_path): if csv_path != None: df = pd.read_csv(csv_path, sep=';') # Leelo con separadores ";" df = df.rename(columns={'latidtud': 'lat', 'longitud': 'lon'}) # Latitud -> // Longitud -> lon st.balloons() return df
[ "jonathansuarezcaceres@gmail.com" ]
jonathansuarezcaceres@gmail.com
cd71c189fbf967e745e42e3248c4421abdfecb06
8b2be934a63fee5e542bb818e81d1452b31d0ecc
/Candidate_models/final_rnn.py
d18abffcc2c1e35e0e6a9f67ccb9f5d595851a7b
[]
no_license
Danny1379/Computational_intelligence_final_project_NLP
06cde2ce40c795c57eb0f31235a9ffc98178e7fa
7c8dc7b69e2f8458959c44b8b1a1e16be300651e
refs/heads/main
2023-05-14T09:23:13.328895
2021-06-01T11:08:50
2021-06-01T11:08:50
338,410,556
1
0
null
null
null
null
UTF-8
Python
false
false
3,850
py
import numpy as np import sklearn as sk import pandas as pd from sklearn.feature_extraction.text import TfidfVectorizer from sklearn.preprocessing import OneHotEncoder from sklearn.preprocessing import LabelEncoder from sklearn.model_selection import train_test_split from sklearn.linear_model import SGDClassifier from sklearn.pipeline import Pipeline import tensorflow as tf from tensorflow.keras.preprocessing.text import Tokenizer from tensorflow.keras.preprocessing.sequence import pad_sequences from keras.utils import np_utils vocab_size = 40000 embedding_dim = 100 trunc_type = 'post' padding_type = 'post' oov_tok = "<OOV>" def load_data(): path = "train.csv" df = pd.read_csv(path) print(np.shape(df)) return df def get_labels_features(df): df.Text = df.Text.str.replace('\d', '') df.Text = df.Text.str.replace('\n', ' ') df.Text = df.Text.str.replace('.', " ") df.Text = df.Text.str.replace(',', " ") x = df['Text'] y = df['Category'] print(np.shape(x), np.shape(y)) return x, y def preprocess_encode(y): label_enc = LabelEncoder() y = label_enc.fit_transform(y) y = np_utils.to_categorical(y) return y def split(x, y): return train_test_split(x, y, test_size=0.10) def tokenize(x_train, x_test): tokenizer = Tokenizer(num_words=vocab_size, oov_token=oov_tok) tokenizer.fit_on_texts(x_train) training_sequences = tokenizer.texts_to_sequences(x_train) max_length = get_sequence_length(training_sequences) training_padded = pad_sequences(training_sequences, maxlen=max_length, padding=padding_type, truncating=trunc_type) testing_sequences = tokenizer.texts_to_sequences(x_test) testing_padded = pad_sequences(testing_sequences, maxlen=max_length, padding=padding_type, truncating=trunc_type) return training_padded, testing_padded, max_length # find average sequence length might go for maximum length ! def get_sequence_length(training_sequences): sum = 0 for i in range(len(training_sequences)): sum += len(training_sequences[i]) max_length = int(sum / len(training_sequences)) print("sequence length is : ", max_length) return max_length def get_array(training_padded, y_train, testing_padded, y_test): training_padded = np.asarray(training_padded) training_labels = np.asarray(y_train) testing_padded = np.asarray(testing_padded) testing_labels = np.asarray(y_test) return training_padded, training_labels, testing_padded, testing_labels def get_model(max_length): model = tf.keras.Sequential([ tf.keras.layers.Embedding(vocab_size, embedding_dim, input_length=max_length), tf.keras.layers.SpatialDropout1D(0.5), tf.keras.layers.Bidirectional(tf.keras.layers.LSTM(max_length)), tf.keras.layers.Bidirectional(tf.keras.layers.LSTM(100)), tf.keras.layers.Dense(34, activation="softmax") ]) model.compile(loss='categorical_crossentropy', optimizer='adam', metrics=['accuracy']) return model def train_model(model, training_padded, training_labels, testing_padded, testing_labels): num_epochs = 10 history = model.fit(training_padded, training_labels, epochs=num_epochs, validation_data=(testing_padded, testing_labels), verbose=1, batch_size=256) def main(): df = load_data() x, y = get_labels_features(df) y = preprocess_encode(y) x_train, x_test, y_train, y_test = split(x, y) x_train, x_test, max_length = tokenize(x_train, x_test) x_train, y_train, x_test, y_test = get_array(x_train, y_train, x_test, y_test) model = get_model(max_length) train_model(model, x_train, y_train, x_test, y_test) if __name__ == '__main__': main()
[ "noreply@github.com" ]
noreply@github.com
6c7fbde29cd1fec7bc805412befb2db644f4048d
a5bb696d63052d6a07d0ca1c48f415e5a5308c8d
/Lesson11/Lesson11_hw02.py
e1d3197e9faaa59f3ec12f3c872a4b8b592b522d
[]
no_license
VitaliiRomaniukKS/python_course
258af6f1a925c5cbc9207ddf0958f30652e84ff8
a530d34ad18c6fcb8e4b573376a21fe34f653f77
refs/heads/master
2020-09-03T17:27:27.892224
2020-03-15T14:44:57
2020-03-15T14:44:57
219,520,871
0
0
null
null
null
null
UTF-8
Python
false
false
1,317
py
# Распарсить файл с информацией о платежах, но использовать только те, # где тип платежа out, также не все строки могут быть в корректном формате. # Кто совершал больше всего покупок? На наибольшую сумму? Файл: out_trans_list = [] with open('payments.txt', 'r') as payments_f: for line in payments_f: new_trans = line.split(';') if (len(new_trans) == 5) and (new_trans[-2] == 'out'): # print(new_trans) new_trans.remove(new_trans[4]) out_trans_list.append(new_trans) print(out_trans_list) payments_d = {} for i in out_trans_list: summa = float (i[1].split()[0].replace(',','.')) if i[0] not in payments_d: payments_d[i[0]] = [summa] else: payments_d[i[0]].append(summa) print() print(payments_d) max_pay = [0,0] max_sum = [0,0] max_price = [0,0] for name, p_count in payments_d.items(): if len(p_count) > max_pay[1]: max_pay = [name,len(p_count)] if sum(p_count) > max_sum[1]: max_sum = [name, sum(p_count)] if max(p_count) > max_price[1]: max_price = [name,max(p_count)] print(max_pay) print(max_sum) print(max_price)
[ "noreply@github.com" ]
noreply@github.com
2ad49bb329c46561a59ca0a5e8fdb28c2b53c900
3f5d387b70ba0c828d9ebe30e6100d8837466b64
/FWUploadThread.py
c431d49b7e06187157e6520f7824eba4c7f76789
[ "MIT" ]
permissive
CsVance82/WIZnet-S2E-Tool-GUI
502eb04841549ff2ad3eeeabe5f3dccb4c6aa9d8
6cadde2c3b37bd3eb403e56e61675ee44e884c5b
refs/heads/master
2020-03-15T02:53:17.775896
2018-04-29T23:38:07
2018-04-29T23:38:07
null
0
0
null
null
null
null
UTF-8
Python
false
false
13,386
py
#!/usr/bin/python import re import sys import io import time import logging import threading import getopt import os import subprocess logging.basicConfig(level=logging.DEBUG) logger = logging.getLogger() import binascii from WIZMSGHandler import WIZMSGHandler from WIZUDPSock import WIZUDPSock from wizsocket.TCPClient import TCPClient from PyQt5.QtCore import QThread, pyqtSignal, pyqtSlot OP_SEARCHALL = 1 OP_SETIP = 2 OP_CHECKIP = 3 OP_FACTORYRESET = 4 OP_GETDETAIL = 5 OP_FWUP = 6 SOCK_CLOSE_STATE = 1 SOCK_OPENTRY_STATE = 2 SOCK_OPEN_STATE = 3 SOCK_CONNECTTRY_STATE = 4 SOCK_CONNECT_STATE = 5 idle_state = 1 datasent_state = 2 class FWUploadThread(QThread): uploading_size = pyqtSignal(int) upload_result = pyqtSignal(int) error_flag = pyqtSignal(int) def __init__(self, conf_sock, dest_mac, idcode, binaryfile, ipaddr, port): QThread.__init__(self) self.dest_mac = None self.bin_filename = None self.fd = None self.data = None self.client = None self.timer1 = None self.istimeout = 0 self.serverip = None self.serverport = None self.sentbyte = 0 self.dest_mac = dest_mac self.bin_filename = binaryfile self.idcode = idcode self.error_noresponse = 0 self.retrycheck = 0 self.conf_sock = conf_sock self.what_sock = '%s' % self.conf_sock # socket config (for TCP unicast) self.ip_addr = ipaddr self.port = port self.cli_sock = None def setparam(self): self.fd = open(self.bin_filename, "rb") self.data = self.fd.read(-1) self.remainbytes = len(self.data) self.curr_ptr = 0 self.fd.close() sys.stdout.write("Firmware file size: %r\n\n" % len(self.data)) def myTimer(self): # sys.stdout.write('timer1 timeout\r\n') self.istimeout = 1 def jumpToApp(self): cmd_list = [] # boot mode change: App boot mode cmd_list.append(["MA", self.dest_mac]) cmd_list.append(["PW", self.idcode]) cmd_list.append(["AB", ""]) if 'TCP' in self.what_sock: self.wizmsghangler = WIZMSGHandler(self.conf_sock, cmd_list, 'tcp', OP_FWUP, 2) elif 'UDP' in self.what_sock: self.wizmsghangler = WIZMSGHandler(self.conf_sock, cmd_list, 'udp', OP_FWUP, 2) self.resp = self.wizmsghangler.run() self.uploading_size.emit(1) self.msleep(1000) def sendCmd(self, command): cmd_list = [] self.resp = None # Send FW UPload request message cmd_list.append(["MA", self.dest_mac]) cmd_list.append(["PW", self.idcode]) cmd_list.append([command, str(len(self.data))]) if 'TCP' in self.what_sock: self.wizmsghangler = WIZMSGHandler(self.conf_sock, cmd_list, 'tcp', OP_FWUP, 2) elif 'UDP' in self.what_sock: self.wizmsghangler = WIZMSGHandler(self.conf_sock, cmd_list, 'udp', OP_FWUP, 2) # sys.stdout.write("cmd_list: %s\r\n" % cmd_list) # if no reponse from device, retry for several times. for i in range(4): # self.resp = self.wizmsghangler.parseresponse() self.resp = self.wizmsghangler.run() if self.resp is not '': break self.msleep(500) self.uploading_size.emit(2) def run(self): self.setparam() self.jumpToApp() if 'UDP' in self.what_sock: pass elif 'TCP' in self.what_sock: self.sock_close() self.SocketConfig() self.sendCmd('FW') if self.resp is not '' and self.resp is not None: resp = self.resp.decode('utf-8') # print('resp', resp) params = resp.split(':') sys.stdout.write('Dest IP: %s, Dest Port num: %r\r\n' % (params[0], int(params[1]))) self.serverip = params[0] self.serverport = int(params[1]) self.uploading_size.emit(3) else: print('No response from device. Check the network or device status.') self.error_flag.emit(-1) self.error_noresponse = -1 try: self.client = TCPClient(2, params[0], int(params[1])) except: pass try: if self.error_noresponse < 0: pass else: # sys.stdout.write("%r\r\n" % self.client.state) while True: if self.retrycheck > 6: break self.retrycheck += 1 if self.client.state is SOCK_CLOSE_STATE: if self.timer1 is not None: self.timer1.cancel() cur_state = self.client.state try: self.client.open() # sys.stdout.write('1 : %r\r\n' % self.client.getsockstate()) # sys.stdout.write("%r\r\n" % self.client.state) if self.client.state is SOCK_OPEN_STATE: sys.stdout.write('[%r] is OPEN\r\n' % (self.serverip)) # sys.stdout.write('[%r] client.working_state is %r\r\n' % (self.serverip, self.client.working_state)) self.msleep(500) except Exception as e: sys.stdout.write('%r\r\n' % e) elif self.client.state is SOCK_OPEN_STATE: self.uploading_size.emit(4) cur_state = self.client.state try: self.client.connect() # sys.stdout.write('2 : %r' % self.client.getsockstate()) if self.client.state is SOCK_CONNECT_STATE: sys.stdout.write('[%r] is CONNECTED\r\n' % (self.serverip)) # sys.stdout.write('[%r] client.working_state is %r\r\n' % (self.serverip, self.client.working_state)) except Exception as e: sys.stdout.write('%r\r\n' % e) elif self.client.state is SOCK_CONNECT_STATE: # if self.client.working_state == idle_state: # sys.stdout.write('3 : %r' % self.client.getsockstate()) try: self.uploading_size.emit(5) while self.remainbytes is not 0: if self.client.working_state == idle_state: if self.remainbytes >= 1024: msg = bytearray(1024) msg[:] = self.data[self.curr_ptr:self.curr_ptr+1024] self.client.write(msg) self.sentbyte = 1024 # sys.stdout.write('1024 bytes sent from at %r\r\n' % (self.curr_ptr)) sys.stdout.write('[%s] 1024 bytes sent from at %r\r\n' % (self.serverip, self.curr_ptr)) self.curr_ptr += 1024 self.remainbytes -= 1024 else : self.uploading_size.emit(6) msg = bytearray(self.remainbytes) msg[:] = self.data[self.curr_ptr:self.curr_ptr+self.remainbytes] self.client.write(msg) # sys.stdout.write('Last %r byte sent from at %r \r\n' % (self.remainbytes, self.curr_ptr)) sys.stdout.write('[%s] Last %r byte sent from at %r \r\n' % (self.serverip, self.remainbytes, self.curr_ptr)) self.curr_ptr += self.remainbytes self.remainbytes = 0 self.sentbyte = self.remainbytes self.client.working_state = datasent_state self.timer1 = threading.Timer(2.0, self.myTimer) self.timer1.start() elif self.client.working_state == datasent_state: # sys.stdout.write('4 : %r' % self.client.getsockstate()) response = self.client.readbytes(2) if response is not None: if int(binascii.hexlify(response), 16): self.client.working_state = idle_state self.timer1.cancel() self.istimeout = 0 else: print('ERROR: No response from device. Stop FW upload...') self.client.close() self.upload_result.emit(-1) self.terminate() if self.istimeout is 1: self.istimeout = 0 self.client.working_state = idle_state self.client.close() self.upload_result.emit(-1) self.terminate() self.uploading_size.emit(7) except Exception as e: sys.stdout.write('%r\r\n' % e) response = "" break print('retrycheck: %d' % self.retrycheck) if self.retrycheck > 6 or self.error_noresponse < 0: sys.stdout.write('Device [%s] firmware upload fail.\r\n' % (self.dest_mac)) self.upload_result.emit(-1) elif self.error_noresponse >= 0: self.uploading_size.emit(8) sys.stdout.write('Device [%s] firmware upload success!\r\n' % (self.dest_mac)) self.upload_result.emit(1) # send FIN packet self.msleep(500) self.client.shutdown() if 'TCP' in self.what_sock: self.conf_sock.shutdown() except Exception as e: self.error_flag.emit(-3) sys.stdout.write('%r\r\n' % e) finally: pass def sock_close(self): # 기존 연결 fin if self.cli_sock is not None: if self.cli_sock.state is not SOCK_CLOSE_STATE: self.cli_sock.shutdown() if self.conf_sock is not None: self.conf_sock.shutdown() def tcpConnection(self, serverip, port): retrynum = 0 self.cli_sock = TCPClient(2, serverip, port) print('sock state: %r' % (self.cli_sock.state)) while True: if retrynum > 6: break retrynum += 1 if self.cli_sock.state is SOCK_CLOSE_STATE: self.cli_sock.shutdown() cur_state = self.cli_sock.state try: self.cli_sock.open() if self.cli_sock.state is SOCK_OPEN_STATE: print('[%r] is OPEN' % (serverip)) time.sleep(0.5) except Exception as e: sys.stdout.write('%r\r\n' % e) elif self.cli_sock.state is SOCK_OPEN_STATE: cur_state = self.cli_sock.state try: self.cli_sock.connect() if self.cli_sock.state is SOCK_CONNECT_STATE: print('[%r] is CONNECTED' % (serverip)) except Exception as e: sys.stdout.write('%r\r\n' % e) elif self.cli_sock.state is SOCK_CONNECT_STATE: break if retrynum > 6: sys.stdout.write('Device [%s] TCP connection failed.\r\n' % (serverip)) return None else: sys.stdout.write('Device [%s] TCP connected\r\n' % (serverip)) return self.cli_sock def SocketConfig(self): # Broadcast if 'UDP' in self.what_sock: self.conf_sock = WIZUDPSock(5000, 50001) self.conf_sock.open() # TCP unicast elif 'TCP' in self.what_sock: print('upload_unicast: ip: %r, port: %r' % (self.ip_addr, self.port)) self.conf_sock = self.tcpConnection(self.ip_addr, self.port) if self.conf_sock is None: # self.isConnected = False print('TCP connection failed!: %s' % self.conf_sock) self.error_flag.emit(-3) self.terminate() else: self.isConnected = True
[ "kyi8907@gmail.com" ]
kyi8907@gmail.com
2d0368b80dbcb9c47412bfcb2b774d5f053f1fb7
aea6cb6ab25a288d9a7143d54523d63aee0d67d9
/alien_invasion/alien.py
0f3173b69a5052510f6efacd2f79a6ddfb1e8ba0
[]
no_license
ivanozz/python-crash-course
13f8fbb00effa37dd513ded6d9d12539af516293
dd889759275c24bddb2ba55c3c4a86af917781a0
refs/heads/master
2021-05-11T01:02:04.620160
2018-02-12T17:33:24
2018-02-12T17:33:24
118,316,519
0
0
null
null
null
null
UTF-8
Python
false
false
1,666
py
import pygame from pygame.sprite import Sprite class Alien(Sprite): """Класс представляющий одного пришельца""" def __init__(self, ai_settings, screen): """Инициализирует пришельца и задает его начальную позицию""" super(Alien, self).__init__() self.screen = screen self.ai_settings = ai_settings # загрузка изображения пришельца и назначение атрибута rect self.image = pygame.image.load('images/alien.bmp') self.rect = self.image.get_rect() # каждый новый пришелец появляется в левом верхнем углу экрана self.rect.x = self.rect.width self.rect.y = self.rect.height # сохранение точной позиции пришельца self.x = float(self.rect.x) def blitme(self): """Выводит пришельца в текущем положении""" self.screen.blit(self.image, self.rect) def update(self): """Перемещение пришельца вправо""" self.x += (self.ai_settings.alien_speed_factor * self.ai_settings.fleet_direction) self.rect.x = self.x def check_edges(self): """Возвращает True, если пришелец находится у края экрана""" screen_rect = self.screen.get_rect() if self.rect.right >= screen_rect.right: return True elif self.rect.left <= 0: return True
[ "iva1989@icloud.com" ]
iva1989@icloud.com
31740dec5203fccc5a4171d951f24d5a9e15aa2a
59a4a6b9a18778566307af6a281d86c3f79531e0
/Trening/TreningDjango/asgi.py
644eee76517ec9b2d15360ca1ad815b2860e5354
[]
no_license
Kasuczi/WebDev-Back-End
d8448ee2ed052518d95d54953f2672e19d997ea7
f3759fbd11d4d572fb6cf960d59dbd6e84692f8f
refs/heads/master
2021-05-26T03:49:11.889975
2020-04-08T09:19:11
2020-04-08T09:19:11
254,039,094
0
0
null
null
null
null
UTF-8
Python
false
false
403
py
""" ASGI config for TreningDjango project. It exposes the ASGI callable as a module-level variable named ``application``. For more information on this file, see https://docs.djangoproject.com/en/3.0/howto/deployment/asgi/ """ import os from django.core.asgi import get_asgi_application os.environ.setdefault('DJANGO_SETTINGS_MODULE', 'TreningDjango.settings') application = get_asgi_application()
[ "janikowski.mateusz96@gmail.com" ]
janikowski.mateusz96@gmail.com
0342556411170d9d8108b8c277f1ca7f02dc2409
45e8df26d895fce1ffced77e2fc7c87aa5fcec71
/leetcode/python-sol/301.Remove_Invalid_Parentheses.py
6bcb271d66e2d15f39cc41fe7eb7f82c787062cb
[]
no_license
mikehung/competitive-programming
5c4b429942357bfbe3c8ff2820d5cb5b7c5dc828
50713dc5973f2ea42220ac0248c3d1a6d90fcc15
refs/heads/master
2021-06-20T08:21:36.837057
2021-01-01T01:29:56
2021-01-01T01:31:55
153,595,870
1
0
null
null
null
null
UTF-8
Python
false
false
1,935
py
class Solution: def removeInvalidParentheses(self, s): def valid(s): left = 0 for c in s: if c == '(': left += 1 elif c == ')': left -= 1 if left < 0: return False return left == 0 def helper(s, num_remove): if s in memo: return memo[s] ret = [] if valid(s): if num_remove == self.max_num_remove: ret.append(s) else: ret = [s] self.max_num_remove = num_remove elif num_remove < self.max_num_remove: for i in range(len(s)): if s[i] in '()': ret += helper(s[:i] + s[i+1:], num_remove+1) memo[s] = ret return ret def trim(s): l = [] found_left = False for c in s: if c == '(': found_left = True if found_left or c != ')': l.append(c) r = [] found_right = False for c in reversed(l): if c == ')': found_right = True if found_right or c != '(': r.append(c) return ''.join(reversed(r)) self.max_num_remove = float('inf') memo = {} s = trim(s) return list(filter(lambda _: len(_) == len(s)-self.max_num_remove, set(helper(s, 0)))) or [''] import time def test(s): print(s) beg = time.time() r = Solution().removeInvalidParentheses(s) print(r, time.time()-beg) test('()())()') test('(a)())()') test(')(') test('()') test('n') test('(a(())()') test("()(((((((()") test("(()()()))((") test("))aaa))s)(()()()))(a((c((") test("((()))((()(()")
[ "mikehung@synology.com" ]
mikehung@synology.com
c058ffd30adadb95fe7dfaf10ca143450a96c2c5
445720e21dce60d8504daeb68a97525343a95639
/PyCybos/pycybos/cybos_EurexJpBid.py
2730aced35e282a67acf7d661199f2898493b0d1
[]
no_license
QuantTraderEd/AQTrader
60826cd71b0fa568852f23be9daeb7f65a13e845
c65ecba53beebce500a2e9cde0bd54374851e980
refs/heads/live_patch
2021-06-01T16:40:30.350977
2020-05-31T07:06:56
2020-05-31T07:06:56
145,120,057
2
0
null
2020-02-06T13:00:22
2018-08-17T12:51:13
Python
UTF-8
Python
false
false
548
py
# -*- coding: utf-8 -*- """ Created on Sat May 31 14:14:25 2014 @author: assa """ from cybos_source import Source class EurexJpBid(Source): """ subscribe index option quote """ def __init__(self, code = None): super(EurexJpBid, self).__init__('CpSysDib.EurexJpbid.1') self.type = 'TAQ' self.data = None if code: self.SetInputValue('0',code) pass def OnSignal(self): self.data = [] for i in xrange(14): self.data.append(self.com.GetHeaderValue(i)) self.Notify() pass
[ "hyojkim79@gmail.com" ]
hyojkim79@gmail.com
112ec58a217bf533202e32fa348ec43b73761ded
f396b2690b62c74dfa6a58a619c9f64828d4cf84
/TreningDjango/news/migrations/0002_alter_post_slug.py
00f42b3df82c567631c5fcfaec9191d44eb476b7
[]
no_license
bart-son/home-budget
7daf7bd56f9bf819e4777a1d1f1588b433df545c
ef9721a099dc55750c42b1437762bb3b206f53c1
refs/heads/master
2023-05-04T23:35:15.796553
2021-05-29T07:42:52
2021-05-29T07:42:52
371,912,274
0
0
null
null
null
null
UTF-8
Python
false
false
378
py
# Generated by Django 3.2 on 2021-04-11 12:30 from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('news', '0001_initial'), ] operations = [ migrations.AlterField( model_name='post', name='slug', field=models.SlugField(unique_for_date='created'), ), ]
[ "bartosz.kwiatk@gmail.com" ]
bartosz.kwiatk@gmail.com
36e68501df2d8e58be3d3faf5755ed0cc380b628
854660b48f63711489eabd96a6d3cbf69f4c1144
/CodingBat/WarmUp1/diff21/diff21.py
1ccef4bbbe8437264f4935a8c30561a58a224abf
[]
no_license
SametSahin10/CodingBat-Exercises
c9727e7d38defeb927d3684263d0d7655b8d8afa
e7371a8b8c71706872c8ba7a0d140d19e7ce20dc
refs/heads/master
2021-05-10T09:12:21.039238
2018-03-05T20:05:54
2018-03-05T20:05:54
118,917,328
2
0
null
2018-03-05T20:05:55
2018-01-25T13:38:52
Java
UTF-8
Python
false
false
92
py
def diff21(n): if(n > 21): return abs(n-21) * 2 else: return abs(n-21)
[ "enesdemirag1@hotmail.com" ]
enesdemirag1@hotmail.com
8652d45d2d668f0a709c1b4574844c3bdb0bca45
e7e4943e67db51791de9f0dbd302a1e6bf6e7446
/Prism_RayTracing.py
6d3a97582d0740e14cc88a6a871a1655fb4c8a2d
[]
no_license
benchizhao/RayTracing-week1-4
7b8949cebd77db81323bbbb686a3a7c11f1eb392
2aa5bc00b5a497018d3e0e8fb37a967375c5e0d4
refs/heads/master
2022-11-08T00:42:07.001028
2020-06-29T03:59:00
2020-06-29T03:59:00
null
0
0
null
null
null
null
UTF-8
Python
false
false
5,663
py
# -*- coding: utf-8 -*- """ Created on Thu Jun 11 09:23:07 2020 This file needs three packages:numpy, matplotlib.pyplt, math. This .py file simulate the behavior of the prism. The geometry of the prism is equilateral triangle. To trace the ray, we use the state of ray to describe. To trace the ray while refracted by the prism, the snell's law is used. In the end we also plot the ray. This file did not take the full reflection into consideration, if the incident angle is smaller than -2 degree, the outcoming ray will disappear. @author: Benchi Zhao """ import numpy as np import matplotlib.pyplot as plt import math class PrismTracing: def __init__(self,x,z,theta): ''' __init__ (self,x,z,theta) Gives the initial state of the ray. Parameters ------------ self.x: folat Initial x-position of the ray. self.z: float Initial z-position of the ray. self.theta: float The angle between the horizontal and the ray path (in degree). To aviod the bug, make sure the input value is greater than -2. self.n_air: float The refractive index of air. self.n_glass: float The refractive index of glass. self.state: list When ray interacting with the optical equipments, the ray state will change, all states are recorded in self.state. self.central_point: float Position of central point of the prism. self.side_length: float Length of each side of the prism. ''' self.x = x self.z = z self.theta = theta self.n_air = 1.0 self.n_glass = 1.5 self.state = [] self.central_point = 0 self.side_length = 0 def ray(self): ''' ray(self) Append the initial ray state into the total ray state. ''' ray_state = np.array([self.x,self.z,self.theta]) self.state.append(ray_state) def prism(self,side_length,central_point): ''' prism(self,side_length,central_point) Simulate the behavior of prism. Append the ray state into self.state after passing the prism. Parameters ------------ side_length: float Length of each side of the prism. self.central_point: float Position of central point of the prism. ''' self.central_point = central_point self.side_length = side_length # The ray incident into the prism incident_slope_1 = np.tan(np.deg2rad(self.state[-1][2])) L = np.array([[-incident_slope_1,1],[-math.sqrt(3),1]]) R = np.array([self.state[-1][1]-incident_slope_1*self.state[-1][0],-math.sqrt(3)*self.central_point+self.side_length/math.sqrt(3)]) result = np.linalg.solve(L,R) # Calculate the position of interacting point incident_angle_1 = 30 + self.state[-1][2] out_angle_1 = np.rad2deg(np.arcsin(self.n_air/self.n_glass * np.sin(np.deg2rad(incident_angle_1)))) ray_state = np.array([result[0],result[1],out_angle_1-30]) self.state.append(ray_state) # The ray come out from the prism incident_slope_2 = np.tan(np.deg2rad(self.state[-1][2])) L = np.array([[-incident_slope_2,1],[math.sqrt(3),1]]) R = np.array([self.state[-1][1]-incident_slope_2*self.state[-1][0],math.sqrt(3)*self.central_point+self.side_length/math.sqrt(3)]) result = np.linalg.solve(L,R) # Calculate the position of interacting point incident_angle_2 = 60- out_angle_1 out_angle_2 = np.rad2deg(np.arcsin(self.n_glass/self.n_air * np.sin(np.deg2rad(incident_angle_2)))) ray_state = np.array([result[0],result[1],30-out_angle_2]) self.state.append(ray_state) def plot_ray(self): ''' plot_ray(self) Plot the prism and the ray path which is described in self.state. ''' # plot the prism x1 = np.linspace(-self.side_length/2+self.central_point,0+self.central_point) y1 = math.sqrt(3)*(x1-self.central_point)+ 2/math.sqrt(3)*self.side_length/2 x2 = np.linspace(0+self.central_point,self.side_length/2+self.central_point) y2 = -math.sqrt(3)*(x2-self.central_point) + 2/math.sqrt(3)*self.side_length/2 x3 = np.linspace(-self.side_length/2+self.central_point,self.side_length/2+self.central_point) y3 = [min(y1)]*len(x3) plt.plot(x1,y1,'k') plt.plot(x2,y2,'k') plt.plot(x3,y3,'k') # plot ray for i in range(len(self.state)): slope = np.tan(np.deg2rad(self.state[i][2])) if i < len(self.state)-1: x = np.linspace(self.state[i][0],self.state[i+1][0]) y = np.linspace(self.state[i][1],self.state[i+1][1],len(x)) plt.plot(x,y) else: x = np.linspace(self.state[i][0],self.state[i][0]+self.state[1][0]) y = slope*x+(self.state[i][1]-slope*self.state[i][0]) plt.plot(x,y) plt.show() if __name__=='__main__': def main(): PT = PrismTracing(0,-1,10) # Three parameters are x, z, angle PT.ray() PT.prism(4,6) # Two parameters are side_length , central position PT.plot_ray() print(PT.state) main()
[ "noreply@github.com" ]
noreply@github.com
690fe2ffb43edf1febae8410ba150129ce00cce0
3419067388879d8a6542df01cb0278ae90b021a2
/py100day/Day01-15/Day04/code/for2.py
22c96886304fa47ecee9b5c39f3f60d9a36a21f1
[]
no_license
oweson/python-river-master
faa31c5248e297a92054cc302e213e2b37fb8bd5
cf9e99e611311b712465eb11dec4bb8f712929b2
refs/heads/master
2021-06-21T15:47:01.755957
2019-10-02T00:08:05
2019-10-02T00:08:05
205,607,518
0
0
null
2021-06-10T21:55:20
2019-08-31T23:39:55
Python
UTF-8
Python
false
false
173
py
""" 用for循环实现1~100之间的偶数求和 Version: 0.1 Author: 骆昊 Date: 2018-03-01 """ # 步长是2 sum = 0 for x in range(2, 101, 2): sum += x print(sum)
[ "570347720@qq.com" ]
570347720@qq.com
066a9bb3b23255dc5f349786bfe1e4b093454a5a
238a0dd7c1bd72b8e241798c8d31ff3cbb2c0c90
/caesar.py
36d0e4ed7074f13b4cd5f3fa51e5d21e4d8bb64d
[]
no_license
Procerus/caesar-converter
7881666ae638ef288af873377436fd482797182c
92a8ce5e764e552351e51048384747683a634c2e
refs/heads/master
2020-09-19T21:21:34.310140
2019-11-26T23:00:48
2019-11-26T23:00:48
224,301,321
0
0
null
null
null
null
UTF-8
Python
false
false
1,730
py
# This program takes in a user input when running the program as an argument # that number is a key that will shift the associated text that the user enters # next and shifts every letter that amount import sys def main(argv): try: sys.argv[1] == None except IndexError: print("Usage: python " + sys.argv[0] + " k") return 0 # key converted to int key = int(argv[1]) # check if the key was entered in properly checks to make sure a number is inputted and # checks if there is extra characters in try: (key == 0 and strncmp(argv[1], "0", true)) or len(sys.argv) > 3 except IndexError: print("Usage: python " + sys.argv[0] + " k") return 0 #ord() convert to in and chr converts to string name = input("plaintext: ") length = len(name) # converts key to modulus of 26 if person typed a larger number key = key % 26 print("ciphertext: ", end="") for i in range(0, length): #checks if the name is lower case if ord(name[i]) > 96 and ord(name[i]) < 123: if (ord(name[i]) + key) % 122 < 97: print(chr(((ord(name[i]) + key) % 122) + 96), end="") else: print(chr(ord(name[i]) + key), end="") # checks if character is uppercase elif ord(name[i]) > 64 and ord(name[i]) < 91: if (ord(name[i]) + key) % 90 < 65: print(chr(((ord(name[i]) + key) % 90) + 65), end="") else: print(chr(ord(name[i]) + key), end="") # if it is non character it will just print else: print(name[i], end="") print("") main(sys.argv)
[ "noreply@github.com" ]
noreply@github.com
26ab05607c4b32f1e825bba9fc453b0a897db9b6
cdf9bba82c5e42758664e1eeb3e157340c29c500
/generator/test.py
b9f9245ed294e3d2fa2ce52879a97137638525f9
[]
no_license
mex3/fizmat-a
ce2802035dd045fcf0f6bfe8d6388ed88c6c2af5
be90332a518f0c3b2113dd967a63f4a056f937ab
refs/heads/master
2021-01-21T13:41:35.050273
2016-05-14T12:13:16
2016-05-14T12:13:16
55,837,428
1
4
null
2016-05-14T12:12:24
2016-04-09T09:59:23
Python
UTF-8
Python
false
false
82
py
inputq = open('test.txt', 'r') s=inputq.readline() print (s) ss=input() print(ss)
[ "gurovic@gmail.com" ]
gurovic@gmail.com
c0bccab0f33fe2f6323731cddd1742ba4d45275c
aa410a95773aeea73e75f0e701db5cdc0eda890b
/weapons.py
cf6e4eb05ba6ad8a453e07637018051ed6eac5f8
[]
no_license
predominant/zombsole
ccc00893b7739c5341c43fc28375415fa628b885
a04ff40a144cb1f63d8aa29ccf0b06ecccc2bc7f
refs/heads/master
2021-01-21T19:29:05.322551
2014-03-26T05:38:15
2014-03-26T05:38:15
null
0
0
null
null
null
null
UTF-8
Python
false
false
710
py
# coding: utf-8 from core import Weapon def _new_weapon_class(name, max_range, damage_range): '''Create new weapon class.''' class NewWeapon(Weapon): def __init__(self): super(NewWeapon, self).__init__(name, max_range, damage_range) NewWeapon.__name__ = name return NewWeapon ZombieClaws = _new_weapon_class('ZombieClaws', 1.5, (5, 10)) Knife = _new_weapon_class('Knife', 1.5, (5, 10)) Axe = _new_weapon_class('Axe', 1.5, (75, 100)) Gun = _new_weapon_class('Gun', 6, (10, 50)) Rifle = _new_weapon_class('Rifle', 10, (25, 75)) Shotgun = _new_weapon_class('Shotgun', 3, (75, 100))
[ "fisadev@gmail.com" ]
fisadev@gmail.com
582bb899d0407eb2070b66f080e8e55395136ae0
5d9d88881abd73cc841f1bc3a523ebbb1c46f6b4
/DeepTrainer/DeepTrainer/state_tracker.py
59f2d525260bf2d1f5fd428196ae7cf7a51ea89f
[]
no_license
zhuMingXu/CarSimRL
3f6e92d73a6eacc9fc311bc5c71e6e909fe79335
bcbb7654f1b68b00edb00ccd6d1480a7db9e6598
refs/heads/master
2022-04-13T14:04:56.596481
2017-02-15T11:57:03
2017-02-15T11:57:03
null
0
0
null
null
null
null
UTF-8
Python
false
false
2,872
py
# -*- coding: utf-8 -*- import constants as CONST import numpy as np import matplotlib.pyplot as plt import matplotlib.patches as mpatches class StateTracker(): def __init__(self): #keeping track of oldest state so I can remove it #before inserting the most recent state self.frame_history = np.zeros(CONST.FRAME_HISTORY_SIZE) self.oldest_state_idx = 0 self.idx_old_to_new = [i for i in range(len(self.frame_history))] #[0,1,2,..n] self.state = np.zeros(CONST.STATE_MATRIX_SIZE) #initalizing gray scale state matrix for scan in self.frame_history: self.state += scan def reset(self): self.frame_history = np.zeros(self.frame_history.shape) self.oldest_state_idx = 0 self.idx_old_to_new = [i for i in range(len(self.frame_history))] #[0,1,2,..n] self.state = np.zeros(self.state.shape) #initalizing gray scale state matrix for scan in self.frame_history: self.state += scan # new_scan is a 2d numpy array representing the lidar one_hot array def update(self, new_scan): #plt.imshow(new_scan, cmap=plt.cm.hot) #new_scan = new_scan.flatten() # sutract oldest scan fron state # self.state -= self.frame_history[self.oldest_state_idx] # superimpose new_scan into state matrix # self.state += new_scan # replace oldest scan with new_scan self.frame_history[self.oldest_state_idx] = new_scan self.state = np.zeros(self.state.shape) weight_idx = 0 for frame in self.frame_history: self.state += frame*CONST.HISTORY_WEIGHTS[weight_idx] weight_idx += 1 # increment olderst_scan_idx self.oldest_state_idx = (self.oldest_state_idx - 1) % len(self.frame_history) for idx in self.idx_old_to_new: idx = (idx + 1) % len(self.idx_old_to_new) def __plotFrame(self, data): values = np.unique(data.ravel()) im = plt.imshow(data, interpolation='none') colors = [im.cmap(im.norm(value)) for value in values] patches = [mpatches.Patch(color=colors[i], label="Level {l}".format(l=values[i])) for i in range(len(values))] plt.legend(handles=patches, bbox_to_anchor=(1.05, 1), loc=2, borderaxespad=0. ) plt.show() def plotState(self, plt_state=True, plt_full_history=False): if plt_state: self.__plotFrame(self.state) if plt_full_history: count = len(self.idx_old_to_new)-1 for idx in self.idx_old_to_new: print("Frame T-{0}: ".format(count)) self.__plotFrame(self.frame_history[idx]) count -= 1
[ "joshua.d.patterson1@gmail.com" ]
joshua.d.patterson1@gmail.com
203f04df3c3f6b979898621a354f1d50daec9fe6
db01067e88324466ba4743e5e53cd53de609c342
/04. Functions Basics Lab/01. Grades.py
0f7ea110f53b23d04e17bca1968d4c40ba88f432
[]
no_license
IlkoAng/-Python-Fundamentals-Softuni
07eaf89d340b2e60214ab5f8e896629ae680dc4a
01a112b13e84ab2f29e6fc4ed39f08f395d54429
refs/heads/main
2023-05-18T02:11:53.676763
2021-06-06T15:39:22
2021-06-06T15:39:22
371,475,022
0
0
null
null
null
null
UTF-8
Python
false
false
368
py
def solve(grade): if 2.00 <= grade <= 2.99: return "Fail" elif 3.00 <= grade <= 3.49: return "Poor" elif 3.50 <= grade <= 4.49: return "Good" elif 4.50 <= grade <= 5.49: return "Very Good" elif 5.50 <= grade <= 6.00: return "Excellent" grade_data = float(input()) print(solve(grade_data))
[ "noreply@github.com" ]
noreply@github.com
1026e1d0f5add5bf40edc076405f2e409f26c5ce
2f2682f778512a75a1ff49d7e267c2f4d355c48e
/geoprocess/controllers.py
7be119b34c9b20b609770261e464a475b5996a9b
[]
no_license
beatcovid/geoprocess
4a44f46b900c2e0ffed0dab18008e7884e759e3b
c2a7b1e4ede06583679db9dadebe2066b0274e54
refs/heads/master
2023-04-13T13:45:48.572825
2020-05-27T03:08:14
2020-05-27T03:08:14
260,215,049
0
1
null
2023-03-29T00:36:19
2020-04-30T13:11:38
Python
UTF-8
Python
false
false
4,538
py
import csv import email.utils import json import logging import os import sys from datetime import datetime from pprint import pprint from dotenv import load_dotenv from pymongo import MongoClient from geoprocess.find_psma import find_lga, find_sa3 from geoprocess.google_geo import google_geocode, lookup_placeid, place_autocomplete from geoprocess.settings import MONGO_CONNECT_URL load_dotenv() logger = logging.getLogger("geoprocess") logging.basicConfig(level=logging.INFO) logger.setLevel(logging.INFO) mongo_connection = MongoClient(MONGO_CONNECT_URL) def flatten_google_place(place, prefix): ac = place["address_components"] flattened = {} for component in ac: for ctype in component["types"]: if not ctype == "political": flattened[prefix + "_" + ctype] = component["short_name"] return flattened def get_granuality(flat_geo, prefix): FIELDS = [ f"{prefix}_postal_code", f"{prefix}_locality", f"{prefix}_administrative_area_level_2", f"{prefix}_administrative_area_level_1", f"{prefix}_country", ] for field in FIELDS: if field in flat_geo: return field[len(prefix) + 1 :] return "country" def update_geoplots(): """ just a simple q """ db = mongo_connection.prod_covid19_api_docdb.instances query = {"_geo_processed": {"$ne": True}} processed = 0 updated = 0 place_fields = ["userdetail_city", "travel_country"] for a in db.find(query).sort("_submission_time", -1): for place_field in place_fields: if place_field in a: if not type(a[place_field]) is str: continue if " " in a[place_field]: continue try: p = lookup_placeid(a[place_field]) except Exception as e: logger.error("Could not find place id for: {}".format(a[place_field])) logger.error(e) continue p_flat = flatten_google_place(p, place_field) if ( place_field + "_country" in p_flat and p_flat[place_field + "_country"] == "AU" and ( place_field + "_locality" in p_flat or place_field + "_postal_code" in p_flat ) ): if not place_field + "_lga_id" in a: lgs = find_lga( p["geometry"]["location"]["lat"], p["geometry"]["location"]["lng"], ) if lgs: p_flat[place_field + "_lga_id"] = lgs if not place_field + "_sa3_id" in a: sa3 = find_sa3( p["geometry"]["location"]["lat"], p["geometry"]["location"]["lng"], ) if sa3: p_flat[place_field + "_sa3_id"] = sa3 p_flat[place_field + "_granuality"] = get_granuality(p_flat, place_field) if ( place_field + "_country" in p_flat and p_flat[place_field + "_country"] == "AU" and ( place_field + "_administrative_area_level_1" in p_flat or "userdetail_city_postal_code" in p_flat ) ): p_flat[place_field + "_state"] = p_flat[ place_field + "_administrative_area_level_1" ] p_flat["_geo_processed"] = True pprint(p_flat) try: db.update_one( {"_id": a["_id"]}, {"$set": p_flat}, ) except Exception as e: logger.error( "Db error on updating place_id: {} {}".format( a["_id"], place_field ) ) logger.error(e) continue logger.info( "Updated {} {} -> {}".format(place_field, a["_id"], a[place_field]) ) updated += 1 processed += 1 print("Processed {} and updated {}".format(processed, updated))
[ "nc9@protonmail.com" ]
nc9@protonmail.com
2eac0fe3402f79f389178ebe792a10a16f7c1a4a
039f2c747a9524daa1e45501ada5fb19bd5dd28f
/AGC001/AGC001c.py
6f3fb7892212fb5a2683a833717ea55a344d0dfd
[ "Unlicense" ]
permissive
yuto-moriizumi/AtCoder
86dbb4f98fea627c68b5391bf0cc25bcce556b88
21acb489f1594bbb1cdc64fbf8421d876b5b476d
refs/heads/master
2023-03-25T08:10:31.738457
2021-03-23T08:48:01
2021-03-23T08:48:01
242,283,632
0
0
null
null
null
null
UTF-8
Python
false
false
168
py
#AGC001c def main(): import sys input=sys.stdin.readline sys.setrecursionlimit(10**6) # map(int, input().split()) if __name__ == '__main__': main()
[ "kurvan1112@gmail.com" ]
kurvan1112@gmail.com
cee887c62ba400532290d58768227c8f8146d85e
e892207dd96f39ee2b9a99761b08381c442bef61
/sources/simple_shop/adapters/message_bus/scheme.py
aa54e44be9cd5c4ff308c76d099b67dd11a68c6a
[ "MIT" ]
permissive
theevs/classic_demo_medium_difficulty
f9196795bd60acea3458c3b83d8a56c79ecdb9e7
335b50f9509c85dc426a76dbef7f7f2d28603e76
refs/heads/main
2023-08-31T21:32:51.001017
2021-10-26T14:38:58
2021-10-26T14:38:58
null
0
0
null
null
null
null
UTF-8
Python
false
false
186
py
from classic.messaging_kombu import BrokerScheme from kombu import Exchange, Queue broker_scheme = BrokerScheme( Queue('PrintOrderPlaced', Exchange('OrderPlaced'), max_length=1) )
[ "variasov@gmail.com" ]
variasov@gmail.com
a61174c4d8077eef4dc25a83b1c32e6f227bcf5d
f0f2d8cb16d494443a678ea24c04be95d1cbf824
/Time_table_generator_/py_ui/Room.py
e40c89794cad5314c9a7d7c2e008ddbbc15889b5
[]
no_license
Automatic-Timetable-Generator/ATG
314e09b2becef67913df0744c094bca4d20635f0
6b1187e0be434346bfdd1a61a30bb57718fb0cbc
refs/heads/master
2021-02-17T10:51:00.604358
2020-04-08T07:30:30
2020-04-08T07:30:30
245,091,233
0
0
null
null
null
null
UTF-8
Python
false
false
3,306
py
# -*- coding: utf-8 -*- # Form implementation generated from reading ui file 'room.ui' # # Created by: PyQt5 UI code generator 5.6 # # WARNING! All changes made in this file will be lost! from PyQt5 import QtCore, QtGui, QtWidgets class Ui_Dialog(object): def setupUi(self, Dialog): Dialog.setObjectName("Dialog") Dialog.resize(716, 553) Dialog.setMinimumSize(QtCore.QSize(716, 553)) Dialog.setMaximumSize(QtCore.QSize(716, 553)) self.verticalLayout = QtWidgets.QVBoxLayout(Dialog) self.verticalLayout.setObjectName("verticalLayout") self.gridLayout = QtWidgets.QGridLayout() self.gridLayout.setSizeConstraint(QtWidgets.QLayout.SetMinimumSize) self.gridLayout.setObjectName("gridLayout") self.lblName = QtWidgets.QLabel(Dialog) self.lblName.setObjectName("lblName") self.gridLayout.addWidget(self.lblName, 0, 0, 1, 1) self.lineEditName = QtWidgets.QLineEdit(Dialog) self.lineEditName.setObjectName("lineEditName") self.gridLayout.addWidget(self.lineEditName, 0, 1, 1, 1) self.groupBox = QtWidgets.QGroupBox(Dialog) self.groupBox.setObjectName("groupBox") self.horizontalLayout_2 = QtWidgets.QHBoxLayout(self.groupBox) self.horizontalLayout_2.setObjectName("horizontalLayout_2") self.radioLec = QtWidgets.QRadioButton(self.groupBox) self.radioLec.setObjectName("radioLec") self.horizontalLayout_2.addWidget(self.radioLec) self.radioLab = QtWidgets.QRadioButton(self.groupBox) self.radioLab.setObjectName("radioLab") self.horizontalLayout_2.addWidget(self.radioLab) self.gridLayout.addWidget(self.groupBox, 0, 2, 1, 1) self.verticalLayout.addLayout(self.gridLayout) self.tableSchedule = QtWidgets.QTableView(Dialog) self.tableSchedule.setObjectName("tableSchedule") self.verticalLayout.addWidget(self.tableSchedule) self.horizontalLayout = QtWidgets.QHBoxLayout() self.horizontalLayout.setObjectName("horizontalLayout") self.btnFinish = QtWidgets.QPushButton(Dialog) self.btnFinish.setObjectName("btnFinish") self.horizontalLayout.addWidget(self.btnFinish) self.btnCancel = QtWidgets.QPushButton(Dialog) self.btnCancel.setObjectName("btnCancel") self.btnFinish.setStyleSheet('background-color:#833471;color:white;') self.btnCancel.setStyleSheet('background-color:#747d8c;color:white;') self.horizontalLayout.addWidget(self.btnCancel) self.verticalLayout.addLayout(self.horizontalLayout) self.retranslateUi(Dialog) QtCore.QMetaObject.connectSlotsByName(Dialog) def retranslateUi(self, Dialog): _translate = QtCore.QCoreApplication.translate Dialog.setWindowTitle(_translate("Dialog", "Room")) self.lblName.setText(_translate("Dialog", "Name")) self.groupBox.setTitle(_translate("Dialog", "Type")) self.radioLec.setText(_translate("Dialog", "Lecture")) self.radioLab.setText(_translate("Dialog", "Laboratory")) self.btnFinish.setText(_translate("Dialog", "Finish")) self.btnCancel.setText(_translate("Dialog", "Cancel"))
[ "noreply@github.com" ]
noreply@github.com
dafc1d3715006b5918f20ffe70c8e774c8a86cf5
1fb9a17e9273a627b090b1f332fab15f14d5baca
/Clustering_Twitter_Data.py
e0a27cd69ed88ca34b7c77cfda65f70fe673eb8e
[]
no_license
Akhilavk1106/Clustering-Twitter-Health-Data
1803481d3f52d7167909c15404894e4a73cd2fd1
8dc2f3a30d68ac287680ecc4318ea2c621b12933
refs/heads/master
2020-03-29T07:34:13.632976
2018-09-21T00:52:54
2018-09-21T00:52:54
149,670,672
0
0
null
null
null
null
UTF-8
Python
false
false
1,983
py
import re from nltk.corpus import stopwords from sklearn.feature_extraction.text import TfidfVectorizer from sklearn.decomposition import PCA from sklearn.cluster import KMeans import matplotlib.pyplot as plt import numpy as np from sklearn import cluster, datasets import pandas as pd import sys original_labels=np.empty(16,dtype='int64') tweet_num=0 tweets=[] listfile = [] for file_name in sys.argv[1:]: list=[] with open(file_name,encoding='ISO-8859-1') as file: list_special=['rt','video','amp','may'] for row in file.readlines(): content=row.split('|') c=content[-1].split(' http') c[0]=c[0].lower() remove_pun = re.sub("[\s+\.\-\!\;\:\/_,$%^(+\"\']+|[+——!,。?、~@#¥%…&…*()]+", " ", c[0]) list.append(remove_pun) tweets.extend(list) original_labels[tweet_num]=len(list) tweet_num +=1 print(original_labels) true_cluster2 = np.empty((len(tweets), 1), dtype='int64') begin = 0 for i in range(len(original_labels)): end = begin + original_labels[i] true_cluster2[begin:end, 0] = i begin = end vectorizer = TfidfVectorizer(stop_words='english',max_features=5000) X = vectorizer.fit_transform(tweets) array_trans=X.toarray() pca2=PCA(n_components=2) newMat = pca2.fit_transform(array_trans) kmeans = KMeans(n_clusters=16,random_state=0).fit(newMat) labels = kmeans.labels_ X_clustered = kmeans.fit_predict(newMat) ind=0 print(true_cluster2) for la in labels: print('OriginalLabel:',true_cluster2[ind],'ClusterLabel',la) ind+=1 #Define our own color map LABEL_COLOR_MAP = {0: 'b', 1: 'c', 2: 'k',3:'m', 4: 'green', 5: 'r',6:'w', 7: 'y', 8: 'ivory',9:'navy', 10: 'orange', 11: 'purple',12:'olive', 13: 'gray', 14: 'maroon',15:'pink', 16: 'tan'} label_color = [LABEL_COLOR_MAP[l] for l in X_clustered] # Plot the scatter digram plt.figure(figsize = (25,25)) plt.scatter(newMat[:,0],newMat[:,1], c= label_color, alpha=0.5) plt.show()
[ "noreply@github.com" ]
noreply@github.com
1955dee98cd2d41ac68b3c467664e2dd537aaf21
354f5a285b0c6f14475aba45a18312995a337de6
/core/admin.py
7a082b38139b57888ec5f982338832dba5f04bf7
[]
no_license
lucasdmarten/navedexAPI
45063cf63a8b7d55cc23a01ada6451a8e726e6f7
712425268a2cbd8f041deac32ea3d25292ea30fe
refs/heads/master
2023-03-23T11:27:19.833069
2021-03-15T22:15:41
2021-03-15T22:15:41
345,178,453
0
0
null
null
null
null
UTF-8
Python
false
false
3,453
py
from django import forms from django.contrib import admin from django.contrib.auth.models import Group from django.contrib.auth.admin import UserAdmin as BaseUserAdmin from django.contrib.auth.forms import ReadOnlyPasswordHashField from django.core.exceptions import ValidationError from core.models import MyUser, Naver, Projeto class UserCreationForm(forms.ModelForm): """A form for creating new users. Includes all the required fields, plus a repeated password.""" username = forms.CharField(max_length=10) password1 = forms.CharField(label='Password', widget=forms.PasswordInput) password2 = forms.CharField(label='Password confirmation', widget=forms.PasswordInput) class Meta: model = MyUser fields = ('username','email',) def clean_password2(self): # Check that the two password entries match password1 = self.cleaned_data.get("password1") password2 = self.cleaned_data.get("password2") if password1 and password2 and password1 != password2: raise ValidationError("Passwords don't match") return password2 def save(self, commit=True): # Save the provided password in hashed format user = super().save(commit=False) user.set_password(self.cleaned_data["password1"]) if commit: user.save() return user class UserChangeForm(forms.ModelForm): """A form for updating users. Includes all the fields on the user, but replaces the password field with admin's password hash display field. """ password = ReadOnlyPasswordHashField() class Meta: model = MyUser fields = ('username','email', 'password', 'is_active', 'is_admin') def clean_password(self): # Regardless of what the user provides, return the initial value. # This is done here, rather than on the field, because the # field does not have access to the initial value return self.initial["password"] class UserAdmin(BaseUserAdmin): # The forms to add and change user instances form = UserChangeForm add_form = UserCreationForm # The fields to be used in displaying the User model. # These override the definitions on the base UserAdmin # that reference specific fields on auth.User. list_display = ('username', 'email', 'is_admin') list_filter = ('is_admin',) fieldsets = ( (None, {'fields': ('username','email', 'password')}), ('Permissions', {'fields': ('is_admin',)}), ) # add_fieldsets is not a standard ModelAdmin attribute. UserAdmin # overrides get_fieldsets to use this attribute when creating a user. add_fieldsets = ( (None, { 'classes': ('wide',), 'fields': ('username','email', 'password1', 'password2'), }), ) search_fields = ('username','email',) ordering = ('username','email',) filter_horizontal = () class NaverAdmin(admin.ModelAdmin): list_display = ('id','user','fullname','birthdate','admission_date','job_role','get_projects') class ProjetoAdmin(admin.ModelAdmin): list_display = ('id','name','get_users_id',) # Now register the new UserAdmin... admin.site.register(MyUser, UserAdmin) admin.site.register(Naver, NaverAdmin) admin.site.register(Projeto, ProjetoAdmin) # ... and, since we're not using Django's built-in permissions, # unregister the Group model from admin. admin.site.unregister(Group)
[ "lucasdmarten@gmail.com" ]
lucasdmarten@gmail.com
008b54483f4cb8901c5fb12ee52cc4dfed87daa6
8900db5096dcf2f8a758f6792cc7c8f2bc1d7785
/src/dataset.py
b23694ff85df42853de900a29099a30f0b2c3931
[]
no_license
PeterZhouSZ/seg2vid
ca5d649a19c47b769fc243ef5da4f95e965c42a9
306ca254ac30620afe48767d02b9cf3fecba1194
refs/heads/master
2020-06-01T19:50:22.737812
2019-03-26T11:37:58
2019-03-26T11:37:58
null
0
0
null
null
null
null
UTF-8
Python
false
false
3,811
py
from datasets.dataset_path import * def get_training_set(opt): assert opt.datset in ['cityscapes', 'cityscapes_two_path', 'kth'] if opt.dataset == 'cityscapes': from datasets.cityscapes_dataset_w_mask import Cityscapes train_Dataset = Cityscapes(datapath=CITYSCAPES_TRAIN_DATA_PATH, datalist=CITYSCAPES_TRAIN_DATA_LIST, size=opt.input_size, split='train', split_num=1, num_frames=opt.num_frames) elif opt.dataset == 'cityscapes_two_path': from datasets.cityscapes_dataset_w_mask_two_path import Cityscpes train_Dataset = Cityscapes(datapath=CITYSCAPES_TRAIN_DATA_PATH, mask_data_path=CITYSCAPES_TRAIN_DATA_SEGMASK_PATH, datalist=CITYSCAPES_TRAIN_DATA_LIST, size=opt.input_size, split='train', split_num=1, num_frames=opt.num_frames, mask_suffix='ssmask.png') elif opt.dataset == 'kth': from datasets.kth_dataset import KTH train_Dataset = KTH(dataset_root=KTH_DATA_PATH, datalist=KTH_DATA_PATH_LIST, size=opt.input_size, num_frames=opt.num_frames) return train_Dataset def get_test_set(opt): assert opt.dataset in ['cityscapes', 'cityscapes_two_path', 'kth', 'ucf101', 'KITTI'] if opt.dataset == 'cityscapes': from datasets.cityscapes_dataset_w_mask import Cityscapes test_Dataset = Cityscapes(datapath=CITYSCAPES_VAL_DATA_PATH, mask_data_path=CITYSCAPES_VAL_DATA_SEGMASK_PATH, datalist=CITYSCAPES_VAL_DATA_LIST, size=opt.input_size, split='train', split_num=1, num_frames=opt.num_frames, mask_suffix='ssmask.png', returnpath=True) elif opt.dataset == 'cityscapes_two_path': from datasets.cityscapes_dataset_w_mask_two_path import Cityscapes test_Dataset = Cityscapes(datapath=CITYSCAPES_VAL_DATA_PATH, mask_data_path=CITYSCAPES_VAL_DATA_SEGMASK_PATH, datalist=CITYSCAPES_VAL_DATA_LIST, size=opt.input_size, split='train', split_num=1, num_frames=opt.num_frames, mask_suffix='ssmask.png', returnpath=True) elif opt.dataset == 'cityscapes_pix2pixHD': from cityscapes_dataloader_w_mask_pix2pixHD import Cityscapes test_Dataset = Cityscapes(datapath=CITYSCAPES_TEST_DATA_PATH, mask_data_path=CITYSCAPES_VAL_DATA_SEGMASK_PATH, datalist=CITYSCAPES_VAL_DATA_MASK_LIST, size= opt.input_size, split='test', split_num=1, num_frames=opt.num_frames, mask_suffix='ssmask.png', returnpath=True) elif opt.dataset == 'kth': from datasets.kth_dataset import KTH test_Dataset = KTH(dataset_root=KTH_DATA_PATH, datalist='./file_list/kth_test_%s_16_ok.txt' % opt.category, size=opt.input_size, num_frames=opt.num_frames) elif opt.dataset == 'KITTI': from datasets.kitti_dataset import KITTI kitti_dataset_list = os.listdir(KITTI_DATA_PATH) test_Dataset = KITTI(datapath=KITTI_DATA_PATH, datalist=kitti_dataset_list, size=opt.input_size, returnpath=True) elif opt.dataset == 'ucf101': from datasets.ucf101_dataset import UCF101 test_Dataset = UCF101(datapath=os.path.join(UCF_101_DATA_PATH, category), datalist=os.path.join(UCF_101_DATA_PATH, 'list/test%s.txt' % (opt.category.lower())), returnpath=True) return test_Dataset
[ "panjunting@sensetime.com" ]
panjunting@sensetime.com
9cc02e6a288eb047e372805fdff7b5c41409b6f1
7b71da9189de3358ef73b37a3083a56c1ab10772
/robobench/calibration/pipette_calibration/classify.py
3f968b169085ed7e6e911dc5b6a88230de32baef
[]
no_license
EndyLab/opentrons
91ff3b8364c9b1746f7d073875651baa5efaf4c3
75d1789ad6ddef556a2c46e6608d5496f9f5ec7d
refs/heads/master
2021-01-13T13:39:32.443502
2019-11-19T22:23:47
2019-11-19T22:23:47
76,410,678
8
0
null
2020-10-27T21:05:49
2016-12-14T00:42:02
Roff
UTF-8
Python
false
false
1,890
py
import cv2 import numpy as np # classifies an array of imgs def knn(img, k=5): # load the data we generated previously training_dir = "C:/Users/gohna/Documents/bioe reu/opentrons/robobench/calibration/pipette_calibration/training" samples = np.loadtxt(training_dir+'/general-samples.data').astype(np.float32) responses = np.loadtxt(training_dir+'/general-responses.data').astype(np.float32) responses = responses.reshape((responses.size,1)) # train the KNN model knn_model = cv2.ml.KNearest_create() knn_model.train(samples,cv2.ml.ROW_SAMPLE,responses) dists = [] img_scaled = cv2.resize(img, (10,25)) sample = img_scaled.reshape((1,250)) sample = np.float32(sample) ret, results, neighbours, dist = knn_model.findNearest(sample, k) identified = int(results[0][0]) dists.append(dist) # print('distance:',neighbours) return identified if __name__ == '__main__': training_dir = "C:/Users/gohna/Documents/bioe reu/opentrons/robobench/calibration/pipette_calibration/training" # load the data we generated previously samples = np.loadtxt(training_dir+'/general-samples.data').astype(np.float32) responses = np.loadtxt(training_dir+'/general-responses.data').astype(np.float32) responses = responses.reshape((responses.size,1)) # train the KNN model print("sample size", samples.shape,"response size:",responses.size) knn_model = cv2.ml.KNearest_create() knn_model.train(samples,cv2.ml.ROW_SAMPLE,responses) test = training_dir + '/9/DIGIT120207.jpg' img = cv2.imread(test) img = cv2.cvtColor(img, cv2.COLOR_BGR2GRAY) img_scaled = cv2.resize(img, (10, 25)) print(img_scaled.shape) sample = img_scaled.reshape((1,250)) sample = np.float32(sample) print("img test size", sample.shape) ret, results, neighbours, dist = knn_model.findNearest(sample, k=2) matches = results==responses string = str(int((results[0][0]))) # print(matches) print(string)
[ "natagoh@stanford.edu" ]
natagoh@stanford.edu
6168b0398ebb09f7c55ed863aca797354077e778
c0b4a1991ad529f162779e60d0af3e22f2468aaa
/cogs/members.py
fa99048e26540f997e676784f27765288f0b4420
[]
no_license
scosco97/apibot
51ae66317a4abfe7bb9380b23679ad476800ad1e
3f714c2daa6a2cd046d417bc0e74e2501ed55959
refs/heads/master
2023-07-29T09:05:36.770121
2021-09-11T20:19:35
2021-09-11T20:19:35
null
0
0
null
null
null
null
UTF-8
Python
false
false
6,015
py
import discord import random from config import settings from discord.ext import commands class MembersCog(commands.Cog): def __init__(self, bot): self.bot = bot @commands.command(name="welcome", hidden=True) async def welcome(self, ctx, member: discord.Member = None): if not member: return await ctx.send("Member does not exist.") channel = self.bot.get_channel(settings['channels']['welcome']) msg = (f"Welcome to the Clash API Developers server, {member.mention}! We're glad to have you!\n" f"First, please let us know what your preferred programming language is. " f"Next, if you've already started working with the API, please tell us a little about your project. " f"If you haven't started a project yet, let us know what you're interested in making.") await channel.send(msg) mod_log = self.bot.get_channel(settings['channels']['mod-log']) msg = f"{member.display_name}#{member.discriminator} just joined the server." await mod_log.send(f"{msg} (This message generated by the `//welcome` command initiated by " f"{ctx.author.display_name}.") @commands.Cog.listener() async def on_member_join(self, member): """Discord listener which is called when a user joins the Discord server.""" if member.guild.id != 566451504332931073: # only act if they are joining API server return if not member.bot: channel = self.bot.get_channel(settings['channels']['welcome']) msg = (f"Welcome to the Clash API Developers server, {member.mention}! We're glad to have you!\n" f"First, please let us know what your preferred programming language is. " f"Next, if you've already started working with the API, please tell us a little about your project. " f"If you haven't started a project yet, let us know what you're interested in making.") await channel.send(msg) else: channel = self.bot.get_channel(settings['channels']['admin']) await channel.send(f"{member.mention} has just been invited to the server. " f"Perhaps it is time to set up a demo channel? Try `//setup {member.mention} @owner`") mod_log = self.bot.get_channel(settings['channels']['mod-log']) msg = f"{member.display_name}#{member.discriminator} just joined the server." await mod_log.send(msg) @commands.Cog.listener() async def on_member_update(self, old_member, new_member): """Discord listener to announce new member with Developer role to #general""" if new_member.guild.id != 566451504332931073: # only act if this is the API server return if old_member.roles == new_member.roles: return developer_role = new_member.guild.get_role(settings['roles']['developer']) if developer_role not in old_member.roles and developer_role in new_member.roles: if new_member.bot: channel = self.bot.get_channel(settings['channels']['admin']) await channel.send(f"Who is the bonehead that assigned the Developer role to a bot? " f"{new_member.name} is a bot.") # At this point, it should be a member on our server that has just received the developers role self.bot.logger.info(f"New member with Developers role: {new_member.display_name}") sql = "SELECT role_id, role_name, emoji_repr FROM bot_language_board" fetch = await self.bot.pool.fetch(sql) language_roles = [[row['role_id'], row['role_name'], row['emoji_repr']] for row in fetch] member_languages = "" member_role_emoji = [] for language_role in language_roles: for role in new_member.roles: if language_role[0] == role.id: member_languages += f"{language_role[1]}\n" member_role_emoji.append(language_role[2]) channel = new_member.guild.get_channel(settings['channels']['general']) embed = discord.Embed(color=discord.Color.blue(), description=f"Please welcome {new_member.display_name} to the Clash API Developers " f"server.") embed.set_thumbnail(url=new_member.avatar_url_as(size=128)) if member_languages: embed.add_field(name="Languages:", value=member_languages) msg = await channel.send(embed=embed) if member_role_emoji: for emoji in member_role_emoji: await msg.add_reaction(emoji) @commands.Cog.listener() async def on_member_remove(self, member): """Discord listener which is called when a user leaves the Discord server.""" if member.guild.id != 566451504332931073: # only act if they are joining API server return # Build random list of messages msg_options = [" just left the server. Buh Bye!", " just left our Discord. I wonder if we will miss them.", " just left. What's up with that?", " went bye-bye. Who will fill the void?", " has left us. A short moment of silence.", " has departed. Hope they learned everything they need!", ] channel = self.bot.get_channel(settings['channels']['general']) msg = member.display_name + random.choice(msg_options) await channel.send(msg) mod_log = self.bot.get_channel(settings['channels']['mod-log']) msg = f"{member.display_name}#{member.discriminator} just left the server." await mod_log.send(msg) def setup(bot): bot.add_cog(MembersCog(bot))
[ "wpmjones@gmail.com" ]
wpmjones@gmail.com
41203f92213b29c8d6459485d713fd321114f4fd
6f33999bb1cc273388bf4d7dfa550bdf428cdf04
/myapp/migrations/0001_initial.py
f97924a41e670ebe398b72ef59bf9e701d396ab3
[]
no_license
xiezongzheng/test9_29
863fa5b85b65b2200b070800c576f41da11d4653
afb8c567f29f60a9e9d28693ceb1bfada967e44a
refs/heads/master
2021-01-10T01:19:05.587110
2015-11-01T04:17:44
2015-11-01T04:17:44
44,243,815
3
0
null
null
null
null
UTF-8
Python
false
false
3,063
py
# -*- coding: utf-8 -*- from __future__ import unicode_literals from django.db import models, migrations class Migration(migrations.Migration): dependencies = [ ] operations = [ migrations.CreateModel( name='Bactname', fields=[ ('id', models.IntegerField(serialize=False, primary_key=True)), ('num', models.TextField(db_column='NUM', blank=True)), ('genus', models.CharField(max_length=50, db_column='GENUS', blank=True)), ('species', models.CharField(max_length=50, db_column='SPECIES', blank=True)), ('subspecies', models.CharField(max_length=50, db_column='SUBSPECIES', blank=True)), ('reference', models.CharField(max_length=50, db_column='REFERENCE', blank=True)), ('status', models.CharField(max_length=50, db_column='STATUS', blank=True)), ('authors', models.CharField(max_length=50, db_column='AUTHORS', blank=True)), ('remarks', models.CharField(max_length=50, db_column='REMARKS', blank=True)), ('risk_grp', models.CharField(max_length=50, db_column='RISK_GRP', blank=True)), ('type_strains', models.CharField(max_length=50, db_column='TYPE_STRAINS', blank=True)), ('taxonid', models.CharField(max_length=50, db_column='taxonId', blank=True)), ('ncbitaxonid', models.CharField(max_length=50, db_column='ncbiTaxonId', blank=True)), ('mclid', models.CharField(max_length=50, db_column='mclId', blank=True)), ('sequence', models.TextField(db_column='SEQUENCE', blank=True)), ], options={ 'db_table': 'bactname', 'managed': False, }, bases=(models.Model,), ), migrations.CreateModel( name='TaxonMapping', fields=[ ('id', models.IntegerField(serialize=False, primary_key=True)), ('speciesname', models.CharField(max_length=100, db_column='speciesName', blank=True)), ('taxonid', models.CharField(max_length=50, db_column='taxonId', blank=True)), ('ncbitaxonid', models.CharField(max_length=50, db_column='ncbiTaxonId', blank=True)), ('mclid', models.CharField(max_length=50, db_column='mclId', blank=True)), ], options={ 'db_table': 'taxon_mapping', 'managed': False, }, bases=(models.Model,), ), migrations.CreateModel( name='User', fields=[ ('id', models.IntegerField(serialize=False, primary_key=True)), ('username', models.CharField(max_length=100, blank=True)), ('password', models.CharField(max_length=100, blank=True)), ], options={ 'db_table': 'user', 'managed': False, }, bases=(models.Model,), ), ]
[ "421489422@qq.com" ]
421489422@qq.com
4bfe6ebbc3532652449f4621355b38f922dd4b06
977eb763cdf049d6cd58b3055bd353e2d93afbed
/readfinstar.py
ae730e393490e5c4d0bf1c6bf0f47bf9db7fea71
[]
no_license
syadav8126/toofanTicker
7a64a3af9b1e73d20ab8b3d6af18c7e99c7a8a19
e731748e54c780d3964a0e8595d0f08e46d1d938
refs/heads/main
2023-02-20T07:36:15.848279
2021-01-26T06:44:01
2021-01-26T06:44:01
332,995,439
0
0
null
null
null
null
UTF-8
Python
false
false
400
py
import finstar import csv import sys import subprocess from subprocess import Popen import time input_file='standalone.csv' with open(input_file, 'r') as f: data = csv.reader(f) for row in data: cmd=[sys.executable, './finstar.py',row[0]] Popen(cmd,shell=False,stdin=None,stdout=None,stderr=None,close_fds=True) time.sleep(0.18) #subprocess.call([sys.executable, './finstar.py',row[0]])
[ "syadav8126@gmail.com" ]
syadav8126@gmail.com
def39a55d547e1131e0f8dcf639f5da81e09bb90
f0d713996eb095bcdc701f3fab0a8110b8541cbb
/cGaTqHsPfR5H6YBuj_0.py
c3936bfae1158025ccd064458e0c9c17ee2d0b5e
[]
no_license
daniel-reich/turbo-robot
feda6c0523bb83ab8954b6d06302bfec5b16ebdf
a7a25c63097674c0a81675eed7e6b763785f1c41
refs/heads/main
2023-03-26T01:55:14.210264
2021-03-23T16:08:01
2021-03-23T16:08:01
350,773,815
0
0
null
null
null
null
UTF-8
Python
false
false
818
py
""" Given a list of ingredients `i` and a flavour `f` as input, create a function that returns the list, but with the elements `bread` around the selected ingredient. ### Examples make_sandwich(["tuna", "ham", "tomato"], "ham") ➞ ["tuna", "bread", "ham", "bread", "tomato"] make_sandwich(["cheese", "lettuce"], "cheese") ➞ ["bread", "cheese", "bread", "lettuce"] make_sandwich(["ham", "ham"], "ham") ➞ ["bread", "ham", "bread", "bread", "ham", "bread"] ### Notes * You will always get valid inputs. * Make two separate sandwiches if two of the same elements are next to each other (see example #3). """ def make_sandwich(ingredients, flavour): sandwich = [] for i in ingredients: sandwich += ['bread', i, 'bread'] if i == flavour else [i] return sandwich
[ "daniel.reich@danielreichs-MacBook-Pro.local" ]
daniel.reich@danielreichs-MacBook-Pro.local
5dc6bd71fa378f65e395229b4201d11a93f1a69b
e55c20745958762f899d79e9fad8fedee0cc2a53
/apps/courses/migrations/0009_course_teacher.py
bb9c3c92de0ba0507b54c908fbc1b9af6b5a71f5
[]
no_license
lisi2016/MxOnline
1b4703fbac6c88c66f0e7b3b5fbcfb1f1ab077ee
0aeca26244b8c446fea096dcdbefcbaee2835dc1
refs/heads/master
2021-08-19T01:25:34.808049
2017-11-24T10:14:06
2017-11-24T10:14:06
null
0
0
null
null
null
null
UTF-8
Python
false
false
649
py
# -*- coding: utf-8 -*- # Generated by Django 1.9 on 2017-11-17 15:58 from __future__ import unicode_literals from django.db import migrations, models import django.db.models.deletion class Migration(migrations.Migration): dependencies = [ ('organization', '0006_teacher_image'), ('courses', '0008_video_learn_times'), ] operations = [ migrations.AddField( model_name='course', name='teacher', field=models.ForeignKey(blank=True, null=True, on_delete=django.db.models.deletion.CASCADE, to='organization.Teacher', verbose_name='\u6388\u8bfe\u8bb2\u5e08'), ), ]
[ "caozhiqiango@foxmail.com" ]
caozhiqiango@foxmail.com
5314e273b1fab0e24e981f1fa735e30171ba22a2
a76ea88df1f28bab17fae40f1c0b4a4471e3da05
/wanderbot-ws/src/teleopbot/src/keys_to_twist.py
6fd8b0342f416ee377d5cad97ce7997915336bc4
[]
no_license
oemergenc/ros-wander-bot
0182b902851f6e896654c72d08e7c86f2bd72311
d78d2a6595882f2f05c468c84a4d102cca8a699b
refs/heads/master
2020-04-14T02:03:03.829173
2019-01-07T10:05:29
2019-01-07T10:05:29
163,575,261
0
0
null
null
null
null
UTF-8
Python
false
false
756
py
#!/usr/bin/env python # BEGIN ALL import rospy from std_msgs.msg import String from geometry_msgs.msg import Twist # BEGIN KEYMAP key_mapping = {'w': [0, 1], 'x': [0, -1], 'a': [-1, 0], 'd': [1, 0], 's': [0, 0]} # END KEYMAP def keys_cb(msg, twist_pub): # BEGIN CB if len(msg.data) == 0 or not key_mapping.has_key(msg.data[0]): return # unknown key. vels = key_mapping[msg.data[0]] # END CB t = Twist() t.angular.z = vels[0] t.linear.x = vels[1] twist_pub.publish(t) if __name__ == '__main__': rospy.init_node('keys_to_twist') twist_pub = rospy.Publisher('cmd_vel', Twist, queue_size=1) rospy.Subscriber('keys', String, keys_cb, twist_pub) rospy.spin() # END ALL
[ "oemergenc@gmail.com" ]
oemergenc@gmail.com
cb5a0612aa59838260c5412af9a13b580b3b489b
28114f0389cc27d34c3a12e241cc4aac73952b3c
/SensorData/SensorData/urls.py
c9ea4cd0562eb2141ab41382d17844dc50c4a54a
[]
no_license
sherlockvind/swlab2_assignment5
57e1feebc377be4dacaf555ac1db1a28798a7fa5
5c8da346e7bbdbcd4942af4bb88fa68d113b861b
refs/heads/main
2023-04-06T15:58:39.086948
2021-04-21T07:17:57
2021-04-21T07:17:57
353,388,229
0
0
null
null
null
null
UTF-8
Python
false
false
815
py
"""SensorData URL Configuration The `urlpatterns` list routes URLs to views. For more information please see: https://docs.djangoproject.com/en/3.1/topics/http/urls/ Examples: Function views 1. Add an import: from my_app import views 2. Add a URL to urlpatterns: path('', views.home, name='home') Class-based views 1. Add an import: from other_app.views import Home 2. Add a URL to urlpatterns: path('', Home.as_view(), name='home') Including another URLconf 1. Import the include() function: from django.urls import include, path 2. Add a URL to urlpatterns: path('blog/', include('blog.urls')) """ from django.contrib import admin from django.urls import path, include urlpatterns = [ path('admin/', admin.site.urls), path('', include('main.urls', namespace='main')), ]
[ "vb.3596@gmail.com" ]
vb.3596@gmail.com
6f53e73452513ff22b1f2c3fd170aeff0d69f4c2
532dc9970b8808fe806aa6a02c0cf9bb7c742b3c
/model_form/forms.py
1a207e3c15e90dcd1a694af226bcfe159e1a27f5
[]
no_license
adkhn777/forms_example
3e604aaa3390d49d2b76c83c07ec5d1d837e56ed
56b78b770bdddbb984585c08f2d9c495345622b1
refs/heads/master
2021-01-24T11:22:11.893363
2016-10-07T07:06:47
2016-10-07T07:06:47
70,223,138
0
0
null
null
null
null
UTF-8
Python
false
false
178
py
from django import forms from .models import ModelFormModel class ModelFormForm(forms.ModelForm): class Meta: model = ModelFormModel fields = ["name", "bio"]
[ "adnan@tuple.tech" ]
adnan@tuple.tech
971bdaf13fb6d5bfbbcd7260417062a0c83db880
f84ecb8178baaad91914ca20212a67d22fdce293
/account/account/settings.py
229e6511004eb60dc7308bd414a52f6cd2a9a762
[]
no_license
TonyMistark/account_statement
33047972fcf2854a973e35a8aea56ec0e051b2a1
aeb66f1ca687e3afe4f0c308889929019209ec4e
refs/heads/master
2021-01-22T03:49:13.483873
2017-02-09T16:22:14
2017-02-09T16:22:14
81,460,715
2
1
null
null
null
null
UTF-8
Python
false
false
3,791
py
""" Django settings for account project. Generated by 'django-admin startproject' using Django 1.9. For more information on this file, see https://docs.djangoproject.com/en/1.9/topics/settings/ For the full list of settings and their values, see https://docs.djangoproject.com/en/1.9/ref/settings/ """ import os # Build paths inside the project like this: os.path.join(BASE_DIR, ...) BASE_DIR = os.path.dirname(os.path.dirname(os.path.abspath(__file__))) # Quick-start development settings - unsuitable for production # See https://docs.djangoproject.com/en/1.9/howto/deployment/checklist/ # SECURITY WARNING: keep the secret key used in production secret! SECRET_KEY = '8fb7nwrenc3tn4j1gcb@%ztui@2gti!*jpdeobe2ip&u36^q3+' # SECURITY WARNING: don't run with debug turned on in production! DEBUG = True ALLOWED_HOSTS = [] # Application definition INSTALLED_APPS = [ 'django.contrib.admin', 'django.contrib.auth', 'django.contrib.contenttypes', 'django.contrib.sessions', 'django.contrib.messages', 'django.contrib.staticfiles', "rest_framework", "account", ] REST_FRAMEWORK = { 'DEFAULT_PERMISSION_CLASSES': [ 'rest_framework.permissions.IsAdminUser', ], 'PAGE_SIZE': 10 } MIDDLEWARE_CLASSES = [ 'django.middleware.security.SecurityMiddleware', 'django.contrib.sessions.middleware.SessionMiddleware', 'django.middleware.common.CommonMiddleware', 'django.middleware.csrf.CsrfViewMiddleware', 'django.contrib.auth.middleware.AuthenticationMiddleware', 'django.contrib.auth.middleware.SessionAuthenticationMiddleware', 'django.contrib.messages.middleware.MessageMiddleware', 'django.middleware.clickjacking.XFrameOptionsMiddleware', ] ROOT_URLCONF = 'account.urls' TEMPLATES = [ { 'BACKEND': 'django.template.backends.django.DjangoTemplates', 'DIRS': [], 'APP_DIRS': True, 'OPTIONS': { 'context_processors': [ 'django.template.context_processors.debug', 'django.template.context_processors.request', 'django.contrib.auth.context_processors.auth', 'django.contrib.messages.context_processors.messages', ], }, }, ] WSGI_APPLICATION = 'account.wsgi.application' # Database # https://docs.djangoproject.com/en/1.9/ref/settings/#databases DATABASES = { 'default': { 'ENGINE': 'django.db.backends.mysql', # Add 'postgresql_psycopg2', 'mysql', 'sqlite3' or 'oracle'. 'NAME': "account", # Or path to database file if using sqlite3. 'USER': 'root', # Not used with sqlite3. 'PASSWORD': 'root', # Not used with sqlite3. 'HOST': 'localhost', # Set to empty string for localhost. Not used with sqlite3. 'PORT': '3306', # Set to empty string for default. Not used with sqlite3. } } # Password validation # https://docs.djangoproject.com/en/1.9/ref/settings/#auth-password-validators AUTH_PASSWORD_VALIDATORS = [ { 'NAME': 'django.contrib.auth.password_validation.UserAttributeSimilarityValidator', }, { 'NAME': 'django.contrib.auth.password_validation.MinimumLengthValidator', }, { 'NAME': 'django.contrib.auth.password_validation.CommonPasswordValidator', }, { 'NAME': 'django.contrib.auth.password_validation.NumericPasswordValidator', }, ] # Internationalization # https://docs.djangoproject.com/en/1.9/topics/i18n/ LANGUAGE_CODE = 'en-us' TIME_ZONE = 'UTC' USE_I18N = True USE_L10N = True USE_TZ = True # Static files (CSS, JavaScript, Images) # https://docs.djangoproject.com/en/1.9/howto/static-files/ STATIC_URL = '/static/'
[ "tony_mistark@163.com" ]
tony_mistark@163.com
efe1f8c522d049fcbb235a250a4ab33ac544503a
47299b9cca902b847371fa78eacbdaae0bae25f4
/webclone/one/urls.py
d67a6f7c91c658bc0feb139b85ba6826aba95130
[]
no_license
kapoorarpit/web_clone-
5c293fc2c10568562fd3c678e2fc8e43bc436b93
3540255fd6336583a9409c51deff0eae92810ee8
refs/heads/master
2023-06-11T04:51:07.138328
2021-06-29T19:50:52
2021-06-29T19:50:52
321,369,324
1
0
null
null
null
null
UTF-8
Python
false
false
194
py
from django.contrib import admin from django.urls import path from .import views urlpatterns = [ path('', views.home, name='home'), path('home/search/', views.search, name='search'), ]
[ "kapoorarpit2000@gmail.com" ]
kapoorarpit2000@gmail.com
7eced97eac47dfd2ce21cee31fe289634f7a5bf7
eac6dc8eb8e5f088500f425a7323cd35a4f99bd6
/src/courses/migrations/0012_course_active.py
af89db3155df4d47be9b84b4c843f0b847c617a6
[]
no_license
aminhp93/django_serverup_2
a14195af756799795282028ba611dbccc3848870
aef31722e882367c731e9e48fc8af8740befc112
refs/heads/master
2020-05-27T01:54:15.268661
2017-02-25T21:58:36
2017-02-25T21:58:36
82,514,017
1
0
null
null
null
null
UTF-8
Python
false
false
451
py
# -*- coding: utf-8 -*- # Generated by Django 1.10.5 on 2017-02-19 18:06 from __future__ import unicode_literals from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('courses', '0011_auto_20170219_1749'), ] operations = [ migrations.AddField( model_name='course', name='active', field=models.BooleanField(default=True), ), ]
[ "minhpn.org.ec@gmail.com" ]
minhpn.org.ec@gmail.com
9ced689eedce285664f086c784d82698560785ff
324ae3f56926da4169598897b4b9f06053d0a78d
/src/satisfy/nonogram.py
a0d242026fbebebb46a473f94765a8a6f23c6589
[ "Apache-2.0" ]
permissive
simone-campagna/satisfy
654152b410f81d0aaa672a8f3629aecce5924498
b5327e937e32c5324c05f6288f59cfaac4a316dc
refs/heads/master
2023-06-29T04:46:00.051479
2019-06-29T17:20:35
2019-06-29T17:20:35
157,007,810
0
0
null
null
null
null
UTF-8
Python
false
false
6,321
py
import collections from .solver import ModelSolver, VarSelectionPolicy __all__ = [ 'NonogramSolver', 'pixmap_shape', 'pixmap_to_nonogram', ] VarInfo = collections.namedtuple('VarInfo', 'size start_value end_value') class NonogramSolver(ModelSolver): def __init__(self, nonogram, **args): if args.get('var_selection_policy', None) is None: args['var_selection_policy'] = VarSelectionPolicy.MIN_BOUND super().__init__(**args) model = self._model rows = nonogram['rows'] cols = nonogram['columns'] num_rows = len(rows) num_cols = len(cols) var_infos = {} # add row vars and constraints: row_vars = {r: [] for r in range(num_rows)} for r, row in enumerate(rows): cur_vars = row_vars[r] if row: start = 0 rem_size = sum(row) + len(row) - 1 for k, size in enumerate(row): offset = size + int(k != len(row) - 1) end = num_cols - rem_size + 1 domain = list(range(start, end)) var = model.add_int_variable(name='r{}_{}'.format(r, k), domain=domain) var_infos[var.name] = VarInfo(size=size, start_value=start, end_value=end + size) # model.add_constraint(var + size <= num_cols) # TODO diff SERVE??? start += offset rem_size -= offset if cur_vars: prev_var = cur_vars[-1] constraint = var > prev_var + var_infos[prev_var.name].size model.add_constraint(constraint) cur_vars.append(var) # add col vars and constraints: col_vars = {c: [] for c in range(num_cols)} for c, col in enumerate(cols): cur_vars = col_vars[c] if col: start = 0 rem_size = sum(col) + len(col) - 1 for k, size in enumerate(col): offset = size + int(k != len(col) - 1) end = num_rows - rem_size + 1 domain = list(range(start, end)) var = model.add_int_variable(name='c{}_{}'.format(c, k), domain=domain) var_infos[var.name] = VarInfo(size=size, start_value=start, end_value=end + size) # model.add_constraint(var + size <= num_rows) # TODO diff SERVE??? start += offset rem_size -= offset if cur_vars: prev_var = cur_vars[-1] constraint = var > prev_var + var_infos[prev_var.name].size model.add_constraint(constraint) cur_vars.append(var) # add row<>col constraints: for r in range(num_rows): for c in range(num_cols): r_expr_list = [] for var in row_vars[r]: size = var_infos[var.name].size var_info = var_infos[var.name] if var_info.start_value <= c < var_info.end_value: r_expr_list.append((var <= c) & (c < var + size)) # else: # print("r: {}: discard {} ({})".format(var.name, c, var_info), model.get_var_domain(var)) c_expr_list = [] for var in col_vars[c]: size = var_infos[var.name].size var_info = var_infos[var.name] if var_info.start_value <= r < var_info.end_value: c_expr_list.append((var <= r) & (r < var + size)) # else: # print("c: {}: discard {} ({})".format(var.name, r, var_info), model.get_var_domain(var)) if r_expr_list or c_expr_list: if r_expr_list: r_expr = sum(r_expr_list) else: r_expr = 0 if c_expr_list: c_expr = sum(c_expr_list) else: c_expr = 0 constraint = (sum(r_expr_list) == sum(c_expr_list)) model.add_constraint(constraint) # instance attributes: self._var_infos = var_infos self._shape = (num_rows, num_cols) self._row_vars = row_vars self._col_vars = col_vars @property def source(self): return self._source @property def expr(self): return self._expr def __iter__(self): model = self._model solver = self._solver num_rows, num_cols = self._shape var_infos = self._var_infos row_vars = self._row_vars for solution in solver.solve(model): pixmap = [[0 for _ in range(num_cols)] for _ in range(num_rows)] for r, cur_vars in row_vars.items(): for var in cur_vars: start = solution[var.name] size = var_infos[var.name].size for c in range(start, start + size): pixmap[r][c] = 1 yield pixmap def pixmap_shape(pixmap): num_rows = len(pixmap) if pixmap: num_cols = max(len(row) for row in pixmap) else: num_cols = 0 return num_rows, num_cols def pixmap_to_nonogram(pixmap): num_rows, num_cols = pixmap_shape(pixmap) rows = [] for r, pixmap_row in enumerate(pixmap): row = [] count = 0 for c, cell in enumerate(pixmap_row): if cell: count += 1 else: if count: row.append(count) count = 0 if count: row.append(count) rows.append(row) cols = [] for c in range(num_cols): col = [] count = 0 for r in range(num_rows): cell = pixmap[r][c] if cell: count += 1 else: if count: col.append(count) count = 0 if count: col.append(count) cols.append(col) return {'rows': rows, 'columns': cols}
[ "simone.campagna11@gmail.com" ]
simone.campagna11@gmail.com
72b1bd0b8b29a08c14a6a75b7ceb058d86883236
39d100d1ed768ab4bdc768dc70e68d4bf943f233
/tgmate/views/__init__.py
a21ee0f14e8e4e8b7a1506789e34fefb9805171c
[]
no_license
ivan-koryshkin/tgmate
702b5c465a3435be134d858cc5fbd0f5ca8fd1f3
7ae1f5125ac19f00c53d557c70dbbdbe99886cac
refs/heads/master
2023-08-30T09:20:04.947011
2021-11-09T13:21:17
2021-11-09T13:21:17
null
0
0
null
null
null
null
UTF-8
Python
false
false
231
py
from .admin_view import UserModelView from .admin_view import TgUserView from .admin_view import MessageView from .admin_view import ChatView __all__ = [ 'UserModelView', 'TgUserView', 'MessageView', 'ChatView' ]
[ "ivan.koryshkin@gmail.com" ]
ivan.koryshkin@gmail.com
472f3e9fe76c99a8fc0b7f48fea8176b6b9b582e
5f9ec375125dae625b5fe169b6f3f836a2431dd1
/core/logger_helper.py
9ce5bc46fbbe8b6e97a9ed7da18f446afd2fbc52
[]
no_license
mumudd/python_weixin
e280b6bdc81f30365b1bb0e4700d9a00e6b99037
144dbedc72c010beae0d243001b82b9f687d0a1f
refs/heads/master
2021-06-23T20:14:36.237386
2020-12-09T08:37:51
2020-12-09T08:37:51
160,120,308
1
0
null
null
null
null
UTF-8
Python
false
false
1,915
py
import logging from logging import Logger from logging.handlers import TimedRotatingFileHandler '''日志管理类''' def init_logger(logger_name): if logger_name not in Logger.manager.loggerDict: logger1 = logging.getLogger(logger_name) logger1.setLevel(logging.INFO) # 设置最低级别 # logger1.setLevel(logging.DEBUG) # 设置最低级别 df = '%Y-%m-%d %H:%M:%S' format_str = '[%(asctime)s]: %(name)s %(levelname)s %(lineno)s %(message)s' formatter = logging.Formatter(format_str, df) # handler all try: handler1 = TimedRotatingFileHandler('/usr/web_wx/log/all.log', when='D', interval=1, backupCount=7) except Exception: handler1 = TimedRotatingFileHandler('F:\program\web_wx\core\log\/all.log', when='D', interval=1, backupCount=7) handler1.setFormatter(formatter) handler1.setLevel(logging.DEBUG) logger1.addHandler(handler1) # handler error try: handler2 = TimedRotatingFileHandler('/usr/web_wx/log/error.log', when='D', interval=1, backupCount=7) except Exception: handler2 = TimedRotatingFileHandler('F:\program\web_wx\core\log\error.log', when='D', interval=1, backupCount=7) handler2.setFormatter(formatter) handler2.setLevel(logging.ERROR) logger1.addHandler(handler2) # console console = logging.StreamHandler() console.setLevel(logging.DEBUG) # 设置日志打印格式 console.setFormatter(formatter) # 将定义好的console日志handler添加到root logger logger1.addHandler(console) logger1 = logging.getLogger(logger_name) return logger1 logger = init_logger('runtime-log') if __name__ == '__main__': logger.debug('test-debug') logger.info('test-info') logger.warn('test-warn') logger.error('test-error')
[ "sufaith@foxmail.com" ]
sufaith@foxmail.com
b2bed29df5eede8d6e01cc0c3ae685153dd0d69a
84750e22e48440a292c305dbd2ae75c4a210e934
/exspider/utils/view_funcs.py
736f3db21f558507697e89f6966b94cbba6307eb
[]
no_license
stonegithubs/exspider
617800a13ec9a1ca9c723d517766d00dcedd36a1
42b5cb0415c90dd60bc2c009a6aa467e71823854
refs/heads/master
2022-08-02T10:02:55.566046
2020-04-29T14:54:58
2020-04-29T14:54:58
null
0
0
null
null
null
null
UTF-8
Python
false
false
1,400
py
#! /usr/bin/python # -*- coding:utf-8 -*- # @zhuchen : 2019-03-06 15:18 import time from django.conf import settings from rest_framework.response import Response # 成功 success_code = { 1: "{}" } # 用户错误 user_error_code = { 1001: "登录失败", 1002: "重复请求验证码", 1003: "验证码错误", 1004: "您已经登录", 1005: "需要登录才能操作", 1006: "验证码过期", 1007: "稍后再试", 1008: "{}" } # 系统错误 http_error_code = { 9001: "必传参数[{}]错误", 9002: "[{}]参数错误", 9003: "[{}]格式错误", 9004: "自定义错误", 9005: "数据不存在", 9006: "数据添加失败,{}", 9007: "数据保存失败", 9008: "{}" # 自定义错误,客户端展示 } def http_response(http_code, http_msg=None, data=None, **kwargs): resp = settings.RESPONSE_FORMAT.copy() resp['code'] = http_code if http_code in user_error_code: resp['message'] = user_error_code[http_code] elif http_code in http_error_code: resp['message'] = http_error_code[http_code] else: resp['message'] = success_code[http_code] if http_msg is not None: resp['message'] = resp['message'].format(http_msg) if data is not None: resp['data'] = data resp['server_time'] = int(time.time()) resp.update(kwargs) return Response(resp)
[ "chen.zhu@blacktail.tech" ]
chen.zhu@blacktail.tech
7637d837e8cb0ba7f81221d92b23e7c92de9f971
f925eae9b78d44f9aa56cff17ef07aab87346a18
/stats/plot_all_roles.py
7e0f598a1128f40a4935de1b80120f39d0da950a
[ "MIT" ]
permissive
wejradford/castminer
f05d965e514c236657142c4db15a5c42db5160d1
6b792ba59621e7d0925d4ed683a51946c5193f3c
refs/heads/master
2020-12-24T08:55:13.205547
2016-09-06T12:46:15
2016-09-06T12:46:15
31,730,686
0
0
null
null
null
null
UTF-8
Python
false
false
705
py
#!/usr/bin/env python from __future__ import print_function import logging import os from utils import argparser_factory, db_factory, plot_role_counts, \ get_all_role_counts, FIGS log = logging.getLogger() p = argparser_factory() p.add_argument('-w', '--window', default=5, type=int) args = p.parse_args() db = db_factory(args.db) c = db.cursor() counts, total = get_all_role_counts(c) log.info('Collected {} data points'.format(total)) fname = os.path.join(FIGS, 'counts.rm-{}.pdf'.format(args.window)) plot_role_counts(counts, 'counts', fname, window=args.window, height=args.height, width=args.width, font_size=args.font_size)
[ "wejradford@gmail.com" ]
wejradford@gmail.com
ba63f7efdf10aab9c7481c9a2bee33143ac12df2
2037235643046608bf883f11c1bc448e2df8a4a3
/HuaYing/practice/test14.py
a18f331036c28c57f36f4079f83d4f9d3c4a6650
[]
no_license
Hardworking-tester/HuaYingAutoTest
7e46dfb0729961cee0da06762fc0be11724ad80b
c1f0cf7aa4433f482bbae88d1a5637b9859359ca
refs/heads/master
2021-01-10T18:38:37.788736
2015-09-05T10:37:10
2015-09-05T10:37:10
41,957,309
0
0
null
null
null
null
UTF-8
Python
false
false
1,368
py
#encoding:utf-8 from selenium import webdriver from selenium.webdriver.common.action_chains import ActionChains br=webdriver.Firefox() # br.maximize_window() br.get("http://www.xebest.com:8000") elements=br.find_elements_by_class_name("nav-arrow") element1=elements[4] if element1.is_displayed(): print ("网站导航链接已定位到") else: print ("网站导航元素未找到,请更换定位方式后重新定位") # if br.find_element_by_xpath("//*[@id='topnav']/ul/li[5]/div[2]/ul[2]/li[2]/a").is_displayed(): # if br.find_element_by_css_selector("div#topnav>ul:first>li:nth(4)>div:nth(1)>ul:nth(1)>li(1)>a").is_displayed(): # if br.find_element_by_css_selector("li#all_menu>ul:nth(0)>li:nth(0)>a>span").is_displayed(): # if br.find_element_by_link_text(u"易支付").is_displayed(): # print ("易支付元素已找到") # else: # print("易支付元素未找到,请更换定位方式后重新定位") # epay=br.find_element_by_css_selector("div#topnav>ul>li:nth(4)>div:nht(1)>ul:nth(1)>li(1)>a") # epay=br.find_element_by_xpath("//*[@id='topnav']/ul/li[5]/div[2]/ul[2]/li[2]/a") # epay=br.find_element_by_xpath("//*[@id='topnav']/ul/li[5]/div[2]/ul[2]/li[2]/a") epay=br.find_element_by_link_text(u"易支付") ActionChains(br).move_to_element(element1).click(element1).perform() ActionChains(br).move_to_element(epay).click(epay).perform()
[ "373391120@qq.com" ]
373391120@qq.com
eaeef1d5a47d3ff5621d988c694458cf63dc39a6
ceab178d446c4ab55951c3d65d99815e9fdee43a
/archive/coding_practice/python/ticks_plot.py
83e7d35370f009514aa95366b78a92f4f61f0afa
[]
no_license
DeneBowdalo/AtmosChem_Tools
01ecedb0df5c5d6e01966a0c3d8055826f5ac447
220c2f697a4f4c1e5443c336ede923b2004fe9f5
refs/heads/master
2021-01-10T18:05:30.800218
2017-02-06T16:08:14
2017-02-06T16:08:14
43,529,442
0
0
null
null
null
null
UTF-8
Python
false
false
183
py
import matplotlib.pyplot as plt x = [5,3,7,2,4,1,11,25,33] plt.plot(x) plt.xticks(range(len(x)), ['a', 'b', 'c', 'd', 'e', 'f', 'g', 'h', 'i']); plt.yticks(range(1,36,2)); plt.show()
[ "db876@earth0.york.ac.uk" ]
db876@earth0.york.ac.uk
d0b7766854f6f8576dbddc9fb4645f233cca2c41
2f3999daf2a359f6677835718958ca6c6e0e4a6a
/example1.py
7a6e8c0a528cd327d451c3a3f6724d5cb11f6fac
[]
no_license
Sangeetha-Naresh/class97
d9402203a5804ecd24d51e4eb6eff2cb8b4802ec
6d36c52368bcc9dd47bf011c48768b5358b4e5c5
refs/heads/main
2023-05-06T22:37:51.255175
2021-05-16T14:59:01
2021-05-16T14:59:01
367,911,176
0
0
null
null
null
null
UTF-8
Python
false
false
142
py
age= int(input("enter your age:")) if age>18: print("you are an adult") elif age >12: print("teenager") else: print("kiddo")
[ "noreply@github.com" ]
noreply@github.com
911507e11b9253bb23a002ed90852dba054ea2f8
9e22cd10e0e89872146b2ced45a8fcff29ae30d2
/module_integration/raspberrypi/manage.py
50d7c99e7444d50ab732838ce32b87b4252cd5e2
[]
no_license
Supriya-Suresh/eYSIP-2017_Vegetable-Identification-Using-Transfer-Learning
ca19833e5a2208252bfcf33515fd7ea0a3480c6d
8c570408b4394789840660fa9123caea8e634f6c
refs/heads/master
2022-11-23T22:59:40.466739
2017-07-08T06:20:49
2017-07-08T06:20:49
107,849,691
0
0
null
2022-11-22T01:06:25
2017-10-22T08:34:22
Python
UTF-8
Python
false
false
10,977
py
import loadcell as lc #import load cell library import RPi.GPIO as GPIO import lcd #import lcd library import kpad #import keypad library import time import os import math import datetime import sys import json #address constant for lines in lcd display LINE_1 = 0x80 LINE_2 = 0xC0 LINE_3 = 0x94 LINE_4 = 0xD4 baseValue = 0 #variable to store the base value of load cell taredWeight = 0 #variable to store tared weight imgName = "" #variable to store image name measuredWeight = 0.0 #variable to store calculated weight DOUT = 22 #constant stores gpio pin used by dout pin of hx711. It will be used to check if hx711 is ready to send data or not troughID = "" #variable to sotre trough ID locationID = "" #variable to sotre location ID cropID = "" #variable to store crop ID cropName = "" #variable to store crop name locationName = "" #variable to store location name #Flag variables troughIDExcepted = 0 #to check if trough id is accepted or not locationIDExcepted = 0 #to check if location id is accepted or not cropIDExcepted = 0 #to check if crop id is accepted or not pictureTaken = 0 #to check if picture is taken or not active = 1 #to check if program is active or not #initialize lcd lcd.lcd_init() ''' * * Function Name: calculateWeight * Input: none * Output: returns the calculated weight from the load cell value * Logic: 1) take the reading from load cell * 2) take the difference between current value and base value * 3) divide the difference with diference got with known weight * 4) finally multiply the division answer with known weight value to get the weight * Example Call: calculateWeight() * ''' def caculateWeight(): global taredWeight global measuredWeight global baseValue val = lc.read_cell_value() #read load cell value weight = ((baseValue - val) / 49000.0) * 230.0 #convert them into weight weight = weight - taredWeight #remove tared weight from calculated weight if weight < 0: #if weight becomes negative then set it back to zero weight = 0 weight = int(weight) measuredWeight = weight #store weight into measuredWeight variable return measuredWeight #return the calculated weight ''' * * Function Name: displayWeight * Input: none * Output: none * Logic: it displays weight on the lcd screen by using calculateWeight function * Example Call: displayWeight() * ''' def displayWeight() : global measuredWeight lcd.string("Object weight is:", LINE_3) weight = caculateWeight() #get calculated weight from the calculateWeight function lcd.string(str(weight) + " grams", LINE_4) #display the weight on the lcd if measuredWeight < 10: lcd.string("Place your object on", LINE_1) lcd.string("the platform", LINE_2) else: lcd.string("Press * button to", LINE_1) lcd.string("continue.", LINE_2) ''' * * Function Name: tare * Input: none * Output: none * Logic: takes the current weight of the object and stores it in variable then it will be subtracted form current weight value * Example Call: tare() * ''' def tare(): global baseValue global taredWeight lcd.clear() lcd.string("Taring weight...", LINE_1) lcval = lc.read_average_value(10) diff = math.fabs(baseValue- lcval) taredWeight = (diff / 49000.0) * 230.0 #store the calculated weight in variable ''' * * Function Name: takePicture * Input: none * Output: none * Logic: takes picture using USB camera using fscam program * Example Call: takePicture() * ''' def takePicture(): global imgName global pictureTaken lcd.string("Taking picture...", LINE_2) if os.path.exists('/dev/video0'): #create image file name with current date imgName = "image-" + datetime.datetime.now().isoformat() + ".jpg" imgName = "/home/pi/ghfarm/images/%s" %imgName #capture image and save in images directory. if image file does not exists in folder then retake the image while os.path.isfile(imgName) == False: os.system("fswebcam -r 640x480 -S 10 --no-banner /home/pi/ghfarm/images/%s" %imgName) pictureTaken = 1 #if picture is successfully taken then set pictureTaken flag to 1 else: #if camera is not attached display error message lcd.clear() lcd.string(" FAILED", LINE_1) lcd.string("No camera attached", LINE_2) time.sleep(2) ###################################################################################################################### pictureTaken = 1 ''' * * Function Name: storeData * Input: none * Output: none * Logic: stores the data into local database * Example Call: storeData() * ''' def storeData(): global imgName lcd.string("Storing data...", LINE_3) f = open('/home/pi/ghfarm/details.txt','a') time = datetime.datetime.now().strftime('%Y-%m-%d %H:%M:%S') crops = {'weight':measuredWeight,'crop_id':cropID,'time': time, 'imagepath':imgName, 'troughid':troughID} crop_details = json.dumps(crops) f.write(crop_details +'\n') ''' * * Function Name: fetchCropInfo * Input: none * Output: none * Logic: fetches the crop name and id from local database * Example Call: fetchCropInfo() * ''' def fetchCropInfo(): global cropName global cropIDExcepted cropName = "Spinach" cropIDExcepted = 1 # "retrieves crop info through cropid info" # global cropID # global cropName # global cropIDExcepted # try: # lcd.clear() # lcd.string("Fetching crop info...", LINE_1) # #create instance of a database with host, username, password and database name # db = sqldb.connect("localhost", "root", "firebird", "maindb") # #create cursor object # cursor = db.cursor() # cid = int(cropID) #convert accepted crop id into integer # sql = "SELECT name FROM cropinfo WHERE id = %d" % (cid) # cursor.execute(sql) # data = cursor.fetchone() # #if there some crop exists with this id # if data > 0: # cropName = data[0] #then assigh cropname to variable # cropIDExcepted = 1 #set cropIDExcepted flag to one # #lcd.clear() # #lcd.string("Successfully fetched", LINE_1) # #lcd.string("crop information", LINE_2) # #time.sleep(0.5) # else: #if no crop exists with entered ID # lcd.clear() # lcd.string("Crop ID does not", LINE_1) # lcd.string("exists!", LINE_2) # lcd.string("Enter valid crop ID", LINE_3) # time.sleep(3) # except: #if database connection Fails # lcd.clear() # lcd.string(" FAILED", LINE_1) # lcd.string("Unable to connect to", LINE_2) # lcd.string("local database", LINE_3) # lcd.string("Try again later", LINE_4) # time.sleep(3) # db.close() ''' * * Function Name: acceptCropID * Input: none * Output: none * Logic: accepts crop ID from user using keypad * Example Call: acceptCropID() * ''' def acceptCropID(): global cropID lcd.clear() cropID = "" key = "" time.sleep(0.1) lcd.string("Enter Crop ID", LINE_1) lcd.string("Press * to continue", LINE_2) lcd.string("and # for backspace", LINE_3) #loop until some crop id is entered and * key is pressed. Following loop will run until valid crop id entered while key != "*": lcd.string(cropID, LINE_4) key = kpad.get_key() if key == '*': if len(cropID) <= 0: lcd.clear() lcd.string("Crop ID cant", LINE_1) lcd.string("be null", LINE_2) time.sleep(1) lcd.clear() lcd.string("Enter Crop ID", LINE_1) lcd.string("Press * to continue", LINE_2) lcd.string("and # for backspace", LINE_3) else: break elif key == '#': #for backspacing if len(cropID) > 0: cropID = cropID[:-1] elif key.isdigit(): cropID += key time.sleep(0.2) key = "" #after accepting crop ID fetch crop information from local database print("Calling fetchcrop") fetchCropInfo() ''' * * Function Name: fetchTroughInfo * Input: none * Output: none * Logic: fetches the trough name and id from local database * Example Call: fetchTroughInfo() * ''' def fetchTroughInfo(): global troughID global troughIDExcepted troughIDExcepted = 1 ''' * * Function Name: acceptTroughID * Input: none * Output: none * Logic: accepts trough ID from user using keypad * Example Call: acceptTroughID() * ''' def acceptTroughID(): global troughID global troughIDExcepted lcd.clear() troughID = "" key = "E" time.sleep(0.1) lcd.string("Enter Trough ID", LINE_1) lcd.string("Press * to continue", LINE_2) lcd.string("and # for backspace", LINE_3) #loop until some trough id is entered and * key is pressed. Following loop will only break when valid trough ID is entered while key != "*": lcd.string(troughID, LINE_4) key = kpad.get_key() if key == '*': if len(troughID) <= 0: lcd.clear() lcd.string("Trough ID can't", LINE_1) lcd.string("be null", LINE_2) time.sleep(1) lcd.clear() lcd.string("Enter Trough ID", LINE_1) lcd.string("Press * to continue", LINE_2) lcd.string("and # for backspace", LINE_3) else: break elif key == '#': #for backspacing if len(troughID) > 0: troughID = troughID[:-1] elif key.isdigit(): troughID += key time.sleep(0.1) key = "" #check if entered trough ID is valid or not by fetching it from local database fetchTroughInfo() ''' * * Function Name: init * Input: none * Output: none * Logic: calculates the baseValue of load cell and fetches the crop info from the server database * Example Call: init() * ''' def init(): print("Initialization") global baseValue lcd.string(" Welcome", LINE_1) lcd.string(" Remove any object", LINE_2) lcd.string(" from the platform.", LINE_3) time.sleep(2) lcd.clear() lcd.string(" Welcome", LINE_1) lcd.string(" Please wait...", LINE_2) baseValue = lc.base_value() try : init() print("Started System") lcd.string("Started System", LINE_1) troughIDExcepted = cropIDExcepted = pictureTaken = 0 key = "E" #while key pressed is not the * key while True: while key is not '*' : displayWeight() key = kpad.get_key() if key == 'D' : tare() elif key == 'A': lcd.clear() lcd.string(" System", LINE_2) lcd.string(" Shutting down...", LINE_3) active = 0 os.system("sudo poweroff") lcd.clear() break elif key == 'B': lcd.clear() lcd.string(" Script", LINE_2) lcd.string(" Restarting", LINE_3) lcd.string(" Please wait...", LINE_4) active = 0 GPIO.cleanup() sys.stdout.flush() os.execv(sys.executable, ['python'] + sys.argv) break elif key == 'C': lcd.clear() lcd.string(" System", LINE_2) lcd.string(" Restarting", LINE_3) lcd.string(" Please wait...", LINE_4) active = 0 os.system("sudo reboot") break acceptCropID() print("Accepted Crop") if cropIDExcepted: print("Calling acceptTrough") acceptTroughID() if troughIDExcepted: print("Calling acceptLocation") takePicture() if pictureTaken: print("Calling Store Data") storeData() key = "E" except KeyboardInterrupt: print("Interrupted by keyboard") finally: lcd.clear() time.sleep(1) GPIO.cleanup()
[ "techieboy.teknas@gmail.com" ]
techieboy.teknas@gmail.com
f42e13027d1095f41cb53f127f04978052b43ba3
8b07bc3844f23054abccae1d50e1bc5ede5943c3
/producto/migrations/0003_producto_disponible.py
0723fa7bfd1f6bfdb2e002ca5efd13b4510feb82
[]
no_license
HedleyPty/PythonAnywhere
9c5ba4913e7f8d188d1fb1a0f6f8a3aa4b96210d
324bdb325db2ecfe22521d82ee3fe6cb2acc927a
refs/heads/master
2021-09-14T08:53:48.957057
2018-05-10T19:02:40
2018-05-10T19:02:40
112,934,212
0
0
null
null
null
null
UTF-8
Python
false
false
458
py
# -*- coding: utf-8 -*- # Generated by Django 1.9.2 on 2016-02-22 08:59 from __future__ import unicode_literals from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('producto', '0002_auto_20160221_1425'), ] operations = [ migrations.AddField( model_name='producto', name='disponible', field=models.BooleanField(default=False), ), ]
[ "hedleypanama@gmail.com" ]
hedleypanama@gmail.com