blob_id
stringlengths
40
40
directory_id
stringlengths
40
40
path
stringlengths
2
616
content_id
stringlengths
40
40
detected_licenses
listlengths
0
69
license_type
stringclasses
2 values
repo_name
stringlengths
5
118
snapshot_id
stringlengths
40
40
revision_id
stringlengths
40
40
branch_name
stringlengths
4
63
visit_date
timestamp[us]
revision_date
timestamp[us]
committer_date
timestamp[us]
github_id
int64
2.91k
686M
star_events_count
int64
0
209k
fork_events_count
int64
0
110k
gha_license_id
stringclasses
23 values
gha_event_created_at
timestamp[us]
gha_created_at
timestamp[us]
gha_language
stringclasses
220 values
src_encoding
stringclasses
30 values
language
stringclasses
1 value
is_vendor
bool
2 classes
is_generated
bool
2 classes
length_bytes
int64
2
10.3M
extension
stringclasses
257 values
content
stringlengths
2
10.3M
authors
listlengths
1
1
author_id
stringlengths
0
212
4644ab0b675a97687d04eb2b5e216bb376972288
2e0c35b02706fb5fb4fa489812deb5135345fe74
/Mundo 2/aula 9.py
abdfe6d9ba4a260467a8cd7d696f62019b9b92c9
[]
no_license
damiati-a/CURSO-DE-PYTHON
86a6d063d34ac7c4835fc03cd96c3e708141556f
cc8973611c0175ad0a2cd9b302287eb49f213c47
refs/heads/main
2023-04-21T09:39:12.806428
2021-05-29T13:01:13
2021-05-29T13:01:13
371,973,352
0
0
null
null
null
null
UTF-8
Python
false
false
952
py
# Estruturas de repetição (for) ''' faço 'c' no intervalo(1,10) for 'c' in range(1,10) ''' ''' laço c no intervalo(0,3) passo pula passo pega ################################# for c in range(0,3) passo pula passo pega ''' ''' for c in range(0,3) if moeda: pega passo pula passo pega ''' # Parte Parática ''' for c in range(0, 30, 3): print(c) print('FIM') ''' ''' n = int(input('Digite um número: ')) for c in range(0, n+1): print(c) print('FIM') ''' ''' i = int(input('Inicio: ')) # qual numero começar f = int(input('Fim: ')) # qual numero terminar p = int(input('Passo: ')) # e de quanto em quanto pular for c in range(i, f + 1, p): print(c) print('fim') ''' for c in range(0, 3): n = int(input('DIgite um numero: ')) # vai solicitar que se digite um numero 3 vezes s = n + n print('O somatorio de tudo foi: {}'.format(s))
[ "noreply@github.com" ]
damiati-a.noreply@github.com
9a6be77d3f1ab6b5515bb83d0b6a6eee5e09b43b
eda7fbf7bbc0614e6fc448d2f6e3fd1918dadcbe
/new-api-tests/applications/create-surface-caps-from-centerlines/create_surface_caps.py
e61f1afbfba81befc17a2e58529183112bb6877e
[]
no_license
SimVascular/SimVascular-Tests
e97c136ad3bf3a7275d40c0323abca7817eb2eca
55018e1edcd070bce77ae5af4caf2105353d3697
refs/heads/master
2023-02-11T02:19:06.755815
2023-02-02T18:26:31
2023-02-02T18:26:31
42,211,398
2
10
null
2023-02-02T18:26:32
2015-09-10T00:06:14
Python
UTF-8
Python
false
false
3,496
py
#!/usr/bin/env python """This script is used to create an SV model from a closed segmentation surface. The """ import argparse import os import sys from centerlines import Centerlines from surface import Surface sys.path.insert(1, '../../graphics/') import graphics as gr def parse_args(): '''Parse command-line arguments. ''' parser = argparse.ArgumentParser() parser.add_argument("--clip-distance", type=float, default=0.0, help="The distance from the end of a centerline branch to clip a surface.") parser.add_argument("--clip-width-scale", type=float, default=1.0, help="The width multiplied by the centerline branch end radius to define the width of the box used to clip a surface.") parser.add_argument("--surface-file", required=True, help="Input surface (.vtp or .vtk) file.") parser.add_argument("--mesh-scale", type=float, default=1.0, help="The factor used to scale the fe volume meshing edge size. A larger scale creates a coarser mesh. The initial edge size is determined from the largest surface triangle.") parser.add_argument("--remesh-scale", type=float, default=1.0, help="The factor used to scale the surface remeshing edge size. A larger scale creates a coarser suface mesh. The initial edge size is determined from the largest surface triangle.") args = parser.parse_args() if len(sys.argv) == 1: parser.print_help() sys.exit(1) return args def main(): # Get command-line arguments. args = parse_args() ## Create renderer and graphics window. win_width = 500 win_height = 500 renderer, renderer_window = gr.init_graphics(win_width, win_height) ## Read in the segmentation surface. surface_file_name = args.surface_file surface = Surface(gr, renderer_window, renderer) surface.read(surface_file_name) gr_geom = gr.add_geometry(renderer, surface.geometry, color=[0.8, 0.8, 1.0]) surface.vtk_actor = gr_geom #gr_geom.GetProperty().SetOpacity(0.5) ## Create a Centerlines object used to clip the surface. centerlines = Centerlines() centerlines.graphics = gr centerlines.surface = surface centerlines.window = renderer_window centerlines.renderer = renderer centerlines.clip_distance = args.clip_distance centerlines.clip_width_scale = args.clip_width_scale centerlines.remesh_scale = args.remesh_scale centerlines.mesh_scale = args.mesh_scale print("---------- Alphanumeric Keys ----------") print("a - Compute model automatically for a three vessel surface with flat ends.") print("c - Compute centerlines.") print("m - Create a model from the surface and centerlines.") print("q - Quit") print("s - Select a centerline source point.") print("t - Select a centerline target point.") print("u - Undo the selection of a centerline source or target point.") ## Create a mouse interactor for selecting centerline points. picking_keys = ['s', 't'] event_table = { 'a': (surface.create_model_automatically, centerlines), 'c': (surface.compute_centerlines, surface), 'm': (centerlines.create_model, surface), 's': surface.add_centerlines_source_node, 't': surface.add_centerlines_target_node } interactor = gr.init_picking(renderer_window, renderer, surface.geometry, picking_keys, event_table) ## Display window. interactor.Start() if __name__ == '__main__': main()
[ "davep@stanford.edu" ]
davep@stanford.edu
fb6e690787905e1481dff64dee25695f22bf3eee
70058f6db76e6adfbb06ed6245a63637b83fe9ef
/opendart/config/__init__.py
e64e9265798874238fe8da1d312e841fe0ab8876
[ "MIT" ]
permissive
JehunYoo/opendart
9c3235d3904e4c43bc5702a409ea65bb97d35ee3
c88105baf85af57d006cc2404d192aaf9baf73cc
refs/heads/main
2023-06-16T01:53:05.654129
2021-07-07T07:17:20
2021-07-07T07:17:20
380,931,447
2
0
null
null
null
null
UTF-8
Python
false
false
36
py
from opendart.config.config import *
[ "percyu98@gmail.com" ]
percyu98@gmail.com
5bbc71c559fd0f6b93689846f8d9309ea7d173fc
da2e4fad2d906c045365b4caa69bbfb559aa64a9
/av.py
a9fbbfbe8480f36c67ff5851acbec8bfcfcc93ff
[]
no_license
AGprofessional/PythonTradingBot
ab9fefca54e5420f087e9158b25de2cec8ff0ade
ecf63abd558d3923e865c91156e577dcc597a7dd
refs/heads/main
2023-07-14T18:18:26.761859
2021-08-21T04:28:17
2021-08-21T04:28:17
398,464,973
0
0
null
null
null
null
UTF-8
Python
false
false
2,738
py
import pandas as pd import requests import io from glom import glom #this works #https://www.alphavantage.co/documentation/#fx-intraday #https://www.youtube.com/watch?v=d2kXmWzfS0w from alpha_vantage.foreignexchange import ForeignExchange from alpha_vantage.timeseries import TimeSeries from alpha_vantage.cryptocurrencies import CryptoCurrencies from alpha_vantage.techindicators import TechIndicators import time api_key = 'YVR9BGPFI9BA1KH9' #forex #fx = ForeignExchange(key=api_key, output_format='pandas') #FXdata, FXmeta_data = fx.get_currency_exchange_intraday(from_symbol='EUR', to_symbol='USD',interval='1min', outputsize='full') #crypto: def getCrypto(): url = 'https://www.alphavantage.co/query?function=CRYPTO_INTRADAY&symbol=ETH&market=USD&interval=1min&outputsize=compact&apikey={}'.format( api_key) r = requests.get(url) Cjson = r.json() CtimeSeries = Cjson['Time Series Crypto (1min)'] #print(CtimeSeries) #Cfirstkey=list(CtimeSeries.keys())[0] #Cfirstval=list(CtimeSeries.values())[0] #print(Cfirstkey, Cfirstval) Cnext=next(iter(CtimeSeries.items())) # print(type(Cnext)) CnextValue = Cnext[1] getCrypto.open = CnextValue['1. open'] print("---",getCrypto.open) high = CnextValue['2. high'] low = CnextValue['3. low'] close = CnextValue['4. close'] volume = CnextValue['5. volume'] #print("c next open-->", CnextOpen) #stocks def getStocks(): stock = TimeSeries(key=api_key, output_format='pandas') Sdata, Smeta_data = stock.get_intraday(symbol='MRIN', interval='1min', outputsize='full') currCandle = Sdata.head(1) high = currCandle['2. high'] low = currCandle['3. low'] close = currCandle['4. close'] open = currCandle['1. open'] def getData(): #getCrypto() #tech indicators tech = TechIndicators(key=api_key, output_format='pandas') TdataEMA, Tmeta_dataEMA = tech.get_ema(symbol='MRIN', interval='1min', time_period='15', series_type='close') currEMA = TdataEMA.tail(1) #print(currEMA) TdataRSI, Tmeta_dataRSI = tech.get_rsi(symbol='MRIN', interval='1min', time_period='14', series_type='close') currRSI = TdataRSI.tail(1) #print(TdataRSI) #print(currRSI) TdataMACD, Tmeta_dataMACD = tech.get_macd(symbol='MRIN', interval='1min', series_type='close', fastperiod='12', slowperiod='26', signalperiod='9') currMACD = TdataMACD.head(1) # print(TdataMACD) # print(currMACD) #getData() #print(data) #time.sleep(60) #minute=3 #while (minute>0): # data, meta_data = ts.get_currency_exchange_intraday(from_symbol='EUR', to_symbol='USD',interval='1min', outputsize='compact') # most_recent=data.iloc[0] # print(most_recent) # time.sleep(60) # minute=minute-1
[ "noreply@github.com" ]
AGprofessional.noreply@github.com
c45b5b7aec7cf94b501dec46b90a47078258aa52
70abec2e6d9b0c8effa27bdacb8bf9f3ae09c3ee
/api/opinion/apps.py
eb8c70b2f0878c1f1dcdd9cc04cb3e44a5c1ea09
[]
no_license
lufepama/django-infojobs-clone-backend
730a336c869c54873a952043241f7909c8be3dfe
79d41f7f78b0ce8d71ae39760366ef4008769ed9
refs/heads/master
2023-08-22T06:17:47.392803
2021-10-16T01:35:14
2021-10-16T01:35:14
null
0
0
null
null
null
null
UTF-8
Python
false
false
89
py
from django.apps import AppConfig class OpinionConfig(AppConfig): name = 'opinion'
[ "lufepama31@gmail.com" ]
lufepama31@gmail.com
fb9b993a15edad35bf61dd1c9b5756ab037434ca
246308810fe62bba6a9cedf1473e90311a934017
/camera/test_images/draw_image.py
9ea6011215f4d85446ab761129381e98c4208a33
[]
no_license
MDB22/MedExpress
6051b84a9fcf2b7371087d8a0afda3af599f6fc4
7c7dcd5f53ae9934b138bbe903726a050b6cedc0
refs/heads/master
2021-01-17T22:54:16.328942
2016-09-08T13:52:08
2016-09-08T13:52:08
47,587,845
1
0
null
null
null
null
UTF-8
Python
false
false
314
py
import cv2 import numpy as np from matplotlib import pyplot as plt #read an image img = cv2.imread('test_image.jpg') font = cv2.FONT_HERSHEY_SIMPLEX cv2.putText(img, 'OpenCV',(10,500), font,4,(0,0,0),2,cv2.LINE_AA) plt.imshow(img, cmap='gray',interpolation='bicubic') plt.xticks([]), plt.yticks([]) plt.show()
[ "nathanmurfey@hotmail.com" ]
nathanmurfey@hotmail.com
7d0eac6bc6769a63f609d726e612586ed47b6af8
e1ae535d8613aae44e8f9eaa4daf50c1e63665b7
/multimedia/south_migrations/0026_auto__chg_field_remotestorage_media.py
e4f8b05c6dae4836b6317150e40ea7eda035d2ed
[]
no_license
teury/django-multimedia
48b8fba9abc101286990b1306d85967bd197f08e
4ddd5e6d9f4f680e2f4f68cc3616ced8f0fc2a43
refs/heads/master
2021-01-16T20:50:24.573686
2015-04-23T21:22:38
2015-04-23T21:22:38
null
0
0
null
null
null
null
UTF-8
Python
false
false
6,388
py
# -*- coding: utf-8 -*- from south.utils import datetime_utils as datetime from south.db import db from south.v2 import SchemaMigration from django.db import models class Migration(SchemaMigration): def forwards(self, orm): # Changing field 'RemoteStorage.media' db.alter_column(u'multimedia_remotestorage', 'media_id', self.gf('django.db.models.fields.related.ForeignKey')(to=orm['multimedia.Media'], null=True, on_delete=models.SET_NULL)) def backwards(self, orm): # User chose to not deal with backwards NULL issues for 'RemoteStorage.media' raise RuntimeError("Cannot reverse this migration. 'RemoteStorage.media' and its values cannot be restored.") # The following code is provided here to aid in writing a correct migration # Changing field 'RemoteStorage.media' db.alter_column(u'multimedia_remotestorage', 'media_id', self.gf('django.db.models.fields.related.ForeignKey')(to=orm['multimedia.Media'])) models = { u'auth.group': { 'Meta': {'object_name': 'Group'}, u'id': ('django.db.models.fields.AutoField', [], {'primary_key': 'True'}), 'name': ('django.db.models.fields.CharField', [], {'unique': 'True', 'max_length': '80'}), 'permissions': ('django.db.models.fields.related.ManyToManyField', [], {'to': u"orm['auth.Permission']", 'symmetrical': 'False', 'blank': 'True'}) }, u'auth.permission': { 'Meta': {'ordering': "(u'content_type__app_label', u'content_type__model', u'codename')", 'unique_together': "((u'content_type', u'codename'),)", 'object_name': 'Permission'}, 'codename': ('django.db.models.fields.CharField', [], {'max_length': '100'}), 'content_type': ('django.db.models.fields.related.ForeignKey', [], {'to': u"orm['contenttypes.ContentType']"}), u'id': ('django.db.models.fields.AutoField', [], {'primary_key': 'True'}), 'name': ('django.db.models.fields.CharField', [], {'max_length': '50'}) }, u'auth.user': { 'Meta': {'object_name': 'User'}, 'date_joined': ('django.db.models.fields.DateTimeField', [], {'default': 'datetime.datetime.now'}), 'email': ('django.db.models.fields.EmailField', [], {'max_length': '75', 'blank': 'True'}), 'first_name': ('django.db.models.fields.CharField', [], {'max_length': '30', 'blank': 'True'}), 'groups': ('django.db.models.fields.related.ManyToManyField', [], {'symmetrical': 'False', 'related_name': "u'user_set'", 'blank': 'True', 'to': u"orm['auth.Group']"}), u'id': ('django.db.models.fields.AutoField', [], {'primary_key': 'True'}), 'is_active': ('django.db.models.fields.BooleanField', [], {'default': 'True'}), 'is_staff': ('django.db.models.fields.BooleanField', [], {'default': 'False'}), 'is_superuser': ('django.db.models.fields.BooleanField', [], {'default': 'False'}), 'last_login': ('django.db.models.fields.DateTimeField', [], {'default': 'datetime.datetime.now'}), 'last_name': ('django.db.models.fields.CharField', [], {'max_length': '30', 'blank': 'True'}), 'password': ('django.db.models.fields.CharField', [], {'max_length': '128'}), 'user_permissions': ('django.db.models.fields.related.ManyToManyField', [], {'symmetrical': 'False', 'related_name': "u'user_set'", 'blank': 'True', 'to': u"orm['auth.Permission']"}), 'username': ('django.db.models.fields.CharField', [], {'unique': 'True', 'max_length': '30'}) }, u'contenttypes.contenttype': { 'Meta': {'ordering': "('name',)", 'unique_together': "(('app_label', 'model'),)", 'object_name': 'ContentType', 'db_table': "'django_content_type'"}, 'app_label': ('django.db.models.fields.CharField', [], {'max_length': '100'}), u'id': ('django.db.models.fields.AutoField', [], {'primary_key': 'True'}), 'model': ('django.db.models.fields.CharField', [], {'max_length': '100'}), 'name': ('django.db.models.fields.CharField', [], {'max_length': '100'}) }, u'multimedia.encodeprofile': { 'Meta': {'object_name': 'EncodeProfile'}, 'command': ('django.db.models.fields.CharField', [], {'max_length': '1024'}), 'container': ('django.db.models.fields.CharField', [], {'max_length': '32'}), 'file_type': ('django.db.models.fields.CharField', [], {'max_length': '32'}), u'id': ('django.db.models.fields.AutoField', [], {'primary_key': 'True'}), 'name': ('django.db.models.fields.CharField', [], {'max_length': '255'}) }, u'multimedia.media': { 'Meta': {'ordering': "(u'-created',)", 'object_name': 'Media'}, 'created': ('django.db.models.fields.DateTimeField', [], {}), 'description': ('django.db.models.fields.TextField', [], {'blank': 'True'}), 'file': ('django.db.models.fields.files.FileField', [], {'max_length': '100'}), u'id': ('django.db.models.fields.AutoField', [], {'primary_key': 'True'}), 'modified': ('django.db.models.fields.DateTimeField', [], {}), 'owner': ('django.db.models.fields.related.ForeignKey', [], {'to': u"orm['auth.User']"}), 'profiles': ('django.db.models.fields.related.ManyToManyField', [], {'to': u"orm['multimedia.EncodeProfile']", 'symmetrical': 'False'}), 'title': ('django.db.models.fields.CharField', [], {'max_length': '255'}) }, u'multimedia.remotestorage': { 'Meta': {'object_name': 'RemoteStorage'}, 'content_hash': ('django.db.models.fields.CharField', [], {'max_length': '64'}), 'created': ('django.db.models.fields.DateTimeField', [], {}), u'id': ('django.db.models.fields.AutoField', [], {'primary_key': 'True'}), 'media': ('django.db.models.fields.related.ForeignKey', [], {'to': u"orm['multimedia.Media']", 'null': 'True', 'on_delete': 'models.SET_NULL', 'blank': 'True'}), 'modified': ('django.db.models.fields.DateTimeField', [], {}), 'profile': ('django.db.models.fields.related.ForeignKey', [], {'to': u"orm['multimedia.EncodeProfile']", 'on_delete': 'models.PROTECT'}) } } complete_apps = ['multimedia']
[ "jason.bittel@gmail.com" ]
jason.bittel@gmail.com
6eae886e1c4593ab201a1e978c02d9c0bad900f6
335e91f503ae34fa4bd267396d1faabdf105fb5c
/data/process_data.py
b59d7a65666846f734e395335df2999201ab8b28
[]
no_license
FRANZKAFKA13/Udacity-Data-Science-Proj-2
d5c6786419796acfca73c62da3500679a4e4e37f
b87d58e18ff06f3dbd0e1a5d6e647ed4c4ff364e
refs/heads/master
2023-03-03T09:33:38.619277
2021-02-14T19:50:38
2021-02-14T19:50:38
334,509,995
1
0
null
null
null
null
UTF-8
Python
false
false
3,363
py
import sys import pandas as pd from sqlalchemy import create_engine def extract_data(messages_filepath, categories_filepath): """ Extract step of the ETL pipeline. :param messages_filepath: The filepath of the message.csv file :param categories_filepath: The filepath of the categories.csv file :return: The two .csv files combined in a pandas dataframe """ messages = pd.read_csv(messages_filepath) categories = pd.read_csv(categories_filepath) print(' ' + str(messages.shape[0]) + ' messages extracted from .csv.') print(' ' + str(categories.shape[0]) + ' categories extracted from .csv.') df = messages.merge(categories, how='left', on='id') return df def transform_data(df): """ Transform step of the ETL pipeline. :param df: The dataframe created from the .csv files :return: Returns a cleaned dataframe with categories transformed to dummy variables """ category_df = df['categories'].str.split(pat=";", expand=True) categories_raw = category_df.loc[0] categories_clean = [x[:-2] for x in categories_raw] category_df.columns = categories_clean print(' ' + str(category_df.shape[1]) + ' categories transformed to dummy variables.') for category in category_df: category_df[category] = category_df[category].str[-1] category_df[category] = category_df[category].astype('int32') df = df.drop(labels=['categories'], axis=1) df = pd.concat([df, category_df], axis=1) duplicate_count = df.duplicated('id').sum() df = df.drop_duplicates('id') print(' ' + str(duplicate_count) + ' duplicates dropped.') faulty_message_count = (df['related'] == 2).sum() df = df[df['related'] != 2] print(' ' + str(faulty_message_count) + ' faulty messages dropped.') return df def load_data(df, database_filename): """ Extract step of the ETL pipeline, saving the cleaned dataset in a SQLite database. :param df: The cleaned dataframe :param database_filename: The target database for the load step of the ETL process :return: """ engine = create_engine('sqlite:///' + str(database_filename)) try: df.to_sql('messages_categorized', engine, index=False, if_exists='replace') print('Cleaned data saved to database!') except ValueError: print("Warning: Database error.") def main(): if len(sys.argv) == 4: messages_filepath, categories_filepath, database_filepath = sys.argv[1:] print('Extracting data...\n MESSAGES: {}\n CATEGORIES: {}' .format(messages_filepath, categories_filepath)) df = extract_data(messages_filepath, categories_filepath) print('Transforming data...') df = transform_data(df) print('Loading data...\n DATABASE: {}'.format(database_filepath)) load_data(df, database_filepath) else: print('Please provide the filepaths of the messages and categories ' 'datasets as the first and second argument respectively, as ' 'well as the filepath of the database to save the cleaned data ' 'to as the third argument. \n\nExample run command:\npython process_data.py disaster_messages.csv ' 'disaster_categories.csv disaster_response.db') if __name__ == '__main__': main()
[ "carstengranig@gmail.com" ]
carstengranig@gmail.com
08d5d500388fe07a98eca25fb7e0cc15ff850948
c9499ae048d31f9e37b143500728ca55d8a3e6e3
/MainPackages/RunDSS.py
5acbcadcea9370cdb1158d70179c69eb13c1866d
[]
no_license
fprystupczuk/HPET_PowerFlow_Model
c7eadf6e3a3942736ae3caff8ccc62f45429d9a6
671f5a3085094bf135753c9cda1edb511dc80f65
refs/heads/main
2023-04-26T21:14:30.700547
2021-05-25T11:15:49
2021-05-25T11:15:49
370,665,533
1
0
null
null
null
null
UTF-8
Python
false
false
13,402
py
#!/usr/bin/env python3 # -*- coding: utf-8 -*- import random from MainPackages.SolvePET import SolvePET from MainPackages.SolveHPET import SolveHPET from MainPackages.Extract_DSSmonitors_data import Extract_DSSmonitors_data def RunDSS(os,np,pd,math,Main_path,model,Network,Feeders,Vcontrol,PV_penetration,ElectricVars,Profiles_path,Main_Results_path,timeArray,DSSObj,Lines_set,Loads_LV,Load_bus,Load_phase,Bus_Vnom,MV_loads,TransformerRating,PowerRatingPET,FractionPowerHPET): DSSText = DSSObj.Text # to excecute OpenDSS text commands DSSCircuit = DSSObj.ActiveCircuit # Use it to access the elements in the circuit (e.g., capacitors, buses, etc.) DSSElement = DSSCircuit.ActiveCktElement DSSSolution = DSSCircuit.Solution DSSMonitors = DSSCircuit.Monitors """ DEMAND AND PV ALLOCATION PROFILES As written, the script randomly allocates demand and PV active power profiles to the network using the profiles provided in NetworkData/Profiles """ # MV Demand MV_Dem_data = {} # defined as dictionary for i in range(3): i_MV = 'MV_' + str(i+1) csv_data = pd.read_csv(Profiles_path + '/Load_profiles/Profile_MV_PZIP_QZIP.csv',index_col=0) MV_Dem_data[i_MV,'P_Profile'] = csv_data['P_' + str(i+1)] # Active power profile [W] MV_Dem_data[i_MV,'Q_Profile'] = csv_data['Q_' + str(i+1)] # Reactive power profile [var] MV_Dem_data[i_MV,'ZIP_P'] = csv_data[['P_Z','P_I','P_P']] MV_Dem_data[i_MV,'ZIP_Q'] = csv_data[['Q_Z','Q_I','Q_P']] MV_Dem_data[i_MV,'ZIP_P'].columns = ['Z','I','P'] MV_Dem_data[i_MV,'ZIP_Q'].columns = ['Z','I','P'] # LV Demand np.random.seed(0) # for repeatability House_Dem_data = {} # defined as dictionary h_profiles = np.random.randint(1,100,size=Loads_LV.size) # define which random profiles will be allocated to each customer for i_house in range(Loads_LV.size): csv_data = pd.read_csv(Profiles_path + '/Load_profiles/Profile_' + str(h_profiles[i_house]) + '_PZIP_QZIP.csv',index_col=0) House_Dem_data[Loads_LV[i_house],'P_Profile'] = csv_data['P'] # Active power profile [W] House_Dem_data[Loads_LV[i_house],'Q_Profile'] = csv_data['Q'] # Reactive power profile [var] - here, assumed a PF of House_Dem_data[Loads_LV[i_house],'ZIP_P'] = csv_data[['P_Z','P_I','P_P']] House_Dem_data[Loads_LV[i_house],'ZIP_Q'] = csv_data[['Q_Z','Q_I','Q_P']] House_Dem_data[Loads_LV[i_house],'ZIP_P'].columns = ['Z','I','P'] House_Dem_data[Loads_LV[i_house],'ZIP_Q'].columns = ['Z','I','P'] # PV generation np.random.seed(1) # for repeatability random.seed(1) PV_Gen_data = {} # defined as dictionary PV_set = np.array(random.sample(list(Loads_LV), int(np.rint(len(Loads_LV) * PV_penetration)))) # Random PV alocation PV_profiles = np.random.randint(1,100,size=PV_set.size) # Random selection of PV profiles from the pool PV_profiles_data = pd.read_excel(Profiles_path + '/PV_profiles/Summer_PV_Profiles.xlsx',header=0,index_col=0) # Read selected profiles (data in kW) PV_rating_data = pd.read_csv(Profiles_path + '/PV_profiles/PV_rating.csv') for i_PV in range(PV_set.size): PV_Gen_data[PV_set[i_PV],'Profile'] = PV_profiles_data[PV_profiles[i_PV]]*1000 # Active power output for this PV [W] PV_Gen_data[PV_set[i_PV],'Rating'] = PV_rating_data.loc[PV_profiles[i_PV],'Rating [kW]'] # PV rating in kW # Create PVs in OpenDSS for PV in PV_set: DSSText.Command = 'New Load.PV_' + PV + ' Phases=1 Bus1=' + Load_bus.loc[PV,'Bus'] + '.' + str(Load_phase.loc[PV,'phase']) + ' kV=' + str(0.4/math.sqrt(3)) + ' kW=0 Kvar=0 Model=1 Vminpu=0.7 Vmaxpu=1.3' """ RUN POWER FLOW AND OBTAIN RESULTS """ # OUTPUT DATA CONTAINERS Data = np.zeros((timeArray.size,ElectricVars.size,Lines_set.size,3)) ## Order of ElectricVars in Data array ## # 0 <- Vmag_send # 1 <- Vang_send # 2 <- Vmag_rec # 3 <- Vang_rec # 4 <- Imag # 5 <- P_rec # 6 <- Q_rec # 7 <- P_send # 8 <- Q_send # Name of variables to be collected for not per line but only for each time-step P_LV = np.zeros(timeArray.size) # P delivered by the transformer's secondary side (sum of phases) Q_LV = np.zeros(timeArray.size) # Q delivered by the transformer's secondary side (sum of phases) TotalLosses = np.zeros((timeArray.size,2)) # Line losses + Transformer losses LFTlosses = np.zeros(timeArray.size) # Line losses + Transformer losses PETlosses = np.zeros(timeArray.size) # Losses of the Power Electronic Module PETefficiency = np.zeros(timeArray.size) # Efficiency of the Power Electronic Module PET_Spu = np.zeros(timeArray.size) # Apparent power of the Power Electronic Module in pu PET_PF_LV = np.zeros(timeArray.size) # Power Factor at the transformer's secondary side PEM_Pc1_pu = np.zeros((timeArray.size,3)) # Active power of Module 1 of the HPET in pu PEM_Qc1_pu = np.zeros((timeArray.size,3)) # Reactive power of Module 1 of the HPET in pu PEM_Pc2_pu = np.zeros((timeArray.size,3)) # Active power of Module 2 of the HPET in pu PEM_Qc2_pu = np.zeros((timeArray.size,3)) # Reactive power of Module 2 of the HPET in pu HPETlosses = np.zeros((timeArray.size,2)) # HPET losses: PEM losses,Transformer losses PEMefficiency1 = np.zeros(timeArray.size) # Efficiency of Module 1 of the HPET PEMefficiency2 = np.zeros(timeArray.size) # Efficiency of Module 2 of the HPET PEM_Sc1_pu = np.zeros(timeArray.size) # Apparent power of Module 1 of the HPET in pu PEM_Sc2_pu = np.zeros(timeArray.size) # Apparent power of Module 2 of the HPET in pu PEM_PFc1 = np.zeros(timeArray.size) # Power Factor at Module 1 of the HPET PEM_PFc2 = np.zeros(timeArray.size) # Power Factor at Module 2 of the HPET # Path to store results fileName_start = Main_Results_path + '/DataResults/' + Network + '/' try: os.makedirs(fileName_start) except FileExistsError: pass # Run time series of Power Flow analysis for t in range(timeArray.size): # OpenDSS MV loads if MV_loads == 'y': for i in range(3): i_MV = 'MV_' + str(i+1) DSSCircuit.Loads.Name = i_MV DSSCircuit.Loads.kW = MV_Dem_data[i_MV,'P_Profile'].loc[timeArray[t]]/1000.0 DSSCircuit.Loads.kvar = MV_Dem_data[i_MV,'Q_Profile'].loc[timeArray[t]]/1000.0 Z_p = float(MV_Dem_data[i_MV,'ZIP_P'].loc[timeArray[t],'Z']) I_p = float(MV_Dem_data[i_MV,'ZIP_P'].loc[timeArray[t],'I']) Z_q = float(MV_Dem_data[i_MV,'ZIP_Q'].loc[timeArray[t],'Z']) I_q = float(MV_Dem_data[i_MV,'ZIP_Q'].loc[timeArray[t],'I']) DSSCircuit.Loads.ZIPV = (Z_p,I_p,1-Z_p-I_p,Z_q,I_q,1-Z_q-I_q,0.8) # Last coefficient: voltage in pu from wich the load model changes to constant impedance to facilitate convergency of OpenDSS # OpenDSS LV feeder house demand for i_house in Loads_LV: DSSCircuit.Loads.Name = i_house DSSCircuit.Loads.kW = House_Dem_data[i_house,'P_Profile'].loc[timeArray[t]]/1000.0 DSSCircuit.Loads.kvar = House_Dem_data[i_house,'Q_Profile'].loc[timeArray[t]]/1000.0 Z_p = float(House_Dem_data[i_house,'ZIP_P'].loc[timeArray[t],'Z']) I_p = float(House_Dem_data[i_house,'ZIP_P'].loc[timeArray[t],'I']) Z_q = float(House_Dem_data[i_house,'ZIP_Q'].loc[timeArray[t],'Z']) I_q = float(House_Dem_data[i_house,'ZIP_Q'].loc[timeArray[t],'I']) DSSCircuit.Loads.ZIPV = (Z_p,I_p,1-Z_p-I_p,Z_q,I_q,1-Z_q-I_q,0.8) # Last coefficient: voltage in pu from wich the load model changes to constant impedance to facilitate convergency of OpenDSS # OpenDSS PVs for PV in PV_set: DSSCircuit.Loads.Name = 'PV_' + PV DSSCircuit.Loads.kW = -1 * PV_Gen_data[PV,'Profile'].loc[timeArray[t]]/1000.0 DSSCircuit.Loads.kvar = -1 * 0.0 # Assumes PF=1 for all the PVs # Solve time step for different transformer models if model == 'LFT': # SOLVE DSSText.Command ='Reset Monitors' DSSSolution.Solve() if not(DSSSolution.Converged): raise ValueError('Solution did not Converge') DSSMonitors.SampleAll() DSSMonitors.SaveAll() DSSMonitors.Name = 'LFT-PQ' for phase in range(3): P_LV[t] = P_LV[t] - DSSMonitors.Channel(phase * 2 + 1)[0] Q_LV[t] = Q_LV[t] + abs(DSSMonitors.Channel(phase * 2 + 2)[0]) # Compute transformer losses and total losses DSSCircuit.Transformers.Name = 'TR1' LFTlosses[t] = DSSElement.Losses[0]/1000 # DSSElement.Losses returns Watts and VAr TotalLosses[t,:] = np.asarray(DSSCircuit.Losses)/1000 # DSSCircuit.Losses returns Watts and VAr if model == 'PET': # Regulate output voltage, compute losses and update active power in the primary side [Pp,Qp,P_LV[t],Q_LV[t],PETlosses[t],PETefficiency[t],PET_Spu[t],PET_PF_LV[t]] = SolvePET(Main_path,np,Vcontrol,DSSObj,Bus_Vnom,Loads_LV,Load_phase,TransformerRating) # Compute total losses and add power electronics losses TotalLosses[t,:] = np.asarray(DSSCircuit.Losses)/1000 # DSSCircuit.Losses returns Watts and VAr TotalLosses[t,0] = TotalLosses[t,0] + PETlosses[t] if model == 'HPET': # compute losses and update active power in the primary side [PEM_Pc1_pu[t,:],PEM_Qc1_pu[t,:],PEM_Pc2_pu[t,:],PEM_Qc2_pu[t,:],HPETlosses[t,0],PEMefficiency1[t],PEMefficiency2[t],PEM_Sc1_pu[t],PEM_Sc2_pu[t],PEM_PFc1[t],PEM_PFc2[t]] = SolveHPET(Main_path,t,np,math,Vcontrol,DSSObj,Loads_LV,Load_phase,TransformerRating,FractionPowerHPET) # Obtain losses at the LFT of the HPET for phase in range(3): DSSCircuit.Transformers.Name = 'Trafo' + str(phase+1) HPETlosses[t,1] = HPETlosses[t,1] + DSSElement.Losses[0]/1000 # HPET losses: PEM losses,Transformer losses # Compute total losses and add power electronics losses TotalLosses[t,:] = np.asarray(DSSCircuit.Losses)/1000 # DSSCircuit.Losses returns Watts and VAr TotalLosses[t,0] = TotalLosses[t,0] + HPETlosses[t,0] # HPET losses: PEM losses,Transformer losses # Obtains the total power delivered by the secondary winding of the transformer DSSMonitors.Name = 'Trafo_sec-PQ' for phase in range(3): P_LV[t] = P_LV[t] + DSSMonitors.Channel(phase * 2 + 1)[0] Q_LV[t] = Q_LV[t] + abs(DSSMonitors.Channel(phase * 2 + 2)[0]) # Store main results into a numpy array Data[t,:,:,:] = Extract_DSSmonitors_data(np,ElectricVars,DSSMonitors,Lines_set) # Define final paths to store results for each transformer model if model == 'LFT': fileName_end = model + '_PV' + str(PV_penetration) + '_' + '-'.join(Feeders) np.save(fileName_start + 'LFTlosses_' + fileName_end,LFTlosses) if model == 'PET': fileName_end = model + str(PowerRatingPET) + '_PV' + str(PV_penetration) + '_' + str(Vcontrol[0]) + str(Vcontrol[1][0]) + '_' + str(Vcontrol[1][1]) + '_' + '-'.join(Feeders) np.save(fileName_start + 'PETlosses_' + fileName_end,PETlosses) np.save(fileName_start + 'PETefficiency_' + fileName_end,PETefficiency) np.save(fileName_start + 'PET_Spu_' + fileName_end,PET_Spu) np.save(fileName_start + 'PET_PF_LV_' + fileName_end,PET_PF_LV) if model == 'HPET': fileName_end = model + str(FractionPowerHPET) + '_PV' + str(PV_penetration) + '_' + str(Vcontrol[0]) + str(Vcontrol[1][0]) + '_' + str(Vcontrol[1][1]) + '_' + '-'.join(Feeders) np.save(fileName_start + 'HPETlosses_' + fileName_end,HPETlosses) # HPET losses: PEM losses,Transformer losses np.save(fileName_start + 'PEM_Pc1_pu_' + fileName_end,PEM_Pc1_pu) np.save(fileName_start + 'PEM_Qc1_pu_' + fileName_end,PEM_Qc1_pu) np.save(fileName_start + 'PEM_Pc2_pu_' + fileName_end,PEM_Pc2_pu) np.save(fileName_start + 'PEM_Qc2_pu_' + fileName_end,PEM_Qc2_pu) np.save(fileName_start + 'PEMefficiency1_' + fileName_end,PEMefficiency1) np.save(fileName_start + 'PEMefficiency2_' + fileName_end,PEMefficiency2) np.save(fileName_start + 'PEM_Sc1_pu_' + fileName_end,PEM_Sc1_pu) np.save(fileName_start + 'PEM_Sc2_pu_' + fileName_end,PEM_Sc2_pu) np.save(fileName_start + 'PEM_PFc1_' + fileName_end,PEM_PFc1) np.save(fileName_start + 'PEM_PFc2_' + fileName_end,PEM_PFc2) # Save final results into the destination npy files if MV_loads == 'y': fileName_end = fileName_end + '_MVloads' np.save(fileName_start + 'Data_' + fileName_end,Data) np.save(fileName_start + 'P_LV_' + fileName_end,P_LV) np.save(fileName_start + 'Q_LV_' + fileName_end,Q_LV) np.save(fileName_start + 'TotalLosses_' + fileName_end,TotalLosses)
[ "nrvfederico@gmail.com" ]
nrvfederico@gmail.com
8cab5cb127fb2a4d84d2c87d1da7e4fe38a79492
dfb306fbc6f85f4ff66ab659911909cf4dc1e7c1
/180104.py
03d231ed48308b93ea26c7137615b101c0937874
[ "MIT" ]
permissive
majaldm/ML
d83e4577361e6d15725eb1d52caf8ead2f828de2
1f99f36c90b226acf151c3d65ed767f635a1469c
refs/heads/master
2021-05-05T21:34:11.524219
2018-01-04T07:17:18
2018-01-04T07:17:18
115,596,139
0
0
null
null
null
null
UTF-8
Python
false
false
2,271
py
# coding: utf-8 # In[14]: get_ipython().system(' pip install --user flask') # In[31]: from flask import Flask, request, make_response, jsonify app = Flask('flask-api') @app.route('/') def hello_world(): message = {'message': 'hello world'} return jsonify(message) from ptt_crawler import PttCrawler @app.route('/ptt_crawler', methods=['GET','POST']) def run_crawler(): if request.method == 'GET': crawler = PttCrawler('Gossiping', page=1) elif request.method == 'POST': board = request.get_json().get('board','Gossiping') page = request.get_json().get('page','1') crawler = PttCrawler(board, page=page) result = crawler.run() return jsonify(result) if __name__ == '__main__': app.run(port=8000) # In[37]: print(app) print('-------------------') print(vars(app)) print('-------------------') print(', '.join("%s: %s" % item for item in vars(app).items())) print('-------------------') print(app.__dict__) print('-------------------') print(dir(app)) # In[43]: print(app.test_client()) print('-------------------') print(vars(app.test_client())) print('-------------------') print(', '.join("%s: %s" % item for item in vars(app.test_client()).items())) print('-------------------') print(app.test_client().__dict__) print('-------------------') print(dir(app)) # In[45]: print(app.test_client().get('/')) print('-------------------') print(vars(app.test_client().get('/'))) print('-------------------') print(', '.join("%s: %s" % item for item in vars(app.test_client().get('/')).items())) print('-------------------') print(app.test_client().get('/').__dict__) print('-------------------') print(dir(app)) # In[13]: import json resp = app.test_client().get('/') print(resp.data) print(resp.data.decode()) print(json.loads(resp.data.decode())) # In[10]: from ptt_crawler import PttCrawler crawler = PttCrawler('Gossiping', page=1) result = crawler.run() print(result) # In[11]: resp = app.test_client().get('/ptt_crawler') print(resp.data) print(json.loads(resp.data.decode())) # In[32]: resp = app.test_client().post( '/ptt_crawler', data=json.dumps({'board':'Baseball','page':3}), content_type='application/json' ) print(json.loads(resp.data.decode()))
[ "noreply@github.com" ]
majaldm.noreply@github.com
b70442fbe888bee246a8aac51eb35d72971386bc
82dc886733bd7999d8189e03f1f7a6201541abaa
/_operators.py
5ceb1071dc7421d674d12265e9b33bfdd80f4ed5
[]
no_license
lucaskaimthelen/python_training_3
b78f4da101627ad2974538bc0926e65f82749e18
4339dd125b7b67d153f8b1f2eac966f4dde5d9bd
refs/heads/master
2020-07-30T12:09:06.007258
2019-09-22T23:37:20
2019-09-22T23:37:20
210,228,607
0
0
null
null
null
null
UTF-8
Python
false
false
612
py
# http://www.tutorialspoint.com/python/python_basic_operators.htm # Python operators. a = 2 b = 2 # Equality print(a == b) # Identity tests if the two objects in memory are the same. is and is not. a = [1, 2, 3] b = [1, 2, 3] print(id(a)) print(id(b)) print(a is b) # Set b object equal to a object. a = b print(id(a)) print(id(b)) print(a is b) a = 10 b = 2 # Not equal. print(a!= b) # Less than. print(a < b) # Less then or equal to. print(a <= b) a = 10 # In operator. a in [10, 11, 13] # Not in operator. a not in [10, 11, 13] # Add and. b = 2 b += 1 print(b) # Subtract AND b = 2 b -= 1 print(b)
[ "lucaskaimthelen@gmail.com" ]
lucaskaimthelen@gmail.com
d75546d084acf87c9f272a76978aafe55c63e533
880495f69099e6c1efbcfd8fb763a5de6df12786
/code/23-merge-k-sorted-lists.py
922f35eca621e5ff82fc343c587b6cbf43d40454
[]
no_license
shen-ee/LC
69209f474e1ec4265d45e072cb65105b2a28c621
8664bde56453dea18f069725470d0401e65d45e2
refs/heads/master
2022-11-07T20:27:24.029396
2019-01-16T08:14:47
2019-01-16T08:14:47
null
0
0
null
null
null
null
UTF-8
Python
false
false
1,097
py
import Queue # Definition for singly-linked list. # class ListNode(object): # def __init__(self, x): # self.val = x # self.next = None class Solution(object): def mergeKLists(self, lists): """ :type lists: List[ListNode] :rtype: ListNode """ # # Solution 1 # q = Queue.PriorityQueue() # root = ListNode(None) # curnode = root # for node in lists: # if node is None : # continue # q.put((node.val,node)) # while q.qsize() > 0: # curnode.next = q.get()[1] # curnode = curnode.next # if curnode.next: # q.put((curnode.next.val, curnode.next)) # return root.next # Solution 2 ans = [] for list in lists: while list: ans.append(list.val) list = list.next head = root = ListNode(None) for i in sorted(ans): head.next = ListNode(i) head = head.next return root.next
[ "jerry31@qq.com" ]
jerry31@qq.com
91aefd24b6e56d4907a756aa0de57697ad4bcc97
cead5aa67c0c45038917f306fefe31b53b5b639f
/temp/PyQt/Qthread.py
b5f82a26132421016342acaef9e4b384791399a7
[]
no_license
songaal/AutoLogin
0a79400a88f1823aebfb819c98b489aeb267a679
778f30dd92dc4dba3d0a632511113bfe145d1b94
refs/heads/master
2022-02-27T23:13:54.152894
2019-09-22T16:27:41
2019-09-22T16:27:41
null
0
0
null
null
null
null
UTF-8
Python
false
false
2,762
py
from PyQt5.QtCore import * from PyQt5.QtWidgets import * ''' https://freeprog.tistory.com/351 ''' class MyMainGUI(QDialog): def __init__(self, parent=None): super().__init__(parent) self.qtxt1 = QTextEdit(self) self.btn1 = QPushButton("Start", self) self.btn2 = QPushButton("Stop", self) self.btn3 = QPushButton("add 100", self) self.btn4 = QPushButton("send instance", self) vbox = QVBoxLayout() vbox.addWidget(self.qtxt1) vbox.addWidget(self.btn1) vbox.addWidget(self.btn2) vbox.addWidget(self.btn3) vbox.addWidget(self.btn4) self.setLayout(vbox) self.setGeometry(100,50,300,300) # self.show() class Test: def __init__(self): name = "" job = "" class MyMain(MyMainGUI): add_sec_signal = pyqtSignal() send_instance_signal = pyqtSignal("PyQt_PyObject") def __init__(self, parent=None): super().__init__(parent) self.btn1.clicked.connect(self.time_start) self.btn2.clicked.connect(self.time_stop) self.btn3.clicked.connect(self.add_sec) self.btn4.clicked.connect(self.send_instance) self.th = Worker(parent=self) self.th.sec_changed.connect(self.time_update) self.add_sec_signal.connect(self.th.add_sec) self.send_instance_signal.connect(self.th.recive_instance_signal) self.show() @pyqtSlot() def time_start(self): self.th.start() self.th.working = True @pyqtSlot() def time_stop(self): self.th.working = False @pyqtSlot() def add_sec(self): print(".... add signal emit ....") self.add_sec_signal.emit() @pyqtSlot(str) def time_update(self, msg): self.qtxt1.append(msg) @pyqtSlot() def send_instance(self): t1 = Test() t1.name = "SuperPower!!!" t1.job = "cta" self.send_instance_signal.emit(t1) class Worker(QThread): sec_changed = pyqtSignal(str) def __init__(self, sec=0, parent=None): super().__init__() self.main = parent self.working = True self.sec = sec def __del__(self): print('end thread.....') self.wait() def run(self): while self.working: self.sec_changed.emit('time (secs) : {}'.format(self.sec)) self.sleep(1) self.sec += 1 @pyqtSlot() def add_sec(self): print('add sec...') self.sec += 100 @pyqtSlot("PyQt_PyObject") def recive_instance_signal(self, inst): print(inst.name) print(inst.job) if __name__ == "__main__": import sys app = QApplication(sys.argv) form = MyMain() app.exec_()
[ "taxkmj@naver.com" ]
taxkmj@naver.com
00f9cf22aeaf7658d0bb303616a50711fe27b39a
f3e3f61d4cec65ad5b924c0564d99c541ccdb625
/src_4/chapter04/migrations/0001_initial.py
6e2758ee1268c74d81ccd745fea58bbf8c096da7
[]
no_license
uxx53/bur1
15b0a96b4c0afdcbd1b4e230237725f2f013b593
1b013f87792204bdbc179e678caa80616eb62279
refs/heads/master
2016-09-05T11:13:21.706850
2015-08-14T10:36:43
2015-08-14T10:36:43
40,514,067
0
0
null
null
null
null
UTF-8
Python
false
false
614
py
# -*- coding: utf-8 -*- from __future__ import unicode_literals from django.db import models, migrations class Migration(migrations.Migration): dependencies = [ ] operations = [ migrations.CreateModel( name='Author', fields=[ ('id', models.AutoField(verbose_name='ID', auto_created=True, serialize=False, primary_key=True)), ('name', models.CharField(max_length=100)), ('title', models.CharField(max_length=3)), ('birth_date', models.DateField(blank=True, null=True)), ], ), ]
[ "uxx@mail.ru" ]
uxx@mail.ru
2d337c23eb9d0b0e67728ffd5b44be81efb0d9cf
1f32a4a5d54e4690b68e8e3e92524535bb1ca8f0
/akai-to-wLightBox.py
adeb80dd9cac282e14f0e22fc0955abcfc30cc9e
[]
no_license
EdwardEisenhauer/AKAI-LPD8
afeb2d193141e8b2f8d72c1ed84cf3273aa17604
2dd6db4926d12e15b1227997d92cb4a1aa52f4d2
refs/heads/master
2023-01-14T09:23:04.085039
2020-03-27T15:31:42
2020-03-27T15:31:42
247,360,448
1
1
null
null
null
null
UTF-8
Python
false
false
2,665
py
from threading import Thread from queue import Queue from time import sleep from blebox import LightBox from akai import Akai import sys ip = sys.argv[1] pasek = LightBox(ip) akai = Akai() """ Read knob values and send the GET to set the Lights. Monitor the knobs values. If anyone changes send a GET to set the lights. Wait 100ms before next packet. TODO: - Detect incoming MIDI data at the mido level? - Do something about this ugly time.sleep(0.1) - Introduce an extrapolating method :return: """ def knob_to_ms(value): knob_min = 0 knob_max = 127 ms_min = 0 ms_max = 1000 return int(value/(knob_max-knob_min)*(ms_max-ms_min)) def knob_to_effect(value): knob_min = 0 knob_max = 127 effect_id_min = 0 effect_id_max = 6 return int(value/(knob_max-knob_min)*(effect_id_max-effect_id_min)) def producer(output_queue): [r, g, b, w] = [i * 2 for i in akai.knobs][:4] w_prev = w while True: # Oh my ficking god do something about this!!! if akai.knobs_color_change: [r, g, b, w] = [i * 2 for i in akai.knobs][:4] # Get RGB values from the AKAI if w == w_prev: output_queue.put([r, g, b]) else: output_queue.put([w, w, w]) w_prev = w akai.knobs_color_change = False elif akai.knobs_durations_change: [color_fade, effect_fade, effect_step] = map(knob_to_ms, akai.knobs[4:7]) effect_id = knob_to_effect(akai.knobs[7]) pasek.color_fade = color_fade pasek.effect_fade = effect_fade pasek.effect_step = effect_step pasek.set_durations() # It shouldn't be like this XD pasek.effect_id = effect_id pasek.set_effect() # Jak to się nie zesra to będzie cud XD akai.knobs_durations_change = False pasek.print_state() elif akai.pads_change: if akai.pads[3]: output_queue.put([255, 255, 255]) else: output_queue.put(map(lambda x: 255 if x else 0,akai.pads[:3])) akai.pads_change = False sleep(0.1) # Do something about it! def consumer(input_queue): while True: [r, g, b] = input_queue.get() pasek.set_colors(r, g, b) input_queue.task_done() # def print_state(): # while True: # akai.print_state() # sleep(3) if __name__ == '__main__': q = Queue() Thread(target=producer, args=(q,)).start() Thread(target=consumer, args=(q,)).start() Thread(target=akai.listen).start()
[ "sergiusz.warga@gmail.com" ]
sergiusz.warga@gmail.com
a19eba7bf0ec98307d379e2f857ae2b25e827d02
c8b64c3c4c0639eae0046f9c6de06a8a3cd7cef7
/Code/Othello/bin/easy_install
359dee08207374ef40fc885f7b3c83642e623946
[]
no_license
nicholaskrieg/MLAI
24bf1c2f72eb0b3c54e109d7f65e46794d4f1846
21fe04af11643d6dd3d3e2363d9d1bb4f0daa45b
refs/heads/master
2021-05-09T06:38:24.669883
2018-03-19T21:29:25
2018-03-19T21:29:25
119,329,977
0
0
null
null
null
null
UTF-8
Python
false
false
279
#!/Users/nicholaskrieg/Documents/CS/MLAI/Orthello/bin/python3.6 # -*- coding: utf-8 -*- import re import sys from setuptools.command.easy_install import main if __name__ == '__main__': sys.argv[0] = re.sub(r'(-script\.pyw?|\.exe)?$', '', sys.argv[0]) sys.exit(main())
[ "nicholas.krieg@gmail.com" ]
nicholas.krieg@gmail.com
d9af383d21bace3659d9e389f6fd471dde4ed386
fd2a7b889809ed302d2b3013cefb2c441dfb642a
/venv/bin/pylint
d519fa039ead9994ac2d7332d516f5a5eb8c57bd
[]
no_license
AnotherCoolDude/hello_flask
f27a976243e7956d9cc1a43953a119e396606089
79069a2975b479cb7479822c5946607783ab37b0
refs/heads/master
2022-10-08T13:10:44.602551
2019-10-02T14:35:21
2019-10-02T14:35:21
210,346,653
0
0
null
2022-09-16T18:10:44
2019-09-23T12:13:33
Python
UTF-8
Python
false
false
267
#!/Users/christianhovenbitzer/python/hello_flask/venv/bin/python3 # -*- coding: utf-8 -*- import re import sys from pylint import run_pylint if __name__ == '__main__': sys.argv[0] = re.sub(r'(-script\.pyw?|\.exe)?$', '', sys.argv[0]) sys.exit(run_pylint())
[ "christianhovenbitzer@Christians-MBP.fritz.box" ]
christianhovenbitzer@Christians-MBP.fritz.box
8af8855e074aad7b7515f888ec0f24f85164debb
5e6d8b9989247801718dd1f10009f0f7f54c1eb4
/sdk/python/pulumi_azure_native/web/v20210115/web_app_relay_service_connection_slot.py
dda61e329470f348eb6bc50714d28126c870113f
[ "BSD-3-Clause", "Apache-2.0" ]
permissive
vivimouret29/pulumi-azure-native
d238a8f91688c9bf09d745a7280b9bf2dd6d44e0
1cbd988bcb2aa75a83e220cb5abeb805d6484fce
refs/heads/master
2023-08-26T05:50:40.560691
2021-10-21T09:25:07
2021-10-21T09:25:07
null
0
0
null
null
null
null
UTF-8
Python
false
false
16,719
py
# coding=utf-8 # *** WARNING: this file was generated by the Pulumi SDK Generator. *** # *** Do not edit by hand unless you're certain you know what you are doing! *** import warnings import pulumi import pulumi.runtime from typing import Any, Mapping, Optional, Sequence, Union, overload from ... import _utilities __all__ = ['WebAppRelayServiceConnectionSlotArgs', 'WebAppRelayServiceConnectionSlot'] @pulumi.input_type class WebAppRelayServiceConnectionSlotArgs: def __init__(__self__, *, name: pulumi.Input[str], resource_group_name: pulumi.Input[str], slot: pulumi.Input[str], biztalk_uri: Optional[pulumi.Input[str]] = None, entity_connection_string: Optional[pulumi.Input[str]] = None, entity_name: Optional[pulumi.Input[str]] = None, hostname: Optional[pulumi.Input[str]] = None, kind: Optional[pulumi.Input[str]] = None, port: Optional[pulumi.Input[int]] = None, resource_connection_string: Optional[pulumi.Input[str]] = None, resource_type: Optional[pulumi.Input[str]] = None): """ The set of arguments for constructing a WebAppRelayServiceConnectionSlot resource. :param pulumi.Input[str] name: Name of the app. :param pulumi.Input[str] resource_group_name: Name of the resource group to which the resource belongs. :param pulumi.Input[str] slot: Name of the deployment slot. If a slot is not specified, the API will create or update a hybrid connection for the production slot. :param pulumi.Input[str] kind: Kind of resource. """ pulumi.set(__self__, "name", name) pulumi.set(__self__, "resource_group_name", resource_group_name) pulumi.set(__self__, "slot", slot) if biztalk_uri is not None: pulumi.set(__self__, "biztalk_uri", biztalk_uri) if entity_connection_string is not None: pulumi.set(__self__, "entity_connection_string", entity_connection_string) if entity_name is not None: pulumi.set(__self__, "entity_name", entity_name) if hostname is not None: pulumi.set(__self__, "hostname", hostname) if kind is not None: pulumi.set(__self__, "kind", kind) if port is not None: pulumi.set(__self__, "port", port) if resource_connection_string is not None: pulumi.set(__self__, "resource_connection_string", resource_connection_string) if resource_type is not None: pulumi.set(__self__, "resource_type", resource_type) @property @pulumi.getter def name(self) -> pulumi.Input[str]: """ Name of the app. """ return pulumi.get(self, "name") @name.setter def name(self, value: pulumi.Input[str]): pulumi.set(self, "name", value) @property @pulumi.getter(name="resourceGroupName") def resource_group_name(self) -> pulumi.Input[str]: """ Name of the resource group to which the resource belongs. """ return pulumi.get(self, "resource_group_name") @resource_group_name.setter def resource_group_name(self, value: pulumi.Input[str]): pulumi.set(self, "resource_group_name", value) @property @pulumi.getter def slot(self) -> pulumi.Input[str]: """ Name of the deployment slot. If a slot is not specified, the API will create or update a hybrid connection for the production slot. """ return pulumi.get(self, "slot") @slot.setter def slot(self, value: pulumi.Input[str]): pulumi.set(self, "slot", value) @property @pulumi.getter(name="biztalkUri") def biztalk_uri(self) -> Optional[pulumi.Input[str]]: return pulumi.get(self, "biztalk_uri") @biztalk_uri.setter def biztalk_uri(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "biztalk_uri", value) @property @pulumi.getter(name="entityConnectionString") def entity_connection_string(self) -> Optional[pulumi.Input[str]]: return pulumi.get(self, "entity_connection_string") @entity_connection_string.setter def entity_connection_string(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "entity_connection_string", value) @property @pulumi.getter(name="entityName") def entity_name(self) -> Optional[pulumi.Input[str]]: return pulumi.get(self, "entity_name") @entity_name.setter def entity_name(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "entity_name", value) @property @pulumi.getter def hostname(self) -> Optional[pulumi.Input[str]]: return pulumi.get(self, "hostname") @hostname.setter def hostname(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "hostname", value) @property @pulumi.getter def kind(self) -> Optional[pulumi.Input[str]]: """ Kind of resource. """ return pulumi.get(self, "kind") @kind.setter def kind(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "kind", value) @property @pulumi.getter def port(self) -> Optional[pulumi.Input[int]]: return pulumi.get(self, "port") @port.setter def port(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "port", value) @property @pulumi.getter(name="resourceConnectionString") def resource_connection_string(self) -> Optional[pulumi.Input[str]]: return pulumi.get(self, "resource_connection_string") @resource_connection_string.setter def resource_connection_string(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "resource_connection_string", value) @property @pulumi.getter(name="resourceType") def resource_type(self) -> Optional[pulumi.Input[str]]: return pulumi.get(self, "resource_type") @resource_type.setter def resource_type(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "resource_type", value) class WebAppRelayServiceConnectionSlot(pulumi.CustomResource): @overload def __init__(__self__, resource_name: str, opts: Optional[pulumi.ResourceOptions] = None, biztalk_uri: Optional[pulumi.Input[str]] = None, entity_connection_string: Optional[pulumi.Input[str]] = None, entity_name: Optional[pulumi.Input[str]] = None, hostname: Optional[pulumi.Input[str]] = None, kind: Optional[pulumi.Input[str]] = None, name: Optional[pulumi.Input[str]] = None, port: Optional[pulumi.Input[int]] = None, resource_connection_string: Optional[pulumi.Input[str]] = None, resource_group_name: Optional[pulumi.Input[str]] = None, resource_type: Optional[pulumi.Input[str]] = None, slot: Optional[pulumi.Input[str]] = None, __props__=None): """ Hybrid Connection for an App Service app. :param str resource_name: The name of the resource. :param pulumi.ResourceOptions opts: Options for the resource. :param pulumi.Input[str] kind: Kind of resource. :param pulumi.Input[str] name: Name of the app. :param pulumi.Input[str] resource_group_name: Name of the resource group to which the resource belongs. :param pulumi.Input[str] slot: Name of the deployment slot. If a slot is not specified, the API will create or update a hybrid connection for the production slot. """ ... @overload def __init__(__self__, resource_name: str, args: WebAppRelayServiceConnectionSlotArgs, opts: Optional[pulumi.ResourceOptions] = None): """ Hybrid Connection for an App Service app. :param str resource_name: The name of the resource. :param WebAppRelayServiceConnectionSlotArgs args: The arguments to use to populate this resource's properties. :param pulumi.ResourceOptions opts: Options for the resource. """ ... def __init__(__self__, resource_name: str, *args, **kwargs): resource_args, opts = _utilities.get_resource_args_opts(WebAppRelayServiceConnectionSlotArgs, pulumi.ResourceOptions, *args, **kwargs) if resource_args is not None: __self__._internal_init(resource_name, opts, **resource_args.__dict__) else: __self__._internal_init(resource_name, *args, **kwargs) def _internal_init(__self__, resource_name: str, opts: Optional[pulumi.ResourceOptions] = None, biztalk_uri: Optional[pulumi.Input[str]] = None, entity_connection_string: Optional[pulumi.Input[str]] = None, entity_name: Optional[pulumi.Input[str]] = None, hostname: Optional[pulumi.Input[str]] = None, kind: Optional[pulumi.Input[str]] = None, name: Optional[pulumi.Input[str]] = None, port: Optional[pulumi.Input[int]] = None, resource_connection_string: Optional[pulumi.Input[str]] = None, resource_group_name: Optional[pulumi.Input[str]] = None, resource_type: Optional[pulumi.Input[str]] = None, slot: Optional[pulumi.Input[str]] = None, __props__=None): if opts is None: opts = pulumi.ResourceOptions() if not isinstance(opts, pulumi.ResourceOptions): raise TypeError('Expected resource options to be a ResourceOptions instance') if opts.version is None: opts.version = _utilities.get_version() if opts.id is None: if __props__ is not None: raise TypeError('__props__ is only valid when passed in combination with a valid opts.id to get an existing resource') __props__ = WebAppRelayServiceConnectionSlotArgs.__new__(WebAppRelayServiceConnectionSlotArgs) __props__.__dict__["biztalk_uri"] = biztalk_uri __props__.__dict__["entity_connection_string"] = entity_connection_string __props__.__dict__["entity_name"] = entity_name __props__.__dict__["hostname"] = hostname __props__.__dict__["kind"] = kind if name is None and not opts.urn: raise TypeError("Missing required property 'name'") __props__.__dict__["name"] = name __props__.__dict__["port"] = port __props__.__dict__["resource_connection_string"] = resource_connection_string if resource_group_name is None and not opts.urn: raise TypeError("Missing required property 'resource_group_name'") __props__.__dict__["resource_group_name"] = resource_group_name __props__.__dict__["resource_type"] = resource_type if slot is None and not opts.urn: raise TypeError("Missing required property 'slot'") __props__.__dict__["slot"] = slot __props__.__dict__["type"] = None alias_opts = pulumi.ResourceOptions(aliases=[pulumi.Alias(type_="azure-nextgen:web/v20210115:WebAppRelayServiceConnectionSlot"), pulumi.Alias(type_="azure-native:web:WebAppRelayServiceConnectionSlot"), pulumi.Alias(type_="azure-nextgen:web:WebAppRelayServiceConnectionSlot"), pulumi.Alias(type_="azure-native:web/v20150801:WebAppRelayServiceConnectionSlot"), pulumi.Alias(type_="azure-nextgen:web/v20150801:WebAppRelayServiceConnectionSlot"), pulumi.Alias(type_="azure-native:web/v20160801:WebAppRelayServiceConnectionSlot"), pulumi.Alias(type_="azure-nextgen:web/v20160801:WebAppRelayServiceConnectionSlot"), pulumi.Alias(type_="azure-native:web/v20180201:WebAppRelayServiceConnectionSlot"), pulumi.Alias(type_="azure-nextgen:web/v20180201:WebAppRelayServiceConnectionSlot"), pulumi.Alias(type_="azure-native:web/v20181101:WebAppRelayServiceConnectionSlot"), pulumi.Alias(type_="azure-nextgen:web/v20181101:WebAppRelayServiceConnectionSlot"), pulumi.Alias(type_="azure-native:web/v20190801:WebAppRelayServiceConnectionSlot"), pulumi.Alias(type_="azure-nextgen:web/v20190801:WebAppRelayServiceConnectionSlot"), pulumi.Alias(type_="azure-native:web/v20200601:WebAppRelayServiceConnectionSlot"), pulumi.Alias(type_="azure-nextgen:web/v20200601:WebAppRelayServiceConnectionSlot"), pulumi.Alias(type_="azure-native:web/v20200901:WebAppRelayServiceConnectionSlot"), pulumi.Alias(type_="azure-nextgen:web/v20200901:WebAppRelayServiceConnectionSlot"), pulumi.Alias(type_="azure-native:web/v20201001:WebAppRelayServiceConnectionSlot"), pulumi.Alias(type_="azure-nextgen:web/v20201001:WebAppRelayServiceConnectionSlot"), pulumi.Alias(type_="azure-native:web/v20201201:WebAppRelayServiceConnectionSlot"), pulumi.Alias(type_="azure-nextgen:web/v20201201:WebAppRelayServiceConnectionSlot"), pulumi.Alias(type_="azure-native:web/v20210101:WebAppRelayServiceConnectionSlot"), pulumi.Alias(type_="azure-nextgen:web/v20210101:WebAppRelayServiceConnectionSlot"), pulumi.Alias(type_="azure-native:web/v20210201:WebAppRelayServiceConnectionSlot"), pulumi.Alias(type_="azure-nextgen:web/v20210201:WebAppRelayServiceConnectionSlot")]) opts = pulumi.ResourceOptions.merge(opts, alias_opts) super(WebAppRelayServiceConnectionSlot, __self__).__init__( 'azure-native:web/v20210115:WebAppRelayServiceConnectionSlot', resource_name, __props__, opts) @staticmethod def get(resource_name: str, id: pulumi.Input[str], opts: Optional[pulumi.ResourceOptions] = None) -> 'WebAppRelayServiceConnectionSlot': """ Get an existing WebAppRelayServiceConnectionSlot resource's state with the given name, id, and optional extra properties used to qualify the lookup. :param str resource_name: The unique name of the resulting resource. :param pulumi.Input[str] id: The unique provider ID of the resource to lookup. :param pulumi.ResourceOptions opts: Options for the resource. """ opts = pulumi.ResourceOptions.merge(opts, pulumi.ResourceOptions(id=id)) __props__ = WebAppRelayServiceConnectionSlotArgs.__new__(WebAppRelayServiceConnectionSlotArgs) __props__.__dict__["biztalk_uri"] = None __props__.__dict__["entity_connection_string"] = None __props__.__dict__["entity_name"] = None __props__.__dict__["hostname"] = None __props__.__dict__["kind"] = None __props__.__dict__["name"] = None __props__.__dict__["port"] = None __props__.__dict__["resource_connection_string"] = None __props__.__dict__["resource_type"] = None __props__.__dict__["type"] = None return WebAppRelayServiceConnectionSlot(resource_name, opts=opts, __props__=__props__) @property @pulumi.getter(name="biztalkUri") def biztalk_uri(self) -> pulumi.Output[Optional[str]]: return pulumi.get(self, "biztalk_uri") @property @pulumi.getter(name="entityConnectionString") def entity_connection_string(self) -> pulumi.Output[Optional[str]]: return pulumi.get(self, "entity_connection_string") @property @pulumi.getter(name="entityName") def entity_name(self) -> pulumi.Output[Optional[str]]: return pulumi.get(self, "entity_name") @property @pulumi.getter def hostname(self) -> pulumi.Output[Optional[str]]: return pulumi.get(self, "hostname") @property @pulumi.getter def kind(self) -> pulumi.Output[Optional[str]]: """ Kind of resource. """ return pulumi.get(self, "kind") @property @pulumi.getter def name(self) -> pulumi.Output[str]: """ Resource Name. """ return pulumi.get(self, "name") @property @pulumi.getter def port(self) -> pulumi.Output[Optional[int]]: return pulumi.get(self, "port") @property @pulumi.getter(name="resourceConnectionString") def resource_connection_string(self) -> pulumi.Output[Optional[str]]: return pulumi.get(self, "resource_connection_string") @property @pulumi.getter(name="resourceType") def resource_type(self) -> pulumi.Output[Optional[str]]: return pulumi.get(self, "resource_type") @property @pulumi.getter def type(self) -> pulumi.Output[str]: """ Resource type. """ return pulumi.get(self, "type")
[ "noreply@github.com" ]
vivimouret29.noreply@github.com
c6ecf3c59e8d315c1650c67532864af71b386c05
4e8b37ca121be19cd3b4e73a6592be2659d8134c
/backend/Techfesia2019/accounts/migrations/0005_auto_20190701_1708.py
a7113d24504a420a0d91b930fb768ac3673981f3
[ "MIT" ]
permissive
masterashu/Techfesia2019
365b9b8dc1cb0bc6b613c72632e8b7a2a2a70905
8fd82c4867c8d870b82a936fc0f9e80f11ae03e7
refs/heads/backend-event-registrations
2020-06-10T20:58:40.850415
2019-07-27T23:00:21
2019-07-27T23:00:21
193,744,800
1
1
MIT
2019-06-29T17:12:31
2019-06-25T16:29:12
Python
UTF-8
Python
false
false
466
py
# Generated by Django 2.2.2 on 2019-07-01 11:38 from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('accounts', '0004_auto_20190701_0956'), ] operations = [ migrations.AlterField( model_name='institute', name='name', field=models.CharField(default='Indian Institute of Information Technology, Sri City', max_length=200, unique=True), ), ]
[ "masterashu@live.in" ]
masterashu@live.in
bc863a6118cbe18297247f263fbc963beca38953
0940b1cdfee8d7deb488f396736b0df7c01b8bb6
/Exam/faculty/forms.py
aa76c4ab78faa88142ef132c0f9d3a64fe8268e3
[]
no_license
srilaasya/Online-examination-system
500767f77c2e2a261ddaf90029542d4e9d5312ae
f6ef76f6569141c02c861c67c2b5f7aa375c719a
refs/heads/main
2023-08-28T21:14:48.884215
2021-11-02T23:57:38
2021-11-02T23:57:38
395,973,123
1
0
null
null
null
null
UTF-8
Python
false
false
860
py
from django import forms from .models import FacultyInfo from django.contrib.auth.models import User class FacultyForm(forms.ModelForm): class Meta(): model = User fields = ['username', 'email', 'password'] widgets = { 'password': forms.PasswordInput(attrs = {'id':'passwordfield','class':'form-control'}), 'email' : forms.EmailInput(attrs = {'id':'emailfield','class':'form-control'}), 'username' : forms.TextInput(attrs = {'id':'usernamefield','class':'form-control'}) } class FacultyInfoForm(forms.ModelForm): class Meta(): model = FacultyInfo fields = ['address','subject','picture'] widgets = { 'address': forms.Textarea(attrs = {'class':'form-control'}), 'subject' : forms.TextInput(attrs = {'class':'form-control'}) }
[ "nutheti.laasya@gmail.com" ]
nutheti.laasya@gmail.com
9c5a610bba7b0c3825198a36fe15ea9d099394bc
172b9f849cbe48ef85cf7cca776a432bdad4587b
/lunchy/sublunchy/shortcuts.py
e702d0e3db66ae63537816ac036361452bb41069
[]
no_license
cpasbanal/lunchy
7ec9a0d1dba9e0dc6ed0b84698da928046829eb8
c0f10c25866640bc8f3c8a34f97bc6fe2397959f
refs/heads/master
2021-06-01T02:49:07.479605
2016-08-17T12:56:50
2016-08-17T12:56:50
null
0
0
null
null
null
null
UTF-8
Python
false
false
587
py
''' Create shortcuts in and out between user message and wit out : message from Wit to user --> add keys to quick replies in : use the keys to force a story in Wit ''' # TODO store in database the mapping and make a web interface of this module # import the logging library import logging # Get an instance of a logger logger = logging.getLogger("lunchy") class Shortcut(): def __init__(self, *args, **kwargs): self.keys_out = { "Quelle est la météo ?" : "weather", "Un déj aléatoire ?" : "randomlunch", }
[ "jcroyere@gmail.com" ]
jcroyere@gmail.com
9b9812e5c6c7e010b1d4272c9d1eb1a2bd8e4047
98be616c9e077db836803ec56452564b291e3ff9
/Python+/Python3/src/module/builtins/collection/lists/list_typed.py
198006e815badc59d1731ae055d876a0bd1c6caf
[]
no_license
grozhnev/yaal_examples
adef871b0f23f866491cc41f6c376964ace3fbe2
fe337ee825b3c063d7ed0c3e7e54bef440421894
refs/heads/master
2021-03-30T12:20:42.402132
2019-12-26T07:01:54
2019-12-26T15:08:11
null
0
0
null
null
null
null
UTF-8
Python
false
false
120
py
# Strongly typed List # Strongly typed list from typing import List typed_list: List[RuntimeError] = [RuntimeError()]
[ "aleksei_iablokov@epam.com" ]
aleksei_iablokov@epam.com
a9d2eeab18066cbc76789aba31dd51329d4f3780
9f0b9a8fe27336b8a231a33c6f693ed019a61b6e
/blacklinetest.py
f6eb1fa445e64a1ab1daa8cf7cc3bd44fcadc93b
[]
no_license
Duong-NVH/tool-set
e2647cf74fa085eab42fe3f19c852634629e956e
e7c5f7f4522e75eefe74e808a07ecf6575c4ebf5
refs/heads/main
2023-06-15T07:37:30.783287
2021-07-09T15:58:12
2021-07-09T15:58:12
382,987,402
0
0
null
null
null
null
UTF-8
Python
false
false
517
py
import cv2 import numpy as np img = cv2.imread('blacklinetest.jpg') gray = cv2.cvtColor(img, cv2.COLOR_BGR2GRAY) edges = cv2.Canny(gray, 50, 150, apertureSize=3) lines = cv2.HoughLines(edges, 1, np.pi/180, 500) for rho, theta in lines[0]: a = np.cos(theta) b = np.sin(theta) x0 = a*rho y0 = b*rho x1 = int(x0 + 1000*(-b)) y1 = int(y0 + 1000*(a)) x2 = int(x0 - 1000*(-b)) y2 = int(y0 - 1000*(a)) cv2.line(img, (x1, y1), (x2, y2), (0, 0, 255), 2) cv2.imwrite('houghlines3.jpg', img)
[ "you@example.com" ]
you@example.com
42e555c359f3334d8e2cea4ef890e0bcdd2716ec
2100ea2f84adcaeb89bd8e49f50d03497a1f87fb
/txtToExcel/txtToExcel.py
747f8dea3e15fc01ffdabad8be75c0c69373802b
[]
no_license
JBPrew/RNG-Generator
d1c8c79710fce5b122a4423b3aab207698bb08ce
20db203e702987fc937660b744b0bcb1ff63b508
refs/heads/main
2023-04-20T18:41:04.067449
2021-05-24T01:56:23
2021-05-24T01:56:23
370,199,459
1
0
null
null
null
null
UTF-8
Python
false
false
1,245
py
import xlsxwriter import pandas as pd import xlrd from xlsxwriter.utility import xl_rowcol_to_cell txtFileName = input("Name of txt File: ") while txtFileName == "": txtFileName = input("Name of txt File: ") workName = input("Excel Worksheet Name: ") if (workName == ""): workName = txtFileName # Create a workbook and add a worksheet. txtFile = open("C:/SciFair/txtFiles/" + txtFileName + ".txt", "r") # txtFileLength = len(txtFile.readlines()) txtString = txtFile.read() workbook = xlsxwriter.Workbook('C:/SciFair/txtToExcel/' + workName + '.xlsx') worksheet = workbook.add_worksheet() print("Excel Sheet Created") # Start from the first cell. Rows and columns are zero indexed. row = 5 col = 0 start = xl_rowcol_to_cell(row, col) # Iterate over the data and write it out row by row. for i in range(len(txtString)): worksheet.write(row, col, int(txtString[i])) col += 1 if (col >= 2500): row += 1 col = 0 print("Excel Sheet Filled") col = 2500 end = xl_rowcol_to_cell(row, col) worksheet.write_formula(0, 0, "=AVERAGE(" + start + ":" + end + ")") worksheet.write_formula(1, 0, "=MEDIAN(" + start + ":" + end + ")") worksheet.write_formula(2, 0, "=MODE(" + start + ":" + end + ")") workbook.close()
[ "jackbprewitt@gmail.com" ]
jackbprewitt@gmail.com
148dd580eb7efdf224a1b60c6aa54dc584d2a4e3
6aa8605b7fa86ec6568247941c8c0dbde3fc9bd1
/bbn/camb_to_class.py
6f6d9dc6a1f804fc4b41f2f7c689bcb787104286
[]
no_license
anirbandas89/BDM_CLASS
dfa3a0a4a24060d622829ce4bd7c9d0e74780bd1
33404bdf7dea762413cb9dd3aff9957fae7221ad
refs/heads/master
2020-05-16T09:31:47.898864
2019-06-08T07:25:09
2019-06-08T07:25:09
182,951,756
0
0
null
null
null
null
UTF-8
Python
false
false
423
py
import numpy as np data = np.loadtxt('BBN_full_Parthenelope_880.3.dat') f = open('BBN_full_Parthenelope_880.3_class_format.dat', 'w') #print data[0] for i in range(len(data)): f.write(str(data[i,0])+' ') f.write(str(data[i,2])+' ') f.write(str(data[i,3])+'\n') print len(data) #f.write( str(data[0,0])+' ' ) #f.write(str(data[0,2])+' ') #f.write(str(data[0,3])+'\n') #f.write(str(data[1,0])) f.close()
[ "noreply@github.com" ]
anirbandas89.noreply@github.com
dba9be4055644dd269dab56629dd4b7349338eb4
10321b56b812da315724c7022e7b0f48a3434883
/07-ejercicios/ejercicio8.py
0962be1baba04bece7c49d64c7a2c341d0fe9ba3
[]
no_license
nisepulvedaa/curso-python
8ef8e6f2b604c10b4549c9046a26be02bfd76ccb
0e1904e71b4cb515920c1610ac6f168dc307d631
refs/heads/master
2023-01-29T21:52:44.912261
2020-12-14T22:47:55
2020-12-14T22:47:55
321,491,748
0
0
null
null
null
null
UTF-8
Python
false
false
399
py
""" Ejercico 8 ¿Cuanto es el x % de z numero ? 20 de 150 """ numero1 = int(input("Ingrese el porcentaje que desea calcular: ")) numero2 = int(input("Ingrese el numero que sea sacar el porcentaje: ")) if numero1 > 100: print("Su porcentaje no puede mayor a 100") else: porcentaje = int((numero1*100)/numero2) print("El {} % de {} es: {} ".format(numero1, numero2, porcentaje))
[ "ni.sepulvedaa@gmail.com" ]
ni.sepulvedaa@gmail.com
d7aab2532f25c287a63c8bd8d282163103684f29
d7567ee75e48bd7872a1c332d471ff3ce7433cb9
/checkout/urls.py
233bfb99df176d4ab47c4bae44affd20f8155e9c
[]
no_license
sarahbarron/ecommerce
30cd0ff26afa5ec9031165b63ecde8c0f7f6086f
aba5370fd731e7ec9e677041504f6c3457b0d405
refs/heads/master
2020-03-17T21:10:56.385918
2020-01-17T18:35:28
2020-01-17T18:35:28
133,947,336
0
0
null
null
null
null
UTF-8
Python
false
false
131
py
from django.conf.urls import url from .views import checkout urlpatterns = [ url(r'^$', checkout, name='checkout'), ]
[ "sarahflavin@yahoo.com" ]
sarahflavin@yahoo.com
ab932c024897c581d9bb5dd95eef2ee759d421c2
bac5ecb5eef06dfe76b9b7bff80faee7485c67dd
/.history/django_vuejs_tutorial/django_vuejs/dataiku/models_20200829125121.py
c7e0a2d229648bf8a2326333ab23d5a72731658d
[]
no_license
MChrys/dataiku
fb1e48401d544cbcc5a80a0a27668dc9d2d196e5
6091b24f565224260a89246e29c0a1cbb72f58ed
refs/heads/master
2022-12-16T11:06:13.896643
2020-09-12T19:03:51
2020-09-12T19:03:51
293,287,620
0
0
null
null
null
null
UTF-8
Python
false
false
5,147
py
from django.db import models from django.utils import timezone class Task(models.Model): ''' All task that could be apply to a specific account ''' name = models.CharField(max_length=60, primary_key=True) description = models.CharField(max_length=510, null=True, blank=True) supertask = models.ForeignKey('self',null=True, blank=True, on_delete=models.SET_NULL) class Dataiku_account(models.Model): STATUS = ( ('in operation', 'in operation'), ('avaible', 'avaible') ) email = models.CharField(max_length=60, primary_key=True) password = models.CharField(max_length=255, null=True, blank=True) #task = models.CharField(max_length=255, null=True, blank=True) status = models.CharField(max_length=255, null=True, blank=True, choices = STATUS) def __str__(self): return self.email class Operation(models.Model): ''' A running Task like : validate this course or take this QCM ''' creation = models.DateTimeField( editable = False) STATUS = ( ('pending', 'pending'), ('running', 'running'), ('done', 'done') ) task = models.OneToOneField(Task, null=True, blank=True,on_delete=models.SET_NULL) account = models.ForeignKey(Dataiku_account, on_delete=models.CASCADE) statut = models.CharField(max_length=255, null=True, blank=True, choices = STATUS) def save(self, *args, **kwargs): if not self.id: self.creation = timezone.now() return super(User, self).save(*args, **kwargs) class QCM(models.Model): LearningPathUrl = models.CharField(max_length=255, null=True, blank=True) LearningPathName = models.CharField(max_length=255, null=True, blank=True) CourseUrl = models.CharField(max_length=255, null=True, blank=True) CourseName = models.CharField(max_length=255, null=True, blank=True) QcmUrl = models.CharField(max_length=255, null=True, blank=True) QcmName = models.CharField(max_length=255, null=True, blank=True) Lenght = models.IntegerField(default =0) Verif = models.IntegerField(default =0) status = models.BooleanField(default = False) def __str__(self): return "{}_{}_{}".format(self.LearningPathName, self.CourseName,self.QcmName) class Session(models.Model): STATUS = ( ('running','running'), ('finish','finish') ) email = models.ForeignKey(Dataiku_account , on_delete=models.CASCADE) start = models.DateTimeField(editable =False) countdown = models.CharField(max_length=10, blank=True, null=True, default= '59:59') score = models.IntegerField(default=0) lenght = models.IntegerField(default=0) def save(self, *args, **kwargs): ''' On save, update timestamps ''' if not self.id: self.start = timezone.now() return super(User, self).save(*args, **kwargs) # Create your models here. class Question(models.Model): STATUS = ( ('pending', 'pending'), ('check', 'check') ) CHOICES_TYPE = ( ('checkbox', 'checkbox'), ('radio', 'radio') ) text = models.CharField(max_length=255, primary_key=True) #session = models.ForeignKey(Session, null=True, blank=True,on_delete = models.SET_NULL) status = models.CharField(max_length=255, null=True, blank=True, choices = STATUS) choice_type = models.CharField(max_length=255, null=True, blank=True, default= "radio" ,choices = CHOICES_TYPE) max_choices = models.IntegerField(default = 0) cursor = models.IntegerField(default = 1) # cursor is the number of answer currently explored qcm_link = models.ForeignKey(QCM, null=True, blank=True, on_delete=models.CASCADE) def __str__(self): return self.text class Run(models.Model): ''' A Run is a try ''' STATUS = ( (True, 'True'), (False, 'False') ) id = models.AutoField(primary_key=True) #creation = models.DateTimeField(editable =False) question_link = models.ForeignKey(Question, null=True, blank=True, on_delete=models.CASCADE) session_link = models.ForeignKey(Session, null=True, blank=True, on_delete=models.CASCADE) status = models.BooleanField(default = False) class Posibility(models.Model): CHOICES = ( (1, '1'), (2, '2'), (3, '3'), (4, '4'), ) rank = models.IntegerField( null=True, default= 1, blank=True, choices = CHOICES) question_link = models.ForeignKey(Question, null=True, blank=True,on_delete=models.CASCADE) text = models.CharField(max_length=255, null=True, blank=True) #rank = models.ForeignKey(Answer, null=True, blank=True,on_delete= models.SET_NULL) def __str__(self): return self.text class Answer(models.Model): #choice = models.IntegerField( null=True, default= 1, blank=True, choices = CHOICES) choice = models.ForeignKey(Posibility,to_field='rank',blank=True,null=True,on_delete= models.SET_NULL) connected_run = models.ForeignKey(Run,to_field='id',blank=True,null=True,on_delete= models.SET_NULL) def __str__(self): return self.choice
[ "cbeltran@umanis.com" ]
cbeltran@umanis.com
18b985fd2a25b161ab12d7f4f4e09fc83c30cc2e
3b21cbe5320137a3d8f7da40558294081211f63f
/Chapter04/AutoencMnist.py
daebd29ec15d7b88a838e6b5aa4a4d8016f69927
[ "MIT" ]
permissive
Evelynatrocks/Python-Machine-Learning-Cookbook-Second-Edition
d06812bba0a32a9bd6e5e8d788769a07d28084cd
99d8b799dbfe1d9a82f0bcc3648aaeb147b7298f
refs/heads/master
2023-04-06T20:23:05.384943
2021-01-18T12:06:36
2021-01-18T12:06:36
null
0
0
null
null
null
null
UTF-8
Python
false
false
2,181
py
from keras.datasets import mnist (XTrain, YTrain), (XTest, YTest) = mnist.load_data() print('XTrain shape = ',XTrain.shape) print('XTest shape = ',XTest.shape) print('YTrain shape = ',YTrain.shape) print('YTest shape = ',YTest.shape) import numpy as np print('YTrain values = ',np.unique(YTrain)) print('YTest values = ',np.unique(YTest)) unique, counts = np.unique(YTrain, return_counts=True) print('YTrain distribution = ',dict(zip(unique, counts))) unique, counts = np.unique(YTest, return_counts=True) print('YTrain distribution = ',dict(zip(unique, counts))) import matplotlib.pyplot as plt plt.figure(1) plt.subplot(121) plt.hist(YTrain, alpha=0.8, ec='black') plt.xlabel("Classes") plt.ylabel("Number of occurrences") plt.title("YTrain data") plt.subplot(122) plt.hist(YTest, alpha=0.8, ec='black') plt.xlabel("Classes") plt.ylabel("Number of occurrences") plt.title("YTest data") plt.show() XTrain = XTrain.astype('float32') / 255 XTest = XTest.astype('float32') / 255 XTrain = XTrain.reshape((len(XTrain), np.prod(XTrain.shape[1:]))) XTest = XTest.reshape((len(XTest), np.prod(XTest.shape[1:]))) from keras.layers import Input from keras.layers import Dense from keras.models import Model InputModel = Input(shape=(784,)) EncodedLayer = Dense(32, activation='relu')(InputModel) DecodedLayer = Dense(784, activation='sigmoid')(EncodedLayer) AutoencoderModel = Model(InputModel, DecodedLayer) AutoencoderModel.summary() AutoencoderModel.compile(optimizer='adadelta', loss='binary_crossentropy') history = AutoencoderModel.fit(XTrain, XTrain, batch_size=256, epochs=100, shuffle=True, validation_data=(XTest, XTest)) DecodedDigits = AutoencoderModel.predict(XTest) n=5 plt.figure(figsize=(20, 4)) for i in range(n): ax = plt.subplot(2, n, i + 1) plt.imshow(XTest[i+10].reshape(28, 28)) plt.gray() ax.get_xaxis().set_visible(False) ax.get_yaxis().set_visible(False) ax = plt.subplot(2, n, i + 1 + n) plt.imshow(DecodedDigits[i+10].reshape(28, 28)) plt.gray() ax.get_xaxis().set_visible(False) ax.get_yaxis().set_visible(False) plt.show()
[ "joecasillas001@gmail.com" ]
joecasillas001@gmail.com
be0eb741b4aaaad5085131454dec219bdd1c93dd
971e0efcc68b8f7cfb1040c38008426f7bcf9d2e
/tests/artificial/transf_Anscombe/trend_LinearTrend/cycle_30/ar_/test_artificial_1024_Anscombe_LinearTrend_30__100.py
70d9b6daa1932fc44ee8f23227fa9317aea8fd0d
[ "BSD-3-Clause", "LicenseRef-scancode-unknown-license-reference" ]
permissive
antoinecarme/pyaf
a105d172c2e7544f8d580d75f28b751351dd83b6
b12db77cb3fa9292e774b2b33db8ce732647c35e
refs/heads/master
2023-09-01T09:30:59.967219
2023-07-28T20:15:53
2023-07-28T20:15:53
70,790,978
457
77
BSD-3-Clause
2023-03-08T21:45:40
2016-10-13T09:30:30
Python
UTF-8
Python
false
false
268
py
import pyaf.Bench.TS_datasets as tsds import tests.artificial.process_artificial_dataset as art art.process_dataset(N = 1024 , FREQ = 'D', seed = 0, trendtype = "LinearTrend", cycle_length = 30, transform = "Anscombe", sigma = 0.0, exog_count = 100, ar_order = 0);
[ "antoine.carme@laposte.net" ]
antoine.carme@laposte.net
3b086ad101b3545f4ba98eab3b0c7bd9ac43e051
d487843a873554db9ea63dac55110591c265c984
/classes/utilities.py
96d08b7a02c479407b1cc21de574a00ef555438d
[]
no_license
simone989/ForwardingDropboxBot
dd7e7215fb2cc3556da843ad3c10b0375fb8a45d
2ba8ec1f36930b355168456c20a45998fd8b9f48
refs/heads/master
2021-01-01T05:22:35.110007
2016-05-10T15:06:54
2016-05-10T15:06:54
57,361,235
0
0
null
null
null
null
UTF-8
Python
false
false
3,274
py
#!/usr/bin/env python import dropbox class Response(object): def __init__(self): self.responseFile = open("commands.txt","r") self.textLines = [] self.fromFileToList() def fromFileToList(self): #Metodo che salve il contenuto del txt in una lista. Più semplice da gestire. with self.responseFile as fileLine: self.textLines = fileLine.read().splitlines() self.responseFile.close() def getList(self): #Metodo temporaneo. Per la prova della stampa della lista for lines in self.textLines: print (lines) def responseText(self,command): for lines in self.textLines: lineSplitted = lines.split("=") if(self.isCommand(command,lineSplitted[0]) == True): return lineSplitted[1] return False def isCommand(self,textCommand, fileCommand): #if(textCommand == fileCommand): #return True #elif print ("Comando del file: /"+fileCommand) if(textCommand == "/"+fileCommand): return True return False def reloadFile(self): try: self.responseFile = open("commands.txt","r") self.textLines = [] self.fromFileToList() return True except Exception as error: print (error) #debug return False class CommandManager(object): def __init__(self): self.errorMessage = "Comando non riconosciuto" self.sessionOpen = False def executeCommand(self,command,fromuser = None): command = self.parseCommand(command) if(self.sessionOpen): if(self.dropboxSession.waitingToken == True and command.split(" ")[0] == "/token"): self.dropboxSession.token = command.split(" ")[1] return self.dropboxSession.startAuth() else: if(command.startswith("/token")): return "Iniziare una sessione con /startsession prima di continuare." if(command == "/listFiles"): return self.dropboxSession.listOfFile() if(command == "/start"): print ("Debug metodo. Per evitare il loop") return "Started" elif (command == "/info"): return "Bot creato da @xVinz e @simone989" elif (command == "/startsession"): self.startSession(fromuser.username) self.dropboxSession.waitingToken = True return "Digita /token <token di autenticazione> per procedere." #self.startSession(fromuser.username) else: return self.errorMessage def parseCommand(self,command): if(command.startswith("/")): return command else: return "/"+command.lower() def startSession(self,user): self.sessionOpen = True self.dropboxSession = NewSession() def DEBUGlog(self,fromuser,command): print("Utente: "+ fromuser +" ha digitato: "+command) class NewSession(object): def __init__(self): self.isAuthenticated = False self.waitingToken = False self.token = None print ("DEBUG Inizializzazione classe NewSession") def startAuth(self): self.client = dropbox.client.DropboxClient(self.token) self.checkAuthentication() if(self.isAuthenticated): self.waitingToken = False return "Autenticazione eseguita correttamente." else: self.waitingToken = True return "Errore nell'inserimento del token." def checkAuthentication(self): try: self.client.account_info() self.isAuthenticated = True except Exception as error: self.isAuthenticated = False def listOfFile(self):#Da continuare folderMetadata = client.metadata('/') return folderMetadata
[ "0xVinz@gmail.com" ]
0xVinz@gmail.com
e333c381e106259eee7a3f4e11f26674dd3a3594
30a8b69bd2e0a3f3c2c1c88fb3bd8a28e6fc4cd0
/Part1/auth_foursquare.py
dc09d963b40958ce2c5e3b9030a232e3dd9ca643
[]
no_license
llord1/Mining-Georeferenced-Data
d49108f443922f02b90431ad7a9626ea17fd0554
c71f2e151ccfc4a1a9c07b5fcf4e95b7f7ba70e9
refs/heads/master
2021-05-30T13:27:57.663015
2015-12-29T09:10:08
2015-12-29T09:10:08
null
0
0
null
null
null
null
UTF-8
Python
false
false
283
py
#!/usr/bin/env python import foursquare from foursquare_accounts import accounts app = accounts["tutorial"] client = foursquare.Foursquare(client_id=app["client_id"], client_secret=app["client_secret"]) client.set_access_token(app["access_token"])
[ "bgoncalves@gmail.com" ]
bgoncalves@gmail.com
ef79b4514dbb5c1740bdf58c252e4cd319457aba
8526007f636dc57a30dbd6ef7805ec27d90481af
/pipelines/error_rate_pipeline_load.py
3ba0129048553611ce344ddc8583461a021c1fd1
[ "MIT", "LicenseRef-scancode-proprietary-license" ]
permissive
alphagov/app-performance-summary
d4fbb26e73de20ef4e41334e4ad599254565de64
e94c63c26dec5da39b8458b1e46bcc4f922ab7dc
refs/heads/master
2021-09-09T03:45:54.147829
2018-03-13T16:50:50
2018-03-13T16:50:50
119,363,307
0
1
MIT
2018-03-13T16:50:51
2018-01-29T09:46:08
Python
UTF-8
Python
false
false
1,907
py
''' Extract step for application error rates pipeline ''' import luigi from pipeline_steps.export_to_google_sheets import ExportToGoogleSheets from pipeline_steps.production_error_rates_source import ProductionErrorRatesSource from error_rate_pipeline_extract import ErrorRatePipelineExtract from pipeline_util.google_sheet_client import GoogleSheetClient, GoogleSheetTarget from base import BaseTask import pandas as pd import os class ErrorRatePipelineLoad(BaseTask): def requires(self): return [ ErrorRatePipelineExtract(application='whitehall-admin', date_interval=self.date_interval), ErrorRatePipelineExtract(application='whitehall-frontend', date_interval=self.date_interval), ] def __init__(self, *args, **kwargs): super().__init__(task_name='error_rate_extract', *args, **kwargs) self.filename = filename = self.resource_manager.output_file_name( step_name=self.task_name, segment=self.segment ).replace('.csv', '') self.glossary = ProductionErrorRatesSource.glossary() self.gsheet_share_email = os.environ['PLATFORM_METRICS_MAILING_LIST'] def output(self): return GoogleSheetTarget(self.filename) def run(self): for application in ('whitehall-admin', 'whitehall-frontend'): df = self.load_from_step('error_rate_extract', application) export_step = ExportToGoogleSheets(GoogleSheetClient()) export_step.validate_input( df, df_name='input from error_rate_extract {}'.format(application) ) export_step.write_data( df, self.filename, self.glossary, share_email=self.gsheet_share_email ) if __name__ == '__main__': luigi.run(main_task_cls=ErrorRatePipelineLoad, local_scheduler=True)
[ "matmoore@users.noreply.github.com" ]
matmoore@users.noreply.github.com
e8728e4ea571e7a40c1306932b395860af5900fd
173c2c265fb719dc71f01283ea525baf3480cf45
/Required Remainder.py
a4346a8c6a483a9ff07c156b404b799dbb2e837d
[]
no_license
Nafsan/codeforces-codes
6098994c86298e187edd9bba28bc933ca983218a
d8df6fd39413e78c77c269b4050afe45505131ed
refs/heads/master
2023-01-02T12:10:29.251287
2020-11-02T11:55:42
2020-11-02T11:55:42
285,189,084
0
0
null
null
null
null
UTF-8
Python
false
false
412
py
t = int(input()) for p in range(t): x, y, n = map(int, input().split()) rem = n % x if rem == y: print(n) elif rem < y: pos = n - rem - (x - y) if pos < 0: pos = n + abs(rem - x) print(pos) else: diff = rem - y pos = n - diff if pos < 0: diff_2 = abs(y - rem) pos = n + diff_2 + y print(pos)
[ "noreply@github.com" ]
Nafsan.noreply@github.com
4c601e56a8d4ebc420b8a94e8588e6635cb33073
4780fd0e2afc40c3bbc8736605ebc5c351f3bc30
/checkMidpoint.py
9a0266319429f7108a0f59d9f435102131a91392
[]
no_license
fpeckert/cbp_database_public
a1e85a80bd21b72d4a6ea5b3c0065b2cc2ffa85e
0d671a613e74218976ba78577c9a8ba004787d2d
refs/heads/master
2022-02-08T15:20:21.611884
2021-02-01T04:12:44
2021-02-01T04:12:44
225,522,376
18
2
null
2022-02-01T14:30:35
2019-12-03T03:27:50
Python
UTF-8
Python
false
false
4,713
py
# -*- coding: utf-8 -*- """ Created on Wed Jul 29 19:04:20 2020 @author: Anubhav Agarwal """ import cbp import pandas as pd def midpointable(year, suffix): is_sic = False if int(year) <= 1997: is_sic = True national_df = pd.read_csv('cbp' + year + 'us' + suffix + '.csv') state_df = pd.read_csv('cbp' + year + 'st' + suffix + '.csv') county_df = pd.read_csv('cbp' + year + 'co' + suffix + '.csv') if is_sic: national_df = national_df.rename(index=str, columns={'ind': 'naics'}) state_df = state_df.rename(index=str, columns={'ind': 'naics'}) county_df = county_df.rename(index=str, columns={'ind': 'naics'}) industry_ref_file = cbp.refFileName(year) naics_codes = cbp.newNaicsCodes(industry_ref_file, year) geo_codes = cbp.geoCodes(state_df, county_df) # ## # Construct tree for NAICS codes # ## # determine level function based on which industry code is used industry_level_function = cbp.naics_level if is_sic: industry_level_function = cbp.sic_level naics_tree = cbp.preorderTraversalToTree(naics_codes, 'naics', industry_level_function) # ## # Construct tree for Geography # ## geo_tree = cbp.preorderTraversalToTree(geo_codes, 'geo', cbp.geo_level) results_df = cbp.merge_dataframes(national_df, state_df, county_df) results_df = results_df.rename(index = str, columns = {'ind' : 'naics'}) results_df = results_df.drop_duplicates(subset = ['naics', 'geo']) ub_matrix = results_df.pivot(index='naics', columns='geo', values='ub').fillna(0).astype(int) lb_matrix = results_df.pivot(index='naics', columns='geo', values='lb').fillna(0).astype(int) for geo_index, geo in enumerate(geo_codes): for naics_index, naics in enumerate(naics_codes): # does the geo code have children in theory if len(geo_tree[geo_index]['children']) != 0: children = list(map(lambda x: geo_codes[x], geo_tree[geo_index]['children'])) # sum of geographical children's lower/upper bounds geo_sum_lower = sum(lb_matrix[geo_codes[child]][naics] for child in geo_tree[geo_index]['children']) geo_sum_upper = sum(ub_matrix[geo_codes[child]][naics] for child in geo_tree[geo_index]['children']) # check if the code has children in data if geo_sum_upper == 0: continue if geo_sum_lower > ub_matrix[geo][naics] or geo_sum_upper < lb_matrix[geo][naics]: print('Found error in geo tree at industry code = %s' % naics) print('children sum (lower, upper): ' + str((geo_sum_lower, geo_sum_upper))) print('parent (lower, upper): ' + str((lb_matrix[geo][naics], ub_matrix[geo][naics]))) print() print('Parent: %s. Children: %s \n' % (str(geo), children)) return False # check if the geo code has children in theory (in the industry tree) if len(naics_tree[naics_index]['children']) != 0: children = list(map(lambda x: naics_codes[x], naics_tree[naics_index]['children'])) # sum of industrial children's lower/upper bounds naics_sum_lower = sum(lb_matrix[geo][naics_codes[child]] for child in naics_tree[naics_index]['children']) naics_sum_upper = sum(ub_matrix[geo][naics_codes[child]] for child in naics_tree[naics_index]['children']) # check if the code has children in data if naics_sum_upper == 0: continue if naics_sum_lower > ub_matrix[geo][naics] or naics_sum_upper < lb_matrix[geo][naics]: # sic does not have exact hierarchy after level 2 (inclusive) if is_sic and cbp.sic_level(naics) >= 2 and naics_sum_upper < lb_matrix[geo][naics]: continue # discrepancy print('Found error in industry tree at geoography (fipstate, fipscty) = %s' % str(geo)) print('children sum (lower, upper): ' + str((naics_sum_lower, naics_sum_upper))) print('parent (lower, upper): ' + str((lb_matrix[geo][naics], ub_matrix[geo][naics]))) print() print('Parent: %s. Children: %s \n' % (naics, children)) return False return True
[ "noreply@github.com" ]
fpeckert.noreply@github.com
d643ba0ffc2781833fc948211a92075a0281833a
89cef15f8c54f16599f461bce55c4c29954fb918
/scrapy_workua/scrapy_workua/settings.py
d393d3570ca2b72aeee684626fbfc9104b9763a3
[]
no_license
KonDob/HT_20_Scrapy
39417738e6fb9ade6f600bb021c323ef1718f51c
b762ea4d26c01b994bed066ffdb011b467220d91
refs/heads/master
2023-03-24T03:53:21.939319
2021-03-09T14:23:34
2021-03-09T14:23:34
343,230,651
0
0
null
null
null
null
UTF-8
Python
false
false
3,140
py
# Scrapy settings for scrapy_workua project # # For simplicity, this file contains only settings considered important or # commonly used. You can find more settings consulting the documentation: # # https://docs.scrapy.org/en/latest/topics/settings.html # https://docs.scrapy.org/en/latest/topics/downloader-middleware.html # https://docs.scrapy.org/en/latest/topics/spider-middleware.html BOT_NAME = 'scrapy_workua' SPIDER_MODULES = ['scrapy_workua.spiders'] NEWSPIDER_MODULE = 'scrapy_workua.spiders' # Crawlresponsibly by identifying yourself (and your website) on the user-agent #USER_AGENT = 'scrapy_workua (+http://www.yourdomain.com)' # Obey robots.txt rules ROBOTSTXT_OBEY = True # Configure maximum concurrent requests performed by Scrapy (default: 16) #CONCURRENT_REQUESTS = 32 # Configure a delay for requests for the same website (default: 0) # See https://docs.scrapy.org/en/latest/topics/settings.html#download-delay # See also autothrottle settings and docs #DOWNLOAD_DELAY = 3 # The download delay setting will honor only one of: #CONCURRENT_REQUESTS_PER_DOMAIN = 16 #CONCURRENT_REQUESTS_PER_IP = 16 # Disable cookies (enabled by default) #COOKIES_ENABLED = False # Disable Telnet Console (enabled by default) #TELNETCONSOLE_ENABLED = False # Override the default request headers: #DEFAULT_REQUEST_HEADERS = { # 'Accept': 'text/html,application/xhtml+xml,application/xml;q=0.9,*/*;q=0.8', # 'Accept-Language': 'en', #} # Enable or disable spider middlewares # See https://docs.scrapy.org/en/latest/topics/spider-middleware.html #SPIDER_MIDDLEWARES = { # 'scrapy_workua.middlewares.ScrapyWorkuaSpiderMiddleware': 543, #} # Enable or disable downloader middlewares # See https://docs.scrapy.org/en/latest/topics/downloader-middleware.html #DOWNLOADER_MIDDLEWARES = { # 'scrapy_workua.middlewares.ScrapyWorkuaDownloaderMiddleware': 543, #} # Enable or disable extensions # See https://docs.scrapy.org/en/latest/topics/extensions.html #EXTENSIONS = { # 'scrapy.extensions.telnet.TelnetConsole': None, #} # Configure item pipelines # See https://docs.scrapy.org/en/latest/topics/item-pipeline.html #ITEM_PIPELINES = { # 'scrapy_workua.pipelines.ScrapyWorkuaPipeline': 300, #} # Enable and configure the AutoThrottle extension (disabled by default) # See https://docs.scrapy.org/en/latest/topics/autothrottle.html #AUTOTHROTTLE_ENABLED = True # The initial download delay #AUTOTHROTTLE_START_DELAY = 5 # The maximum download delay to be set in case of high latencies #AUTOTHROTTLE_MAX_DELAY = 60 # The average number of requests Scrapy should be sending in parallel to # each remote server #AUTOTHROTTLE_TARGET_CONCURRENCY = 1.0 # Enable showing throttling stats for every response received: #AUTOTHROTTLE_DEBUG = False # Enable and configure HTTP caching (disabled by default) # See https://docs.scrapy.org/en/latest/topics/downloader-middleware.html#httpcache-middleware-settings # noqa: E501 #HTTPCACHE_ENABLED = True #HTTPCACHE_EXPIRATION_SECS = 0 #HTTPCACHE_DIR = 'httpcache' #HTTPCACHE_IGNORE_HTTP_CODES = [] #HTTPCACHE_STORAGE = 'scrapy.extensions.httpcache.FilesystemCacheStorage'
[ "konstantin.dobro@gmail.com" ]
konstantin.dobro@gmail.com
2c49843004e5af3d4f6f1f7a39f56f8280186b0e
ef2dfe0f97cf78e520fc6e30d795691c75dac840
/kin/util.py
d7ce97fe299697f2714c827d5559b248764a8f7e
[ "MIT" ]
permissive
smartdolphin/ai-hackathon-2018
bbcf142865fdf05f0aacd070e8d7f8f99e513dfb
2dce1870ea5a5e10917bb5a018890973e48df886
refs/heads/master
2021-09-14T22:27:54.631329
2018-05-19T17:29:05
2018-05-19T17:30:56
131,360,706
1
0
MIT
2018-04-28T01:41:53
2018-04-28T01:41:53
null
UTF-8
Python
false
false
667
py
import os import tensorflow as tf PATH = '/tmp/kin/models' def local_save(sess, epoch, *args): os.makedirs(PATH, exist_ok=True) saver = tf.train.Saver() saver.save(sess, os.path.join(PATH, str(epoch))) train_writer = tf.summary.FileWriter(os.path.join(PATH, str(epoch)), sess.graph) def local_load(sess): saver = tf.train.Saver() ckpt = tf.train.get_checkpoint_state(PATH) if ckpt and ckpt.model_checkpoint_path: checkpoint = os.path.basename(ckpt.model_checkpoint_path) saver.restore(sess, os.path.join(PATH, checkpoint)) else: raise NotImplemented('No checkpoint!') print('Model loaded')
[ "smartdolphin07@gmail.com" ]
smartdolphin07@gmail.com
63a89a990c8d496437077cb8b6227666b7df1a26
9b7192094d347fb8b506e3a2a67806cab67749f2
/stackoverflow/src/question/migrations/0005_auto_20151117_1839.py
903fa8b5d16878b7e57aa0a21410cd1b36488f24
[]
no_license
ElSacramento/django
830c100907e6d5936761121dec1f8e9ad7d4c286
018f9814318df1b30c02bfeb55562debace5bd5b
refs/heads/master
2021-01-20T19:30:20.440173
2016-06-24T00:03:51
2016-06-24T00:03:51
61,843,133
0
0
null
null
null
null
UTF-8
Python
false
false
477
py
# -*- coding: utf-8 -*- from __future__ import unicode_literals from django.db import migrations, models import django.utils.timezone class Migration(migrations.Migration): dependencies = [ ('question', '0004_auto_20151117_1835'), ] operations = [ migrations.AlterField( model_name='question', name='creation_date', field=models.DateTimeField(default=django.utils.timezone.now, blank=True), ), ]
[ "catherine.zamaraeva@gmail.com" ]
catherine.zamaraeva@gmail.com
8808eb29a9c4f90993a64cfb9ff593217d62ad7e
1da96c1ffecc565307779e3a011e2af2bc75dbcf
/venv/Scripts/django-admin.py
825265bd4c5465fc94f5a2bd2dc2ea1b1afd5c95
[]
no_license
yfs666/py3workspace
058809e11e6d8e6a8c4f96e5ade679c82c5538b2
d64a9165fac239fb7e49f4d4b99921bdc2ed07d2
refs/heads/master
2020-09-03T04:43:13.072897
2019-11-10T14:32:38
2019-11-10T14:32:38
219,387,092
0
0
null
null
null
null
UTF-8
Python
false
false
154
py
#!D:\tools\py3workspace\venv\Scripts\python.exe from django.core import management if __name__ == "__main__": management.execute_from_command_line()
[ "yangfengshuai@xiangshang360.com" ]
yangfengshuai@xiangshang360.com
ee6bca645dc84757845a11e614e0667477bad197
2c0d0a3757f4a21b2de2afc3b8293fe1a2292d68
/test_algoritmos.py
8af9e316f6bd27299e801a0481ee7037c91a2be4
[]
no_license
iErick99/Tarea2-Python
ad0fbfdc24babf456d739df86da6442e9e633c36
b3939cdd415ded50a434cbfa96070167716c81d6
refs/heads/master
2020-04-08T02:40:11.984527
2018-11-24T14:30:24
2018-11-24T14:30:24
158,943,955
0
0
null
null
null
null
UTF-8
Python
false
false
458
py
from Algoritmos import isanigrama, ispalindromo def test_isanigrama(): a = 'hola' b = 'carro' assert isanigrama(a,b) == False, 'Error crítico' def test_ispalindromo(): a = 'hola' assert ispalindromo(a) == False, 'Error crítico' def test_isanigrama2(): a = 'car' b = 'arc' assert isanigrama(a, b) == True, 'Error crítico' def test_ispalindromo2(): a = 'arenera' assert ispalindromo(a) == True, 'Error critico'
[ "lolerick1@hotmail.com" ]
lolerick1@hotmail.com
da2965f9791254063164c9442a3f7a753f730ebb
aa6422117b534e4f4eed197b71f9fcf00eb1983a
/build/flexbe_behavior_engine/flexbe_input/catkin_generated/pkg.installspace.context.pc.py
1aee055bb7df44fed4d20a68d3c6beee1540c93f
[]
no_license
Sinchiguano/StateMachineFlexBe
c07385b09e1ab15e88e894da8fd021d1cbf0de28
d637acf2f26a3f0d83ef4f2d34a2636dff2515f6
refs/heads/master
2020-09-26T02:49:46.882388
2019-12-05T16:44:23
2019-12-05T16:44:23
226,146,554
0
0
null
null
null
null
UTF-8
Python
false
false
449
py
# generated from catkin/cmake/template/pkg.context.pc.in CATKIN_PACKAGE_PREFIX = "" PROJECT_PKG_CONFIG_INCLUDE_DIRS = "/home/casch/catkin_ws/install/include".split(';') if "/home/casch/catkin_ws/install/include" != "" else [] PROJECT_CATKIN_DEPENDS = "".replace(';', ' ') PKG_CONFIG_LIBRARIES_WITH_PREFIX = "".split(';') if "" != "" else [] PROJECT_NAME = "flexbe_input" PROJECT_SPACE_DIR = "/home/casch/catkin_ws/install" PROJECT_VERSION = "1.2.2"
[ "cesarsinchiguano@hotamil.es" ]
cesarsinchiguano@hotamil.es
fcf2843bba0bc1f598ab442734dbe565c35f7c93
bb67740f39e896f4bef0677a3bc164781e533195
/clt/admin.py
c0a932b6d407645297af6575967a97701a90df6f
[]
no_license
gabik/clt
c1fb61fb3ebf40091d68d03b54d744140056761b
76786c0a65c7bbf9b2f4ca6dc16408a37b537666
refs/heads/master
2021-01-23T09:33:19.584041
2013-12-14T13:25:29
2013-12-14T13:25:29
null
0
0
null
null
null
null
UTF-8
Python
false
false
391
py
from django.contrib import admin from clt.models import xml_model, contact_group, group_members, contact_element, contact_phones #class areaAdmin(admin.ModelAdmin): #list_display = ['id', 'parent', 'name'] admin.site.register(xml_model) admin.site.register(contact_group) admin.site.register(contact_phones) admin.site.register(contact_element) admin.site.register(group_members)
[ "gabi@ip-10-190-129-144.ec2.internal" ]
gabi@ip-10-190-129-144.ec2.internal
98c418f97a9b856ecb5507a37fe70ad0085b224a
39996673e85e32497f1a771e6d55fe317fd4e9f2
/Beginner level/strpalindrome.py
46a2d25e65555e3ddb038ddfde76bf88d5216c24
[]
no_license
hariharan31998/Python-Programming
638d1be2d59ad8a6a675b94d1c08c04139344de9
17aba448c36596eea1ab59b5346e53ed644b66de
refs/heads/master
2021-04-27T13:13:49.961069
2019-02-06T05:25:57
2019-02-06T05:25:57
122,435,676
0
2
null
null
null
null
UTF-8
Python
false
false
87
py
a=str(input()) print(a) b=a[::-1] print(b) if a==b: print('yes') else: print('no')
[ "noreply@github.com" ]
hariharan31998.noreply@github.com
c35024eb1eed9b0da1bdde17899977fd5b9b5c96
0201ac814d825cac1030dfe1ccdb7ef1657c205b
/__init__.py
a403709aa7de47dca868813496d90679f83afbc3
[ "BSD-3-Clause" ]
permissive
karttur/geoimagine03-timeseries
c99be449dccaab767d470cfaa2b71d9dae339fba
aa8e1642fd4a8bc196ad6fce9f90b80066d54dac
refs/heads/main
2023-08-22T14:12:50.791746
2021-10-20T10:54:37
2021-10-20T10:54:37
null
0
0
null
null
null
null
UTF-8
Python
false
false
361
py
""" timeseries ========================================== Package belonging to Karttur´s GeoImagine Framework. Author ------ Thomas Gumbricht (thomas.gumbricht@karttur.com) """ from .version import __version__, VERSION, metadataD from .timeseries import ProcessTimeSeries from .numbautil import TimeSeriesNumba #from .tsgraph import ProcessTimeSeriesGraph
[ "thomas.gumbricht@gmail.com" ]
thomas.gumbricht@gmail.com
33f930a46d509398e633fffe7486592b06d47845
3fc8a38257d855e95fe16c1f117790fdfcc665e5
/contrib/spendfrom/spendfrom.py
d06c624a29f6d532e1e2b6600d07ac3aedc5ce09
[ "MIT" ]
permissive
MichaelHDesigns/Aezora
b139e6b0a0886dc8fe4f1e0ed1a702cb26ca6986
fc7170fb5e0e016379f98d4691b014e8ca0746ad
refs/heads/master
2023-03-13T00:31:17.894219
2021-03-08T14:29:27
2021-03-08T14:29:27
null
0
0
null
null
null
null
UTF-8
Python
false
false
9,996
py
#!/usr/bin/env python # # Use the raw transactions API to spend AZRs received on particular addresses, # and send any change back to that same address. # # Example usage: # spendfrom.py # Lists available funds # spendfrom.py --from=ADDRESS --to=ADDRESS --amount=11.00 # # Assumes it will talk to a aezorad or aezora-Qt running # on localhost. # # Depends on jsonrpc # from decimal import * import getpass import math import os import os.path import platform import sys import time from jsonrpc import ServiceProxy, json BASE_FEE=Decimal("0.001") def check_json_precision(): """Make sure json library being used does not lose precision converting BTC values""" n = Decimal("20000000.00000003") satoshis = int(json.loads(json.dumps(float(n)))*1.0e8) if satoshis != 2000000000000003: raise RuntimeError("JSON encode/decode loses precision") def determine_db_dir(): """Return the default location of the aezora data directory""" if platform.system() == "Darwin": return os.path.expanduser("~/Library/Application Support/AEZORA/") elif platform.system() == "Windows": return os.path.join(os.environ['APPDATA'], "AEZORA") return os.path.expanduser("~/.aezora") def read_bitcoin_config(dbdir): """Read the aezora.conf file from dbdir, returns dictionary of settings""" from ConfigParser import SafeConfigParser class FakeSecHead(object): def __init__(self, fp): self.fp = fp self.sechead = '[all]\n' def readline(self): if self.sechead: try: return self.sechead finally: self.sechead = None else: s = self.fp.readline() if s.find('#') != -1: s = s[0:s.find('#')].strip() +"\n" return s config_parser = SafeConfigParser() config_parser.readfp(FakeSecHead(open(os.path.join(dbdir, "aezora.conf")))) return dict(config_parser.items("all")) def connect_JSON(config): """Connect to a aezora JSON-RPC server""" testnet = config.get('testnet', '0') testnet = (int(testnet) > 0) # 0/1 in config file, convert to True/False if not 'rpcport' in config: config['rpcport'] = 14727 if testnet else 14724 connect = "http://%s:%s@127.0.0.1:%s"%(config['rpcuser'], config['rpcpassword'], config['rpcport']) try: result = ServiceProxy(connect) # ServiceProxy is lazy-connect, so send an RPC command mostly to catch connection errors, # but also make sure the aezorad we're talking to is/isn't testnet: if result.getmininginfo()['testnet'] != testnet: sys.stderr.write("RPC server at "+connect+" testnet setting mismatch\n") sys.exit(1) return result except: sys.stderr.write("Error connecting to RPC server at "+connect+"\n") sys.exit(1) def unlock_wallet(aezorad): info = aezorad.getinfo() if 'unlocked_until' not in info: return True # wallet is not encrypted t = int(info['unlocked_until']) if t <= time.time(): try: passphrase = getpass.getpass("Wallet is locked; enter passphrase: ") aezorad.walletpassphrase(passphrase, 5) except: sys.stderr.write("Wrong passphrase\n") info = aezorad.getinfo() return int(info['unlocked_until']) > time.time() def list_available(aezorad): address_summary = dict() address_to_account = dict() for info in aezorad.listreceivedbyaddress(0): address_to_account[info["address"]] = info["account"] unspent = aezorad.listunspent(0) for output in unspent: # listunspent doesn't give addresses, so: rawtx = aezorad.getrawtransaction(output['txid'], 1) vout = rawtx["vout"][output['vout']] pk = vout["scriptPubKey"] # This code only deals with ordinary pay-to-aezora-address # or pay-to-script-hash outputs right now; anything exotic is ignored. if pk["type"] != "pubkeyhash" and pk["type"] != "scripthash": continue address = pk["addresses"][0] if address in address_summary: address_summary[address]["total"] += vout["value"] address_summary[address]["outputs"].append(output) else: address_summary[address] = { "total" : vout["value"], "outputs" : [output], "account" : address_to_account.get(address, "") } return address_summary def select_coins(needed, inputs): # Feel free to improve this, this is good enough for my simple needs: outputs = [] have = Decimal("0.0") n = 0 while have < needed and n < len(inputs): outputs.append({ "txid":inputs[n]["txid"], "vout":inputs[n]["vout"]}) have += inputs[n]["amount"] n += 1 return (outputs, have-needed) def create_tx(aezorad, fromaddresses, toaddress, amount, fee): all_coins = list_available(aezorad) total_available = Decimal("0.0") needed = amount+fee potential_inputs = [] for addr in fromaddresses: if addr not in all_coins: continue potential_inputs.extend(all_coins[addr]["outputs"]) total_available += all_coins[addr]["total"] if total_available < needed: sys.stderr.write("Error, only %f BTC available, need %f\n"%(total_available, needed)); sys.exit(1) # # Note: # Python's json/jsonrpc modules have inconsistent support for Decimal numbers. # Instead of wrestling with getting json.dumps() (used by jsonrpc) to encode # Decimals, I'm casting amounts to float before sending them to aezorad. # outputs = { toaddress : float(amount) } (inputs, change_amount) = select_coins(needed, potential_inputs) if change_amount > BASE_FEE: # don't bother with zero or tiny change change_address = fromaddresses[-1] if change_address in outputs: outputs[change_address] += float(change_amount) else: outputs[change_address] = float(change_amount) rawtx = aezorad.createrawtransaction(inputs, outputs) signed_rawtx = aezorad.signrawtransaction(rawtx) if not signed_rawtx["complete"]: sys.stderr.write("signrawtransaction failed\n") sys.exit(1) txdata = signed_rawtx["hex"] return txdata def compute_amount_in(aezorad, txinfo): result = Decimal("0.0") for vin in txinfo['vin']: in_info = aezorad.getrawtransaction(vin['txid'], 1) vout = in_info['vout'][vin['vout']] result = result + vout['value'] return result def compute_amount_out(txinfo): result = Decimal("0.0") for vout in txinfo['vout']: result = result + vout['value'] return result def sanity_test_fee(aezorad, txdata_hex, max_fee): class FeeError(RuntimeError): pass try: txinfo = aezorad.decoderawtransaction(txdata_hex) total_in = compute_amount_in(aezorad, txinfo) total_out = compute_amount_out(txinfo) if total_in-total_out > max_fee: raise FeeError("Rejecting transaction, unreasonable fee of "+str(total_in-total_out)) tx_size = len(txdata_hex)/2 kb = tx_size/1000 # integer division rounds down if kb > 1 and fee < BASE_FEE: raise FeeError("Rejecting no-fee transaction, larger than 1000 bytes") if total_in < 0.01 and fee < BASE_FEE: raise FeeError("Rejecting no-fee, tiny-amount transaction") # Exercise for the reader: compute transaction priority, and # warn if this is a very-low-priority transaction except FeeError as err: sys.stderr.write((str(err)+"\n")) sys.exit(1) def main(): import optparse parser = optparse.OptionParser(usage="%prog [options]") parser.add_option("--from", dest="fromaddresses", default=None, help="addresses to get AZRs from") parser.add_option("--to", dest="to", default=None, help="address to get send AZRs to") parser.add_option("--amount", dest="amount", default=None, help="amount to send") parser.add_option("--fee", dest="fee", default="0.0", help="fee to include") parser.add_option("--datadir", dest="datadir", default=determine_db_dir(), help="location of aezora.conf file with RPC username/password (default: %default)") parser.add_option("--testnet", dest="testnet", default=False, action="store_true", help="Use the test network") parser.add_option("--dry_run", dest="dry_run", default=False, action="store_true", help="Don't broadcast the transaction, just create and print the transaction data") (options, args) = parser.parse_args() check_json_precision() config = read_bitcoin_config(options.datadir) if options.testnet: config['testnet'] = True aezorad = connect_JSON(config) if options.amount is None: address_summary = list_available(aezorad) for address,info in address_summary.iteritems(): n_transactions = len(info['outputs']) if n_transactions > 1: print("%s %.8f %s (%d transactions)"%(address, info['total'], info['account'], n_transactions)) else: print("%s %.8f %s"%(address, info['total'], info['account'])) else: fee = Decimal(options.fee) amount = Decimal(options.amount) while unlock_wallet(aezorad) == False: pass # Keep asking for passphrase until they get it right txdata = create_tx(aezorad, options.fromaddresses.split(","), options.to, amount, fee) sanity_test_fee(aezorad, txdata, amount*Decimal("0.01")) if options.dry_run: print(txdata) else: txid = aezorad.sendrawtransaction(txdata) print(txid) if __name__ == '__main__': main()
[ "59030118+Aezora@users.noreply.github.com" ]
59030118+Aezora@users.noreply.github.com
fd96a157748820ddb00179c606ac48e6ff091f69
0e0545c05ed9e019d070f732b9f96b7cec1c64f3
/student/schedule.py
49b9800c7b087672e1a76d910c9d93fbeffebb03
[]
no_license
czarhao/sync_crawler
8b635a62b04518d968b4f41f82e6014966d4e94e
f8c7641618d3befe028195e6d26de964b7ad33d5
refs/heads/master
2023-05-27T19:37:48.184658
2021-09-14T10:47:25
2021-09-14T10:47:25
203,547,407
3
0
null
2023-05-22T22:26:26
2019-08-21T09:05:58
Python
UTF-8
Python
false
false
4,844
py
#!/usr/bin/env python3 import json from other import return_soup from student.models import Course def get_schedule(session, now_url, soup, use_header): # 获取用户课程表 try: url = now_url + soup.find(onclick="GetMc('学生个人课表');")["href"] schedule_info = session.get(url, headers=use_header) schedule_soup, _ = return_soup(schedule_info) tmp_time = 0 tmp_info = [] for tmp_tr in schedule_soup.find(id="Table1").find_all("tr"): if tmp_time > 1: for tmp_td in tmp_tr.find_all("td"): if tmp_td.text in ["上午", "下午", "晚上"]: pass elif tmp_td.text == "": pass else: tmp_info = tmp_info + get_info_td(str(tmp_td)) tmp_time = tmp_time + 1 return return_json_schedule(True, "", tmp_info) except Exception as err: return return_json_schedule(False, str(err), []) def return_start_end(start_end): # 获取上课时间,第几节,第几周 if "|" in start_end: return start_end[1], start_end[start_end.find('第') + 1: start_end.find("{") - 1], \ start_end[start_end.find("{") + 2: start_end.find("-")], start_end[start_end.find("-") + 1:-5] return start_end[1], start_end[start_end.find('第') + 1: start_end.find("{") - 1], \ start_end[start_end.find("{") + 2: start_end.find("-")], start_end[start_end.find("-") + 1:-2] def return_start_end_jud(start_end): # 获取上课时间,第几节,第几周,单双周 if '单周' in start_end: return start_end[1], start_end[start_end.find('第') + 1: start_end.find("{") - 1], \ start_end[start_end.find("{") + 2: start_end.find("-")], start_end[start_end.find("-") + 1:-5], 1 elif '双周' in start_end: return start_end[1], start_end[start_end.find('第') + 1: start_end.find("{") - 1], \ start_end[start_end.find("{") + 2: start_end.find("-")], start_end[start_end.find("-") + 1:-5], 2 else: return start_end[1], start_end[start_end.find('第') + 1: start_end.find("{") - 1], \ start_end[start_end.find("{") + 2: start_end.find("-")], start_end[start_end.find("-") + 1:-2], 0 def get_info_td(td): # 对td进行分词,得到学生课表 return_list = [] start = True r_list = [] tmp_info = "" for value in td: if value == ">": start = not start elif value == "<": start = not start r_list.append(tmp_info) tmp_info = "" elif start: tmp_info = tmp_info + value if len(r_list) > 4: if len(r_list) > 10: week, sec, star, end, jud = return_start_end_jud(r_list[3]) new_course = Course(r_list[1] + "(" + r_list[2] + ")", r_list[5], week, sec, int(star), int(end), r_list[4], jud) return_list.append(new_course) week, sec, star, end, jud = return_start_end_jud(r_list[9]) other_course = Course(r_list[7] + "(" + r_list[8] + ")", r_list[11], week, sec, int(star), int(end), r_list[10], jud) return_list.append(other_course) else: week, sec, star, end = return_start_end(r_list[3]) new_course = Course(r_list[1] + "(" + r_list[2] + ")", r_list[5], week, sec, int(star), int(end), r_list[4], 0) return_list.append(new_course) return return_list def return_json_schedule(jud, err_info, info_list): # 返回json maps = [] return_json = { "success": jud, "info": err_info, "courses": maps } if jud: time = 0 for i in info_list: if "(学必)" in i.cname: cour_type = 0 elif "(学选)" in i.cname: cour_type = 1 else: cour_type = 2 tmp = i.section.find(',') if tmp == -1: cour_start = int(i.section) cour_length = 1 else: cour_start = int(i.section[:tmp]) cour_length = int(i.section[tmp + 1:]) - cour_start + 1 tmp_map = { "id": time, "type": cour_type, "start": i.start, "end": i.end, "day_week": i.day_week, "cour_start": cour_start, "cour_length": cour_length, "cour_name": i.cname, "teacher_name": i.tname, "cour_where": i.where, "jud": i.jud } maps.append(tmp_map) time += 1 return return_json
[ "czarhao@foxmail.com" ]
czarhao@foxmail.com
6d553104dce5c9fcd4b3086daec02a36d90a0e92
644bde16761bebf341847dc302bcc195e33e664f
/data/scrape.py
f891f46811de9347b9f31c1d04058fe977e4774a
[]
no_license
lukevs/charity-explorer
797c14d813518869d54cfb3fca74825394646083
452d8bf802b05a28eb46bf117dad0a4a8fa97051
refs/heads/master
2022-12-10T21:08:27.561555
2019-09-17T07:23:58
2019-09-17T07:23:58
203,031,399
5
1
null
2022-12-09T00:43:36
2019-08-18T16:43:36
JavaScript
UTF-8
Python
false
false
1,211
py
from bs4 import BeautifulSoup as bs import requests import csv BASE_URL = 'https://en.wikipedia.org' CHARITY_URL = 'https://en.wikipedia.org/wiki/List_of_charitable_foundations' def get_description(url): res = requests.get(url) soup = bs(res.text, 'html.parser') body = [] for item in soup.find_all("p"): if item.text.startswith("The history"): break body.append(item.text) return ' '.join(body) res = requests.get(CHARITY_URL) soup = bs(res.text, "html.parser") charities = {} content = soup.find('div', {'id': 'content'}) for link in content.find_all("a"): url = link.get("href", "") if "/wiki" in url and url.count('/') == 2 and ':' not in url: charities[link.text.strip()] = "%s%s" % (BASE_URL, url) with open('charities.csv', 'w') as f: csv_file = csv.writer(f, delimiter='\t') names = list(charities.keys()) print('found', len(names)) for i, charity_name in enumerate(names): url = charities[charity_name] description = get_description(url) row = [charity_name, url, description] print('row', row) csv_file.writerow(row) f.close()
[ "chrisdistrict@gmail.com" ]
chrisdistrict@gmail.com
0eb944d3d4b625dd58953dcd5ad39efa5bcaeaa1
9c14bd53c8629262b1310962c1663a3c503ba3a0
/projects/golem/tests/project/add_directory_to_pages.py
7b7d283d6d17e02f6e630f4b7d7aad6a000fea95
[]
no_license
ShubhamAnand/golem-demo
b083d44b5d2d5db79eae96aa5bb1f3307272d64b
a40ced5500b3bfdb54351393eeb8ccba19a50564
refs/heads/master
2021-07-16T00:44:57.663282
2017-10-22T22:56:25
2017-10-22T22:56:25
null
0
0
null
null
null
null
UTF-8
Python
false
false
521
py
description = 'Verify that the user can add a directory in the pages section by appending \'\\\' at the end' pages = ['login', 'index', 'project'] def setup(data): navigate(data.env.url) login.do_login('admin', 'admin') index.create_access_project('test') def test(data): store('directory_name', random('ccccc')) project.add_page_directory(data.directory_name) project.verify_page_directory_exists(data.directory_name) def teardown(data): close()
[ "feo.luciano@gmail.com" ]
feo.luciano@gmail.com
e3e98cd7aaf3e9e63df6dd813cbada7c0adbbda5
5e6a67ae273bcbd55dfe0a77d4c839a2e554d3bf
/tree_parsing.py
ee094aed389950a95db950770158e1ce95218d4e
[]
no_license
Prabhanjan-Upadhya/Quora_Challenges
d54dd1e20c9c70f07b68fffe0c67e0d71290a786
d26ddce869205d90b367f88a2dd9c3c11b2466a7
refs/heads/master
2021-01-09T20:33:31.220380
2016-06-27T17:33:31
2016-06-27T17:33:31
62,074,863
2
0
null
null
null
null
UTF-8
Python
false
false
832
py
from collections import deque def parser(string1, tree_dict): start = 0 root = None prev = None lst1 = [] for i in range (0, len(string1)): if string1[i] == ' ': print tree_dict topic = string1[start: i] if topic == '(': if len(tree_dict[prev]) > 0: prev = tree_dict[prev][-1] lst1.append(prev) start = i+1 continue if topic == ')': del lst1[-1] if len(lst1) > 0: prev = lst1[-1] start = i+1 continue if root == None: root = topic tree_dict[topic]= [] prev = root lst1.append(root) else: tree_dict[topic] = [] tree_dict[prev].append(topic) start = i+1 return tree_dict def main(): string1 = "Animals ( Reptiles Birds ( Eagles Pigeons Crows ) )" tree_dict = {} tree_dict = parser(string1, tree_dict) print tree_dict main()
[ "pupadhy@clemson.edu" ]
pupadhy@clemson.edu
6f25ab872e193cdb10b2e82ee3a0967273771d8c
09e57dd1374713f06b70d7b37a580130d9bbab0d
/data/p3BR/R1/benchmark/startQiskit_noisy99.py
63b18b1983cf2ea49aadbaea3bb289271ef74cae
[ "BSD-3-Clause" ]
permissive
UCLA-SEAL/QDiff
ad53650034897abb5941e74539e3aee8edb600ab
d968cbc47fe926b7f88b4adf10490f1edd6f8819
refs/heads/main
2023-08-05T04:52:24.961998
2021-09-19T02:56:16
2021-09-19T02:56:16
405,159,939
2
0
null
null
null
null
UTF-8
Python
false
false
5,183
py
# qubit number=3 # total number=18 import numpy as np from qiskit import QuantumCircuit, execute, Aer, QuantumRegister, ClassicalRegister, transpile, BasicAer, IBMQ from qiskit.visualization import plot_histogram from typing import * from pprint import pprint from math import log2 from collections import Counter from qiskit.test.mock import FakeVigo, FakeYorktown kernel = 'circuit/bernstein' def bitwise_xor(s: str, t: str) -> str: length = len(s) res = [] for i in range(length): res.append(str(int(s[i]) ^ int(t[i]))) return ''.join(res[::-1]) def bitwise_dot(s: str, t: str) -> str: length = len(s) res = 0 for i in range(length): res += int(s[i]) * int(t[i]) return str(res % 2) def build_oracle(n: int, f: Callable[[str], str]) -> QuantumCircuit: # implement the oracle O_f # NOTE: use multi_control_toffoli_gate ('noancilla' mode) # https://qiskit.org/documentation/_modules/qiskit/aqua/circuits/gates/multi_control_toffoli_gate.html # https://quantumcomputing.stackexchange.com/questions/3943/how-do-you-implement-the-toffoli-gate-using-only-single-qubit-and-cnot-gates # https://quantumcomputing.stackexchange.com/questions/2177/how-can-i-implement-an-n-bit-toffoli-gate controls = QuantumRegister(n, "ofc") target = QuantumRegister(1, "oft") oracle = QuantumCircuit(controls, target, name="Of") for i in range(2 ** n): rep = np.binary_repr(i, n) if f(rep) == "1": for j in range(n): if rep[j] == "0": oracle.x(controls[j]) oracle.mct(controls, target[0], None, mode='noancilla') for j in range(n): if rep[j] == "0": oracle.x(controls[j]) # oracle.barrier() # oracle.draw('mpl', filename=(kernel + '-oracle.png')) return oracle def build_circuit(n: int, f: Callable[[str], str]) -> QuantumCircuit: # implement the Bernstein-Vazirani circuit zero = np.binary_repr(0, n) b = f(zero) # initial n + 1 bits input_qubit = QuantumRegister(n+1, "qc") classicals = ClassicalRegister(n, "qm") prog = QuantumCircuit(input_qubit, classicals) # inverse last one (can be omitted if using O_f^\pm) prog.x(input_qubit[n]) # circuit begin prog.h(input_qubit[1]) # number=1 prog.rx(-0.09738937226128368,input_qubit[2]) # number=2 prog.h(input_qubit[1]) # number=3 # apply H to get superposition for i in range(n): prog.h(input_qubit[i]) prog.h(input_qubit[n]) prog.barrier() # apply oracle O_f oracle = build_oracle(n, f) prog.append( oracle.to_gate(), [input_qubit[i] for i in range(n)] + [input_qubit[n]]) # apply H back (QFT on Z_2^n) for i in range(n): prog.h(input_qubit[i]) prog.barrier() # measure return prog def get_statevector(prog: QuantumCircuit) -> Any: state_backend = Aer.get_backend('statevector_simulator') statevec = execute(prog, state_backend).result() quantum_state = statevec.get_statevector() qubits = round(log2(len(quantum_state))) quantum_state = { "|" + np.binary_repr(i, qubits) + ">": quantum_state[i] for i in range(2 ** qubits) } return quantum_state def evaluate(backend_str: str, prog: QuantumCircuit, shots: int, b: str) -> Any: # Q: which backend should we use? # get state vector quantum_state = get_statevector(prog) # get simulate results # provider = IBMQ.load_account() # backend = provider.get_backend(backend_str) # qobj = compile(prog, backend, shots) # job = backend.run(qobj) # job.result() backend = Aer.get_backend(backend_str) # transpile/schedule -> assemble -> backend.run results = execute(prog, backend, shots=shots).result() counts = results.get_counts() a = Counter(counts).most_common(1)[0][0][::-1] return { "measurements": counts, # "state": statevec, "quantum_state": quantum_state, "a": a, "b": b } def bernstein_test_1(rep: str): """011 . x + 1""" a = "011" b = "1" return bitwise_xor(bitwise_dot(a, rep), b) def bernstein_test_2(rep: str): """000 . x + 0""" a = "000" b = "0" return bitwise_xor(bitwise_dot(a, rep), b) def bernstein_test_3(rep: str): """111 . x + 1""" a = "111" b = "1" return bitwise_xor(bitwise_dot(a, rep), b) if __name__ == "__main__": n = 2 a = "11" b = "1" f = lambda rep: \ bitwise_xor(bitwise_dot(a, rep), b) prog = build_circuit(n, f) sample_shot =4000 writefile = open("../data/startQiskit_noisy99.csv", "w") # prog.draw('mpl', filename=(kernel + '.png')) backend = FakeYorktown() circuit1 = transpile(prog, FakeYorktown()) circuit1.h(qubit=2) circuit1.x(qubit=3) circuit1.measure_all() info = execute(circuit1,backend=backend, shots=sample_shot).result().get_counts() print(info, file=writefile) print("results end", file=writefile) print(circuit1.depth(), file=writefile) print(circuit1, file=writefile) writefile.close()
[ "wangjiyuan123@yeah.net" ]
wangjiyuan123@yeah.net
0c27bf906d4b831769bc6b193c48e89a8aafe32d
445368830cc2918f71cdca166e1363c4ac08c56e
/D12_Recursive_power.py
9ebb7a2c5176e9bdbb0be2f62e51f63275e48409
[]
no_license
Chehlarov/Python-Advanced
fe9cea7c600ca064db89bc11af43d944a4385b74
4bfddcfb02d59385d0083a5acbf338beab136a8e
refs/heads/main
2023-03-13T08:53:03.936039
2021-03-03T17:16:02
2021-03-03T17:16:02
336,066,373
0
0
null
null
null
null
UTF-8
Python
false
false
171
py
def recursive_power(number, power): if power == 1: return number return number * recursive_power(number, power - 1) print(recursive_power(2, 10))
[ "noreply@github.com" ]
Chehlarov.noreply@github.com
ef6d97f0f5f530a0ebddca253af70a73492955c8
61b83fb5d4e9e09bddf1a90f2d2ac34962df02db
/elitedata/fixtures/__init__.py
d9d0fb4b2ca6383d00f26ea2fb1dbfa8ffdf7b22
[ "MIT" ]
permissive
jingyuyao/elitetraderoutes
0577b9cfa06e041507a091b751bea03a2e419143
a5b9ac9fba89b594843e527962b0b8539157cfd2
refs/heads/master
2021-01-21T09:28:40.626322
2016-03-27T00:12:44
2016-03-27T00:12:44
36,675,658
0
0
null
null
null
null
UTF-8
Python
false
false
26
py
__author__ = 'Jingyu_Yao'
[ "yjy.emails@gmail.com" ]
yjy.emails@gmail.com
6981c7c8e8c5f58f7815adf008a36b87eb6bc2e3
cd93eed33b120d632890a19e692ad318223162ed
/django_doc/manage.py
dd6b50b80fb33de2cc53a024a30dd68a74614073
[]
no_license
tuyuqi/Django_git_project
0c78bafccf3091284ea325ee8872b452472158d1
01d68a16b8f8b66c5dddda2672f05917b397e230
refs/heads/master
2020-04-21T20:08:54.418385
2019-02-10T07:35:51
2019-02-10T07:35:51
169,833,029
0
0
null
null
null
null
UTF-8
Python
false
false
542
py
#!/usr/bin/env python import os import sys if __name__ == '__main__': os.environ.setdefault('DJANGO_SETTINGS_MODULE', 'django_doc.settings') try: from django.core.management import execute_from_command_line except ImportError as exc: raise ImportError( "Couldn't import Django. Are you sure it's installed and " "available on your PYTHONPATH environment variable? Did you " "forget to activate a virtual environment?" ) from exc execute_from_command_line(sys.argv)
[ "yuqi.tu@gmail.com" ]
yuqi.tu@gmail.com
d3a08515917f144e92f34f6bfbc54f112d6bbd36
d58d8affb44c98db7af8264ebbda303dc0271bcb
/python/selenium/Browser_ipvt.py
4f1a908b608a1fce8052fd1163890bd06d324d78
[]
no_license
jijunjun1112/python
5946504ffcc6235f57a535dc6a978e0f8e3ba43f
5753fcf8d1f20d9cdb441cbf44509edb32c5fcec
refs/heads/master
2021-01-01T04:56:39.374901
2016-06-02T11:44:20
2016-06-02T11:44:20
56,660,838
0
0
null
null
null
null
UTF-8
Python
false
false
1,631
py
#coding:utf-8 from selenium import webdriver from selenium.webdriver.common.action_chains import ActionChains #引入ActionChains鼠标操作类 from selenium.webdriver.common.keys import Keys #引入keys类操作 import time import sys class Browser_ipvt(object): """This is user class.""" url="" browser="" conf_url="" def scheduleconfer(self): self.browser.find_element_by_id("scheduleMeeting").click() time.sleep(1) self.browser.find_element_by_id("meetingTheme").send_keys("sunny") self.browser.find_element_by_id("proAccount").send_keys("8200433") self.browser.find_element_by_xpath(".//*[@class='btn btn-blue meetingSubmit']").click() time.sleep(1) def __init__(self, url): self.url = url def showClassName(self): print self.__class__.__name__ def showClassDoc(self): print self.__class__.__doc__ def quitBrowser(self): self.browser.quit() print "Now quit the browser" def openBrowser(self): self.browser= webdriver.Firefox() self.browser.get(self.url) self.browser.maximize_window() time.sleep(1) print "Open gvc client web!" def login(self): self.browser.find_element_by_id("uerName").send_keys("jijunjun1112") self.browser.find_element_by_id("password").send_keys("Jun13676830606") self.browser.find_element_by_xpath(".//*[@id='loginForm']/div[4]/button").click() time.sleep(1) print "Has already login in as jijunjun1112!" def main(): browser_ipvt=Browser_ipvt('http://account.ipvideotalk.com/login/') browser_ipvt.openBrowser() browser_ipvt.login() browser_ipvt.scheduleconfer() browser_ipvt.quitBrowser() if __name__ == '__main__': main()
[ "jjji@grandstream.cn" ]
jjji@grandstream.cn
3563587f0871cbdbbb1a1a1c326f78a1f00057a5
247342c16db02f52f52e781afae8aaa2ccc467d7
/translate.py
71e049c68d51d4af88c600d91b1534d21b18fd4e
[]
no_license
twadada/Machine-Translation
fb7cd6599132efbdd1f6cb7ea4adf4f8bd4c033a
5619f1276ae2a8954edacc5b6532cfe380921837
refs/heads/master
2020-03-25T19:36:25.347588
2018-08-14T11:42:29
2018-08-14T11:42:29
142,395,971
0
0
null
null
null
null
UTF-8
Python
false
false
6,978
py
import os import pickle import argparse import chainer from models.preprocessing import Convert_word2id,np from tqdm import tqdm from train import NMT # src_test=/cl/work/takashi-w/ASPEC-JE/test/test.txt.tok.low.ja # model=Results/ASPEC.encoder_decoder_epoch3_ppl_56.599_best_model # python translate.py -gpuid 1 -model $model -src_test $src_test -out_attention_weight -k 3 -beam_size 5 parser = argparse.ArgumentParser() parser.add_argument( '-gpuid', type=int, default=-1, help='gpuid; -1 means using cpu (default)') parser.add_argument( '-src_test', type=str, required=True, help='source test data path') parser.add_argument( '-model', type=str, required=True, help='model path') parser.add_argument( '-beam_size', type=int, default=1, help='beam size: 1 means greedy decoding (default)') parser.add_argument( '-normalize', action='store_true', help='normalize decoding probability by length') parser.add_argument( '-out_attention_weight', action='store_true', help='output attention weight heatmap') parser.add_argument( '-k', type=int, default=0, help='output additional txt file that lists top k translation candidates for each soruce sentence' '\n\n; set k larger than 0 to enable this option; (defalut: disabled)') class Translator(chainer.Chain): def __init__(self, encoder, decoder, id2vocab, model): super().__init__() with self.init_scope(): self.encoder = encoder self.decoder = decoder self.id2vocab = id2vocab self.model = model def translate_base(self, s_id, s_lengths, beam_size, normalize, *args): EOS_id = 0 translation_best_list = [] translation_all_list = [] attention_weight_list = [] print("translating") for k in tqdm(range(len(s_id))): h_last, c_last, hs = self.encoder([s_id[k]], [s_lengths[k]]) # (bt * s_len * enc_size) translation_best, prob_best, translation, translation_prob, attention_p_out = self.decoder.translate(h_last, c_last, hs, beam_size, EOS_id, normalize, *args) translation_best_list.append(translation_best) translation_all_list.append(translation) attention_weight_list.append(attention_p_out[0]) return translation_best_list,translation_all_list, attention_weight_list def translate(self, s_id, s_lengths, beam_size, normalize, out_attention_weight, k, *args): translation_best_list, translation_all_list, attention_weight_list \ = self.translate_base(s_id, s_lengths, beam_size, normalize,*args) #save outputs print("save translations") translation_word = translator.save_transation(translation_best_list) if (k>1): print("save top k translation candidates") translator.save_topk_translations(translation_all_list, k) if (out_attention_weight): print("save attention weight heatmap") translator.save_attn_weight(s_id, translation_word, attention_weight_list) def save_transation(self, translation_list): translation_word = [] tgt_id2vocab = self.id2vocab[1] f = open(self.model + ".translation.txt", "w+") for i in range(len(translation_list)): translation = translation_list[i] translation = " ".join([tgt_id2vocab[word_id] for word_id in translation]) translation_word.append(translation) f.write(translation + "\n") f.close() return translation_word def save_topk_translations(self, topk_translation, k): tgt_id2vocab = self.id2vocab[1] f = open(self.model + ".translation_top"+str(k)+".txt", "w+") for i in range(len(topk_translation)): translation_list = topk_translation[i][0:k] for sentence in translation_list: translation = " ".join([tgt_id2vocab[word_id] for word_id in sentence]) f.write(translation + "\n") f.write('\n') f.close() def save_attn_weight(self, s_id, translation_word, attn_wight): src_id2vocab = self.id2vocab[0] pdf_pages = PdfPages(self.model+ ".attn_W.pdf") for k in tqdm(range(len(attn_wight))): s_id[k] =s_id[k].tolist() #np/cupy_array -> list attn_wight_tmp = attn_wight[k] # t_len, s_len translation = translation_word[k].split() + ["<\s>"] attn_wight_tmp = np.round(attn_wight_tmp, 2) # s_len * bt * 1 * 5 (= window +1) x_labels = [src_id2vocab[s_id[k][j]] for j in range(len(s_id[k]))] y_labels = translation plt.figure(figsize=(len(x_labels) * 0.3, len(y_labels) * 0.3)) ax = sns.heatmap(attn_wight_tmp, cbar=False, vmin=0, vmax=1, cmap="Reds") ax.set_xticklabels(x_labels, rotation=90) ax.set_yticklabels(y_labels, rotation=0) plt.tight_layout() pdf_pages.savefig() #plt.savefig(self.model+ ".attn_W" + str(k) + ".pdf") plt.close('all') pdf_pages.close() # merger = PdfFileMerger() #merge all pdfs into one file # for k in range(len(s_id)): # merger.append(open(self.model+ ".attn_W" + str(k) + ".pdf", 'rb')) # os.remove(self.model+ ".attn_W" + str(k) + ".pdf") # with open(self.model+ ".attn_W.pdf", 'wb') as fout: # merger.write(fout) if __name__ == '__main__': opt = parser.parse_args() if(opt.out_attention_weight): import matplotlib matplotlib.use('Agg') font = {'family': 'IPAexGothic'} matplotlib.rc('font', **font) import seaborn as sns import matplotlib.pyplot as plt from matplotlib.backends.backend_pdf import PdfPages from PyPDF2 import PdfFileMerger print("beam_size: ",opt.beam_size) print("normalize: ",opt.normalize) if(opt.k > opt.beam_size): raise Exception("k must not be larger than beam size") file = open(opt.model, 'rb') model = pickle.load(file) test_lines_id, test_sent_length = Convert_word2id(opt.src_test, model.vocab2id[0]) model.to_cpu() if opt.gpuid >= 0: chainer.cuda.get_device_from_id(opt.gpuid).use() to_gpu = chainer.cuda.to_gpu model.to_gpu() ####numpy->cupy#### for i in range(len(test_lines_id)): test_lines_id[i] = to_gpu(test_lines_id[i]) # list of xp.array translator = Translator(model.encoder,model.decoder, model.id2vocab, opt.model) translator.translate(test_lines_id, test_sent_length, opt.beam_size, opt.normalize, opt.out_attention_weight, opt.k)
[ "takashi.wada.764@gmail.com" ]
takashi.wada.764@gmail.com
4034bde7a9e06c5d7487997a7acb9e10b85cca2b
0f1001169c4f229c253a6f1dc1c9aff51c797cca
/docs/markdown_to_html.py
ffacda661ea31a8286a001a77d5178f08b9a1fd3
[ "Apache-2.0" ]
permissive
alencon/dash-bootstrap-components
f40e360787c96a1d9f7827cf042872b2f9cffcac
4f39856c13f66730512c57ed6dc0a819e8629293
refs/heads/master
2023-01-22T13:07:05.880865
2020-12-03T21:25:50
2020-12-03T21:25:50
318,998,227
1
0
Apache-2.0
2020-12-06T09:42:13
2020-12-06T09:42:13
null
UTF-8
Python
false
false
1,541
py
from pathlib import Path import markdown from markdown.extensions.fenced_code import FencedBlockPreprocessor # highlightJS expects the class "language-*" but markdown default is "*" FencedBlockPreprocessor.LANG_TAG = ' class="language-%s"' CONTENT = Path(__file__).parent / "content" DEST = Path(__file__).parent / "templates" / "generated" DOCS_HTML_TEMPLATE = """{% extends "docs.html" %} {% block title %}<title><TITLE></title>{% endblock %} {% block content %}<CONTENT>{% endblock %} """ CHANGELOG_HTML_TEMPLATE = """{% extends "changelog.html" %} {% block title %}<title><TITLE></title>{% endblock %} {% block content %}<CONTENT>{% endblock %} """ def convert_all_markdown_files(): for path in CONTENT.glob("docs/*.md"): template = template_from_markdown(path, title_suffix=" - dbc docs") with open(DEST / "docs" / path.name.replace(".md", ".html"), "w") as f: f.write(template) for path in CONTENT.glob("*.md"): template = template_from_markdown( path, template=CHANGELOG_HTML_TEMPLATE ) with open(DEST / path.name.replace(".md", ".html"), "w") as f: f.write(template) def template_from_markdown(path, title_suffix="", template=DOCS_HTML_TEMPLATE): md = markdown.Markdown(extensions=["fenced_code", "meta"]) text = path.read_text() template = template.replace("<CONTENT>", md.convert(text)) return template.replace("<TITLE>", f"{md.Meta['title'][0]} - dbc docs") if __name__ == "__main__": convert_all_markdown_files()
[ "tomcbegley@gmail.com" ]
tomcbegley@gmail.com
20e3b3314110b5f76c4cbd58f4de848416ac8082
35b5ab3c9863f1f151b600f040904ef83a7ba4d0
/test.py
a08e862b5bcd80c024840892c084bf79e7c09232
[]
no_license
aaron-alphonsus/style-transfer-flask
5d9120f93aa16656d9f59062e268cba977d8dadf
615d5a2a64ada75352018708a626d7abce5a0467
refs/heads/master
2020-06-24T21:15:19.997826
2019-07-29T20:26:05
2019-07-29T20:26:05
199,092,710
0
0
null
2019-07-26T23:42:42
2019-07-26T23:42:42
null
UTF-8
Python
false
false
8,329
py
# from google.cloud import storage import os from flask import request, send_file import tempfile # import cloudstorage as gcs # from google.appengine.api import images, app_identity # STYLE_URL = "/style-input-images-1/Vassily_Kandinsky,_1913_-_Composition_7.jpg" # f = gcs.open(STYLE_URL) # data = f.read() # print(data) # from object_detection import load_object # from style_transfer import run_style_transfer import os import numpy as np from six.moves.urllib.request import urlopen import tarfile # def download_blob(bucket_name, source_blob_name, destination_file_name): # """Downloads a blob from the bucket.""" # storage_client = storage.Client() # bucket = storage_client.get_bucket(bucket_name) # blob = bucket.blob(source_blob_name) # blob.download_to_filename(destination_file_name) # print('Blob {} downloaded to {}.'.format( # source_blob_name, # destination_file_name)) # def DownloadCheckpointFiles(checkpoint_dir=os.path.abspath("")): # """Download checkpoint files if necessary.""" # full_checkpoint = "https://download.magenta.tensorflow.org/models/arbitrary_style_transfer.tar.gz" # # url_prefix = 'http://download.magenta.tensorflow.org/models/' # # checkpoints = ['multistyle-pastiche-generator-monet.ckpt', 'multistyle-pastiche-generator-varied.ckpt'] # # for checkpoint in checkpoints: # # full_checkpoint = os.path.join(checkpoint_dir, checkpoint) # # checkpoint_dir = 'arbitrary_style_transfer' # if not os.path.exists(checkpoint_dir): # print('Downloading {}'.format(full_checkpoint)) # filename = full_checkpoint.split("/")[-1] # with open(filename, "wb") as f: # r = requests.get(full_checkpoint) # f.write(r.content) def DownloadCheckpointFiles(checkpoint_dir=os.path.abspath("")): """Download checkpoint files if necessary.""" url_prefix = 'http://download.magenta.tensorflow.org/models/' checkpoints = ['arbitrary_style_transfer.tar.gz'] path = 'arbitrary_style_transfer' for checkpoint in checkpoints: full_checkpoint = os.path.join(checkpoint_dir, checkpoint) if not os.path.exists(path): print('Downloading {}'.format(full_checkpoint)) response = urlopen(url_prefix + checkpoint) data = response.read() with open(full_checkpoint, 'wb') as fh: fh.write(data) unzip_tar_gz() def unzip_tar_gz(): tf = tarfile.open('arbitrary_style_transfer.tar.gz',"r:gz") tf.extractall() tf.close() DownloadCheckpointFiles() # unzip_tar_gz() # DownloadCheckpointFiles() # detections =[[ 0.5210921 0.68086916 0.6457864 0.81711274 55 0.99862325] # [ 0.40780655 0.58433187 0.52803445 0.71685034 55 0.99838376] # [ 0.41141948 0.30821186 0.5301136 0.43577647 55 0.9982496 ] # [ 0.5196079 0.53558177 0.64213705 0.66894466 55 0.99801254] # [ 0.41874808 0.44184786 0.5332428 0.57064563 55 0.99788 ] # [ 0.51584554 0.37531173 0.65214545 0.5233197 55 0.9976165 ] # [ 0.3246436 0.3725551 0.42766646 0.48901382 55 0.9958526 ] # [ 0. 0. 0. 0. 0. 0. ] # [ 0. 0. 0. 0. 0. 0. ] # [ 0. 0. 0. 0. 0. 0. ] # [ 0. 0. 0. 0. 0. 0. ] # [ 0. 0. 0. 0. 0. 0. ] # [ 0. 0. 0. 0. 0. 0. ] # [ 0. 0. 0. 0. 0. 0. ] # [ 0. 0. 0. 0. 0. 0. ] # [ 0. 0. 0. 0. 0. 0. ] # [ 0. 0. 0. 0. 0. 0. ] # [ 0. 0. 0. 0. 0. 0. ] # [ 0. 0. 0. 0. 0. 0. ] # [ 0. 0. 0. 0. 0. 0. ] # [ 0. 0. 0. 0. 0. 0. ] # [ 0. 0. 0. 0. 0. 0. ] # [ 0. 0. 0. 0. 0. 0. ] # [ 0. 0. 0. 0. 0. 0. ] # [ 0. 0. 0. 0. 0. 0. ] # [ 0. 0. 0. 0. 0. 0. ] # [ 0. 0. 0. 0. 0. 0. ] # [ 0. 0. 0. 0. 0. 0. ] # [ 0. 0. 0. 0. 0. 0. ] # ] # zero_index = np.where(detections[:, 4] == 0)[0] # print(zero_index) # # print(os.path.basename( # # 'gs: // style-input-images-1/all_objects.jpg')) # # style = "static/style_images/Vassily_Kandinsky,_1913_-_Composition_7.jpg" # # content = "static/input_images/styled.jpg" # # test = "arbitrary_image_stylization_with_weights \ # # --checkpoint=arbitrary_style_transfer/model.ckpt \ # # --output_dir=outputs \ # # --style_images_paths="+style+"\ # # --content_images_paths="+content+"\ # # --image_size=256 \ # # --content_square_crop=False \ # # --style_image_size=256 \ # # --style_square_crop=False \ # # --logtostderr" # # path = os.system(test) # # print(path) # # from object_detection import blending # crop_path = 'static/blending/crop.jpg' # original_path = 'static/blending/original.jpg' # style_path = 'static/blending/original_stylized_Vassily_Kandinsky,_1913_-_Composition_7_0.jpg' # print(os.path.join(os.path.abspath(""), style_path)) # print(blending(crop_path, original_path, style_path)) # test1 = None # def test(): # global test1 # test1 = 'TEST' # return # if __name__ == "__main__": # test() # print(test1) # style_file = os.path.join(os.path.abspath(''), 'stylize.jpg') # model = os.path.join(os.path.abspath(''), 'rcnn_model.pkl') # show_objects = load_object(style_file, model) # style_file = request.files['styled.jpg'] # print(style_file) # style_file = os.path.abspath('../Flask-STWA/stylize.jpg') # print(os.path.join(os.path.abspath(''), 'stylize.jpg')) # CLOUD_STORAGE_BUCKET = 'style-input-images-1' # def upload_to_gcloud(file, destination_blob_name): # # style_path = request.files['style_file'] # storage_client = storage.Client(project='amli-245518') # bucket = storage_client.get_bucket(CLOUD_STORAGE_BUCKET) # blob = bucket.blob(destination_blob_name) # blob.upload_from_filename(file) # return blob.public_url # style_file = os.path.join(os.path.abspath(''), 'styled.jpg') # url = upload_to_gcloud(style_file, 'styled.jpg') # print(url) # style_file = os.path.join(os.path.abspath(''), 'stylize.jpg') # CLOUD_STORAGE_BUCKET = 'style-input-images-1' # best, best_loss = run_style_transfer( # content_path, style_path, num_iterations=3) # storage_client = storage.Client(project='amli-245518') # bucket = storage_client.get_bucket(CLOUD_STORAGE_BUCKET) # destination_blob_name = 'styled-test.jpg' # blob = bucket.blob(destination_blob_name) # blob.upload_from_filename(style_file) # def upload(): # style_file = os.path.join(os.path.abspath(''), 'stylize.jpg') # CLOUD_STORAGE_BUCKET = 'style-input-images-1' # best, best_loss = run_style_transfer( # content_path, style_path, num_iterations=3) # storage_client = storage.Client(project='amli-245518') # bucket = storage_client.get_bucket(CLOUD_STORAGE_BUCKET) # destination_blob_name = 'stylize.jpg' # blob = bucket.blob(destination_blob_name) # blob.upload_from_filename(style_file) # def download(): # CLOUD_STORAGE_BUCKET = 'style-input-images-1' # style_file = 'stylize.jpg' # # style_file = os.path.abspath('../Flask-STWA/stylize.jpg') # client = storage.Client() # bucket = client.get_bucket(CLOUD_STORAGE_BUCKET) # blob = bucket.blob(style_file) # filename = 'new.jpg' # blob.download_to_filename(filename) # # return send_file(temp.name, attachment_filename=style_file)
[ "huhuang@cs.hmc.edu" ]
huhuang@cs.hmc.edu
17c59ae366737608237c203cd7a63e511f3e1002
13da61043e182789cfe8d67e0e1f8a4d4d983e96
/src/cms/views/settings/user_settings_view.py
6e83e61cde7860e8a32ec2c28cfd3c1b2c0324eb
[ "Apache-2.0" ]
permissive
digitalfabrik/coldaid-backend
b9af0f9129a974e92bc89003c4266d8e48539d0f
b769510570d5921e30876565263813c0362994e2
refs/heads/develop
2020-09-25T21:48:20.846809
2020-06-17T15:48:53
2020-06-17T15:48:53
226,094,918
4
2
Apache-2.0
2020-06-17T15:48:54
2019-12-05T12:15:01
HTML
UTF-8
Python
false
false
3,005
py
from django.contrib import messages from django.contrib.auth import update_session_auth_hash from django.contrib.auth.decorators import login_required from django.utils.decorators import method_decorator from django.utils.translation import ugettext as _ from django.views.generic import TemplateView from django.shortcuts import render, redirect from django.views.decorators.cache import never_cache from ...forms.users import UserEmailForm, UserPasswordForm @method_decorator(login_required, name='dispatch') class UserSettingsView(TemplateView): template_name = 'settings/user.html' @never_cache def get(self, request, *args, **kwargs): user = request.user user_email_form = UserEmailForm(instance=user) user_password_form = UserPasswordForm(instance=user) return render( request, self.template_name, { 'keys': user.mfa_keys.all(), 'user_email_form': user_email_form, 'user_password_form': user_password_form, } ) # pylint: disable=unused-argument, too-many-branches def post(self, request, *args, **kwargs): user = request.user if request.POST.get('submit_form') == 'email_form': user_email_form = UserEmailForm( request.POST, instance=user ) if not user_email_form.is_valid(): # Add error messages for field in user_email_form: for error in field.errors: messages.error(request, _(error)) for error in user_email_form.non_field_errors(): messages.error(request, _(error)) elif not user_email_form.has_changed(): messages.info(request, _('No changes detected.')) else: user_email_form.save() messages.success(request, _('E-mail-address was successfully saved.')) elif request.POST.get('submit_form') == 'password_form': user_password_form = UserPasswordForm( request.POST, instance=user ) if not user_password_form.is_valid(): # Add error messages for field in user_password_form: for error in field.errors: messages.error(request, _(error)) for error in user_password_form.non_field_errors(): messages.error(request, _(error)) elif not user_password_form.has_changed(): messages.info(request, _('No changes detected.')) else: user = user_password_form.save() # Prevent user from being logged out after password has changed update_session_auth_hash(request, user) messages.success(request, _('Password was successfully saved.')) return redirect('user_settings')
[ "ludwig+secure@integreat-app.de" ]
ludwig+secure@integreat-app.de
2400de35f3a6c6902ae173e097d54b31040a551a
2cbf3aaad62f4922d827af658fb5dbb7ac651bef
/teledusite/teledu/models/conceptAttribute.py
fc12d964e90614a6ff7813077017d177a3c7fecb
[]
no_license
tctimmeh/teledu
0266240aa864cd2eed75857e66eaeb8270f44c1a
04135ffb04f397f29152ca48f868a957b18d504a
refs/heads/master
2021-01-23T08:52:32.817693
2013-10-29T01:34:41
2013-10-29T01:34:41
2,566,893
0
0
null
null
null
null
UTF-8
Python
false
false
674
py
from django.db import models from concept import Concept from attribute import Attribute class ConceptAttribute(Attribute): concept = models.ForeignKey(Concept, related_name = 'attributes') class Meta: app_label = 'teledu' unique_together = (('concept', 'name')) def __unicode__(self): return '%s - %s' % (self.concept.name, self.name) def conceptName(self): return self.concept.name def gameSystem(self): return self.concept.gameSystem def getAttributeValuesForInstance(self, instance): from conceptAttributeValue import ConceptAttributeValue return ConceptAttributeValue.objects.filter(attribute = self, instance = instance)
[ "tctimmeh@gmail.com" ]
tctimmeh@gmail.com
7e625fa71bfa5efc76d0547775d9fb3a2dad6b0b
a1b308d27c922bfa66137151c93b817c7667c4b2
/src/chatbot/cfg/Chatbot.cfg
db50fd43913f3a9ef49e554ad40a58952d093c60
[ "MIT" ]
permissive
dergkat/HEAD
64e81390ffcd04795ef8414376b2d12813e4e42c
0e1a034dbdc052e5e573787f49c9f36a01459373
refs/heads/master
2021-05-01T05:45:17.810469
2017-06-07T09:01:19
2017-06-07T09:01:19
79,799,750
0
1
null
2017-01-23T11:50:33
2017-01-23T11:50:33
null
UTF-8
Python
false
false
1,484
cfg
#!/usr/bin/env python PACKAGE = 'chatbot' from dynamic_reconfigure.parameter_generator_catkin import * gen = ParameterGenerator() gen.add("enable", bool_t, 0, "Enable Chatbot", True) gen.add("chatbot_url", str_t, 0, "Chatbot Server URL", 'http://localhost:8001') gen.add("sentiment", bool_t, 0, "Enable Sentiment", True) gen.add("delay_response", bool_t, 0, "Delay Chatbot Response", False) gen.add("delay_time", double_t, 0, "Response Delay Time", 5, 0, 10) gen.add("ignore_indicator", bool_t, 0, "Ignore the indicator from response", False) gen.add("set_that", str_t, 0, "Set 'that' tag", '') gen.add("set_context", str_t, 0, "Add Chatbot Context (k=v,k2=v2,...)", '') gen.add("mute", bool_t, 0, "Mute the Chatbot", False) weights = gen.add_group("Weights", state=True) weights.add("sophia", double_t, 0, "Weight for Sophia tier", 0.9, 0, 1) weights.add("cs", double_t, 0, "Weight for ChatScript tier", 1, 0, 1) weights.add("generic", double_t, 0, "Weight for generic tier", 1, 0, 1) weights.add("early_pickup", double_t, 0, "Weight for early pickup tier", 0.2, 0, 1) weights.add("mid_pickup", double_t, 0, "Weight for middle pickup tier", 0.2, 0, 1) weights.add("ddg", double_t, 0, "Weight for duck duck go tier", 0.3, 0, 1) weights.add("markov", double_t, 0, "Weight for random sentense generator tier", 0.1, 0, 1) weights.add("reset", bool_t, 0, "Reset the weight to defaults", False) # package name, node name, config name exit(gen.generate(PACKAGE, "chatbot", "Chatbot"))
[ "dut.hww@gmail.com" ]
dut.hww@gmail.com
ccc456d17a7c5c5b509e388397e01ad74e2f0559
00a9295409b78a53ce790f7ab44931939f42c0e0
/FPGA/apio/iCEBreaker/FIR_Filter/sympy/venv/lib/python3.8/site-packages/sympy/solvers/tests/test_pde.py
1b43eb0b0886235a8a8b1b4a593e4e7d486fcfae
[ "Apache-2.0" ]
permissive
klei22/Tech-OnBoarding-Class
c21f0762d2d640d5e9cb124659cded5c865b32d4
960e962322c37be9117e0523641f8b582a2beceb
refs/heads/master
2022-11-10T13:17:39.128342
2022-10-25T08:59:48
2022-10-25T08:59:48
172,292,871
2
3
Apache-2.0
2019-05-19T00:26:32
2019-02-24T03:50:35
C
UTF-8
Python
false
false
9,057
py
from sympy import (Derivative as D, Eq, exp, sin, Function, Symbol, symbols, cos, log) from sympy.core import S from sympy.solvers.pde import (pde_separate, pde_separate_add, pde_separate_mul, pdsolve, classify_pde, checkpdesol) from sympy.testing.pytest import raises a, b, c, x, y = symbols('a b c x y') def test_pde_separate_add(): x, y, z, t = symbols("x,y,z,t") F, T, X, Y, Z, u = map(Function, 'FTXYZu') eq = Eq(D(u(x, t), x), D(u(x, t), t)*exp(u(x, t))) res = pde_separate_add(eq, u(x, t), [X(x), T(t)]) assert res == [D(X(x), x)*exp(-X(x)), D(T(t), t)*exp(T(t))] def test_pde_separate(): x, y, z, t = symbols("x,y,z,t") F, T, X, Y, Z, u = map(Function, 'FTXYZu') eq = Eq(D(u(x, t), x), D(u(x, t), t)*exp(u(x, t))) raises(ValueError, lambda: pde_separate(eq, u(x, t), [X(x), T(t)], 'div')) def test_pde_separate_mul(): x, y, z, t = symbols("x,y,z,t") c = Symbol("C", real=True) Phi = Function('Phi') F, R, T, X, Y, Z, u = map(Function, 'FRTXYZu') r, theta, z = symbols('r,theta,z') # Something simple :) eq = Eq(D(F(x, y, z), x) + D(F(x, y, z), y) + D(F(x, y, z), z), 0) # Duplicate arguments in functions raises( ValueError, lambda: pde_separate_mul(eq, F(x, y, z), [X(x), u(z, z)])) # Wrong number of arguments raises(ValueError, lambda: pde_separate_mul(eq, F(x, y, z), [X(x), Y(y)])) # Wrong variables: [x, y] -> [x, z] raises( ValueError, lambda: pde_separate_mul(eq, F(x, y, z), [X(t), Y(x, y)])) assert pde_separate_mul(eq, F(x, y, z), [Y(y), u(x, z)]) == \ [D(Y(y), y)/Y(y), -D(u(x, z), x)/u(x, z) - D(u(x, z), z)/u(x, z)] assert pde_separate_mul(eq, F(x, y, z), [X(x), Y(y), Z(z)]) == \ [D(X(x), x)/X(x), -D(Z(z), z)/Z(z) - D(Y(y), y)/Y(y)] # wave equation wave = Eq(D(u(x, t), t, t), c**2*D(u(x, t), x, x)) res = pde_separate_mul(wave, u(x, t), [X(x), T(t)]) assert res == [D(X(x), x, x)/X(x), D(T(t), t, t)/(c**2*T(t))] # Laplace equation in cylindrical coords eq = Eq(1/r * D(Phi(r, theta, z), r) + D(Phi(r, theta, z), r, 2) + 1/r**2 * D(Phi(r, theta, z), theta, 2) + D(Phi(r, theta, z), z, 2), 0) # Separate z res = pde_separate_mul(eq, Phi(r, theta, z), [Z(z), u(theta, r)]) assert res == [D(Z(z), z, z)/Z(z), -D(u(theta, r), r, r)/u(theta, r) - D(u(theta, r), r)/(r*u(theta, r)) - D(u(theta, r), theta, theta)/(r**2*u(theta, r))] # Lets use the result to create a new equation... eq = Eq(res[1], c) # ...and separate theta... res = pde_separate_mul(eq, u(theta, r), [T(theta), R(r)]) assert res == [D(T(theta), theta, theta)/T(theta), -r*D(R(r), r)/R(r) - r**2*D(R(r), r, r)/R(r) - c*r**2] # ...or r... res = pde_separate_mul(eq, u(theta, r), [R(r), T(theta)]) assert res == [r*D(R(r), r)/R(r) + r**2*D(R(r), r, r)/R(r) + c*r**2, -D(T(theta), theta, theta)/T(theta)] def test_issue_11726(): x, t = symbols("x t") f = symbols("f", cls=Function) X, T = symbols("X T", cls=Function) u = f(x, t) eq = u.diff(x, 2) - u.diff(t, 2) res = pde_separate(eq, u, [T(x), X(t)]) assert res == [D(T(x), x, x)/T(x),D(X(t), t, t)/X(t)] def test_pde_classify(): # When more number of hints are added, add tests for classifying here. f = Function('f') eq1 = a*f(x,y) + b*f(x,y).diff(x) + c*f(x,y).diff(y) eq2 = 3*f(x,y) + 2*f(x,y).diff(x) + f(x,y).diff(y) eq3 = a*f(x,y) + b*f(x,y).diff(x) + 2*f(x,y).diff(y) eq4 = x*f(x,y) + f(x,y).diff(x) + 3*f(x,y).diff(y) eq5 = x**2*f(x,y) + x*f(x,y).diff(x) + x*y*f(x,y).diff(y) eq6 = y*x**2*f(x,y) + y*f(x,y).diff(x) + f(x,y).diff(y) for eq in [eq1, eq2, eq3]: assert classify_pde(eq) == ('1st_linear_constant_coeff_homogeneous',) for eq in [eq4, eq5, eq6]: assert classify_pde(eq) == ('1st_linear_variable_coeff',) def test_checkpdesol(): f, F = map(Function, ['f', 'F']) eq1 = a*f(x,y) + b*f(x,y).diff(x) + c*f(x,y).diff(y) eq2 = 3*f(x,y) + 2*f(x,y).diff(x) + f(x,y).diff(y) eq3 = a*f(x,y) + b*f(x,y).diff(x) + 2*f(x,y).diff(y) for eq in [eq1, eq2, eq3]: assert checkpdesol(eq, pdsolve(eq))[0] eq4 = x*f(x,y) + f(x,y).diff(x) + 3*f(x,y).diff(y) eq5 = 2*f(x,y) + 1*f(x,y).diff(x) + 3*f(x,y).diff(y) eq6 = f(x,y) + 1*f(x,y).diff(x) + 3*f(x,y).diff(y) assert checkpdesol(eq4, [pdsolve(eq5), pdsolve(eq6)]) == [ (False, (x - 2)*F(3*x - y)*exp(-x/S(5) - 3*y/S(5))), (False, (x - 1)*F(3*x - y)*exp(-x/S(10) - 3*y/S(10)))] for eq in [eq4, eq5, eq6]: assert checkpdesol(eq, pdsolve(eq))[0] sol = pdsolve(eq4) sol4 = Eq(sol.lhs - sol.rhs, 0) raises(NotImplementedError, lambda: checkpdesol(eq4, sol4, solve_for_func=False)) def test_solvefun(): f, F, G, H = map(Function, ['f', 'F', 'G', 'H']) eq1 = f(x,y) + f(x,y).diff(x) + f(x,y).diff(y) assert pdsolve(eq1) == Eq(f(x, y), F(x - y)*exp(-x/2 - y/2)) assert pdsolve(eq1, solvefun=G) == Eq(f(x, y), G(x - y)*exp(-x/2 - y/2)) assert pdsolve(eq1, solvefun=H) == Eq(f(x, y), H(x - y)*exp(-x/2 - y/2)) def test_pde_1st_linear_constant_coeff_homogeneous(): f, F = map(Function, ['f', 'F']) u = f(x, y) eq = 2*u + u.diff(x) + u.diff(y) assert classify_pde(eq) == ('1st_linear_constant_coeff_homogeneous',) sol = pdsolve(eq) assert sol == Eq(u, F(x - y)*exp(-x - y)) assert checkpdesol(eq, sol)[0] eq = 4 + (3*u.diff(x)/u) + (2*u.diff(y)/u) assert classify_pde(eq) == ('1st_linear_constant_coeff_homogeneous',) sol = pdsolve(eq) assert sol == Eq(u, F(2*x - 3*y)*exp(-S(12)*x/13 - S(8)*y/13)) assert checkpdesol(eq, sol)[0] eq = u + (6*u.diff(x)) + (7*u.diff(y)) assert classify_pde(eq) == ('1st_linear_constant_coeff_homogeneous',) sol = pdsolve(eq) assert sol == Eq(u, F(7*x - 6*y)*exp(-6*x/S(85) - 7*y/S(85))) assert checkpdesol(eq, sol)[0] eq = a*u + b*u.diff(x) + c*u.diff(y) sol = pdsolve(eq) assert checkpdesol(eq, sol)[0] def test_pde_1st_linear_constant_coeff(): f, F = map(Function, ['f', 'F']) u = f(x,y) eq = -2*u.diff(x) + 4*u.diff(y) + 5*u - exp(x + 3*y) sol = pdsolve(eq) assert sol == Eq(f(x,y), (F(4*x + 2*y) + exp(x/S(2) + 4*y)/S(15))*exp(x/S(2) - y)) assert classify_pde(eq) == ('1st_linear_constant_coeff', '1st_linear_constant_coeff_Integral') assert checkpdesol(eq, sol)[0] eq = (u.diff(x)/u) + (u.diff(y)/u) + 1 - (exp(x + y)/u) sol = pdsolve(eq) assert sol == Eq(f(x, y), F(x - y)*exp(-x/2 - y/2) + exp(x + y)/S(3)) assert classify_pde(eq) == ('1st_linear_constant_coeff', '1st_linear_constant_coeff_Integral') assert checkpdesol(eq, sol)[0] eq = 2*u + -u.diff(x) + 3*u.diff(y) + sin(x) sol = pdsolve(eq) assert sol == Eq(f(x, y), F(3*x + y)*exp(x/S(5) - 3*y/S(5)) - 2*sin(x)/S(5) - cos(x)/S(5)) assert classify_pde(eq) == ('1st_linear_constant_coeff', '1st_linear_constant_coeff_Integral') assert checkpdesol(eq, sol)[0] eq = u + u.diff(x) + u.diff(y) + x*y sol = pdsolve(eq) assert sol == Eq(f(x, y), -x*y + x + y + F(x - y)*exp(-x/S(2) - y/S(2)) - 2) assert classify_pde(eq) == ('1st_linear_constant_coeff', '1st_linear_constant_coeff_Integral') assert checkpdesol(eq, sol)[0] eq = u + u.diff(x) + u.diff(y) + log(x) assert classify_pde(eq) == ('1st_linear_constant_coeff', '1st_linear_constant_coeff_Integral') def test_pdsolve_all(): f, F = map(Function, ['f', 'F']) u = f(x,y) eq = u + u.diff(x) + u.diff(y) + x**2*y sol = pdsolve(eq, hint = 'all') keys = ['1st_linear_constant_coeff', '1st_linear_constant_coeff_Integral', 'default', 'order'] assert sorted(sol.keys()) == keys assert sol['order'] == 1 assert sol['default'] == '1st_linear_constant_coeff' assert sol['1st_linear_constant_coeff'] == Eq(f(x, y), -x**2*y + x**2 + 2*x*y - 4*x - 2*y + F(x - y)*exp(-x/S(2) - y/S(2)) + 6) def test_pdsolve_variable_coeff(): f, F = map(Function, ['f', 'F']) u = f(x, y) eq = x*(u.diff(x)) - y*(u.diff(y)) + y**2*u - y**2 sol = pdsolve(eq, hint="1st_linear_variable_coeff") assert sol == Eq(u, F(x*y)*exp(y**2/2) + 1) assert checkpdesol(eq, sol)[0] eq = x**2*u + x*u.diff(x) + x*y*u.diff(y) sol = pdsolve(eq, hint='1st_linear_variable_coeff') assert sol == Eq(u, F(y*exp(-x))*exp(-x**2/2)) assert checkpdesol(eq, sol)[0] eq = y*x**2*u + y*u.diff(x) + u.diff(y) sol = pdsolve(eq, hint='1st_linear_variable_coeff') assert sol == Eq(u, F(-2*x + y**2)*exp(-x**3/3)) assert checkpdesol(eq, sol)[0] eq = exp(x)**2*(u.diff(x)) + y sol = pdsolve(eq, hint='1st_linear_variable_coeff') assert sol == Eq(u, y*exp(-2*x)/2 + F(y)) assert checkpdesol(eq, sol)[0] eq = exp(2*x)*(u.diff(y)) + y*u - u sol = pdsolve(eq, hint='1st_linear_variable_coeff') assert sol == Eq(u, exp((-y**2 + 2*y + 2*F(x))*exp(-2*x)/2))
[ "kaunalei@gmail.com" ]
kaunalei@gmail.com
9d07edf3f66ec3474be91599c596f9476549ed59
4f60de93ad66f3ff95a5b86c41293dc95ecc4a2a
/hr/models.py
09d2094f1645475c50fb3e6fabdabcb064e8e6d0
[]
no_license
xxoxx/Luer
87a541f40a12f4aff262d03b9e8e3b4b78a41511
24333edd06dc5e52775329db1f86cf862c68c56e
refs/heads/master
2021-01-20T17:27:13.914273
2015-12-09T15:18:35
2015-12-09T15:18:35
null
0
0
null
null
null
null
UTF-8
Python
false
false
1,762
py
#coding:utf-8 from django.db import models from django.contrib.auth.models import User from django.template.defaultfilters import default # Create your models here. class Department(models.Model): department_name = models.CharField(max_length = 100) department_manager = models.CharField(max_length = 100) def __unicode__(self): return self.department_name class Employee_extra(models.Model): user = models.OneToOneField(User) name = models.CharField(max_length = 100) password = models.CharField(max_length = 100) sex = models.CharField(max_length = 100) email = models.EmailField() department = models.ForeignKey(Department) position = models.CharField(max_length = 100) base_salary = models.FloatField() extra_salary = models.FloatField(default=0) hiredate = models.DateField() pic = models.ImageField(upload_to='upload') class Meta: permissions = (('department_manager','department_manager'), ('personnel_assistant','personnel_assistant'), ('personnel_manager','personnel_manager'), ('admin','admin'), ) class Attendence(models.Model): attendence_name = models.CharField(max_length = 100) is_early_or_late = models.BooleanField() attendence_date = models.DateField() class Salary(models.Model): name = models.CharField(max_length = 100) year = models.CharField(max_length = 100) month = models.CharField(max_length = 100) salary = models.FloatField() class Announce(models.Model): title = models.CharField(max_length = 100) content = models.TextField() date = models.DateField() def __unicode__(self): return self.title
[ "lhlsec@gmail.com" ]
lhlsec@gmail.com
a45dacabb65a8b878d1cb07374fde8bc5ac07d6d
f305f84ea6f721c2391300f0a60e21d2ce14f2a5
/7_graph/bfs求无权图的最短路径/广义的bfs/488. 祖玛游戏-bfs剪枝.py
12413155d1f8a0da0d66c30102d92f4f104f18a7
[]
no_license
981377660LMT/algorithm-study
f2ada3e6959338ae1bc21934a84f7314a8ecff82
7e79e26bb8f641868561b186e34c1127ed63c9e0
refs/heads/master
2023-09-01T18:26:16.525579
2023-09-01T12:21:58
2023-09-01T12:21:58
385,861,235
225
24
null
null
null
null
UTF-8
Python
false
false
2,038
py
# 1 <= board.length <= 16 # 1 <= hand.length <= 5 from collections import deque from functools import lru_cache import re # 为什么使用广度优先搜索? # 因为只需要找出需要回合数最少的方案,因此使用广度优先搜索可以得到可以消除桌面上所有球的方案时就直接返回结果,而不需要继续遍历更多需要回合数更多的方案。 class Solution: def findMinStep(self, board: str, hand: str) -> int: """请你按上述操作步骤移除掉桌上所有球,计算并返回所需的 最少 球数。如果不能移除桌上所有的球,返回 -1 。""" @lru_cache(None) def clean(s: str) -> str: """碰到三个就删除整个""" count = 1 while count: s, count = re.subn(r'(\w)\1{2,}', '', s) return s hand = ''.join(sorted(hand)) queue = deque([(board, hand, 0)]) visited = set([(board, hand)]) while queue: b, h, step = queue.popleft() if not b: return step # 插入位置 for i in range(len(b)): # 删除那个元素 for j in range(len(h)): # 最重要的剪枝是,当手上的球 h[j] 和插入位置 i 前后的球 b[i-1], b[i] 三个球各不相同时,插入是不必要的: sequence = [b[i - 1], b[i], h[j]] if i else [b[i], h[j]] if len(set(sequence)) < len(sequence): nextB = clean(b[:i] + h[j] + b[i:]) nextH = h[:j] + h[j + 1 :] if (nextB, nextH) not in visited: visited.add((nextB, nextH)) queue.append((nextB, nextH, step + 1)) return -1 print(Solution().findMinStep(board="WRRBBW", hand="RB")) print(Solution().findMinStep(board="WWRRBBWW", hand="WRBRW")) # re.subn返回一个元组
[ "lmt2818088@gmail.com" ]
lmt2818088@gmail.com
c560a98412f1f79c8b28518349b9281f419d3cd1
5f313d8fce26a8ecfff8817ff566b7e1810fcba7
/timethings.py
4d68c2cbbbfce64ba5da5943421cca52b094884d
[]
no_license
luispedro/mahotas-paper
cd2769a264149cac74ce8c694ca4f02e3f4a6c93
698f2a8640feba4e285318e2cd866db3705ec2c3
refs/heads/master
2020-03-30T16:26:20.362126
2013-08-26T09:03:17
2013-08-26T09:03:17
4,877,058
1
0
null
null
null
null
UTF-8
Python
false
false
3,733
py
import skimage.morphology import skimage.filter import skimage.feature import numpy as np import timeit import mahotas import cv2 from os import path luispedro_image = path.join( path.dirname(mahotas.__file__), 'demos', 'data', 'luispedro.jpg') f = mahotas.imread(luispedro_image, as_grey=True) markers = np.zeros_like(f) markers[100,100] = 1 markers[200,200] = 2 f = f.astype(np.uint8) markers = markers.astype(int) otsu = mahotas.otsu(f.astype(np.uint8)) fbin = f > otsu fbin8 = fbin.astype(np.uint8) Bc = np.eye(3) Bc = Bc.astype(bool) Bc8 = Bc.astype(np.uint8) f3 = np.dstack([f,f,f]) f3 = f3.astype(np.uint8) f3 = f3.copy() filt = np.array([ [1,0,-1,0], [2,2,3,-2], [-1,0,0,1] ]) markers32 = markers.astype(np.int32) def octagon(r): octagon = np.ones((r*2+1, r*2+1), dtype=np.bool) lim = r//2 for y in xrange(lim): octagon[y,:lim-y] = 0 octagon &= octagon[::-1] octagon &= octagon[:,::-1] return octagon pre =''' import skimage.filter import skimage.morphology import skimage.feature import numpy as np import mahotas import pymorph import cv2 import timethings octagon = timethings.octagon f = timethings.f f3 = timethings.f3 fbin = timethings.fbin fbin8 = timethings.fbin8 f64 = f.astype(np.float64) Bc = timethings.Bc Bc8 = timethings.Bc8 markers = timethings.markers markers32 = timethings.markers32 filt = timethings.filt ''' def t(s): return min(timeit.timeit(s, setup=pre, number=24) for i in xrange(3)) tests = [ ('convolve', [ 'mahotas.convolve(f, filt)', None, None, None, ]), ('erode', [ 'mahotas.erode(fbin, Bc)', 'pymorph.erode(fbin, Bc)', 'skimage.morphology.erosion(fbin8, Bc8)', 'cv2.erode(fbin8, Bc8)', ]), ('dilate', [ 'mahotas.dilate(fbin, Bc)', 'pymorph.dilate(fbin, Bc)', 'skimage.morphology.dilation(fbin8, Bc8)', 'cv2.dilate(fbin8, Bc8)', ]), ('open', [ 'mahotas.open(fbin, Bc)', 'pymorph.open(fbin, Bc)', 'skimage.morphology.opening(fbin8, Bc8)', None, ]), ('median filter (2)', [ 'mahotas.median_filter(f, octagon(2))', None, 'skimage.filter.median_filter(f, 2)', None, ]), ('median filter (10)', [ 'mahotas.median_filter(f, octagon(10))', None, 'skimage.filter.median_filter(f, 10)', None, ]), ('center mass', [ 'mahotas.center_of_mass(f)', None, 'skimage.measure.regionprops(np.ones(f.shape, np.intc), ["WeightedCentroid"], intensity_image=f)', None, ]), ('sobel', [ 'mahotas.sobel(f, just_filter=True)', None, 'skimage.filter.sobel(f)', 'cv2.Sobel(f, cv2.CV_32F, 1, 1)', ]), ('cwatershed', [ 'mahotas.cwatershed(f, markers)', 'pymorph.cwatershed(f, markers)', 'skimage.morphology.watershed(f, markers)', 'cv2.watershed(f3, markers32.copy())', ]), ('daubechies', [ 'mahotas.daubechies(f, "D4")', None, None, None, ]), ('haralick', [ 'mahotas.features.haralick(f)', None, 'skimage.feature.greycoprops(skimage.feature.greycomatrix(f, [1], [0]))', None, ]), ] if __name__ == '__main__': base = t('np.max(f)') for name,statements in tests: print r'{0:<20} &'.format(name), for st in statements: if st is None: result = 'NA' else: result = '{:.1f}'.format( t(st)/base ) print '{0:>8} &'.format(result), print r'\\'
[ "luis@luispedro.org" ]
luis@luispedro.org
8157a66920a4d18df210932be00e7ae648e9bfec
760ef8fa1fa34f7a0621ed1a72cd4070d5825e34
/src/tray_world_creator.py
030feced0d4b97e56653925469dea38f8d799f2b
[]
no_license
michaelarfreed/MotionTaskPlanningProject
9a2670565fe2923504d678407d3dcb4659bc06df
384f9801f031f0d413854bc844a452d2681b34af
refs/heads/master
2023-04-10T12:08:35.538339
2021-04-04T23:12:37
2021-04-04T23:12:37
354,664,871
0
0
null
null
null
null
UTF-8
Python
false
false
6,318
py
import numpy as np import openravepy import utils TARGET_FILE = '../environments/rll_tray_world.dae' TABLE_HEIGHT = 0.657 def create_dest(env, destname, pos): dim = 0.1 THICKNESS = 0.001 surface = openravepy.KinBody.GeometryInfo() surface._type = openravepy.GeometryType.Box surface._vGeomData = [dim/2, dim/2, THICKNESS/2] surface._vDiffuseColor = [1, 0, 0] dest = openravepy.RaveCreateKinBody(env, '') dest.InitFromGeometries([surface]) dest.SetName(destname) t = openravepy.matrixFromPose([1, 0, 0, 0] + list(pos)) dest.SetTransform(t) return dest def create_tray(env, t): dim1 = 0.3048 dim2 = 0.6096 THICKNESS = 0.0095 TRAY_HEIGHT = 0.148 surface = openravepy.KinBody.GeometryInfo() surface._type = openravepy.GeometryType.Box surface._vGeomData = [dim1/2, dim2/2, THICKNESS/2] surface._vDiffuseColor = [0.4, 0.2, 0.4] surface._t = openravepy.matrixFromPose((1, 0, 0, 0, 0, 0, TRAY_HEIGHT - THICKNESS/2)) standoff = openravepy.KinBody.GeometryInfo() standoff._type = openravepy.GeometryType.Box standoff._vGeomData = [0.6*dim1/2, 0.6*dim2/2, TRAY_HEIGHT/2] standoff._vDiffuseColor = [0.4, 0.2, 0.4] standoff._t = openravepy.matrixFromPose((1, 0, 0, 0, 0, 0, TRAY_HEIGHT/2)) tray = openravepy.RaveCreateKinBody(env, '') tray.InitFromGeometries([surface, standoff]) tray.SetName('tray') pose = openravepy.poseFromMatrix(t) pose[:4] = [0.7071, 0, 0, 0.7071] t = openravepy.matrixFromPose(pose) tray.SetTransform(t) return tray def create_cylinder(env, body_name, pos, radius, height): infocylinder = openravepy.KinBody.GeometryInfo() infocylinder._type = openravepy.GeometryType.Cylinder infocylinder._vGeomData = [radius, height] infocylinder._bVisible = True infocylinder._vDiffuseColor = [0, 1, 1] cylinder = openravepy.RaveCreateKinBody(env, '') cylinder.InitFromGeometries([infocylinder]) cylinder.SetName(body_name) pos[2] += height/2 cylinder.SetTransform(openravepy.matrixFromPose([1, 0, 0, 0] + pos)) return cylinder def create_table(env, body_name, dim1, dim2, pos): THICKNESS = 0.2 LEGDIM1 = 1.3 LEGDIM2 = 0.6 LEGHEIGHT = 0.6 # Doesn't actually determine the height of the table in the env tabletop = openravepy.KinBody.GeometryInfo() tabletop._type = openravepy.GeometryType.Box tabletop._vGeomData = [dim1/2, dim2/2, THICKNESS/2] tabletop._t[2, 3] = -THICKNESS/2 tabletop._vDiffuseColor = [0.5, 0.2, 0.1] leg1 = openravepy.KinBody.GeometryInfo() leg1._type = openravepy.GeometryType.Box leg1._vGeomData = [LEGDIM1/2, LEGDIM2/2, LEGHEIGHT/2] leg1._t[0, 3] = dim1/2 - LEGDIM1/2 leg1._t[1, 3] = dim2/2 - LEGDIM2/2 leg1._t[2, 3] = -LEGHEIGHT/2 - THICKNESS/2 leg1._vDiffuseColor = [0.5, 0.2, 0.1] leg2 = openravepy.KinBody.GeometryInfo() leg2._type = openravepy.GeometryType.Box leg2._vGeomData = [LEGDIM1/2, LEGDIM2/2, LEGHEIGHT/2] leg2._t[0, 3] = dim1/2 - LEGDIM1/2 leg2._t[1, 3] = -dim2/2 + LEGDIM2/2 leg2._t[2, 3] = -LEGHEIGHT/2 - THICKNESS/2 leg2._vDiffuseColor = [0.5, 0.2, 0.1] leg3 = openravepy.KinBody.GeometryInfo() leg3._type = openravepy.GeometryType.Box leg3._vGeomData = [LEGDIM1/2, LEGDIM2/2, LEGHEIGHT/2] leg3._t[0, 3] = -dim1/2 + LEGDIM1/2 leg3._t[1, 3] = dim2/2 - LEGDIM2/2 leg3._t[2, 3] = -LEGHEIGHT/2 - THICKNESS/2 leg3._vDiffuseColor = [0.5, 0.2, 0.1] leg4 = openravepy.KinBody.GeometryInfo() leg4._type = openravepy.GeometryType.Box leg4._vGeomData = [LEGDIM1/2, LEGDIM2/2, LEGHEIGHT/2] leg4._t[0, 3] = -dim1/2 + LEGDIM1/2 leg4._t[1, 3] = -dim2/2 + LEGDIM2/2 leg4._t[2, 3] = -LEGHEIGHT/2 - THICKNESS/2 leg4._vDiffuseColor = [0.5, 0.2, 0.1] table = openravepy.RaveCreateKinBody(env, '') table.InitFromGeometries([tabletop, leg1, leg2, leg3, leg4]) table.SetName(body_name) table.SetTransform(openravepy.matrixFromPose([1, 0, 0, 0] + pos)) return table env = openravepy.Environment() env.SetViewer('qtcoin') # plot origin for sanity t = np.eye(4) o = utils.plot_transform(env, t, 1.0) # spawn PR2 robot = env.ReadRobotXMLFile("robots/pr2-beta-sim.robot.xml") env.Add(robot) # spawn main table env.AddKinBody(create_table(env, 'table', 2.25, 0.94, [2.25, 0, TABLE_HEIGHT])) # spawn tray table env.AddKinBody(create_table(env, 'tray_table', 1.525, 0.61, [0.95, 1.0, TABLE_HEIGHT])) # spawn glasses (object2) GLASS_HEIGHT = 0.15 GLASS_RADIUS = 0.04 env.AddKinBody(create_cylinder(env, 'object21', [1.4, 0, TABLE_HEIGHT], GLASS_RADIUS, GLASS_HEIGHT)) env.AddKinBody(create_cylinder(env, 'object22', [1.4, 0.2, TABLE_HEIGHT], GLASS_RADIUS, GLASS_HEIGHT)) env.AddKinBody(create_cylinder(env, 'object23', [1.4, -0.2, TABLE_HEIGHT], GLASS_RADIUS, GLASS_HEIGHT)) # spawn bowls (object1) BOWL_HEIGHT = 0.08 BOWL_RADIUS = 0.06 env.AddKinBody(create_cylinder(env, 'object11', [1.2, 0, TABLE_HEIGHT], BOWL_RADIUS, BOWL_HEIGHT)) env.AddKinBody(create_cylinder(env, 'object12', [1.2, 0.2, TABLE_HEIGHT], BOWL_RADIUS, BOWL_HEIGHT)) env.AddKinBody(create_cylinder(env, 'object13', [1.2, -0.2, TABLE_HEIGHT], BOWL_RADIUS, BOWL_HEIGHT)) # # spawn plates (object3) # PLATE_HEIGHT = 0.04 # PLATE_RADIUS = 0.08 # env.AddKinBody(create_cylinder(env, 'object31', [1.4, -0.2, TABLE_HEIGHT], PLATE_RADIUS, PLATE_HEIGHT)) # env.AddKinBody(create_cylinder(env, 'object32', [1.2, -0.2, TABLE_HEIGHT], PLATE_RADIUS, PLATE_HEIGHT)) # spawn tray env.AddKinBody(create_tray(env, openravepy.matrixFromPose((1, 0, 0, 0, 0.5, 0.85, TABLE_HEIGHT)))) # adding destinations env.AddKinBody(create_dest(env, 'destobject11', (1.8, -0.3, TABLE_HEIGHT))) env.AddKinBody(create_dest(env, 'destobject12', (2.2, -0.3, TABLE_HEIGHT))) env.AddKinBody(create_dest(env, 'destobject13', (2.6, -0.3, TABLE_HEIGHT))) env.AddKinBody(create_dest(env, 'destobject21', (2.0, -0.1, TABLE_HEIGHT))) env.AddKinBody(create_dest(env, 'destobject22', (2.4, -0.1, TABLE_HEIGHT))) env.AddKinBody(create_dest(env, 'destobject23', (2.8, -0.1, TABLE_HEIGHT))) # env.AddKinBody(create_dest(env, 'destobject31', (2.6, -0.3, TABLE_HEIGHT))) # env.AddKinBody(create_dest(env, 'destobject32', (2.8, -0.1, TABLE_HEIGHT))) tl1 = create_dest(env, 'trayloc1', (0.5, 0.85, TABLE_HEIGHT)) env.AddKinBody(tl1) tl2 = create_dest(env, 'trayloc2', (3.2, -0.4, TABLE_HEIGHT)) env.AddKinBody(tl2) env.Save(TARGET_FILE) raw_input("Press ENTER to exit!")
[ "michaelarfreed@gmail.com" ]
michaelarfreed@gmail.com
34acd628169af951f3576d45c718a7a03b92b726
6919841828cf32f3e4b0f6f12f6d54be23321888
/exp/face24.py
db4baf6b7976aeb66a2d63d236bbd38f6c51d276
[]
no_license
ayanc/rpgan
bea6d283278dcead5fa4c63815a70162d5fe3173
cb4d4ab226fa78e4d6f6dcae9fa715952345a0bb
refs/heads/master
2021-01-22T04:23:39.000116
2018-06-27T21:43:45
2018-06-27T21:43:45
92,459,728
20
6
null
null
null
null
UTF-8
Python
false
false
340
py
# Generator Parameters ksz=4 zlen = 100 # Dimensionality of z f1 = 1024 # Features in first layer of Gen output # Discrimnator Parameters df = 128 # No. of hidden features (at first layer of D) # Training set imsz = 64 bsz = 64 lfile='data/faces.txt' crop=False # Learning parameters wts_dir='models/face24' SAVEFREQ=1e3 MAXITER=1e5
[ "ayanc@ttic.edu" ]
ayanc@ttic.edu
736b25d9dd6ccb8c04ea88d667c2c6ec69bf359c
6aee5e2775a99de940f9deef3d34e9833d44819c
/lesson/day03/test.py
c202ccce8b5f53b0a53bb44e0089f589ac030e8b
[]
no_license
lyl0521/lesson-0426
c7669767cf5f6bae54765f2246764e1a06acee7a
07f114af276932c252478d5a739ca88b4bae0fa2
refs/heads/master
2020-05-17T17:21:00.876376
2019-05-07T01:54:56
2019-05-07T01:54:56
183,849,196
0
0
null
null
null
null
UTF-8
Python
false
false
866
py
# all() print(all([0,1,2,3])) # False print(all([1,2,3])) # True print(all(['a','b','c',''])) # False print(all(['a','b','c','d'])) #True print(all(('a','b','c','d'))) #True print(all(('a','b','','c'))) #False print(all((0,1,2,3))) #False print(all((1,2,3))) #True print(all([])) # True empty list print(all(())) # True empty tuple # any() # 如果都为空、0、false,则返回false,如果不都为空、0、false,则返回true。 print(any(['a', 'b', 'c', 'd'])) #True print(any(['a', 'b', '', 'd'])) #True print(any([0, '', False])) #False print(any(('a', 'b', 'c', 'd'))) #True print(any(('a', 'b', '', 'd'))) #True print(any((0, '', False))) #False print(any([])) #True print(any(())) #True # bin() # 返回整数的二进制表示 print(bin(int(10))) print(bin(10))
[ "1247702289@qq.com" ]
1247702289@qq.com
38c9d1b686c768cb8aa82bee7276bc14769b92c1
b4d1de38688da22e64026b0b46a1e0ae3efa97ed
/server_random_data_generator.py
8fb1dd5ec17ffda8aa8e0965dc3111af27b0e041
[]
no_license
misrori/streaming_with_mtcars
a8f57e25f7a5389dc593a1e0e72bf944a7dc0abf
bb459983c7bfd6e648bf46685809e5e7b48c8d36
refs/heads/master
2021-01-18T22:40:54.167528
2017-04-05T07:36:57
2017-04-05T07:36:57
87,066,569
0
0
null
null
null
null
UTF-8
Python
false
false
1,323
py
#!/home/mihaly/hadoop/anaconda3/bin/python from random import * import random import time from datetime import datetime def client_thread(conn, ip, port, MAX_BUFFER_SIZE = 4096): mtcars = open('mtcars.csv') my_line = str(random.sample(mtcars.readlines(),1)).replace('[','').replace(']','').replace('\'','').replace('\"','').replace('\\n','') mtcars.close() my_st = datetime.now().strftime('%H:%M:%S') +','+ my_line + '\n' try: conn.send(my_st.encode("utf8")) return 0 except: return -1 def start_server(): import socket soc = socket.socket(socket.AF_INET, socket.SOCK_STREAM) # this is for easy starting/killing the app soc.setsockopt(socket.SOL_SOCKET, socket.SO_REUSEADDR, 1) print('Socket created') soc.bind(("localhost", 9000)) print('Socket bind complete') #Start listening on socket soc.listen(10) print('Socket now listening') while True: conn, addr = soc.accept() ip, port = str(addr[0]), str(addr[1]) print('Accepting connection from ' + ip + ':' + port) time.sleep(1) while True: #conn.__getstate__: if client_thread(conn, ip, port) != 0: break time.sleep(randint(0,1)*0.1) print(".", end='') soc.close() start_server()
[ "ormraat.pte@gmail.com" ]
ormraat.pte@gmail.com
64f8f8d3a25cb6ac09de771d500d0936684177e1
fc155434e4bb00eb1b6f09ccda3cd73235bce156
/CreateVideo.py
4d8d947f846bb1a053d68f0ebd03f54d0aac8ad3
[]
no_license
alter-sachin/buildar_email
03ac84b2d391cd8951e9311847a97e7d21795a76
8dbe19dfe8893747e35051515568133c6079dcf1
refs/heads/main
2023-04-21T20:33:48.917393
2021-05-16T14:01:17
2021-05-16T14:01:17
367,877,851
0
0
null
null
null
null
UTF-8
Python
false
false
868
py
import json import requests import constants api_endpoint = constants.API_ENDPOINT header_data={'Authorization': constants.API_AUTHKEY} class Video: def __init__(self,actorId,audioUrl): self.actorId = actorId self.audioUrl = audioUrl def create_request_body(self): data = { "actorId": self.actorId, "audioUrl": self.audioUrl, } return json.dumps(data) def get_speaker_details(self): print("implement later") def create_video(self): response = requests.post(api_endpoint+"video", headers=header_data, data = self.create_request_body()) videoUrl = str(json.loads(response.content.decode('utf-8'))['videoUrl']) return videoUrl # videoObj = Video(constants.ACTOR1,"https://dreal.in/aud/1621100562.873771.wav") # #get_users() # videoObj.create_video() # # mailObj = Mailer(from_email,list_of_emails) # # mailObj.send_email()
[ "guitarplayersachin@gmail.com" ]
guitarplayersachin@gmail.com
f8cff679622d038367e9ba166acddf2c25db3acc
8126a1246a7fadcba2a57afdec210e6c00dd85ff
/tests/data_objects/constraints/constraints_object_test.py
556692a6d6f5a40717a7b875d031a5c4615430be
[]
no_license
TF-185/cra-princess-cp1
0961f0bb140a89efb427e4e497318a563a564b75
cbe5dab43983f794a817cd891973f68789ebd3b4
refs/heads/master
2023-05-30T10:45:24.635862
2019-10-11T19:55:00
2019-10-11T19:55:00
null
0
0
null
null
null
null
UTF-8
Python
false
false
8,433
py
# """ # constraints_object_test.py # # Module to test constraints_object.py. # Author: Tameem Samawi (tsamawi@cra.com) # """ # # import unittest # from cp1.common.exception_class import ConstraintsObjectInitializationException # from cp1.data_objects.processing.constraints_object import ConstraintsObject # from cp1.data_objects.processing.channel import Channel # from cp1.data_objects.mdl.txop_timeout import TxOpTimeout # from cp1.data_objects.mdl.bandwidth_types import BandwidthTypes # from cp1.data_objects.mdl.bandwidth_rate import BandwidthRate # from cp1.data_objects.mdl.kbps import Kbps # from cp1.data_objects.mdl.mdl_id import MdlId # from cp1.data_objects.mdl.frequency import Frequency # from cp1.data_objects.processing.ta import TA # # # class ConstraintsObjectTest(unittest.TestCase): # # @classmethod # def setUpClass(cls): # cls.goal_throughput_bulk = BandwidthRate(BandwidthTypes.BULK, Kbps(10)) # cls.goal_throughput_voice = BandwidthRate(BandwidthTypes.VOICE, Kbps(10)) # cls.goal_throughput_safety = BandwidthRate(BandwidthTypes.SAFETY, Kbps(10)) # cls.latency = time(microsecond=50000) # cls.guard_band = Milliseconds(1000) # cls.epoch = Milliseconds(100000) # cls.txop_timeout = TxOpTimeout(255) # # channel_frequency = Frequency(4919500000) # channel_length = Milliseconds(100) # channel_latency = Milliseconds(50) # channel_capacity = Kbps(100) # cls.channels = [Channel(channel_frequency, channel_length, channel_latency, channel_capacity)] # # ta_id_ = MdlId('TA1') # ta_minimum_voice_bandwidth = Kbps(100) # ta_minimum_safety_bandwidth = Kbps(75) # ta_scaling_factor = 1 # ta_c = 0.05 # ta_min_value = 65 # ta = TA(ta_id_, ta_minimum_voice_bandwidth, ta_minimum_safety_bandwidth, ta_scaling_factor, # ta_c) # cls.candidate_tas = [ta] # # def test_valid_constraints_object_init(self): # constraints_object = ConstraintsObject( # self.goal_throughput_bulk, # self.goal_throughput_voice, # self.goal_throughput_safety, # self.latency, # self.guard_band, # self.epoch, # self.txop_timeout, # self.candidate_tas, # self.channels) # self.assertEqual(self.goal_throughput_bulk, # constraints_object.goal_throughput_bulk) # self.assertEqual(self.goal_throughput_voice, # constraints_object.goal_throughput_voice) # self.assertEqual(self.goal_throughput_safety, # constraints_object.goal_throughput_safety) # self.assertEqual(self.latency, # constraints_object.latency) # self.assertEqual(self.guard_band, # constraints_object.guard_band) # self.assertEqual(self.epoch, # constraints_object.epoch) # # tas_equal = True # for i in range(0, len(self.candidate_tas)): # if self.candidate_tas[i] != constraints_object.candidate_tas[i]: # tas_equal = False # self.assertTrue(tas_equal) # # channels_equal = True # for i in range(0, len(self.channels)): # if self.channels[i] != constraints_object.channels[i]: # channels_equal = False # self.assertTrue(channels_equal) # # def test___eq__(self): # c1 = ConstraintsObject( # self.goal_throughput_bulk, # self.goal_throughput_voice, # self.goal_throughput_safety, # self.latency, # self.guard_band, # self.epoch, # self.txop_timeout, # self.candidate_tas, # self.channels) # c2 = ConstraintsObject( # self.goal_throughput_bulk, # self.goal_throughput_voice, # self.goal_throughput_safety, # self.latency, # self.guard_band, # self.epoch, # self.txop_timeout, # self.candidate_tas, # self.channels) # self.assertEqual(c1, c2) # # def test_invalid_goal_throughput_bulk(self): # self.assertRaises( # ConstraintsObjectInitializationException, # ConstraintsObject, # 1, # self.goal_throughput_voice, # self.goal_throughput_safety, # self.latency, # self.guard_band, # self.epoch, # self.txop_timeout, # self.candidate_tas, # self.channels) # # def test_invalid_goal_throughput_voice(self): # self.assertRaises( # ConstraintsObjectInitializationException, # ConstraintsObject, # self.goal_throughput_bulk, # 1, # self.goal_throughput_safety, # self.latency, # self.guard_band, # self.epoch, # self.txop_timeout, # self.candidate_tas, # self.channels) # # def test_invalid_goal_throughput_safety(self): # self.assertRaises( # ConstraintsObjectInitializationException, # ConstraintsObject, # self.goal_throughput_bulk, # self.goal_throughput_voice, # 1, # self.latency, # self.guard_band, # self.epoch, # self.txop_timeout, # self.candidate_tas, # self.channels) # # def test_invalid_guard_band_type(self): # self.assertRaises( # ConstraintsObjectInitializationException, # ConstraintsObject, # self.goal_throughput_bulk, # self.goal_throughput_voice, # self.goal_throughput_safety, # self.latency, # 1, # self.epoch, # self.txop_timeout, # self.candidate_tas, # self.channels) # # def test_invalid_epoch_type(self): # self.assertRaises( # ConstraintsObjectInitializationException, # ConstraintsObject, # self.goal_throughput_bulk, # self.goal_throughput_voice, # self.goal_throughput_safety, # self.latency, # self.guard_band, # 1, # self.txop_timeout, # self.candidate_tas, # self.channels) # # def test_invalid_txop_timeout(self): # self.assertRaises( # ConstraintsObjectInitializationException, # ConstraintsObject, # self.goal_throughput_bulk, # self.goal_throughput_voice, # self.goal_throughput_safety, # self.latency, # self.guard_band, # self.epoch, # 1, # self.candidate_tas, # self.channels) # # def test_invalid_candidate_tas_type(self): # self.assertRaises( # ConstraintsObjectInitializationException, # ConstraintsObject, # self.goal_throughput_bulk, # self.goal_throughput_voice, # self.goal_throughput_safety, # self.latency, # self.guard_band, # self.epoch, # self.txop_timeout, # 1, # self.channels) # # def test_invalid_channels_type(self): # self.assertRaises( # ConstraintsObjectInitializationException, # ConstraintsObject, # self.goal_throughput_bulk, # self.goal_throughput_voice, # self.goal_throughput_safety, # self.latency, # self.guard_band, # self.epoch, # self.txop_timeout, # self.candidate_tas, # 1)
[ "tsamawi@cra.com" ]
tsamawi@cra.com
286c6510e842c109cd1fabfbe090e84a978c9b28
fab14fae2b494068aa793901d76464afb965df7e
/benchmarks/f3_wrong_hints/scaling_ltl_timed_transition_system/19-sender_receiver_7.py
46038b17bbf43428b545dfc234384f2bb5c2c34d
[ "MIT" ]
permissive
teodorov/F3
673f6f9ccc25acdfdecbfc180f439253474ba250
c863215c318d7d5f258eb9be38c6962cf6863b52
refs/heads/master
2023-08-04T17:37:38.771863
2021-09-16T07:38:28
2021-09-16T07:38:28
null
0
0
null
null
null
null
UTF-8
Python
false
false
19,406
py
from typing import FrozenSet from collections import Iterable from math import log, ceil from mathsat import msat_term, msat_env from mathsat import msat_make_constant, msat_declare_function from mathsat import msat_get_integer_type, msat_get_rational_type, msat_get_bool_type from mathsat import msat_make_and, msat_make_not, msat_make_or, msat_make_iff from mathsat import msat_make_leq, msat_make_equal, msat_make_true from mathsat import msat_make_number, msat_make_plus, msat_make_times from pysmt.environment import Environment as PysmtEnv import pysmt.typing as types from ltl.ltl import TermMap, LTLEncoder from utils import name_next, symb_to_next from hint import Hint, Location delta_name = "delta" def decl_consts(menv: msat_env, name: str, c_type) -> tuple: assert not name.startswith("_"), name s = msat_declare_function(menv, name, c_type) s = msat_make_constant(menv, s) x_s = msat_declare_function(menv, name_next(name), c_type) x_s = msat_make_constant(menv, x_s) return s, x_s def make_enum(menv, v_name: str, enum_size: int): bool_type = msat_get_bool_type(menv) num_bits = ceil(log(enum_size, 2)) b_vars = [] for idx in range(num_bits): c_name = "{}{}".format(v_name, idx) b_vars.append(tuple(decl_consts(menv, c_name, bool_type))) vals = [] x_vals = [] for enum_val in range(enum_size): bit_val = format(enum_val, '0{}b'.format(num_bits)) assert len(bit_val) == num_bits assert all(c in {'0', '1'} for c in bit_val) assign = [b_vars[idx] if c == '1' else (msat_make_not(menv, b_vars[idx][0]), msat_make_not(menv, b_vars[idx][1])) for idx, c in enumerate(reversed(bit_val))] pred = assign[0][0] x_pred = assign[0][1] for it in assign[1:]: pred = msat_make_and(menv, pred, it[0]) x_pred = msat_make_and(menv, x_pred, it[1]) vals.append(pred) x_vals.append(x_pred) assert len(vals) == enum_size assert len(x_vals) == enum_size return b_vars, vals, x_vals def msat_make_minus(menv: msat_env, arg0: msat_term, arg1: msat_term): m_one = msat_make_number(menv, "-1") arg1 = msat_make_times(menv, arg1, m_one) return msat_make_plus(menv, arg0, arg1) def msat_make_lt(menv: msat_env, arg0: msat_term, arg1: msat_term): geq = msat_make_geq(menv, arg0, arg1) return msat_make_not(menv, geq) def msat_make_geq(menv: msat_env, arg0: msat_term, arg1: msat_term): return msat_make_leq(menv, arg1, arg0) def msat_make_gt(menv: msat_env, arg0: msat_term, arg1: msat_term): leq = msat_make_leq(menv, arg0, arg1) return msat_make_not(menv, leq) def msat_make_impl(menv: msat_env, arg0: msat_term, arg1: msat_term): n_arg0 = msat_make_not(menv, arg0) return msat_make_or(menv, n_arg0, arg1) def diverging_symbs(menv: msat_env) -> frozenset: real_type = msat_get_rational_type(menv) delta = msat_declare_function(menv, delta_name, real_type) delta = msat_make_constant(menv, delta) return frozenset([delta]) def check_ltl(menv: msat_env, enc: LTLEncoder) -> (Iterable, msat_term, msat_term, msat_term): assert menv assert isinstance(menv, msat_env) assert enc assert isinstance(enc, LTLEncoder) int_type = msat_get_integer_type(menv) real_type = msat_get_rational_type(menv) r2s, x_r2s = decl_consts(menv, "r2s", int_type) s2r, x_s2r = decl_consts(menv, "s2r", int_type) delta, x_delta = decl_consts(menv, delta_name, real_type) sender = Sender("s", menv, enc, r2s, x_r2s, s2r, x_s2r, delta) receiver = Receiver("r", menv, enc, s2r, x_s2r, r2s, x_r2s, delta) curr2next = {r2s: x_r2s, s2r: x_s2r, delta: x_delta} for comp in [sender, receiver]: for s, x_s in comp.symb2next.items(): curr2next[s] = x_s zero = msat_make_number(menv, "0") init = msat_make_and(menv, receiver.init, sender.init) trans = msat_make_and(menv, receiver.trans, sender.trans) # invar delta >= 0 init = msat_make_and(menv, init, msat_make_geq(menv, delta, zero)) trans = msat_make_and(menv, trans, msat_make_geq(menv, x_delta, zero)) # delta > 0 -> (r2s' = r2s & s2r' = s2r) lhs = msat_make_gt(menv, delta, zero) rhs = msat_make_and(menv, msat_make_equal(menv, x_r2s, r2s), msat_make_equal(menv, x_s2r, s2r)) trans = msat_make_and(menv, trans, msat_make_impl(menv, lhs, rhs)) # (G F !s.stutter) -> G (s.wait_ack -> F s.send) lhs = enc.make_G(enc.make_F(msat_make_not(menv, sender.stutter))) rhs = enc.make_G(msat_make_impl(menv, sender.wait_ack, enc.make_F(sender.send))) ltl = msat_make_impl(menv, lhs, rhs) return TermMap(curr2next), init, trans, ltl class Module: def __init__(self, name: str, menv: msat_env, enc: LTLEncoder, *args, **kwargs): self.name = name self.menv = menv self.enc = enc self.symb2next = {} true = msat_make_true(menv) self.init = true self.trans = true def _symb(self, v_name, v_type): v_name = "{}_{}".format(self.name, v_name) return decl_consts(self.menv, v_name, v_type) def _enum(self, v_name: str, enum_size: int): c_name = "{}_{}".format(self.name, v_name) return make_enum(self.menv, c_name, enum_size) class Sender(Module): def __init__(self, name: str, menv: msat_env, enc: LTLEncoder, in_c, x_in_c, out_c, x_out_c, delta): super().__init__(name, menv, enc) bool_type = msat_get_bool_type(menv) int_type = msat_get_integer_type(menv) real_type = msat_get_rational_type(menv) loc, x_loc = self._symb("l", bool_type) evt, x_evt = self._symb("evt", bool_type) msg_id, x_msg_id = self._symb("msg_id", int_type) timeout, x_timeout = self._symb("timeout", real_type) c, x_c = self._symb("c", real_type) self.move = evt self.stutter = msat_make_not(menv, evt) self.x_move = x_evt self.x_stutter = msat_make_not(menv, x_evt) self.send = loc self.wait_ack = msat_make_not(menv, loc) self.x_send = x_loc self.x_wait_ack = msat_make_not(menv, x_loc) self.symb2next = {loc: x_loc, evt: x_evt, msg_id: x_msg_id, timeout: x_timeout, c: x_c} zero = msat_make_number(menv, "0") one = msat_make_number(menv, "1") base_timeout = one # send & c = 0 & msg_id = 0 self.init = msat_make_and(menv, msat_make_and(menv, self.send, msat_make_equal(menv, c, zero)), msat_make_equal(menv, msg_id, zero)) # invar: wait_ack -> c <= timeout self.init = msat_make_and( menv, self.init, msat_make_impl(menv, self.wait_ack, msat_make_leq(menv, c, timeout))) self.trans = msat_make_impl(menv, self.x_wait_ack, msat_make_leq(menv, x_c, x_timeout)) # delta > 0 | stutter -> l' = l & msg_id' = msg_id & timeout' = timeout & # c' = c + delta & out_c' = out_c lhs = msat_make_or(menv, msat_make_gt(menv, delta, zero), self.stutter) rhs = msat_make_and( menv, msat_make_and(menv, msat_make_iff(menv, x_loc, loc), msat_make_equal(menv, x_msg_id, msg_id)), msat_make_and(menv, msat_make_equal(menv, x_timeout, timeout), msat_make_equal(menv, x_c, msat_make_plus(menv, c, delta)))) rhs = msat_make_and(menv, rhs, msat_make_equal(menv, x_out_c, out_c)) self.trans = msat_make_and(menv, self.trans, msat_make_impl(menv, lhs, rhs)) disc_t = msat_make_and(menv, self.move, msat_make_equal(menv, delta, zero)) # (send & send') -> # (msg_id' = msg_id & timeout' = base_timeout & c' = 0 & out_c' = out_c) lhs = msat_make_and(menv, disc_t, msat_make_and(menv, self.send, self.x_send)) rhs = msat_make_and( menv, msat_make_and(menv, msat_make_equal(menv, x_msg_id, msg_id), msat_make_equal(menv, x_timeout, base_timeout)), msat_make_and(menv, msat_make_equal(menv, x_c, zero), msat_make_equal(menv, x_out_c, out_c))) self.trans = msat_make_and(menv, self.trans, msat_make_impl(menv, lhs, rhs)) # (send & wait_ack') -> # (msg_id' = msg_id + 1 & timeout' = base_timeout & c' = 0 & out_c' = out_c) lhs = msat_make_and(menv, disc_t, msat_make_and(menv, self.send, self.x_wait_ack)) rhs = msat_make_and( menv, msat_make_and(menv, msat_make_equal(menv, x_msg_id, msat_make_plus(menv, msg_id, one)), msat_make_equal(menv, x_timeout, base_timeout)), msat_make_and(menv, msat_make_equal(menv, x_c, zero), msat_make_equal(menv, x_out_c, out_c))) self.trans = msat_make_and(menv, self.trans, msat_make_impl(menv, lhs, rhs)) # (wait_ack) -> (c' = 0 & out_c' = out_c & # (wait_ack' <-> (in_c != msg_id & c > timeout)) lhs = msat_make_and(menv, disc_t, self.wait_ack) rhs_iff = msat_make_and(menv, msat_make_not(menv, msat_make_equal(menv, in_c, msg_id)), msat_make_geq(menv, c, timeout)) rhs_iff = msat_make_iff(menv, self.x_wait_ack, rhs_iff) rhs = msat_make_and(menv, msat_make_and(menv, msat_make_equal(menv, x_c, zero), msat_make_equal(menv, x_out_c, out_c)), rhs_iff) self.trans = msat_make_and(menv, self.trans, msat_make_impl(menv, lhs, rhs)) # (wait_ack & wait_ack') -> (timeout' > timeout) lhs = msat_make_and(menv, disc_t, msat_make_and(menv, self.wait_ack, self.x_wait_ack)) rhs = msat_make_gt(menv, x_timeout, timeout) self.trans = msat_make_and(menv, self.trans, msat_make_impl(menv, lhs, rhs)) # (wait_ack) -> (send' <-> (in_c = msg_id & c < timeout)) lhs = msat_make_and(menv, disc_t, self.wait_ack) rhs = msat_make_iff(menv, self.x_send, msat_make_and(menv, msat_make_equal(menv, in_c, msg_id), msat_make_lt(menv, c, timeout))) self.trans = msat_make_and(menv, self.trans, msat_make_impl(menv, lhs, rhs)) # (wait_ack & send') -> (timeout' = base_timeout) lhs = msat_make_and(menv, disc_t, msat_make_and(menv, self.wait_ack, self.x_send)) rhs = msat_make_equal(menv, x_timeout, base_timeout) self.trans = msat_make_and(menv, self.trans, msat_make_impl(menv, lhs, rhs)) class Receiver(Module): def __init__(self, name: str, menv: msat_env, enc: LTLEncoder, in_c, x_in_c, out_c, x_out_c, delta): super().__init__(name, menv, enc) bool_type = msat_get_bool_type(menv) loc, x_loc = self._symb("l", bool_type) self.wait = loc self.work = msat_make_not(menv, loc) self.x_wait = x_loc self.x_work = msat_make_not(menv, x_loc) self.symb2next = {loc: x_loc} zero = msat_make_number(menv, "0") # wait self.init = self.wait # delta > 0 -> loc' = loc & out_c' = out_c lhs = msat_make_gt(menv, delta, zero) rhs = msat_make_and(menv, msat_make_iff(menv, x_loc, loc), msat_make_equal(menv, x_out_c, out_c)) self.trans = msat_make_impl(menv, lhs, rhs) disc_t = msat_make_equal(menv, delta, zero) # wait -> (wait' <-> in_c = out_c) lhs = msat_make_and(menv, disc_t, self.wait) rhs = msat_make_iff(menv, self.x_wait, msat_make_equal(menv, in_c, out_c)) self.trans = msat_make_and(menv, self.trans, msat_make_impl(menv, lhs, rhs)) # (wait & wait') -> (out_c' = out_c) lhs = msat_make_and(menv, disc_t, msat_make_and(menv, self.wait, self.x_wait)) rhs = msat_make_equal(menv, x_out_c, out_c) self.trans = msat_make_and(menv, self.trans, msat_make_impl(menv, lhs, rhs)) # (wait & work') -> out_c' = in_c lhs = msat_make_and(menv, disc_t, msat_make_and(menv, self.wait, self.x_work)) rhs = msat_make_equal(menv, x_out_c, in_c) self.trans = msat_make_and(menv, self.trans, msat_make_impl(menv, lhs, rhs)) # work -> out_c' = out_c lhs = msat_make_and(menv, disc_t, self.work) rhs = msat_make_equal(menv, x_out_c, out_c) self.trans = msat_make_and(menv, self.trans, msat_make_impl(menv, lhs, rhs)) def hints(env: PysmtEnv) -> FrozenSet[Hint]: assert isinstance(env, PysmtEnv) mgr = env.formula_manager delta = mgr.Symbol(delta_name, types.REAL) r2s = mgr.Symbol("r2s", types.INT) s2r = mgr.Symbol("r2s", types.INT) s_l = mgr.Symbol("s_l", types.BOOL) s_evt = mgr.Symbol("s_evt", types.BOOL) s_msg_id = mgr.Symbol("s_msg_id", types.INT) s_timeout = mgr.Symbol("s_timeout", types.REAL) s_c = mgr.Symbol("s_c", types.REAL) r_l = mgr.Symbol("r_l", types.BOOL) symbs = frozenset([delta, r2s, s2r, s_l, s_evt, s_msg_id, s_timeout, s_c, r_l]) x_delta = symb_to_next(mgr, delta) x_r2s = symb_to_next(mgr, r2s) x_s2r = symb_to_next(mgr, s2r) x_s_l = symb_to_next(mgr, s_l) x_s_evt = symb_to_next(mgr, s_evt) x_s_msg_id = symb_to_next(mgr, s_msg_id) x_s_timeout = symb_to_next(mgr, s_timeout) x_s_c = symb_to_next(mgr, s_c) x_r_l = symb_to_next(mgr, r_l) res = [] r0 = mgr.Real(0) r1 = mgr.Real(1) i0 = mgr.Int(0) i1 = mgr.Int(1) loc0 = Location(env, mgr.Equals(delta, r0)) loc0.set_progress(0, mgr.Equals(x_delta, r0)) hint = Hint("h_delta0", env, frozenset([delta]), symbs) hint.set_locs([loc0]) res.append(hint) loc0 = Location(env, mgr.Equals(s2r, i0)) loc0.set_progress(0, mgr.Equals(x_s2r, i0)) hint = Hint("h_s2r0", env, frozenset([s2r]), symbs) hint.set_locs([loc0]) res.append(hint) loc0 = Location(env, mgr.Equals(r2s, i0)) loc0.set_progress(0, mgr.Equals(x_r2s, i0)) hint = Hint("h_r2s0", env, frozenset([r2s]), symbs) hint.set_locs([loc0]) res.append(hint) loc0 = Location(env, s_l) loc0.set_progress(0, x_s_l) hint = Hint("h_s_l0", env, frozenset([s_l]), symbs) hint.set_locs([loc0]) res.append(hint) loc0 = Location(env, s_evt) loc0.set_progress(0, x_s_evt) hint = Hint("h_s_evt0", env, frozenset([s_evt]), symbs) hint.set_locs([loc0]) res.append(hint) loc0 = Location(env, mgr.Equals(s_msg_id, i0)) loc0.set_progress(0, mgr.Equals(x_s_msg_id, i0)) hint = Hint("h_s_msg_id0", env, frozenset([s_msg_id]), symbs) hint.set_locs([loc0]) res.append(hint) loc0 = Location(env, mgr.Equals(s_timeout, r0)) loc0.set_progress(0, mgr.Equals(x_s_timeout, r0)) hint = Hint("h_s_timeout0", env, frozenset([s_timeout]), symbs) hint.set_locs([loc0]) res.append(hint) loc0 = Location(env, mgr.Equals(s_c, r0)) loc0.set_progress(0, mgr.Equals(x_s_c, r0)) hint = Hint("h_s_c0", env, frozenset([s_c]), symbs) hint.set_locs([loc0]) res.append(hint) loc0 = Location(env, r_l) loc0.set_progress(0, x_r_l) hint = Hint("h_r_l0", env, frozenset([r_l]), symbs) hint.set_locs([loc0]) res.append(hint) loc0 = Location(env, mgr.GE(delta, r0)) loc0.set_progress(0, mgr.Equals(x_delta, r1)) hint = Hint("h_delta1", env, frozenset([delta]), symbs) hint.set_locs([loc0]) res.append(hint) loc0 = Location(env, mgr.GE(s2r, i0)) loc0.set_progress(0, mgr.Equals(x_s2r, i1)) hint = Hint("h_s2r1", env, frozenset([s2r]), symbs) hint.set_locs([loc0]) res.append(hint) loc0 = Location(env, mgr.GE(r2s, i0)) loc0.set_progress(0, mgr.Equals(x_r2s, i1)) hint = Hint("h_r2s1", env, frozenset([r2s]), symbs) hint.set_locs([loc0]) res.append(hint) loc0 = Location(env, s_evt) loc0.set_progress(1, mgr.Not(x_s_evt)) loc1 = Location(env, mgr.Not(s_evt)) loc1.set_progress(0, x_s_evt) hint = Hint("h_s_evt1", env, frozenset([s_evt]), symbs) hint.set_locs([loc0, loc1]) res.append(hint) loc0 = Location(env, mgr.GE(s_msg_id, i0)) loc0.set_progress(0, mgr.Equals(x_s_msg_id, mgr.Plus(s_msg_id, i1))) hint = Hint("h_s_msg_id1", env, frozenset([s_msg_id]), symbs) hint.set_locs([loc0]) res.append(hint) loc0 = Location(env, mgr.GE(s_timeout, r0)) loc0.set_progress(0, mgr.Equals(x_s_timeout, mgr.Plus(s_timeout, r1))) hint = Hint("h_s_timeout1", env, frozenset([s_timeout]), symbs) hint.set_locs([loc0]) res.append(hint) loc0 = Location(env, mgr.GE(s_c, r0)) loc0.set_progress(0, mgr.Equals(x_s_c, mgr.Plus(s_c, r1))) hint = Hint("h_s_c1", env, frozenset([s_c]), symbs) hint.set_locs([loc0]) res.append(hint) loc0 = Location(env, r_l) loc0.set_progress(1, mgr.Not(x_r_l)) loc1 = Location(env, mgr.Not(r_l)) loc1.set_progress(0, x_r_l) hint = Hint("h_r_l1", env, frozenset([r_l]), symbs) hint.set_locs([loc0, loc1]) res.append(hint) loc0 = Location(env, mgr.GE(delta, r0)) loc0.set_progress(0, mgr.Equals(x_delta, mgr.Plus(delta, r1))) hint = Hint("h_delta2", env, frozenset([delta]), symbs) hint.set_locs([loc0]) res.append(hint) loc0 = Location(env, mgr.GE(s2r, i0)) loc0.set_progress(0, mgr.Equals(x_s2r, mgr.Plus(s2r, i1))) hint = Hint("h_s2r2", env, frozenset([s2r]), symbs) hint.set_locs([loc0]) res.append(hint) return frozenset(res)
[ "en.magnago@gmail.com" ]
en.magnago@gmail.com
5e795c4e386973917ef112bcb637b06aa09e031e
4a31e84a4a6cf5c4cdf4757cca63ed6ff92088bc
/Python/optimization/ga.py
bc5e41cfed7bdb031e3b8adfb4fe881261d36a7f
[]
no_license
aldo-fsm/tests
e91ac45709a7dbffdd7984f1bc799a68d1c86d9f
9d0317236f459a00af56aaee738b2a46365a7b99
refs/heads/master
2021-01-20T08:44:12.225578
2018-07-17T17:48:05
2018-07-17T17:48:05
101,570,618
0
0
null
null
null
null
UTF-8
Python
false
false
5,329
py
import numpy as np import re def optimize(costFunction, chromSize, popSize, selectionRate, mutRate, **kwargs): kwargs.setdefault('pairing', 'rank_weighting') kwargs.setdefault('pairingTounamentSize', 3) kwargs.setdefault('crossover', 'single_point') kwargs.setdefault('uniformCrossoverProb', 0.5) popKeepSize = round(selectionRate*popSize) pop = np.round(np.random.rand(popSize, chromSize)) numPairs = np.ceil((popSize-popKeepSize)/2) generation = 0 while True: print("Generation: {}".format(generation)) costs = np.array([costFunction(chrom) for chrom in pop]) sortIndexes = np.argsort(costs) costs = costs[sortIndexes] pop = pop[sortIndexes] for (chrom, cost) in zip(pop, costs): print(" {0} ..... {1}".format(re.sub(r'\D', '', str(chrom)), cost)) popKeep = pop[:popKeepSize] parents1, parents2 = selectPairs(popKeep, costs, numPairs, kind=kwargs['pairing'], tournamentSize=kwargs['pairingTounamentSize']) children = crossover(parents1, parents2, kind=kwargs['crossover'], uniformProb=kwargs['uniformCrossoverProb'])[:popSize-popKeepSize] pop = np.concatenate([popKeep,children], axis=0) yield pop mutation(pop, np.ceil((popSize-1)*chromSize*mutRate)) generation+=1 def selectPairs(popKeep, costs, numberPairs, kind='rank_weighting', **kwargs): kwargs.setdefault('tournamentSize', 3) popKeepSize = len(popKeep) numberPairs = int(numberPairs) if kind == 'top_to_bottom': p1 = np.arange(0, popKeepSize, 2)[:numberPairs] p2 = p1 + 1 elif kind == 'random': p1 = np.random.choice(popKeepSize, numberPairs) p2 = np.random.choice(popKeepSize, numberPairs) elif kind == 'cost_weighting': costs = (costs-costs[popKeepSize])[:popKeepSize] costsSum = np.sum(costs[:popKeepSize]) if costsSum == 0: return selectPairs(popKeep, costs, numberPairs, kind='random') prob = costs/costsSum cumProb = np.cumsum(prob) random1 = np.random.rand(numberPairs) random2 = np.random.rand(numberPairs) p1 = [argfirst(lambda n : n > r, cumProb) for r in random1] p2 = [argfirst(lambda n : n > r, cumProb) for r in random2] for i in range(len(p1)): if p1[i] == p2[i] : p2[i] = np.random.choice(popKeepSize) elif kind == 'rank_weighting': ranks = list(range(popKeepSize)) ranksSum = np.sum(ranks) prob = [(popKeepSize-i-1)/ranksSum for i in ranks] cumProb = np.cumsum(prob) random1 = np.random.rand(numberPairs) random2 = np.random.rand(numberPairs) p1 = [argfirst(lambda n : n > r, cumProb) for r in random1] p2 = [argfirst(lambda n : n > r, cumProb) for r in random2] for i in range(len(p1)): if p1[i] == p2[i] : p2[i] = np.random.choice(popKeepSize) elif kind == 'tournament': p1 = [] p2 = [] for _ in range(numberPairs): group1 = np.random.choice(popKeepSize, kwargs['tournamentSize']) group2 = np.random.choice(popKeepSize, kwargs['tournamentSize']) p1.append(np.argmin(costs[group1])) p2.append(np.argmin(costs[group2])) return (popKeep[p1], popKeep[p2]) def crossover(parents1, parents2, kind='single_point', **kwargs): kwargs.setdefault('uniformProb', 0.5) children = [] if kind == 'single_point': for p1, p2 in zip(parents1,parents2): crossoverPoint = np.random.choice(len(p1)-1)+1 # pylint: disable=unbalanced-tuple-unpacking p1Left, p1Right = np.split(p1, [crossoverPoint]) p2Left, p2Right = np.split(p2, [crossoverPoint]) c1 = np.concatenate([p1Left, p2Right]) c2 = np.concatenate([p2Left, p1Right]) children += [c1, c2] elif kind == 'double_point': for p1, p2 in zip(parents1,parents2): crossoverPoints = np.sort(np.random.choice(len(p1)-1, 2, replace=False)+1) # pylint: disable=unbalanced-tuple-unpacking l1, m1, r1 = np.split(p1, crossoverPoints) l2, m2, r2 = np.split(p2, crossoverPoints) c1 = np.concatenate([l1, m2, r1]) c2 = np.concatenate([l2, m1, r2]) children += [c1, c2] elif kind == 'uniform': for p1, p2 in zip(parents1, parents2): prob = kwargs['uniformProb'] aux = list(map(lambda x: int(x >= prob), np.random.rand(len(p1)))) pair = [p1, p2] c1 = np.array([pair[aux[i]][i] for i in range(len(p1))]) c2 = np.array([pair[::-1][aux[i]][i] for i in range(len(p1))]) children += [c1, c2] return children def mutation(pop, numMut): popSize = len(pop) chromSize = len(pop[0]) chrom = np.random.choice(popSize-1, int(numMut))+1 bit = np.random.choice(chromSize, int(numMut)) for i, j in zip(chrom, bit): pop[i,j] = np.abs(pop[i,j]-1) def argfirst(condition, iterable): return next(x[0] for x in enumerate(iterable) if condition(x[1]))
[ "aldo.fsmonteiro@gmail.com" ]
aldo.fsmonteiro@gmail.com
32a624033c4fcd4b0dab2f56ea427738fac85532
0fd5793e78e39adbfe9dcd733ef5e42390b8cc9a
/python3/16_Web_Services/k_Projects/b_twitter/twitter_scrapping.py
e22fe2652cae147f89fc3a8955b3336f6f812e4b
[]
no_license
udhayprakash/PythonMaterial
3ea282ceb4492d94d401e3bc8bad9bf6e9cfa156
e72f44e147141ebc9bf9ec126b70a5fcdbfbd076
refs/heads/develop
2023-07-08T21:07:33.154577
2023-07-03T10:53:25
2023-07-03T10:53:25
73,196,374
8
5
null
2023-05-26T09:59:17
2016-11-08T14:55:51
Jupyter Notebook
UTF-8
Python
false
false
2,215
py
#!/usr/bin/python """ Purpose: Twitter data scrapping """ import tweepy class TwitterLogin: def __init__(self): consumer_key = "xxxxxxxxxxxxxxxxxxxxx" consumer_secret = "xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx" access_token = "00000-xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx" access_token_secret = "xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx" auth = tweepy.OAuthHandler(consumer_key, consumer_secret) auth.set_access_token(access_token, access_token_secret) self.api = tweepy.API( auth, wait_on_rate_limit=True, wait_on_rate_limit_notify=True ) def credentials_verification(self): result = vars(self.api.verify_credentials())["_json"] print( f"""Account User : {result['name']} Screen Name : {result['screen_name']} Location : {result['location']} Profile description : {result['description']} Account Created at : {result['created_at']} Display URL : {result['entities']['url']['urls'][0]['display_url']} """ ) class TwitterScrapping(TwitterLogin): def __init__(self): TwitterLogin.__init__(self) # twtr = TwitterLogin() # twtr.credentials_verification() twrt_scrp = TwitterScrapping() twrt_scrp.credentials_verification() # Latest Public Timeline tweet = twrt_scrp.api.home_timeline()[0] print( f""" tweet.text : {tweet.text} tweet.contributors : {tweet.contributors} tweet.created_at : {tweet.created_at} tweet.lang : {tweet.lang} tweet.source : {tweet.source} tweet.source_url : {tweet.source_url} tweet.truncated : {tweet.truncated} tweet.retweet_count : {tweet.retweet_count} tweet.retweeted : {tweet.retweeted} tweet.retweet : {tweet.retweet} tweet.retweets : {tweet.retweets} tweet.possibly_sensitive : {tweet.possibly_sensitive} tweet.possibly_sensitive_appealable : {tweet.possibly_sensitive_appealable} """ )
[ "uday3prakash@gmail.com" ]
uday3prakash@gmail.com
0154452e8030cb69acdc0d908d2e28193782d8cc
1f9f0e49c2dd5104e24a63416a4f773c017793db
/parse_dir.py
eee4110418bbbda9f46243fcea474adbbd76cad9
[]
no_license
sameeul/duplication_detector
ca6cb957e24e443da92cccea1342d0d4484d8e93
3e374ea0274c1983e78e416a623cd05da3f270a5
refs/heads/master
2020-09-13T07:33:30.400569
2019-11-22T03:19:45
2019-11-22T03:19:45
222,696,689
0
0
null
null
null
null
UTF-8
Python
false
false
3,137
py
import argparse import operator #set up command line processing parser = argparse.ArgumentParser(description = "Find duplication in a directory") parser.add_argument("--name", required = True, type = str, help = "Text file containing file listing") args = parser. parse_args() input_file_name=args.name # now work on finding duplicates file_dict = {} status_code = {"release":2, "review":3, "in-process":5} with open(input_file_name) as fp: for line in fp: line = line.strip() file_name_data = line.split("\\") if len(file_name_data) == 3: file_name = file_name_data[-1].lower() status = file_name_data[1].lower() if status in status_code: if file_name in file_dict: file_dict[file_name] = file_dict[file_name]*status_code[status] else: file_dict[file_name] = status_code[status] #processing done, now print file_in_two_dir = [] file_in_three_dir = [] for file_name in file_dict: if file_dict[file_name] > 5 and file_dict[file_name] < 30: file_in_two_dir.append(file_name) if file_dict[file_name] == 30: file_in_three_dir.append(file_name) print("Total %d files are unique"%(len(file_dict)-len(file_in_two_dir)-len(file_in_three_dir))) print("Total %d files are duplicate across two directories"%(len(file_in_two_dir))) print("Total %d files are duplicate across three directories"%(len(file_in_three_dir))) print("Total %d files are duplicate"%(len(file_in_two_dir)+len(file_in_three_dir))) duplicate_files = file_in_two_dir+file_in_three_dir if len(duplicate_files) != 0: duplicate_files.sort() print("\nList of duplicate files:\n") print("Review | Release | In-Process | Filename") print("----------------------------------------") for file_name in duplicate_files: line = "" if file_dict[file_name] % status_code["review"] == 0: line = line + " X " else: line = line + " " if file_dict[file_name] % status_code["release"] == 0: line = line + " X " else: line = line + " " if file_dict[file_name] % status_code["in-process"] == 0: line = line + " X " else: line = line + " " line = line+file_name print(line) #list top 10 duplication series #optional series_dict = {} for file in duplicate_files: file_comp = file.split("-") series_start_code = file_comp[0] if len(file_comp) > 2: if series_start_code in series_dict: series_dict[series_start_code] = series_dict[series_start_code]+1 else: series_dict[series_start_code] = 1 sorted_series_list = sorted(series_dict.items(), key=operator.itemgetter(1), reverse=True) print("\nMost Duplicated Series:\n") print("Series Code Number of Duplicates") print("------------------------------------------") for data in sorted_series_list[0:10]: print("%s %d"%(data[0], data[1]))
[ "noreply@github.com" ]
sameeul.noreply@github.com
277eefec07dfbb8604106fa7fb9dbdc5dff111a0
4819d5b68ac2708cd9f7e238790568f6fb62bf4a
/shop/ShopSerializers.py
c96c2565256fbd1eb6e73cbd9da8599cb60a0676
[]
no_license
suyashgithub/django-rest
f8c60efaa95ad74220be6d44c766181614b55adc
bb41c41713983ee95fab5d1a799b96231b817305
refs/heads/master
2022-11-29T06:02:04.119874
2020-08-13T22:08:21
2020-08-13T22:08:21
287,384,902
0
0
null
null
null
null
UTF-8
Python
false
false
1,009
py
from rest_framework import serializers from shop.models import Shop class ShopSerializer(serializers.ModelSerializer): created_at = serializers.DateTimeField(format="%Y-%m-%d %H:%M:%S", read_only=True) username = serializers.SerializerMethodField("get_username") class Meta: model = Shop fields = ['id', 'status', 'name', 'user','username', 'created_at'] def get_username(self, obj): return obj.user.username def create(self, validated_data): """ Create and return a new `Shop` instance, given the validated data. """ meetings = Shop.objects.create(**validated_data) return meetings def update(self, instance, validated_data): """ Update and return an existing `Shop` instance, given the validated data. """ instance.name = validated_data.get('name', instance.name) instance.status = validated_data.get('status', instance.status) instance.save() return instance
[ "suyash.mishra@thinkmantra.com" ]
suyash.mishra@thinkmantra.com
04cb69832818c5318fc8810a89edf011e8167b66
29be7382240c4f5ae23a5c06c75b8dda8043800c
/PRO-99/removeFiles.py
1c67cc9e2e7c73e21edf8f6268542f638458bd93
[]
no_license
ABHINAV-KUMAR213/PRO-99
d05e37e91d1305c7a4494475f05bf0d47adc48e2
f282ed9429b20bc7ab87e066c1ceb5cab1edd734
refs/heads/main
2023-06-10T11:52:16.112611
2021-06-29T12:52:54
2021-06-29T12:52:54
381,363,744
0
0
null
null
null
null
UTF-8
Python
false
false
1,615
py
import os import shutil import time def main(): deleted_folders_count = 0 deleted_files_count = 0 path = "e:\Coding\Python\Project99" days = 30 seconds = time.time() - (days * 24 * 60 * 60) if os.path.exists(path): for root_folder, folders, files in os.walk(path): if seconds >= get_file_or_folder_age(root_folder): remove_folder(root_folder) deleted_folders_count += 1 break else: for folder in folders: folder_path = os.path.join(root_folder, folder) if seconds >= get_file_or_folder_age(folder_path): remove_folder(folder_path) deleted_folders_count += 1 for file in files: file_path = os.path.join(root_folder, file) if seconds >= get_file_or_folder_age(file_path): remove_file(file_path) deleted_files_count += 1 else: if seconds >= get_file_or_folder_age(path): remove_file(path) deleted_files_count += 1 else: print(f'"{path}" is not found') deleted_files_count += 1 print(f"Total folders deleted: {deleted_folders_count}") print(f"Total files deleted: {deleted_files_count}") def remove_folder(path): if not shutil.rmtree(path): print(f"{path} is removed successfully") else: print(f"Unable to delete the "+path) def remove_file(path): if not os.remove(path): print(f"{path} is removed successfully") else: print("Unable to delete the "+path) def get_file_or_folder_age(path): ctime = os.stat(path).st_ctime return ctime if __name__ == '__main__': main()
[ "noreply@github.com" ]
ABHINAV-KUMAR213.noreply@github.com
1b5acb3b130ca25e0213d632d2799da91b52eb84
3021d12339945053e89f9d9571f8054d86963088
/scripts/sma-client.py
205f5fd47aa8db61eb60b6e5fe90ad80c3ead6e0
[ "Intel", "BSD-2-Clause", "BSD-3-Clause" ]
permissive
tmakatos/spdk
9d18a586000fc25f659bb185ebb30b68e0073aa7
a82d6f009f1d43cb8107ec0637d9e458c319b369
refs/heads/master
2022-11-15T06:32:52.922360
2022-05-11T16:24:09
2022-10-21T16:42:58
220,963,857
0
1
NOASSERTION
2020-09-16T20:46:56
2019-11-11T11:14:11
C
UTF-8
Python
false
false
2,551
py
#!/usr/bin/env python3 from argparse import ArgumentParser import grpc import google.protobuf.json_format as json_format import importlib import json import logging import os import sys sys.path.append(os.path.dirname(__file__) + '/../python') import spdk.sma.proto.sma_pb2 as sma_pb2 # noqa import spdk.sma.proto.sma_pb2_grpc as sma_pb2_grpc # noqa import spdk.sma.proto.nvmf_tcp_pb2 as nvmf_tcp_pb2 # noqa import spdk.sma.proto.nvmf_tcp_pb2_grpc as nvmf_tcp_pb2_grpc # noqa class Client: def __init__(self, addr, port): self._service = sma_pb2.DESCRIPTOR.services_by_name['StorageManagementAgent'] self.addr = addr self.port = port def _get_message_type(self, descriptor): return getattr(sma_pb2, descriptor.name) def _get_method_types(self, method_name): method = self._service.methods_by_name.get(method_name) return (self._get_message_type(method.input_type), self._get_message_type(method.output_type)) def call(self, method, params): with grpc.insecure_channel(f'{self.addr}:{self.port}') as channel: stub = sma_pb2_grpc.StorageManagementAgentStub(channel) func = getattr(stub, method) input, output = self._get_method_types(method) response = func(request=json_format.ParseDict(params, input())) return json_format.MessageToDict(response, preserving_proto_field_name=True) def load_plugins(plugins): for plugin in plugins: logging.debug(f'Loading external plugin: {plugin}') module = importlib.import_module(plugin) def parse_argv(): parser = ArgumentParser(description='Storage Management Agent client') parser.add_argument('--address', '-a', default='localhost', help='IP address of SMA instance to connect to') parser.add_argument('--port', '-p', default=8080, type=int, help='Port number of SMA instance to connect to') return parser.parse_args() def main(args): argv = parse_argv() logging.basicConfig(level=os.environ.get('SMA_LOGLEVEL', 'WARNING').upper()) load_plugins(filter(None, os.environ.get('SMA_PLUGINS', '').split(':'))) client = Client(argv.address, argv.port) request = json.loads(sys.stdin.read()) result = client.call(request['method'], request.get('params', {})) print(json.dumps(result, indent=2)) if __name__ == '__main__': main(sys.argv[1:])
[ "tomasz.zawadzki@intel.com" ]
tomasz.zawadzki@intel.com
6ebf11b3f019ebe0338ba4e09bbe5dcd2b7fbd4f
ec4e153f3bf1b335bc1b31b85e6f9db4a6c4faa9
/wd_extractor/Document.py
6ed52824a553446bd88f07562c5ca97fb6fb3529
[ "Apache-2.0", "CC-BY-3.0" ]
permissive
DuaneNielsen/wd_extractor
7936ac29ae97972cfe74973108aaad1efa5054b6
128a189bacd0cd2d7f1fa598202b9c4e55f48e2f
refs/heads/master
2021-01-19T14:13:42.441554
2017-09-19T02:16:08
2017-09-19T02:16:08
100,887,646
0
0
null
null
null
null
UTF-8
Python
false
false
903
py
from .Graminator import Graminator class Document: def __init__(self, corpus, path, grams): self.corpus = corpus self.grams = grams self.graminator = None self.path = path self.tokens = corpus.tokenizer.tokens(self) def getText(self): if self.path is not None: handle = open(self.path, "r") text = handle.read() return text def length(self): return len(self.tokens) def nGrams(self, gramsize): return Graminator(self, gramsize) def hasNext(self, index): index += 1 return (index > 0) and index < len(self.tokens) def nextToken(self, index): return self.tokens[index + 1] def hasPrev(self, index): index -= 1 return (index > 0) and index < len(self.tokens) def prevToken(self, index): return self.tokens[index-1]
[ "duane.nielsen.rocks@gmail.com" ]
duane.nielsen.rocks@gmail.com
6cf4963fd3bf6b82bb24b9ffda645ecbcba72071
b10163d7036e0bd2bd2a59f1a8a96951ed169c21
/arvore de natal.py
6ac49eb9524abea4b33adf4c6ca5d714d0eb8552
[]
no_license
YK0L0DIY/Python
4e97e3e180b51e6571d667a3996b04851461fd62
1991cdb8669c6e532766231ce63854956ec5b0e8
refs/heads/master
2020-03-19T13:59:46.415839
2019-02-15T20:27:06
2019-02-15T20:27:06
136,604,461
0
0
null
null
null
null
UTF-8
Python
false
false
269
py
def arvore(a): v=a cont=2 u=a-1 print(" "*u,"*") while(v!=1): u=u-1 print(" "*u,"*"*(1+cont)) cont=cont+2 v=v-1 print(" "*(a-1),"*") print(" "*(a-1),"*") h=int(input("Indique a altura da arvore: ")) arvore(h)
[ "noreply@github.com" ]
YK0L0DIY.noreply@github.com
12eca4b3e8ae4bf6f27c07a03bbc58a313b36f5f
d668209e9951d249020765c011a836f193004c01
/tools/pnnx/tests/test_torch_fft_irfft.py
8f92dd551a1f5c2f0b5ff9c8894b75b1b122d362
[ "BSD-3-Clause", "Zlib", "BSD-2-Clause" ]
permissive
Tencent/ncnn
d8371746c00439304c279041647362a723330a79
14b000d2b739bd0f169a9ccfeb042da06fa0a84a
refs/heads/master
2023-08-31T14:04:36.635201
2023-08-31T04:19:23
2023-08-31T04:19:23
95,879,426
18,818
4,491
NOASSERTION
2023-09-14T15:44:56
2017-06-30T10:55:37
C++
UTF-8
Python
false
false
1,804
py
# Tencent is pleased to support the open source community by making ncnn available. # # Copyright (C) 2022 THL A29 Limited, a Tencent company. All rights reserved. # # Licensed under the BSD 3-Clause License (the "License"); you may not use this file except # in compliance with the License. You may obtain a copy of the License at # # https://opensource.org/licenses/BSD-3-Clause # # Unless required by applicable law or agreed to in writing, software distributed # under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR # CONDITIONS OF ANY KIND, either express or implied. See the License for the # specific language governing permissions and limitations under the License. import torch import torch.nn as nn import torch.nn.functional as F class Model(nn.Module): def __init__(self): super(Model, self).__init__() def forward(self, x, y, z): x = torch.fft.irfft(x, norm="backward") y = torch.fft.irfft(y, dim=(1), norm="forward") z = torch.fft.irfft(z, norm="ortho") return x, y, z def test(): net = Model() net.eval() torch.manual_seed(0) x = torch.rand(1, 3, 120, 120) y = torch.rand(1, 100, 2, 120) z = torch.rand(1, 20, 20) a = net(x, y, z) # export torchscript mod = torch.jit.trace(net, (x, y, z)) mod.save("test_torch_fft_irfft.pt") # torchscript to pnnx import os os.system("../src/pnnx test_torch_fft_irfft.pt inputshape=[1,3,120,120],[1,100,2,120],[1,20,20]") # pnnx inference import test_torch_fft_irfft_pnnx b = test_torch_fft_irfft_pnnx.test_inference() for a0, b0 in zip(a, b): if not torch.equal(a0, b0): return False return True if __name__ == "__main__": if test(): exit(0) else: exit(1)
[ "noreply@github.com" ]
Tencent.noreply@github.com
0d3ffa1d191e50eef4a95f5452a805a99887284a
c65028cd7520d6ac4b109eb4fafc470bfafcabfb
/logit.py
2970708f6eefa3453ef031a0e0a254360824535d
[]
no_license
devsunb/commit-log-exporter
17ca29eb2ede7cfffb140101fa002bffc6567549
5c58b9d5d902ac017a05bf6a9451390f65c5560d
refs/heads/master
2023-07-26T14:36:12.141223
2021-08-31T08:04:56
2021-08-31T08:04:56
401,622,142
0
0
null
null
null
null
UTF-8
Python
false
false
1,654
py
import logging import re import subprocess logger = logging.getLogger('LOGIT') class Logit: @staticmethod def parse(commit_lines): commits = [] commit = {} for line in commit_lines: if line == '' or line == '\n' or re.match('merge:', line, re.IGNORECASE): continue elif re.match('commit', line, re.IGNORECASE): if len(commit) != 0: commits.append(commit) commit = {'hash': re.match('commit (.*)', line, re.IGNORECASE).group(1)} elif re.match('Author:', line, re.IGNORECASE): m = re.compile('Author: (.*) <(.*)>').match(line) commit['author'] = m.group(1) commit['email'] = m.group(2) elif re.match('Date:', line, re.IGNORECASE): m = re.compile('Date: (.*)').match(line) commit['date'] = m.group(1) elif re.match(' ', line, re.IGNORECASE): if 'message' not in commit: commit['message'] = line.strip() else: commit['message'] += '\n' + line.strip() else: logger.error('Unexpected Line: ' + line) if len(commit) != 0: commits.append(commit) return commits def log(self, cwd, path): output = '' with subprocess.Popen(['git', 'log', '--', path], cwd=cwd, stdout=subprocess.PIPE, bufsize=1, universal_newlines=True) as p: for line in p.stdout: output += line return self.parse(output.split('\n'))
[ "devsunb@gmail.com" ]
devsunb@gmail.com
406f62c74fe702355e8c72a821dc3983799b0ff2
2477ba1cd7e04b641f95b800e188d7525d39ab46
/Monochrome_sketches/stripe/stripe.pyde
31e0e239c6a31336735627ca8ac6824354f8348a
[]
no_license
lsm35/ProcessingSketchbook
ec6ddf875caa87f1ccfefefde690c9dd6a0260f8
d15ae1dcdcc920f583bc911c0a5cf72f58821dd1
refs/heads/master
2021-03-31T02:00:24.209133
2018-03-17T00:48:40
2018-03-17T00:48:40
124,572,119
0
0
null
null
null
null
UTF-8
Python
false
false
299
pyde
size(400,400); background(60); Ncols = width; Nrows = height; # Set up 2D array myArray = [[0]*Ncols]*Nrows; for i in range(0,Ncols,20): for j in range(0,Nrows,20): myArray[i][j] = int(random(255)); stroke(myArray[i][j]); rect(i,j,1/2,j/2); save("stripe.tif")
[ "lsm35@cam.ac.uk" ]
lsm35@cam.ac.uk
e8a5e9cff601be65ad6eb8f089f1c1bbc68ad329
5c6238e52a3b58cfa56df2b0376fe972d4eda055
/time/clock_back/clock_back.py
9e3ea9bad8288a2a23031157d5bae86b63553156
[]
no_license
yhoazk/LF331_V2
25258da02654726a79e645f73616978dedf01064
5ca1c2620fbc295cc735f38acf7a6027fc4b9ae2
refs/heads/master
2023-06-30T15:18:31.283875
2023-06-18T11:28:18
2023-06-18T11:28:18
109,063,211
0
0
null
null
null
null
UTF-8
Python
false
false
512
py
#!/usr/bin/env python3 from matplotlib import pyplot as plt import csv def main(): print("sdfsdf") data = [] with open("/tmp/samples.log") as csv_file: samples = csv.reader(csv_file) for row in samples: data.append(int(row[0])) data_diff = [j-i for i, j in zip(data[:-1], data[1:])] for d in data_diff: if d < 0: print("NEGGGGGG") plt.plot(data, marker="x") plt.show() if __name__ == "__main__": main()
[ "yhoazk@gmail.com" ]
yhoazk@gmail.com
429cff803102a46433f095f3ba8e41199c5dcae2
47257562cd6db4185c34ab582f8ea311c3155c45
/DDPG_HER_VIME/Smoothing.py
cf7379093370f65b961a366d14338f46d0e9a24e
[]
no_license
JuliusVsi/HVDDPG_Project
f25ebe429c0ed0ea806b2f41d108f60e798b9a39
a21c6b89ef8e9ab2092a2dcc329b83f6b7813fcd
refs/heads/master
2020-07-17T02:09:38.593852
2019-09-02T19:25:32
2019-09-02T19:25:32
205,919,232
0
0
null
null
null
null
UTF-8
Python
false
false
16,970
py
import matplotlib.pyplot as plt import numpy as np def smooth(data, weight=0.85): last = data[0] smoothed = [] for point in data: smoothed_val = last * weight + (1 - weight) * point smoothed.append(smoothed_val) last = smoothed_val return smoothed if __name__ == '__main__': ######################################################################## # The Reach Figure ######################################################################## plt.figure('Reach') reach_her = np.loadtxt('Plot_Data/Reach_HER.txt', delimiter=',') x_reach_her = np.arange(len(reach_her)) smoothed_reach_her = np.array(smooth(reach_her, weight=0.4)) smoothed_reach_her_bar = np.array(smooth(reach_her, weight=0.1)) np.savetxt('Plot_Data/Smoothed_Reach_HER.txt', smoothed_reach_her, fmt='%f', delimiter=',') d_reach_her = smoothed_reach_her - smoothed_reach_her_bar plt.plot(x_reach_her, smoothed_reach_her, label='DDPG+HER') plt.fill_between(x_reach_her, smoothed_reach_her + d_reach_her, smoothed_reach_her - d_reach_her, color='blue', alpha=0.1) reach_vime = np.loadtxt('Plot_Data/Reach_VIME.txt', delimiter=',') x_reach_vime = np.arange(len(reach_vime)) smoothed_reach_vime = np.array(smooth(reach_vime, weight=0.4)) smoothed_reach_vime_bar = np.array(smooth(reach_vime, weight=0.1)) np.savetxt('Plot_Data/Smoothed_Reach_VIME.txt', smoothed_reach_vime, fmt='%f', delimiter=',') d_reach_vime = smoothed_reach_vime - smoothed_reach_vime_bar plt.plot(x_reach_vime, smoothed_reach_vime, label='HVDDPG') plt.fill_between(x_reach_vime, smoothed_reach_vime + d_reach_vime, smoothed_reach_vime - d_reach_vime, color='orange', alpha=0.1) reach_ddpg = np.loadtxt('Plot_Data/Reach_DDPG.txt', delimiter=',') x_reach_ddpg = np.arange(len(reach_ddpg)) smoothed_reach_ddpg = np.array(smooth(reach_ddpg, weight=0.6)) smoothed_reach_ddpg_bar = np.array(smooth(reach_ddpg, weight=0.3)) np.savetxt('Plot_Data/Smoothed_Reach_DDPG.txt', smoothed_reach_ddpg, fmt='%f', delimiter=',') d_reach_ddpg = smoothed_reach_ddpg - smoothed_reach_ddpg_bar plt.plot(x_reach_ddpg, smoothed_reach_ddpg, label='DDPG') plt.fill_between(x_reach_ddpg, smoothed_reach_ddpg + d_reach_ddpg, smoothed_reach_ddpg - d_reach_ddpg, color='orange', alpha=0.1) reach_dqn = np.loadtxt('Plot_Data/Reach_DQN.txt', delimiter=',') x_reach_dqn = np.arange(len(reach_dqn)) smoothed_reach_dqn = np.array(smooth(reach_dqn, weight=0.7)) smoothed_reach_dqn_bar = np.array(smooth(reach_dqn, weight=0.0)) np.savetxt('Plot_Data/Smoothed_Reach_DDPG.txt', smoothed_reach_dqn, fmt='%f', delimiter=',') d_reach_dqn = smoothed_reach_dqn - smoothed_reach_dqn_bar plt.plot(x_reach_dqn, smoothed_reach_dqn, label='DQN') plt.fill_between(x_reach_dqn, smoothed_reach_dqn + d_reach_dqn, smoothed_reach_dqn - d_reach_ddpg, color='orange', alpha=0.1) plt.grid() plt.ylim((-0.05, 1.05)) plt.xlabel('Episodes') plt.ylabel('Success Rate') plt.legend(loc='lower right') ######################################################################## # The Push Figure ######################################################################## plt.figure('Push') push_her = np.loadtxt('Plot_Data/Push_HER.txt', delimiter=',') x_push_her = np.arange(len(push_her)) smoothed_push_her = np.array(smooth(push_her, weight=0.9)) smoothed_push_her_bar = np.array(smooth(push_her, weight=0.7)) np.savetxt('Plot_Data/Smoothed_Push_HER.txt', smoothed_push_her, fmt='%f', delimiter=',') d_push_her = smoothed_push_her - smoothed_push_her_bar temp_bar_1 = np.maximum(smoothed_push_her + d_push_her, 0) temp_bar_2 = np.maximum(smoothed_push_her - d_push_her, 0) plt.plot(x_push_her, smoothed_push_her, label='DDPG+HER') plt.fill_between(x_push_her, temp_bar_1, temp_bar_2, color='blue', alpha=0.1) push_vime = np.loadtxt('Plot_Data/Push_VIME.txt', delimiter=',') push_vime = push_vime[0:100] x_push_vime = np.arange(len(push_vime)) smoothed_push_vime = np.array(smooth(push_vime, weight=0.9)) smoothed_push_vime_bar = np.array(smooth(push_vime, weight=0.7)) np.savetxt('Plot_Data/Smoothed_Push_VIME.txt', smoothed_push_vime, fmt='%f', delimiter=',') d_push_vime = smoothed_push_vime - smoothed_push_vime_bar temp_bar_1 = np.maximum(smoothed_push_vime + d_push_vime, 0) temp_bar_2 = np.maximum(smoothed_push_vime - d_push_vime, 0) plt.plot(x_push_vime, smoothed_push_vime, label='HVDDPG') plt.fill_between(x_push_vime, temp_bar_1, temp_bar_2, color='orange', alpha=0.1) push_ddpg = np.loadtxt('Plot_Data/Push_DDPG.txt', delimiter=',') x_push_ddpg = np.arange(len(push_ddpg)) smoothed_push_ddpg = np.array(smooth(push_ddpg, weight=0.9)) smoothed_push_ddpg_bar = np.array(smooth(push_ddpg, weight=0.7)) np.savetxt('Plot_Data/Smoothed_Push_DDPG.txt', smoothed_push_ddpg, fmt='%f', delimiter=',') d_push_ddpg = smoothed_push_ddpg - smoothed_push_ddpg_bar temp_bar_1 = np.maximum(smoothed_push_ddpg + d_push_ddpg, 0) temp_bar_2 = np.maximum(smoothed_push_ddpg - d_push_ddpg, 0) plt.plot(x_push_ddpg, smoothed_push_ddpg, label='DDPG') plt.fill_between(x_push_ddpg, temp_bar_1, temp_bar_2, color='green', alpha=0.1) push_dqn = np.loadtxt('Plot_Data/Push_DQN.txt', delimiter=',') x_push_dqn = np.arange(len(push_dqn)) smoothed_push_dqn = np.array(smooth(push_dqn, weight=0.9)) smoothed_push_dqn_bar = np.array(smooth(push_dqn, weight=0.7)) np.savetxt('Plot_Data/Smoothed_Push_DQN.txt', smoothed_push_dqn, fmt='%f', delimiter=',') d_push_dqn = smoothed_push_dqn - smoothed_push_dqn_bar temp_bar_1 = np.maximum(smoothed_push_dqn + d_push_dqn, 0) temp_bar_2 = np.maximum(smoothed_push_dqn - d_push_dqn, 0) plt.plot(x_push_dqn, smoothed_push_dqn, label='DQN') plt.fill_between(x_push_dqn, temp_bar_1, temp_bar_2, color='red', alpha=0.1) plt.grid() plt.ylim((-0.05, 1.05)) plt.xlabel('Episodes') plt.ylabel('Success Rate') plt.legend(loc='upper left') ######################################################################## # The Pick Figure ######################################################################## plt.figure('Pick') pick_her = np.loadtxt('Plot_Data/Pick_HER.txt', delimiter=',') x_pick_her = np.arange(len(pick_her)) smoothed_pick_her = np.array(smooth(pick_her, weight=0.98)) smoothed_pick_her_bar = np.array(smooth(pick_her, weight=0.7)) np.savetxt('Plot_Data/Smoothed_Pick_HER.txt', smoothed_pick_her, fmt='%f', delimiter=',') d_pick_her = smoothed_pick_her - smoothed_pick_her_bar temp_bar_1 = np.maximum(smoothed_pick_her + d_pick_her, 0) temp_bar_2 = np.maximum(smoothed_pick_her - d_pick_her, 0) plt.plot(x_pick_her, smoothed_pick_her, label='DDPG+HER') plt.fill_between(x_pick_her, temp_bar_1, temp_bar_2, color='blue', alpha=0.1) pick_vime = np.loadtxt('Plot_Data/Pick_VIME.txt', delimiter=',') x_pick_vime = np.arange(len(pick_vime)) smoothed_pick_vime = np.array(smooth(pick_vime, weight=0.98)) smoothed_pick_vime_bar = np.array(smooth(pick_vime, weight=0.7)) np.savetxt('Plot_Data/Smoothed_Pick_VIME.txt', smoothed_pick_vime, fmt='%f', delimiter=',') d_pick_vime = smoothed_pick_vime - smoothed_pick_vime_bar temp_bar_1 = np.maximum(smoothed_pick_vime + d_pick_vime, 0) temp_bar_2 = np.maximum(smoothed_pick_vime - d_pick_vime, 0) plt.plot(x_pick_vime, smoothed_pick_vime, label='HVDDPG') plt.fill_between(x_pick_vime, temp_bar_1, temp_bar_2, color='orange', alpha=0.1) pick_ddpg = np.loadtxt('Plot_Data/Pick_DDPG.txt', delimiter=',') x_pick_ddpg = np.arange(len(pick_ddpg)) smoothed_pick_ddpg = np.array(smooth(pick_ddpg, weight=0.9)) smoothed_pick_ddpg_bar = np.array(smooth(pick_ddpg, weight=0.0)) np.savetxt('Plot_Data/Smoothed_Pick_DDPG.txt', smoothed_pick_ddpg, fmt='%f', delimiter=',') d_pick_ddpg = smoothed_pick_ddpg - smoothed_pick_ddpg_bar temp_bar_1 = np.maximum(smoothed_pick_ddpg + d_pick_ddpg, 0) temp_bar_2 = np.maximum(smoothed_pick_ddpg - d_pick_ddpg, 0) plt.plot(x_pick_ddpg, smoothed_pick_ddpg, label='DDPG') plt.fill_between(x_pick_ddpg, temp_bar_1, temp_bar_2, color='green', alpha=0.1) pick_dqn = np.loadtxt('Plot_Data/Pick_DQN.txt', delimiter=',') x_pick_dqn = np.arange(len(pick_dqn)) smoothed_pick_dqn = np.array(smooth(pick_dqn, weight=0.9)) smoothed_pick_dqn_bar = np.array(smooth(pick_dqn, weight=0.7)) np.savetxt('Plot_Data/Smoothed_Pick_DQN.txt', smoothed_pick_dqn, fmt='%f', delimiter=',') d_pick_dqn = smoothed_pick_dqn - smoothed_pick_dqn_bar temp_bar_1 = np.maximum(smoothed_pick_dqn + d_pick_dqn, 0) temp_bar_2 = np.maximum(smoothed_pick_dqn - d_pick_dqn, 0) plt.plot(x_pick_dqn, smoothed_pick_dqn, label='DQN') plt.fill_between(x_pick_dqn, temp_bar_1, temp_bar_2, color='red', alpha=0.1) plt.grid() plt.ylim((-0.05, 1.05)) plt.xlabel('Episodes') plt.ylabel('Success Rate') plt.legend(loc='upper left') ######################################################################## # The Slide Figure ######################################################################## plt.figure('Slide') # her slide_her = np.loadtxt('Plot_Data/Slide_HER.txt', delimiter=',') x_slide_her = np.arange(len(slide_her)) smoothed_slide_her = np.array(smooth(slide_her, weight=0.98)) smoothed_slide_her_bar = np.array(smooth(slide_her, weight=0.8)) np.savetxt('Plot_Data/Smoothed_Slide_HER.txt', smoothed_slide_her, fmt='%f', delimiter=',') d_slide_her = smoothed_slide_her - smoothed_slide_her_bar temp_bar_1 = np.maximum(smoothed_slide_her + d_slide_her, 0) temp_bar_2 = np.maximum(smoothed_slide_her - d_slide_her, 0) plt.plot(x_slide_her, smoothed_slide_her, label='DDPG+HER') plt.fill_between(x_slide_her, temp_bar_1, temp_bar_2, color='blue', alpha=0.1) # hvddpg slide_vime = np.loadtxt('Plot_Data/Slide_VIME.txt', delimiter=',') x_slide_vime = np.arange(len(slide_vime)) smoothed_slide_vime = np.array(smooth(slide_vime, weight=0.98)) smoothed_slide_vime_bar = np.array(smooth(slide_vime, weight=0.7)) np.savetxt('Plot_Data/Smoothed_Slide_VIME.txt', smoothed_slide_vime, fmt='%f', delimiter=',') d_slide_vime = smoothed_slide_vime - smoothed_slide_vime_bar temp_bar_1 = np.maximum(smoothed_slide_vime + d_slide_vime, 0) temp_bar_2 = np.maximum(smoothed_slide_vime - d_slide_vime, 0) plt.plot(x_slide_vime, smoothed_slide_vime, label='HVDDPG') plt.fill_between(x_slide_vime, temp_bar_1, temp_bar_2, color='orange', alpha=0.1) # ddpg slide_ddpg = np.loadtxt('Plot_Data/Slide_DDPG.txt', delimiter=',') x_slide_ddpg = np.arange(len(slide_ddpg)) smoothed_slide_ddpg = np.array(smooth(slide_ddpg, weight=0.98)) smoothed_slide_ddpg_bar = np.array(smooth(slide_ddpg, weight=0.5)) np.savetxt('Plot_Data/Smoothed_Slide_DDPG.txt', smoothed_slide_ddpg, fmt='%f', delimiter=',') d_slide_ddpg = smoothed_slide_ddpg - smoothed_slide_ddpg_bar temp_bar_1 = np.maximum(smoothed_slide_ddpg + d_slide_ddpg, 0) temp_bar_2 = np.maximum(smoothed_slide_ddpg - d_slide_ddpg, 0) plt.plot(x_slide_ddpg, smoothed_slide_ddpg, label='DDPG') plt.fill_between(x_slide_ddpg, temp_bar_1, temp_bar_2, color='green', alpha=0.1) slide_dqn = np.loadtxt('Plot_Data/Slide_DQN.txt', delimiter=',') x_slide_dqn = np.arange(len(slide_dqn)) smoothed_slide_dqn = np.array(smooth(slide_dqn, weight=0.98)) smoothed_slide_dqn_bar = np.array(smooth(slide_dqn, weight=0.0)) np.savetxt('Plot_Data/Smoothed_Slide_DQN.txt', smoothed_slide_dqn, fmt='%f', delimiter=',') d_slide_dqn = smoothed_slide_dqn - smoothed_slide_dqn_bar temp_bar_1 = np.maximum(smoothed_slide_dqn + d_slide_dqn, 0) temp_bar_2 = np.maximum(smoothed_slide_dqn - d_slide_dqn, 0) plt.plot(x_slide_dqn, smoothed_slide_dqn, label='DQN') plt.fill_between(x_slide_dqn, temp_bar_1, temp_bar_2, color='red', alpha=0.1) plt.grid() plt.ylim((-0.05, 1.05)) plt.xlabel('Episodes') plt.ylabel('Success Rate') plt.legend(loc='upper left') ######################################################################## # The Egg Figure ######################################################################## plt.figure('Egg') # her egg_her = np.loadtxt('Plot_Data/Egg_HER.txt', delimiter=',') x_egg_her = np.arange(len(egg_her)) smoothed_egg_her = np.array(smooth(egg_her, weight=0.98)) smoothed_egg_her_bar = np.array(smooth(egg_her, weight=0.3)) np.savetxt('Plot_Data/Smoothed_Egg_HER.txt', smoothed_egg_her, fmt='%f', delimiter=',') d_egg_her = smoothed_egg_her - smoothed_egg_her_bar temp_bar_1 = np.maximum(smoothed_egg_her + d_egg_her, 0) temp_bar_2 = np.maximum(smoothed_egg_her - d_egg_her, 0) plt.plot(x_egg_her, smoothed_egg_her, label='DDPG+HER') plt.fill_between(x_egg_her, temp_bar_1, temp_bar_2, color='blue', alpha=0.1) # hvddpg egg_vime = np.loadtxt('Plot_Data/Egg_VIME.txt', delimiter=',') x_egg_vime = np.arange(len(egg_vime)) smoothed_egg_vime = np.array(smooth(egg_vime, weight=0.98)) smoothed_egg_vime_bar = np.array(smooth(egg_vime, weight=0.5)) np.savetxt('Plot_Data/Smoothed_Egg_VIME.txt', smoothed_egg_vime, fmt='%f', delimiter=',') d_egg_vime = smoothed_egg_vime - smoothed_egg_vime_bar temp_bar_1 = np.maximum(smoothed_egg_vime + d_egg_vime, 0) temp_bar_2 = np.maximum(smoothed_egg_vime - d_egg_vime, 0) plt.plot(x_egg_vime, smoothed_egg_vime, label='HVDDPG') plt.fill_between(x_egg_vime, temp_bar_1, temp_bar_2, color='orange', alpha=0.1) # ddpg egg_ddpg = np.loadtxt('Plot_Data/Egg_DDPG.txt', delimiter=',') x_egg_ddpg = np.arange(len(egg_ddpg)) smoothed_egg_ddpg = np.array(smooth(egg_ddpg, weight=0.98)) smoothed_egg_ddpg_bar = np.array(smooth(egg_ddpg, weight=0.2)) np.savetxt('Plot_Data/Smoothed_Egg_DDPG.txt', smoothed_egg_ddpg, fmt='%f', delimiter=',') d_egg_ddpg = smoothed_egg_ddpg - smoothed_egg_ddpg_bar temp_bar_1 = np.maximum(smoothed_egg_ddpg + d_egg_ddpg, 0) temp_bar_2 = np.maximum(smoothed_egg_ddpg - d_egg_ddpg, 0) plt.plot(x_egg_ddpg, smoothed_egg_ddpg, label='DDPG') plt.fill_between(x_egg_ddpg, temp_bar_1, temp_bar_2, color='green', alpha=0.1) ######################################################################## # The pen Figure ######################################################################## plt.figure('pen') # her pen_her = np.loadtxt('Plot_Data/Pen_HER.txt', delimiter=',') pen_her = pen_her[0:800] x_pen_her = np.arange(len(pen_her)) smoothed_pen_her = np.array(smooth(pen_her, weight=0.99)) smoothed_pen_her_bar = np.array(smooth(pen_her, weight=0.3)) np.savetxt('Plot_Data/Smoothed_Pen_HER.txt', smoothed_pen_her, fmt='%f', delimiter=',') d_pen_her = smoothed_pen_her - smoothed_pen_her_bar temp_bar_1 = np.maximum(smoothed_pen_her + d_pen_her, 0) temp_bar_2 = np.maximum(smoothed_pen_her - d_pen_her, 0) plt.plot(x_pen_her, smoothed_pen_her, label='DDPG+HER') plt.fill_between(x_pen_her, temp_bar_1, temp_bar_2, color='blue', alpha=0.1) # hvddpg pen_vime = np.loadtxt('Plot_Data/Pen_VIME.txt', delimiter=',') x_pen_vime = np.arange(len(pen_vime)) smoothed_pen_vime = np.array(smooth(pen_vime, weight=0.98)) smoothed_pen_vime_bar = np.array(smooth(pen_vime, weight=0.5)) np.savetxt('Plot_Data/Smoothed_Pen_VIME.txt', smoothed_pen_vime, fmt='%f', delimiter=',') d_pen_vime = smoothed_pen_vime - smoothed_pen_vime_bar temp_bar_1 = np.maximum(smoothed_pen_vime + d_pen_vime, 0) temp_bar_2 = np.maximum(smoothed_pen_vime - d_pen_vime, 0) plt.plot(x_pen_vime, smoothed_pen_vime, label='HVDDPG') plt.fill_between(x_pen_vime, temp_bar_1, temp_bar_2, color='orange', alpha=0.1) # ddpg pen_ddpg = np.loadtxt('Plot_Data/Pen_DDPG.txt', delimiter=',') x_pen_ddpg = np.arange(len(pen_ddpg)) smoothed_pen_ddpg = np.array(smooth(pen_ddpg, weight=0.98)) smoothed_pen_ddpg_bar = np.array(smooth(pen_ddpg, weight=0.5)) np.savetxt('Plot_Data/Smoothed_Pen_DDPG.txt', smoothed_pen_ddpg, fmt='%f', delimiter=',') d_pen_ddpg = smoothed_pen_ddpg - smoothed_pen_ddpg_bar temp_bar_1 = np.maximum(smoothed_pen_ddpg + d_pen_ddpg, 0) temp_bar_2 = np.maximum(smoothed_pen_ddpg - d_pen_ddpg, 0) plt.plot(x_pen_ddpg, smoothed_pen_ddpg, label='DDPG') plt.fill_between(x_pen_ddpg, temp_bar_1, temp_bar_2, color='green', alpha=0.1) plt.grid() plt.ylim((-0.05, 1.05)) plt.xlabel('Episodes') plt.ylabel('Success Rate') plt.legend(loc='upper left') plt.show()
[ "noreply@github.com" ]
JuliusVsi.noreply@github.com
fff29da02d95309713cc9a0f7a86f69832ba5220
83a506a501561602ad3b259341225ddfbddab160
/GameServer/matchGames/Match_PK_DouDiZhu/redis_instance.py
3fe50de16f52f543bb74fc19e6b8dcc7b80828c3
[]
no_license
daxingyou/SouYouJi_Game
9dc5f02eb28b910efb229653a8d0bffe425a7911
7311a994c9aba15b7234331709975ebc37e8453d
refs/heads/master
2023-03-28T01:36:48.955107
2020-04-05T01:24:17
2020-04-05T01:24:17
null
0
0
null
null
null
null
UTF-8
Python
false
false
519
py
# -*- coding:utf-8 -*- # !/bin/python """ Author: Pipo Date: $Date$ Revision: $Revision$ Description: Redis """ import redis from configs import CONFIGS redisdb = None def getInst(dbNum=CONFIGS['redis']['db']): global redisdb redisdb = redis.ConnectionPool( host=CONFIGS['redis']['host'], port=CONFIGS['redis']['port'], db=dbNum, password=CONFIGS['redis']['password'] ) redisData = redis.Redis(connection_pool=redisdb) return redisData
[ "ronnyzh@yeah.net" ]
ronnyzh@yeah.net
ce43314f310b9b65dee36e38c84db5b49f04c6b6
2fd194196f21f51c881ec88d736bb8b2ee1937ce
/RaspberryPi/interface.py
fadbf12345a8799e9d5a53d86879de1f32419990
[]
no_license
MV-Automation/MV_Pandemic_Solution
6d609ab8ab2efff7fadfe2b09e27be9bdeaf72e1
a95c2b090aa5625ee9be407294735eccf93c3869
refs/heads/main
2023-02-18T02:23:56.248935
2021-01-20T17:55:14
2021-01-20T17:55:14
312,030,877
0
0
null
null
null
null
UTF-8
Python
false
false
9,648
py
from tkinter import Tk, Label, Button, Frame from PIL import ImageTk, Image import cv2 from camera import camera import config as cfg import mv_rasp_upload_snaps_to_cloud as awsupload import AmazonKeys as key import boto3 """ Secuencia: 1. Se inicializan todos los componentes de la GUI y se añaden a la ventana de Tkinter. 2. Se crea el proceso paralelo de la cámara y se inicia el stream de rtsp. 3. Se inicia el mainloop de Tkinter para que funcione al interfaz. 4. Del frame de portada se va al frame con las instrucciones para tomar el snapshot de la cara. -El retraso actual del stream de RTSP es de alrededor de 2 segundos. 4. Se muestra la pantalla con el contenido del stream de RTSP y un overlay para indicar dónde colocarse. 5. Se toma el snapshot y se pasa al frame con las instrucciones para el snapshot de la identificación. 6. Se muestra la pantalla con el contenido del stream de RTSP y un overlay para indicar dónde colocar la identificación. 7. Se toma el snapshot y se pasa a un frame indicando que el resultado se está procesando. 8. Se muestra una imagen promocional mientras los resultados son calculados. 9. Se muestra el resultado al usuario. 10. La aplicación regresa al frame portada. """ #status flags and snapshot variables face_snapshot = False id_snapshot = False face_img = None id_img = None #Load images for instructions cover_img = Image.open("img/cover.jpg") face_ins_img = Image.open("img/face_instructions.jpg") id_ins_img = Image.open("img/id_instructions.jpg") processing_img = Image.open("img/processing.jpg") positive_result_img = Image.open("img/positive_result.jpg") negative_result_img = Image.open("img/negative_result.jpg") promo_img = Image.open(cfg.promo_img['location']) promo_img = promo_img.resize((800,480)) #Function to initialize the rtsp stream def init_stream(): global cam cam = camera(cfg.rtsp['address']) face_stream() #Raise a frame to show it def raise_frame(frame): frame.tkraise() #Function to get the rtsp stream for the face snapshot def face_stream(): #Get the next frame of the video frame = cam.get_frame(resize=None) #Resize the frame and draw the rectangle for where the face should go resize_img = cv2.resize(frame, (800,480)) cv2.rectangle(resize_img, (200,80), (600,400), (255,255,255), 3) #Add the image to the Tkinter interface img = Image.fromarray(resize_img) imgtk = ImageTk.PhotoImage(img) global bface bface.imgtk = imgtk bface.configure(image=imgtk) #Repeat the process after 30ms in order to achieve a video of around ~30 FPS faceloop = window.after(30, face_stream) if face_snapshot == True: global face_img face_img = Image.fromarray(frame) window.after_cancel(faceloop) bface.configure(image='') #Function to get the rtsp stream for the id snapshot def id_stream(): #Get the next frame of the video frame = cam.get_frame(resize=None) #Resize the frame and draw the rectangle for where the id should go resize_img = cv2.resize(frame, (800,480)) cv2.rectangle(resize_img, (150,120), (650,360), (255,0,0), 3) #Add the image to the Tkinter interface img = Image.fromarray(resize_img) imgtk = ImageTk.PhotoImage(image=img) global bid bid.imgtk = imgtk bid.configure(image=imgtk) #Repeat the process after 30ms in order to achieve a video of around ~30 FPS idloop = window.after(30, id_stream) if id_snapshot == True: global id_img id_img = Image.fromarray(frame) window.after_cancel(idloop) bid.configure(image='') print("Images are ready") upload_images() #Upload the images to AWS, raise the frame with the promo image and open the sqs notification function def upload_images(): print("Uploading images") cv2.destroyAllWindows() awsupload.main(face_img, id_img) window.after(1000, change_promo, promo_img) try: window.after(15000, sqs_notif) global cam cam.end() cam = None except: print("No message in the queue") #SQS notifications, raising the results frame and redirecting to the application cover frame def sqs_notif(): global face_snapshot global id_snapshot global face_img global id_img face_snapshot = False id_snapshot = False face_img = None id_img = None init_stream() # Create SQS client sqs = boto3.client('sqs', aws_access_key_id=key.aws_access, aws_secret_access_key=key.aws_secret, region_name="us-east-2") queue_url = 'https://sqs.us-east-2.amazonaws.com/727103842412/MyQueueTest' # Receive message from SQS queue response = sqs.receive_message( QueueUrl=queue_url, AttributeNames=[ 'SentTimestamp' ], MaxNumberOfMessages=1, MessageAttributeNames=[ 'All' ], VisibilityTimeout=0, WaitTimeSeconds=0 ) if 'Messages' not in response: print("No hay mensaje") raise_frame(negative_results_frame) window.after(10000, raise_frame, cover_frame) else: message = response['Messages'][0] print("LOG: Hay mensaje.") body = message['Body'] receipt_handle = message['ReceiptHandle'] print('Mensaje desde lambda: %s' % body) if body == 'Allowed': raise_frame(positive_results_frame) else: print('not allowed') raise_frame(negative_results_frame) window.after(10000, raise_frame, cover_frame) # Delete received message from queue sqs.delete_message( QueueUrl=queue_url, ReceiptHandle=receipt_handle ) #Change the promo image def change_promo(promo_img): global bpromo promo_img = ImageTk.PhotoImage(image=promo_img) bpromo.img = promo_img bpromo.configure(image=promo_img) bpromo.grid(sticky='news') raise_frame(promo_frame) #Switch to the id snapshot instructions frame def switch_to_id_ins(): global face_snapshot id_stream() face_snapshot = True raise_frame(id_ins_frame) #Switch to the face snapshot instructions fame def switch_to_face_ins(): raise_frame(face_ins_frame) #Switch to the id snapshot frame def switch_to_id(): raise_frame(id_frame) #Switch to the face snapshot frame def switch_to_face(): raise_frame(face_frame) #Switch to the processing results frame def switch_to_processing(): global id_snapshot id_snapshot = True raise_frame(processing_frame) #Create the root component, named window and its attributes window = Tk() window.geometry("800x480") window.attributes('-fullscreen', True) #alt + F4 allows you to exit the fullscreen app # Create the initial instructions frame and button cover_frame = Frame(window, bg="white") bcover = Button(cover_frame, command=switch_to_face_ins) cover_img = ImageTk.PhotoImage(image=cover_img) bcover.img = cover_img bcover.configure(image=cover_img) bcover.grid(sticky='news') # Create the face snapshot instructions frame and button face_ins_frame = Frame(window, bg="white") bface_ins = Button(face_ins_frame, command=switch_to_face) face_ins_img = ImageTk.PhotoImage(image=face_ins_img) bface_ins.img = face_ins_img bface_ins.configure(image=face_ins_img) bface_ins.grid(sticky='news') # Create the id snapshot instructions frame and button id_ins_frame = Frame(window, bg="white") bid_ins = Button(id_ins_frame, command=switch_to_id) id_ins_img = ImageTk.PhotoImage(image=id_ins_img) bid_ins.img = id_ins_img bid_ins.configure(image=id_ins_img) bid_ins.grid(sticky='news') # Create a frame and button for the face snapshot face_frame = Frame(window, bg="white") bface = Button(face_frame, command=switch_to_id_ins) bface.grid(sticky='news') # Create a frame and button for the id snapshot id_frame = Frame(window, bg="white") bid = Button(id_frame, command=switch_to_processing) bid.grid(sticky='news') # Create a frame and button for the processing results screen processing_frame = Frame(window, bg="white") bprocessing = Button(processing_frame) processing_img = ImageTk.PhotoImage(image=processing_img) bprocessing.img = processing_img bprocessing.configure(image=processing_img) bprocessing.grid(sticky='news') # Create a frame and button for the positive results screen positive_results_frame = Frame(window, bg="white") bresults = Button(positive_results_frame) positive_result_img = ImageTk.PhotoImage(image=positive_result_img) bresults.img = positive_result_img bresults.configure(image=positive_result_img) bresults.grid(sticky='news') # Create a frame and button for the negative results screen negative_results_frame = Frame(window, bg="white") bnegresults = Button(negative_results_frame) negative_result_img = ImageTk.PhotoImage(image=negative_result_img) bnegresults.img = negative_result_img bnegresults.configure(image=negative_result_img) bnegresults.grid(sticky='news') # Create a frame and button for the promo screen promo_frame = Frame(window, bg="white") bpromo = Label(promo_frame) #Add the frames to the grid for frame in (cover_frame, face_frame, id_frame, face_ins_frame, id_ins_frame, processing_frame, positive_results_frame, negative_results_frame, promo_frame): frame.grid(row=0, column=0, sticky='news') if __name__ == "__main__": #Raise the cover frame so we start with the cover raise_frame(cover_frame) # initialize the cam object cam = None init_stream() window.mainloop() cam.end() cv2.destroyAllWindows()
[ "70815913+anarrode@users.noreply.github.com" ]
70815913+anarrode@users.noreply.github.com
1570433f75a24869c25727245eb54167af8b00ab
bdbb20034786685601d0a424f29ea8bd04755419
/stopWait/client/client.py
9bcb24ee87990c4551f147c2c65755c851585681
[]
no_license
s19-nets/udp-file-transfer-mc
48df5efc60af4b41caf48a06381c45c17e640f4f
8a5aff2b3142589f1e1ebb4702bdafe1ef0ed4ae
refs/heads/master
2020-04-27T19:41:42.532207
2019-03-09T07:14:55
2019-03-09T07:14:55
174,629,686
0
0
null
null
null
null
UTF-8
Python
false
false
2,963
py
from socket import * serverAddress = ('localhost', 50000) ack = 0 last = 0 doneNum = 0 seqnum = 0 done = False retry = False #Get user input input = input("What would you like to do?") input_list = input.split() #Create socket and send user input clientSocket = socket(AF_INET, SOCK_DGRAM) clientSocket.sendto(input.encode(), serverAddress) #GET file if(input_list[0].upper() == "GET"): with open('get_file', 'wb') as f: print("Receiving:", input_list[1]) while True and not done: #Retrieve seqNum and done number (0 = not done, 1 = done) message, serverAddrPort = clientSocket.recvfrom(1000) splitMessage = message.decode('utf_8', 'ignore').split() seqNum = splitMessage[0] doneNum = splitMessage[1] if(int(doneNum) == 1): done = True #skip over the seqNum & doneNum if(int(seqNum) <10 ): text = message.decode('utf_8', 'ignore')[4:] elif(int(seqNum) >=10 and int(seqNum) <100): text = message.decode('utf_8', 'ignore')[5:] else: text = message.decode('utf_8', 'ignore')[6:] #Check to see if it's retrying if(int(seqNum) != ack): ack= ack+1 retry = False else: retry = True #send ack num sACK = str(ack) clientSocket.sendto(sACK.encode(), serverAddress) if not message: done = True break #write message to a file if not retrying if not retry: f.write(text.encode()) print('Done Receiving') f.close() #PUT file if(input_list[0].upper() == "PUT"): print("Sending", (input_list[1])) #Open requested file file = open(input_list[1].encode(),'rb') l = file.read(90); while (l): #increment seqnum if not retrying if(not retry): seqnum = seqnum +1 #prepare and insert into package sSeqnum = str(seqnum) + " " doneNum = str(doneNum) emptyString = " " package = sSeqnum.encode()+ doneNum.encode() + emptyString.encode() + l #send to client and receive acknowledgement clientSocket.sendto(package, serverAddress) ACK, cAddr = clientSocket.recvfrom(100) #Check to see if ack number is correct if(seqnum != int(ACK.decode())): retry = True #if it is, keep reading file if(not retry): l = file.read(80) #check to see if end of file if(len(l) < 80): doneNum = 1 print('Done Sending') file.close()
[ "noreply@github.com" ]
s19-nets.noreply@github.com
acd320e9b5670c9b4b981d67159ec8153117268f
6a020d2afc402b443013a1a65b961d21e868ad8b
/tools/mdtable.py
92d401420be8a0b8255ba25b904ef3829c2fbe13
[]
no_license
scientist1642/anagram_benchmarks
fd5645f44e8454596ac3f06768a586819fe12403
ccbc728544ce86028e0a2d9f8de0490c8ccecc72
refs/heads/master
2020-04-27T19:26:06.883530
2020-02-27T12:36:32
2020-02-27T12:36:32
174,617,771
4
1
null
2019-03-09T10:42:27
2019-03-08T22:15:14
Python
UTF-8
Python
false
false
3,600
py
#!/usr/bin/python # -*- coding: utf-8 -*- import sys def get_matrix_from_csv(stream, separator=";"): """Return a matrix of values from an input stream containing csv data.""" for line in stream: yield [col.strip("\r\n") for col in line.split(separator)] def get_matrix_from_md(stream): """Return a matrix of values from an input stream containing a markdown table.""" for idx, line in enumerate(stream): if idx == 1: continue cols = line.split("|") yield [col.strip(" \r\n") for col in cols][1:len(cols) - 1] def get_column_store_from_matrix(matrix, escape=False): """Return input stream as a column-store.""" columns = [] for row in matrix: if not len(columns): columns = [[get_escaped_string(col, escape)] for col in row] else: for idx, col in enumerate(row): columns[idx].append(get_escaped_string(col, escape)) return columns MARKDOWN_ESCAPE_CHAR = "\\`*_{}[]()#+-.!" def get_escaped_string(text, escape=True): """Return text with escaped characters for markdown.""" if escape: for esc_chr in MARKDOWN_ESCAPE_CHAR: text = text.replace(esc_chr, "\\" + esc_chr) return text def get_formatted_string(text, width, formatted=True): """Return text with trailing spaces.""" return " " + text + " " + (" " * (width - len(text))) if formatted else text def get_md_table(column_store, formatted=True): """Return a formatted markdown table from a column-store.""" widths = [max(len(value) for value in column) for column in column_store] md_table = "" for idx in range(len(column_store[0])): columns = (column[idx] for column in column_store) md_table += "|" + "|".join((get_formatted_string(column, width, formatted) for column, width in zip(columns, widths))) + "|\n" if idx == 0: md_table += "|" + "|".join(("-" * (width + 2 if formatted else 3) for width in widths)) + "|\n" return md_table def get_csv_table(column_store, separator=";"): """Return a csv table from a column-store.""" csv_table = "" for idx in range(len(column_store[0])): columns = (column[idx] for column in column_store) csv_table += separator.join(columns) + "\n" return csv_table def main(args): """Main function.""" formatted = True escape = False input_type = "csv" output_type = "md" separator = ";" for arg in args: if arg == "-mini": formatted = False elif arg.startswith("-in:"): input_type = arg[4:].lower() elif arg.startswith("-out:"): output_type = arg[5:].lower() elif arg.startswith("-separator:"): separator = arg[11:] elif arg.startswith("-escape"): escape = True if separator == "tab": separator = "\t" if input_type == "csv": matrix = get_matrix_from_csv(sys.stdin, separator) elif input_type == "md": matrix = get_matrix_from_md(sys.stdin) else: raise Exception("Invalid input type: %s (csv or md expected)" % input_type) if output_type == "csv": table = get_csv_table(get_column_store_from_matrix(matrix), separator) elif output_type == "md": table = get_md_table(get_column_store_from_matrix(matrix, escape), formatted) else: raise Exception("Invalid output type: %s (csv or md expected)" % output_type) sys.stdout.write(table) sys.stdout.flush() if __name__ == "__main__": main(sys.argv[1:])
[ "z.isakadze@gmail.com" ]
z.isakadze@gmail.com
48dfd03ed0706d7a711f1e304f10e7e7b6b13b01
0d59d36b45143e8710d6a19dce41c7a8fc14fc3b
/tasker/libs/client_desktop/event.py
9d9d31fa183128fc1573f381b098c9c7f4010c83
[]
no_license
nkitsan/trello_clone
4d09b8f5bfe1b77ce2d4a5eb6eb49492b229b1fc
c86c83f4fe76323bc558911cb81bbfe234bfe3ce
refs/heads/master
2020-04-13T12:06:29.129019
2018-12-26T15:29:07
2018-12-26T15:29:07
null
0
0
null
null
null
null
UTF-8
Python
false
false
4,251
py
import requests import click from tasker.libs.client_desktop.helper import HOST, date_validation from tasker.libs.client_desktop.access import read_api @click.group() def event_operations(): """Here is commands which allow to work with calendar events""" @event_operations.command(short_help='Add the event to the calendar') @click.option('--name', default='', help='Name of the event which will be added') @click.option('--date', required=True, help='Date of event') def add_event(name, date): if not date_validation(date): click.echo('Format of date should be Y-M-D H:M') return api = read_api() if api is None: click.echo('Use login --api to register your api key and work further') return url = HOST + api + '/events' data = {'event_name': name, 'event_date': date} event_response = requests.post(url=url, data=data).json() if 'error' in event_response: click.echo(event_response['error']) return for event_id in event_response: click.echo(event_id + ': ' + event_response[event_id]['name'] + '\ndate: ' + ' '.join(event_response[event_id]['date'][0:-1].split('T'))) return @event_operations.command(short_help='Change the event to the calendar') @click.option('--event_id', required=True, type=click.INT, help='ID of the event to change') @click.option('--name', default=None, help='Name of the event on which will be renewed') @click.option('--date', default=None, help='Date of event on which will be renewed') def change_event(event_id, name, date): api = read_api() if api is None: click.echo('Use login --api to register your api key and work further') return event_id = str(event_id) url = HOST + api + '/events/' + event_id data = {} if name is not None: data.update({'event_name': name}) if date is not None: data.update({'event_date': date}) event_response = requests.put(url=url, data=data).json() if 'error' in event_response: click.echo(event_response['error']) return click.echo(event_id + ': ' + event_response[event_id]['name'] + '\ndate: ' + ' '.join(event_response[event_id]['date'][0:-1].split('T'))) return @event_operations.command(short_help='Delete the event from the calendar') @click.option('--event_id', required=True, type=click.INT, help='ID of the event to delete') def delete_event(event_id): api = read_api() if api is None: click.echo('Use login --api to register your api key and work further') return url = HOST + api + '/events' data = {'event_id': str(event_id)} event_response = requests.delete(url=url, data=data).json() if 'error' in event_response: click.echo(event_response['error']) return click.echo('event was deleted') return @event_operations.command(short_help='Show events in the calendar') def show_events(): api = read_api() if api is None: click.echo('Use login --api to register your api key and work further') return url = HOST + api + '/events' event_response = requests.get(url=url).json() if 'error' in event_response: click.echo(event_response['error']) return events = '' for event_id in event_response: events += (event_id + ' ' + event_response[event_id]['name'] + '\ndate:' + ' '.join(event_response[event_id]['date'][0:-1].split('T')) + '\n') click.echo(events.rstrip('\n')) return @event_operations.command(short_help='Show the event from the calendar') @click.option('--event_id', required=True, type=click.INT, help='ID of the event to see') def show_event(event_id): api = read_api() if api is None: click.echo('Use login --api to register your api key and work further') return event_id = str(event_id) url = HOST + api + '/events/' + event_id event_response = requests.get(url=url).json() if 'error' in event_response: click.echo(event_response['error']) return click.echo(event_id + ': ' + event_response[event_id]['name'] + '\ndate: ' + ' '.join(event_response[event_id]['date'][0:-1].split('T'))) return
[ "nastik.kitsan@gmail.com" ]
nastik.kitsan@gmail.com
0e8a3dffcb9fc9ad6d8d25500665e90fa65b0b3d
189996bd337adf575352201ccddc5bc9fe7550cd
/Incomplete_tensorflow_implementation/Test.py
7ffecf73a67e41dee8aaca4fcba66e534c62f55e
[]
no_license
Ru0628/LUNA16_Challange
07ab83cb2f9d09207c38559ebb8a9c12d43c1aff
cd380f6b44fa5cfd45de377112b7770b94a1f451
refs/heads/master
2022-04-19T05:40:22.791855
2020-04-15T15:45:44
2020-04-15T15:45:44
null
0
0
null
null
null
null
UTF-8
Python
false
false
4,964
py
import numpy as np from scipy import ndimage from os import listdir import matplotlib.pyplot as plt from PIL import Image # ****************** Connected Component Analysis # testMatrix = np.array([[[1, 1, 0], # [0, 0, 0], # [1, 0, 0]], # [[0, 1, 0], # [0, 0, 0], # [1, 0, 0]], # [[0, 1, 1], # [0, 0, 0], # [1, 0, 1]]]) # # outMatrix = ndimage.label(testMatrix) # # matrix = outMatrix[0] # # print(matrix) # print('***') # # matrix[matrix != 2] = 0 # matrix[matrix == 2] = 1 # # print(matrix) # ****************** Erosion Operation # testMatrix = np.array([[[1, 0, 0, 0, 1], # [0, 1, 1, 1, 0], # [0, 1, 1, 1, 0], # [0, 1, 1, 1, 0], # [1, 0, 0, 0, 1]], # [[0, 1, 0, 1, 1], # [0, 1, 1, 1, 0], # [0, 1, 1, 1, 0], # [0, 1, 1, 1, 0], # [0, 0, 0, 0, 0]], # [[0, 0, 0, 0, 0], # [0, 1, 1, 1, 1], # [0, 1, 1, 1, 1], # [0, 1, 1, 1, 1], # [0, 0, 1, 1, 1]]]) # # kernel = np.ones((2,2,2)) # # print(kernel) # # outputMatrix = ndimage.binary_erosion(testMatrix, structure=kernel).astype(testMatrix.dtype) # # print(outputMatrix) # ************* Dilation Operration # testMatrix = np.array([[[0, 0, 0, 0, 0], # [0, 1, 1, 1, 0], # [0, 1, 1, 1, 0], # [0, 1, 1, 1, 0], # [0, 0, 0, 0, 0]], # [[0, 0, 0, 0, 0], # [0, 1, 1, 1, 0], # [0, 1, 1, 1, 0], # [0, 1, 1, 1, 0], # [0, 0, 0, 0, 0]], # [[0, 0, 0, 0, 0], # [0, 1, 1, 1, 0], # [0, 1, 1, 1, 0], # [0, 1, 1, 1, 0], # [0, 0, 0, 0, 1]]]) # # kernel = np.ones((2,2,2)) # # print(kernel) # # outputMatrix = ndimage.binary_dilation(testMatrix, structure=kernel).astype(testMatrix.dtype) # # print(outputMatrix) # kernel = ndimage.generate_binary_structure(3,1) # print(kernel) # ************ Get File List # fileNames = listdir(r'C:\Users\polat\Desktop\TEZ\LUNA\subset0') # # print(fileNames[1]) # ************ Image Show # image = np.array([[0.1, 0.2, 0.7, 0.5, 0.9], # [0.2, 0.4, 0.5, 0.5, 0.7], # [0.2, 0.3, 0.1, 0.5, 0.7], # [0.7, 0.7, 0.1, 0.5, 0.3], # [0.9, 0.7, 0.9, 0.3, 0.1]]) # # img = Image.fromarray(np.uint8(255 * image)) # img.show() # ************* Save Numpy Array # testMatrix = np.array([[[0, 0, 0, 0, 0], # [0, 1, 1, 1, 0], # [0, 1, 1, 1, 0], # [0, 1, 1, 1, 0], # [0, 0, 0, 0, 0]], # [[0, 0, 0, 0, 0], # [0, 1, 1, 1, 0], # [0, 1, 1, 1, 0], # [0, 1, 1, 1, 0], # [0, 0, 0, 0, 0]], # [[0, 0, 0, 0, 0], # [0, 1, 1, 1, 0], # [0, 1, 1, 1, 0], # [0, 1, 1, 1, 0], # [0, 0, 0, 0, 1]]]) # # np.save('3darr',testMatrix) # *************** Read Numpy Array # readArr = np.load('3darr.npy') # # print(readArr) # *************** Resize 3D volume # WIDTH = 6 # ZWIDTH = 4 # testMatrix = np.array([[[0, 0, 0, 0, 0], # [0, 0.5, 1, 0.5, 0], # [0, 1, 1, 1, 0], # [0, 0.5, 1,0.5, 0], # [0, 0, 0, 0, 0]], # [[0, 0, 0, 0, 0], # [0, 1, 1, 1, 0], # [0, 1, 1, 1, 0], # [0, 1, 1, 1, 0], # [0, 0, 0, 0, 0]], # [[0, 0, 0, 0, 0], # [0, 1, 1, 1, 0], # [0, 1, 1, 1, 0], # [0, 1, 1, 1, 0], # [0, 0, 0, 0, 1]]]) # # zoomFactor = [ZWIDTH / float(testMatrix.shape[0]), WIDTH / float(testMatrix.shape[1]), # WIDTH / float(testMatrix.shape[2])] # resizedImage = ndimage.zoom(testMatrix, zoom=zoomFactor) # # print(resizedImage) # print(resizedImage.shape) testMatrix = np.array([0,1.0,2.0,1.0,0]) resizedImage = ndimage.zoom(testMatrix,zoom=2, mode='nearest') print(resizedImage)
[ "noreply@github.com" ]
Ru0628.noreply@github.com
8e8d989c5ae7bb188f444018f73a3c7200c98092
6220fd6bfd24626ea3f15b862606ecffe090e211
/Course_2_Python_to_Access Web Data/chapter15/roster.py
ef6818724c91f27ab94ca4d4a5c59ce06d7dc212
[]
no_license
Amithabh/Python4Informatics
ffa459a5340673d8df488aff3be4cef43499a6ef
3faad370928f5bd2a17edaa4caf3d1b8fb759710
refs/heads/master
2020-05-30T12:27:13.802610
2016-10-07T21:04:04
2016-10-07T21:04:04
70,282,664
0
0
null
null
null
null
UTF-8
Python
false
false
1,422
py
import json import sqlite3 conn = sqlite3.connect('rosterdb.sqlite') cur = conn.cursor() # Do some setup cur.executescript(''' DROP TABLE IF EXISTS User; DROP TABLE IF EXISTS Member; DROP TABLE IF EXISTS Course; CREATE TABLE User ( id INTEGER NOT NULL PRIMARY KEY AUTOINCREMENT UNIQUE, name TEXT UNIQUE ); CREATE TABLE Course ( id INTEGER NOT NULL PRIMARY KEY AUTOINCREMENT UNIQUE, title TEXT UNIQUE ); CREATE TABLE Member ( user_id INTEGER, course_id INTEGER, role INTEGER, PRIMARY KEY (user_id, course_id) ) ''') fname = input('Enter file name: ') if ( len(fname) < 1 ) : fname = 'roster_data.json' # [ # [ "Charley", "si110", 1 ], # [ "Mea", "si110", 0 ], str_data = open(fname).read() json_data = json.loads(str_data) for entry in json_data: name = entry[0]; title = entry[1]; print name, title cur.execute('''INSERT OR IGNORE INTO User (name) VALUES ( ? )''', ( name, ) ) cur.execute('SELECT id FROM User WHERE name = ? ', (name, )) user_id = cur.fetchone()[0] cur.execute('''INSERT OR IGNORE INTO Course (title) VALUES ( ? )''', ( title, ) ) cur.execute('SELECT id FROM Course WHERE title = ? ', (title, )) course_id = cur.fetchone()[0] cur.execute('''INSERT OR REPLACE INTO Member (user_id, course_id) VALUES ( ?, ? )''', ( user_id, course_id ) ) conn.commit()
[ "amithabh5@gmail.com" ]
amithabh5@gmail.com
95e9f1d292ccffad970294b2b502147429f71198
23b5337bf410415b7b150e3ad60cafc1578a0441
/07-User-Authentication/01-Flask-Login/myproject/__init__.py
54b954d72924a39c7987de9eda326bbc04bd3512
[]
no_license
VerdantFox/flask_course
b8de13ad312c14229f0c3bc2af70e8609a3b00fb
47b167b54bc580734fa69fc1a2d7e724adfb9610
refs/heads/master
2021-09-10T05:01:47.385859
2020-02-24T21:07:05
2020-02-24T21:07:05
241,973,705
0
0
null
2021-09-08T01:40:59
2020-02-20T19:40:42
Python
UTF-8
Python
false
false
761
py
import os from flask import Flask from flask_sqlalchemy import SQLAlchemy from flask_migrate import Migrate from flask_login import LoginManager # Create a login manager object login_manager = LoginManager() app = Flask(__name__) # Often people will also separate these into a separate config.py file app.config["SECRET_KEY"] = "mysecretkey" basedir = os.path.abspath(os.path.dirname(__file__)) app.config["SQLALCHEMY_DATABASE_URI"] = "sqlite:///" + os.path.join( basedir, "data.sqlite" ) app.config["SQLALCHEMY_TRACK_MODIFICATIONS"] = False db = SQLAlchemy(app) Migrate(app, db) # We can now pass in our app to the login manager login_manager.init_app(app) # Tell users what view to go to when they need to login. login_manager.login_view = "login"
[ "verdantfoxx@gmail.com" ]
verdantfoxx@gmail.com
d9d48cd7afade834ffcb015e029a3826b4194562
c351bc7388ece0e36b0c91cdc91dbce821940066
/TC/redes-tp1/herramientaS2.py
990187e7cca4726d39693846d36f107fa1d7c098
[]
no_license
julianlen/tps_facultad
059d62dc6a8c017688511e3d30ee93f3e0baca60
9b776e685036d34de5cf9dc86ad019e2f20f572d
refs/heads/master
2021-09-02T06:16:00.274517
2017-12-31T00:02:01
2017-12-31T00:02:01
115,832,205
0
0
null
null
null
null
UTF-8
Python
false
false
4,740
py
#!/usr/bin/env python # encoding: utf-8 from sys import argv, exit import math import matplotlib.pyplot as plt from matplotlib import pylab import logging logging.getLogger("scapy.runtime").setLevel(logging.ERROR) from scapy .all import * #import graphviz as gv import functools DECIMALES = 3 def PlotBars(simbolos, entropia, entropiaMax): info_set = set() filtered_simbolos = {} if (len(simbolos) > 10): for host in simbolos: info_host = simbolos[host] if info_host not in info_set: info_set.add(info_host) filtered_simbolos[host] = simbolos[host] else: filtered_simbolos = simbolos fig, ax = plt.subplots() #Dibuja entrop_line = ax.axvline(entropia, color='blue', linewidth=2,alpha = 0.7) maxEntrop_line = ax.axvline(entropiaMax, color='red', linewidth=2,alpha = 0.7) bars = ax.barh(range(len(filtered_simbolos)), filtered_simbolos.values(), align='center', alpha=0.4, color='green') plt.yticks(range(len(filtered_simbolos)), filtered_simbolos.keys(), fontsize=7, rotation=45) #Rótulos y títulos ax.legend((bars[0], entrop_line, maxEntrop_line), ('I(S(i))', 'H(S)', 'HMAX(S)')) pylab.xlabel("INFORMACION") pylab.ylabel("S(i)") pylab.title("Cantidad de INFORMACION por SIMBOLO") #Lo muestra plt.show() def PrintTable(titulos, tabla): row_format ="{:>20}" * len(titulos) print(row_format.format(*titulos)) for row in tabla: print(row_format.format(*row)) print("\n") def CreateRow(simbolo, s_prob, s_info, cant): return (simbolo, round(s_prob, DECIMALES), round(s_info, DECIMALES), cant) def GetSymbolFromFrame(paquete): return paquete[ARP].pdst def Condition(paquete): if ARP in paquete and paquete.op == 1 and paquete[ARP].psrc != paquete[ARP].pdst: nodos.add(paquete[ARP].psrc) nodos.add(paquete[ARP].pdst) aristas.add((paquete[ARP].psrc,paquete[ARP].pdst)) return True else: return False def PrintResults(tabla, tablaTitulos, entropia, entropiaMax): PrintTable(tablaTitulos, tabla) print("Entropía: {0} ({1:.2f})").format(int(math.ceil(entropia)), entropia) print("Entropía Máxima: {0} ({1:.2f})\n").format(int(math.ceil(entropiaMax)), entropiaMax) def AddNodes(graph, nodes): for n in nodes: if isinstance(n, tuple): graph.node(n[0], **n[1]) else: graph.node(n) return graph def AddEdges(graph, edges): for e in edges: if isinstance(e[0], tuple): graph.edge(*e[0], **e[1]) else: graph.edge(*e) return graph def herramienta(fnObtenerSimbolo, fnCondicion, fnGenerarItemDeTabla): simbolos = set() cantidadPorSimbolo = dict() infoPorSimbolos = dict() cantidadTotal = 0 for paquete in pkts: if fnCondicion(paquete): simbolo = fnObtenerSimbolo(paquete) simbolos.add(simbolo) if simbolo in cantidadPorSimbolo: cantidadPorSimbolo[simbolo] += 1 else: cantidadPorSimbolo[simbolo] = 1 cantidadTotal += 1 tabla = set() entropia = 0 entropiaMax = 0 for simbolo in simbolos: s_prob = float(cantidadPorSimbolo[simbolo]) / cantidadTotal s_info = math.log(float(1) / s_prob, 2) tabla.add(fnGenerarItemDeTabla(simbolo, s_prob, s_info, cantidadPorSimbolo[simbolo])) entropia += (s_prob * s_info) infoPorSimbolos[simbolo] = s_info tabla = sorted(tabla, key=lambda x: x[len(next(iter(tabla)))-2]) entropiaMax = math.log(len(simbolos), 2) return (simbolos, cantidadPorSimbolo, tabla, cantidadTotal, entropia, entropiaMax, infoPorSimbolos) if __name__ == '__main__': #Leer una captura desde el archivo de entrada if len(argv) == 2: pkts = rdpcap(argv[1]) else: print("Invalid Parameters") print("python file.py file.pcap") exit() #Creamos la lista de nodos y aristas para el grafo de la red nodos = set() aristas = set() #Para los paquetes de la captura, correr la herramienta (S2, apariciones, tabla, cantidad, entropia, entropiaMax, informacion) = herramienta(GetSymbolFromFrame, Condition, CreateRow) #Imprimir la tabla para S2 PrintResults(tabla, ["IP", "PROBABILIDAD", "INFORMACIÓN", "APARICIONES"], entropia, entropiaMax) #Gráfico de barras para S2 PlotBars(informacion, int(math.ceil(entropia)), entropiaMax) # #Grafo de la red # digraph = functools.partial(gv.Digraph, format='png') # graph = digraph() # AddEdges(AddNodes(graph, nodos),aristas).render(argv[1] + ".graph")
[ "julianlen@gmail.com" ]
julianlen@gmail.com
2044349675b019719547202cd026f4a96865b6c6
b4ad07eba142d4c2cbec643a3f679b7590a69a78
/FinalProject_CombatSim/sim_combat/army.py
a5d357af331dd04353adc6197bbd71e1f4d11ae3
[]
no_license
Heraclito1107/D3Labs_Heraclito
af344b4692ef3504e604ab025451d8f456a8e46f
b89a8bc405bc89ba5ef9730ba5a2f99a202a8030
refs/heads/master
2022-08-27T15:06:00.380790
2020-05-30T02:40:20
2020-05-30T02:40:20
266,623,516
0
0
null
null
null
null
UTF-8
Python
false
false
699
py
""" from characters.character import Character from characters.character_type import CharacterType class Army: def __init__(self, army_id: int): self._id = army_id self._size = 0 self._army = [0 for x in range(10)] def assignArmy(self, size: int, army_type: CharacterType) -> None: self._size = size if army_type == CharacterType.SOLDIER: self._army = [Character() for x in range(size)] def getCenter(self) -> (float, float): avg_x = 0 avg_y = 0 for x in range(self._size): avg_x += self._army[x].x avg_y += self._army[x].y return (avg_x/self._size), (avg_y/self._size) """
[ "heraclito.diaz@hotmail.com" ]
heraclito.diaz@hotmail.com
e8e634391f303d0b6ceac4ace6299bdd2bf13d16
92caa71c5a5135603cd875ce56dcd28a7dac7655
/GHER/common/vec_env/subproc_vec_env.py
44ac0f946d3087a480df1d7ca5ba9e3acb077a7b
[]
no_license
MrDadaGuy/GHER
286343d66815971afd5bd47f6ee0e1940b144610
8b1b2a2aba6a684db5921d2328c79e7f36aa8eae
refs/heads/master
2022-03-25T08:48:43.359759
2019-07-19T14:59:53
2019-07-19T14:59:53
173,989,952
0
0
null
2019-03-05T17:21:49
2019-03-05T17:21:49
null
UTF-8
Python
false
false
2,859
py
import numpy as np from multiprocessing import Process, Pipe from GHER.common.vec_env import VecEnv, CloudpickleWrapper def worker(remote, parent_remote, env_fn_wrapper): parent_remote.close() env = env_fn_wrapper.x() while True: cmd, data = remote.recv() if cmd == 'step': ob, reward, done, info = env.step(data) if done: ob = env.reset() remote.send((ob, reward, done, info)) elif cmd == 'reset': ob = env.reset() remote.send(ob) elif cmd == 'reset_task': ob = env.reset_task() remote.send(ob) elif cmd == 'close': remote.close() break elif cmd == 'get_spaces': remote.send((env.observation_space, env.action_space)) else: raise NotImplementedError class SubprocVecEnv(VecEnv): def __init__(self, env_fns, spaces=None): """ envs: list of gym environments to run in subprocesses """ self.waiting = False self.closed = False nenvs = len(env_fns) self.remotes, self.work_remotes = zip(*[Pipe() for _ in range(nenvs)]) self.ps = [Process(target=worker, args=(work_remote, remote, CloudpickleWrapper(env_fn))) for (work_remote, remote, env_fn) in zip(self.work_remotes, self.remotes, env_fns)] for p in self.ps: p.daemon = True # if the main process crashes, we should not cause things to hang p.start() for remote in self.work_remotes: remote.close() self.remotes[0].send(('get_spaces', None)) observation_space, action_space = self.remotes[0].recv() VecEnv.__init__(self, len(env_fns), observation_space, action_space) def step_async(self, actions): for remote, action in zip(self.remotes, actions): remote.send(('step', action)) self.waiting = True def step_wait(self): results = [remote.recv() for remote in self.remotes] self.waiting = False obs, rews, dones, infos = zip(*results) return np.stack(obs), np.stack(rews), np.stack(dones), infos def reset(self): for remote in self.remotes: remote.send(('reset', None)) return np.stack([remote.recv() for remote in self.remotes]) def reset_task(self): for remote in self.remotes: remote.send(('reset_task', None)) return np.stack([remote.recv() for remote in self.remotes]) def close(self): if self.closed: return if self.waiting: for remote in self.remotes: remote.recv() for remote in self.remotes: remote.send(('close', None)) for p in self.ps: p.join() self.closed = True
[ "baichenjia@163.com" ]
baichenjia@163.com
108cc53743d0761bb5612f81ccca8e5ac6f491ff
cb6f6888859567727e7a4c03f29ad0cf549d465a
/main1.py
a0729ed3f708bbdac13d7b5c2d969d3c76f373e8
[]
no_license
Parth-AI/file-uploader
bfe3781a440cd276192b847732cbb219fc8a8016
b7fe462d8d7f10b6c4200cede8af0f105e3657fa
refs/heads/main
2023-04-25T17:12:41.785404
2021-05-17T23:58:33
2021-05-17T23:58:33
368,351,186
0
0
null
null
null
null
UTF-8
Python
false
false
774
py
import dropbox class TransferData: def __init__(self, access_token): self.access_token = access_token def upload_file(self, file_from, file_to): """upload a file to Dropbox using API v2 """ dbx = dropbox.Dropbox(self.access_token) with open(file_from, 'rb') as f: dbx.files_upload(f.write('overwrite'), file_to) def main(): access_token = 'ayJVxaSAXWQAAAAAAAAAAX2ufOgVcoK7yEVIO1BSD8s38jdii7m25yBCLWm1M25I' transferData = TransferData(access_token) file_from = 'test.txt' file_to = '/Python101/test.txt' # The full path to upload the file to, including the file name # API v2 transferData.upload_file(file_from, file_to) if __name__ == '__main__': main()
[ "noreply@github.com" ]
Parth-AI.noreply@github.com
2da8ab85dde17e654b3141983d5e3f549d3c1c70
3bf41ea64766b4f39f90db90aea8a97ccf915c30
/Insertionsort.py
9fc1b47b8388b81deedee8edaea0ba5d43ced68a
[]
no_license
utkarsht724/Pythonprograms
b461eebd1414f00f45f37343df212ae02abe91c8
2613544a974b34d3939466482fe4ef7b32a9918d
refs/heads/master
2020-12-29T17:48:43.728569
2020-02-22T00:26:49
2020-02-22T00:26:49
238,688,221
0
0
null
null
null
null
UTF-8
Python
false
false
553
py
#sort the string using insertion sort def insertion_sort(names): for i in range(1,len(names)): value=names[i] #select value to be inserted hole=i while hole>0 and names[hole-1]>value: #locate hole position for the element to be inserted names[hole]=names[hole-1] hole-=1 names[hole]=value #insert the element at hole position #driver_class names = input("enter the name u want to sort: ").split() insertion_sort(names) print(names)
[ "utkarsht724@gmail.com" ]
utkarsht724@gmail.com
1511174dcb89eb47693c4902e477237c79e49570
a1dd53f96c79955a6723aca2627f207020d215eb
/bin/pyrsa-verify
2de038a1db0ff9eb17a4bce4d5390230411f2e16
[ "MIT" ]
permissive
linuxg33k76/speech
bfbf20efb84ba4bb1113960d9f287c30c047aad8
9e5c131a0f8b76756d52dacb764fb35bc8302dc4
refs/heads/main
2023-05-17T06:00:28.220754
2021-06-06T20:04:07
2021-06-06T20:04:07
374,419,223
0
0
null
null
null
null
UTF-8
Python
false
false
232
#!/home/linuxg33k/dev/speech/bin/python3 # -*- coding: utf-8 -*- import re import sys from rsa.cli import verify if __name__ == '__main__': sys.argv[0] = re.sub(r'(-script\.pyw|\.exe)?$', '', sys.argv[0]) sys.exit(verify())
[ "linuxg33k76@gmail.com" ]
linuxg33k76@gmail.com
40944c5fac9d6218fbb3fccd6144a95baa133e07
8bef4e20c773ebd6ccb6af0cb81eb788db920cb0
/Routes/addNote.py
bd5398cafada0dd39b1819418c86ad88086ec1f0
[]
no_license
animoffa/PhytonProj
44207cc3f52826c136b8ea2d204f443f0e604fd7
1e62c94af510631fb9ad3b46f6dd882687d183ee
refs/heads/master
2023-01-28T00:22:56.258766
2020-12-09T14:41:15
2020-12-09T14:41:15
319,985,055
0
0
null
null
null
null
UTF-8
Python
false
false
2,203
py
from flask.blueprints import Blueprint from flask import render_template from flask import request from models.lectгurers import Lect from models.subjects import Subj from models.intervals import Intervals from models.groups import Group from managers.DatabaseManager import DatabaseManager from extensions import db db_manager = DatabaseManager(db) addNote = Blueprint('addNote', __name__, template_folder='templates', static_folder='static') @addNote.route('/addNote') def index4(): return render_template('addNote.html', lecturers=Lect.query.all(), groups=Group.query.all(), intervals=Intervals.query.all(), subjects = Subj.query.all()) @addNote.route('/addNote',methods=['post', 'get']) def addN(): if request.method == 'POST': interval = request.form.get('interval') chet = request.form.get('chet') day = request.form.get('day') subject = request.form.get('subject') lecturer = request.form.get('lecturer') group = request.form.get('group') if interval and subject and lecturer and group and chet and day: message = "Note added successfully" new_dict = {} new_dict["lect"] = lecturer new_dict["subj"] = subject new_dict["interv"] = interval new_dict["group"] = group new = lecturer.split() name = new[1] l_name = new[0] s_name = new[2] lecturer_id = db.session.query(Lect.id).filter(Lect.name == name, Lect.last_name == l_name, Lect.surname == s_name).first()[0] group_id = db.session.query(Group.id).filter(Group.name==group).first()[0] subject_id = db.session.query(Subj.id).filter(Subj.subject_name == subject).first()[0] interval_id = db.session.query(Intervals.id).filter(Intervals.interval == interval).first()[0] db_manager.add_schedule(day=day,chet=chet,group_id=group_id, interval_id = interval_id, subject_id = subject_id, lecturer_id= lecturer_id) else: message = "error" return render_template('addNote.html', message=message, prob=" ", lecturers=Lect.query.all(), groups=Group.query.all(), intervals=Intervals.query.all(), subjects = Subj.query.all())
[ "animoffa1352@gmail.com" ]
animoffa1352@gmail.com
b449c65779fdc5cecf2dd164f7627c6d1568c411
b2bbd4ab93467ed9971a73e5393108943a3424e3
/programing/object_oriented_programing.py
57a124d982255a99e31046dc75886ac6dd2ffa0b
[]
no_license
ajaykr26/learning-python-programing
57c00325f1150897679e8c2071af31c1c55ba35a
5341727592677bd214de5b22a600a0c546c6d7ab
refs/heads/master
2022-04-08T12:07:25.407114
2020-03-18T04:52:49
2020-03-18T04:52:49
null
0
0
null
null
null
null
UTF-8
Python
false
false
8,743
py
# oops: inheritance, polymorphism, encapsulation, abstraction # inheritance: single level inheritance, multi level inheritance, hierarchical inheritance, multiple inheritance # class: A class is a blueprint for the third.We can think of class as an sketch of a parrot with labels. # It contains all the details about the name, colors, size etc. # Based on these descriptions, we can study about the parrot. Here, parrot is an third. # third: An third (instance) is an instantiation of a class. # When class is defined, only the description for the third is defined. Therefore, no memory or storage is allocated. # Example 1: Creating Class and Object in Python # class Parrot: # # class attribute # species = "bird" # # # instance attribute # def __init__(self, name, age): # self.name = name # self.age = age # # # blu = Parrot("Blu", 10) # instantiation of the Parrot class i.e creating third of parrot class # woo = Parrot("Woo", 15) # instantiation of the Parrot class i.e creating third of parrot class # # # accessing the class attributes # print("Blu is a {}".format(blu.__class__.species)) # print("Woo is a {}".format(woo.__class__.species)) # # # accessing the instance attributes # print("{} is {} years old".format(blu.name, blu.age)) # print("{} is {} years old".format(woo.name, woo.age)) # method: functions defined inside the body of a class. They are used to define the behaviors of an third. # Example 2 : Creating Methods in Python # class Parrot: # # # instance attributes # def __init__(self, name, age): # self.name = name # self.age = age # # # instance method # def sing(self, song): # return "{} is now {}".format(self.name, song) # # def dance(self): # return "{} is now dancing".format(self.name) # # # # instantiate the third # blu = Parrot("Blu", 10) # # # call our instance methods # print(blu.sing("singing")) # print(blu.dance()) # inheritance: Inheritance is a way of creating new class for using details of existing class without modifying it. # The newly formed class is a child class. Similarly, the existing class is a parent class. # single level inheritance, multi level inheritance, hierarchical inheritance, multiple inheritance # Example 3: Use of Inheritance in Python # single level inheritance # parent class # class ParentClass: # # class attribute # species = "animals" # # def __init__(self): # print("this is instantiation of parent class") # # def common_method(self): # print("this is overriden method in parent class") # # @staticmethod # def parent_method_one(arg): # print("this is static method in parent class: ", arg) # # # # child class # class ChildClass(ParentClass): # # class attribute # animal = "dogs" # # def __init__(self): # # call super() function # super().__init__() # # super(ChildClass, self).__init__() # print("this is instantiation of child class") # # def common_method(self): # print("this is overriding method in child class") # # @staticmethod # def child_method_one(arg): # print("this is static method in child class: ", arg) # # # parent = ParentClass() # print(parent.species) # parent.common_method() # parent.parent_method_one('parent') # # child = ChildClass() # print(child.species, child.animal) # child.common_method() # child.child_method_one('child') # multi level inheritance # first class # class FirstClass: # # class attribute # first_attribute = "first attribute" # # def __init__(self): # print("this is instantiation of first class") # # # def common_method(self): # # print("this is overriden method in first class") # # @staticmethod # def first_method_one(arg): # print("this is static method in first class: ", arg) # # # # second class # class SecondClass(FirstClass): # # class attribute # second_attribute = "second attribute" # # def __init__(self): # # call super() function # # super().__init__() # super(SecondClass, self).__init__() # print("this is instantiation of second class") # # def common_method(self): # print("this is overriding method in second class") # # @staticmethod # def second_method_one(arg): # print("this is static method in second class: ", arg) # # # # third class # class ThirdClass(SecondClass): # # class attribute # third_attribute = "third attribute" # # def __init__(self): # # call super() function # # super().__init__() # super(ThirdClass, self).__init__() # print("this is instantiation of third class") # # def common_method(self): # print("this is overriding method in third class") # # @staticmethod # def third_method_one(arg): # print("this is static method in third class: ", arg) # # # third = ThirdClass() # print(third.first_attribute, third.second_attribute, third.third_attribute) # third.first_method_one('first') # third.second_method_one('second') # third.third_method_one('third') # third.common_method() # multiple inheritance # first class # class FirstClass: # # class attribute # first_attribute = "first attribute" # # def __init__(self): # print("this is instantiation of first class") # # def common_method_one(self): # print("this is overriden method in first class") # # @staticmethod # def first_method_one(arg): # print("this is static method in first class: ", arg) # # # # second class # class SecondClass: # # class attribute # second_attribute = "second attribute" # # def __init__(self): # # call super() function # # super().__init__() # super(SecondClass, self).__init__() # print("this is instantiation of second class") # # def common_method_two(self): # print("this is overriding method in second class") # # @staticmethod # def second_method_one(arg): # print("this is static method in second class: ", arg) # # # # third class # class ThirdClass(FirstClass, SecondClass): # # class attribute # third_attribute = "third attribute" # # def __init__(self): # # call super() function # super(ThirdClass, self).__init__() # print("this is instantiation of third class") # # def common_method_one(self): # print("this is overriding method in first class") # # def common_method_two(self): # print("this is overriding method in second class") # # @staticmethod # def third_method_one(arg): # print("this is static method in third class: ", arg) # # # third = ThirdClass() # print(third.first_attribute, third.second_attribute, third.third_attribute) # third.first_method_one('first') # third.second_method_one('second') # third.third_method_one('third') # third.common_method_one() # third.common_method_two() # hierarchical inheritance # first class # class FirstClass: # # class attribute # first_attribute = "first attribute" # # def __init__(self): # print("this is instantiation of first class") # # def common_method_one(self): # print("this is overriden method in first class") # # @staticmethod # def first_method_one(arg): # print("this is static method in first class: ", arg) # # # # second class # class SecondClass(FirstClass): # # class attribute # second_attribute = "second attribute" # # def __init__(self): # # call super() function # # super().__init__() # super(SecondClass, self).__init__() # print("this is instantiation of second class") # # def common_method_one(self): # print("this is overriding method in second class") # # @staticmethod # def second_method_one(arg): # print("this is static method in second class: ", arg) # # # # third class # class ThirdClass(FirstClass): # # class attribute # third_attribute = "third attribute" # # def __init__(self): # # call super() function # super(ThirdClass, self).__init__() # print("this is instantiation of third class") # # def common_method_one(self): # print("this is overriding method in first class") # # @staticmethod # def third_method_one(arg): # print("this is static method in third class: ", arg) # # # third = ThirdClass() # second = SecondClass() # print(third.first_attribute, third.third_attribute) # print(second.first_attribute, second.second_attribute) # second.first_method_one('second') # third.first_method_one('third') # second.common_method_one() # third.common_method_one()
[ "ajaykr.mailbox@gmail.com" ]
ajaykr.mailbox@gmail.com