blob_id
stringlengths
40
40
directory_id
stringlengths
40
40
path
stringlengths
3
281
content_id
stringlengths
40
40
detected_licenses
listlengths
0
57
license_type
stringclasses
2 values
repo_name
stringlengths
6
116
snapshot_id
stringlengths
40
40
revision_id
stringlengths
40
40
branch_name
stringclasses
313 values
visit_date
timestamp[us]
revision_date
timestamp[us]
committer_date
timestamp[us]
github_id
int64
18.2k
668M
star_events_count
int64
0
102k
fork_events_count
int64
0
38.2k
gha_license_id
stringclasses
17 values
gha_event_created_at
timestamp[us]
gha_created_at
timestamp[us]
gha_language
stringclasses
107 values
src_encoding
stringclasses
20 values
language
stringclasses
1 value
is_vendor
bool
2 classes
is_generated
bool
2 classes
length_bytes
int64
4
6.02M
extension
stringclasses
78 values
content
stringlengths
2
6.02M
authors
listlengths
1
1
author
stringlengths
0
175
3c27fc64a1e079739ba16dd4a648e0eb9ab172c6
89b5966506cb7dcbdbce8f686d920b7146cceebe
/Exponential_Experiments/HMC_runner.py
bcd2d38ec90d000a0d82ded4bbb9be8f48c89a81
[]
no_license
matthewwicker/CertifiableBayesianInference
b4f796b7ea9cdcffb067ab8896ecfaafb1687b62
289b87b0332de81b2132f6461af24f7e2a6310de
refs/heads/main
2022-12-30T17:38:17.111368
2020-10-22T04:11:54
2020-10-22T04:11:54
302,975,908
5
1
null
null
null
null
UTF-8
Python
false
false
3,064
py
# Author: Matthew Wicker # Description: Minimal working example of training and saving # a BNN trained with Bayes by backprop (BBB) # can handle any Keras model import sys, os from pathlib import Path path = Path(os.getcwd()) sys.path.append(str(path.parent)) import BayesKeras import BayesKeras.optimizers as optimizers import tensorflow as tf from tensorflow.keras.models import * from tensorflow.keras.layers import * #tf.debugging.set_log_device_placement(True) #os.environ['CUDA_VISIBLE_DEVICES'] = '-1' import argparse parser = argparse.ArgumentParser() parser.add_argument("--eps") parser.add_argument("--lam") parser.add_argument("--rob") parser.add_argument("--gpu", nargs='?', default='0,1,2,3,4,5') parser.add_argument("--opt") args = parser.parse_args() eps = float(args.eps) lam = float(args.lam) optim = str(args.opt) rob = int(args.rob) gpu = str(args.gpu) os.environ['CUDA_VISIBLE_DEVICES'] = gpu (X_train, y_train), (X_test, y_test) = tf.keras.datasets.mnist.load_data() X_train = X_train/255. X_test = X_test/255. X_train = X_train.astype("float32").reshape(-1, 28*28) X_test = X_test.astype("float32").reshape(-1, 28* 28) #X_train = X_train[0:10000] #y_train = y_train[0:10000] model = Sequential() model.add(Dense(512, activation="relu", input_shape=(1, 28*28))) model.add(Dense(10, activation="softmax")) inf = 2 full_covar = False if(optim == 'VOGN'): # was 0.25 for a bit inf = 2 learning_rate = 0.35; decay=0.0 opt = optimizers.VariationalOnlineGuassNewton() elif(optim == 'BBB'): learning_rate = 0.5; decay=0.0 opt = optimizers.BayesByBackprop() elif(optim == 'SWAG'): learning_rate = 0.01; decay=0.0 opt = optimizers.StochasticWeightAveragingGaussian() elif(optim == 'SWAG-FC'): learning_rate = 0.01; decay=0.0; full_covar=True opt = optimizers.StochasticWeightAveragingGaussian() elif(optim == 'SGD'): learning_rate = 1.0; decay=0.0 opt = optimizers.StochasticGradientDescent() elif(optim == 'NA'): inf = 2 learning_rate = 0.001; decay=0.0 opt = optimizers.NoisyAdam() elif(optim == 'ADAM'): learning_rate = 0.00001; decay=0.0 opt = optimizers.Adam() elif(optim == 'HMC'): # learning_rate = 0.075; decay=0.0; inf=250 # used 25 steps learning_rate = 0.01; decay=0.0; inf=200 linear_schedule = False opt = optimizers.HamiltonianMonteCarlo() # Compile the model to train with Bayesian inference if(rob == 0 or rob >=4): loss = tf.keras.losses.SparseCategoricalCrossentropy() elif(rob != 0): loss = BayesKeras.optimizers.losses.robust_crossentropy_loss bayes_model = opt.compile(model, loss_fn=loss, epochs=20, learning_rate=learning_rate, decay=decay, robust_train=rob, inflate_prior=inf, burn_in=1, steps=25, b_steps=20, epsilon=eps, rob_lam=lam , preload="SGD_FCN_Posterior_%s"%(rob)) #steps was 50 # Train the model on your data bayes_model.train(X_train, y_train, X_test, y_test) # Save your approxiate Bayesian posterior bayes_model.save("%s_FCN_Posterior_%s"%(optim, rob))
[ "matthewwicker@cs.ox.ac.uk" ]
matthewwicker@cs.ox.ac.uk
f236d30755d31dd21adfeefd650ed057583684a9
084191aab1a6f975515fc33b7c7e55f7e2fb90fb
/DataRace/DataFountain/消费金融场景下的用户购买预测/git方案学习/rank1/extract_feature.py
ecfd7818044085ce984ee27e3eae86d562e76d44
[]
no_license
whoami-zy/StudyML
481829be126a789faa79524a2cc5acee28aa3174
5f12d6cb1b97ba4b2dbabe83062c6435b7729fd4
refs/heads/master
2020-04-05T20:30:05.789389
2018-11-13T07:14:33
2018-11-13T07:14:33
157,183,037
1
0
null
null
null
null
UTF-8
Python
false
false
16,700
py
import pandas as pd import numpy as np from collections import Counter import scipy.stats as sp import time import datetime def get_continue_launch_count(strs,parm): time = strs.split(":") time = dict(Counter(time)) time = sorted(time.items(), key=lambda x: x[0], reverse=False) key_list = [] value_list = [] if len(time) == 1: return -2 for key,value in dict(time).items(): key_list.append(int(key)) value_list.append(int(value)) if np.mean(np.diff(key_list, 1)) == 1: if parm == '1': return np.mean(value_list) elif parm == '2': return np.max(value_list) elif parm == '3': return np.min(value_list) elif parm == '4': return np.sum(value_list) elif parm == '5': return np.std(value_list) else: return -1 def get_time_gap(strs,parm): time = strs.split(":") time = list(set(time)) time = sorted(list(map(lambda x:int(x),time))) time_gap = [] #用户只在当天活跃 if len(time) == 1: return -20 for index, value in enumerate(time): if index <= len(time) - 2: gap = abs(time[index] - time[index + 1]) time_gap.append(gap) if parm == '1': return np.mean(time_gap) elif parm == '2': return np.max(time_gap) elif parm == '3': return np.min(time_gap) elif parm == '4': return np.std(time_gap) elif parm == '5': return sp.stats.skew(time_gap) elif parm == '6': return sp.stats.kurtosis(time_gap) def get_week(day): day = int(day) if day >= 1 and day <= 7: return 1 if day >= 8 and day <= 14: return 2 if day >= 15 and day <= 21: return 3 if day >= 22 and day <= 28: return 4 if day >= 28: return 5 def cur_day_repeat_count(strs): time = strs.split(":") time = dict(Counter(time)) time = sorted(time.items(), key=lambda x: x[1], reverse=False) # 一天一次启动 if (len(time) == 1) & (time[0][1] == 1): return 0 # 一天多次启动 elif (len(time) == 1) & (time[0][1] > 1): return 1 # 多天多次启动 elif (len(time) > 1) & (time[0][1] >= 2): return 2 else: return 3 def get_lianxu_day(day_list): time = day_list.split(":") time = list(map(lambda x:int(x),time)) m = np.array(time) if len(set(m)) == 1: return -1 m = list(set(m)) if len(m) == 0: return -20 n = np.where(np.diff(m) == 1)[0] i = 0 result = [] while i < len(n) - 1: state = 1 while n[i + 1] - n[i] == 1: state += 1 i += 1 if i == len(n) - 1: break if state == 1: i += 1 result.append(2) else: i += 1 result.append(state + 1) if len(n) == 1: result.append(2) if len(result) != 0: # print(result) return np.max(result) def load_csv(): train_agg = pd.read_csv('../orig_data/train_agg.csv',sep='\t') train_log = pd.read_csv('../orig_data/train_log.csv', sep='\t') train_flg = pd.read_csv('../orig_data/train_flg.csv', sep='\t') test_agg = pd.read_csv('../orig_data/test_agg.csv', sep='\t') test_log = pd.read_csv('../orig_data/test_log.csv', sep='\t') return train_agg,train_log,train_flg,test_agg,test_log def merge_table(train_agg, train_log, train_flg, test_agg, test_log): train_log['label'] = 1 test_log['label'] = 0 data = pd.concat([train_log,test_log],axis=0) data = extract_feature(data) train_log = data[data.label == 1] test_log = data[data.label == 0] del train_log['label'] del test_log['label'] all_train = pd.merge(train_flg, train_agg, on=['USRID'], how='left') train = pd.merge(all_train,train_log,on='USRID',how='left') test = pd.merge(test_agg,test_log,on='USRID',how='left') return train,test def extract_feature(data): data['cate_1'] = data['EVT_LBL'].apply(lambda x: int(x.split('-')[0])) data['cate_2'] = data['EVT_LBL'].apply(lambda x: int(x.split('-')[1])) data['cate_3'] = data['EVT_LBL'].apply(lambda x: int(x.split('-')[2])) data['day'] = data['OCC_TIM'].apply(lambda x: int(x[8:10])) data['hour'] = data['OCC_TIM'].apply(lambda x: int(x[11:13])) data['week'] = data['day'].apply(get_week) feat1 = data.groupby(['USRID'], as_index=False)['OCC_TIM'].agg({"user_count": "count"}) feat2 = data.groupby(['USRID'], as_index=False)['day'].agg({"user_act_day_count": "nunique"}) feat3 = data[['USRID', 'day']] feat3['day'] = feat3['day'].astype('str') feat3 = feat3.groupby(['USRID'])['day'].agg(lambda x: ':'.join(x)).reset_index() feat3.rename(columns={'day': 'act_list'}, inplace=True) # 用户是否多天有多次启动(均值) feat3['time_gap_mean'] = feat3['act_list'].apply(get_time_gap,args=('1')) # 最大 feat3['time_gap_max'] = feat3['act_list'].apply(get_time_gap,args=('2')) # 最小 feat3['time_gap_min'] = feat3['act_list'].apply(get_time_gap,args=('3')) # 方差 feat3['time_gap_std'] = feat3['act_list'].apply(get_time_gap,args=('4')) # 锋度 feat3['time_gap_skew'] = feat3['act_list'].apply(get_time_gap, args=('5')) # 偏度 feat3['time_gap_kurt'] = feat3['act_list'].apply(get_time_gap, args=('6')) # 平均行为次数 feat3['mean_act_count'] = feat3['act_list'].apply(lambda x: len(x.split(":")) / len(set(x.split(":")))) # 平均行为日期 feat3['act_mean_date'] = feat3['act_list'].apply(lambda x: np.sum([int(ele) for ele in x.split(":")]) / len(x.split(":"))) # 活动天数占当月的比率 # feat3['act_rate'] = feat3['act_list'].apply(lambda x: len(list(set(x.split(":")))) / 31) # 用户是否当天有多次启动 feat3['cur_day_repeat_count'] = feat3['act_list'].apply(cur_day_repeat_count) # 连续几天启动次数的均值, feat3['con_act_day_count_mean'] = feat3['act_list'].apply(get_continue_launch_count, args=('1')) # 最大值, feat3['con_act_day_count_max'] = feat3['act_list'].apply(get_continue_launch_count, args=('2')) # 最小值 feat3['con_act_day_count_min'] = feat3['act_list'].apply(get_continue_launch_count, args=('3')) # 次数 feat3['con_act_day_count_total'] = feat3['act_list'].apply(get_continue_launch_count, args=('4')) # 方差 feat3['con_act_day_count_std'] = feat3['act_list'].apply(get_continue_launch_count, args=('5')) feat3['con_act_max'] = feat3['act_list'].apply(get_lianxu_day) del feat3['act_list'] # 用户发生行为的天数 feat4 = data.groupby(['USRID'], as_index=False)['cate_1'].agg({'user_cate_1_count': "count"}) feat5 = data.groupby(['USRID'], as_index=False)['cate_2'].agg({'user_cate_2_count': "count"}) feat6 = data.groupby(['USRID'], as_index=False)['cate_3'].agg({'user_cate_3_count': "count"}) # 判断时期是否为高峰日 higt_act_day_list = [7, 14, 21, 28] feat8 = data[['USRID', 'day']] feat8['is_higt_act'] = feat8['day'].apply(lambda x: 1 if x in higt_act_day_list else 0) feat8 = feat8.drop_duplicates(subset=['USRID']) feat10 = data.groupby(['USRID','day'], as_index=False)['TCH_TYP'].agg({'user_per_count': "count"}) feat10_copy = feat10.copy() # 用户平均每天启动次数 feat11 = feat10_copy.groupby(['USRID'],as_index=False)['user_per_count'].agg({"user_per_count_mean":"mean"}) # 用户启动次数最大值 feat12 = feat10_copy.groupby(['USRID'], as_index=False)['user_per_count'].agg({"user_per_count_max": "max"}) # 用户启动次数最小值 feat13 = feat10_copy.groupby(['USRID'], as_index=False)['user_per_count'].agg({"user_per_count_min": "min"}) # 用户每天启动次数的众值 feat14 = feat10_copy.groupby(['USRID'], as_index=False)['user_per_count'].agg({"user_mode_count":lambda x: x.value_counts().index[0]}) # 方差 feat15 = feat10_copy.groupby(['USRID'], as_index=False)['user_per_count'].agg({"user_std_count":np.std}) # 峰度 feat16 = feat10_copy.groupby(['USRID'], as_index=False)['user_per_count'].agg({"user_skew_count": sp.stats.skew}) # 偏度 feat17 = feat10_copy.groupby(['USRID'], as_index=False)['user_per_count'].agg({"user_kurt_count": sp.stats.kurtosis}) # 中位数 feat18 = feat10_copy.groupby(['USRID'], as_index=False)['user_per_count'].agg({"user_median_count": np.median}) feat27 = data[['USRID', 'OCC_TIM']] feat27['OCC_TIM'] = feat27['OCC_TIM'].apply(lambda x: time.mktime(time.strptime(x, "%Y-%m-%d %H:%M:%S"))) log = feat27.sort_values(['USRID', 'OCC_TIM']) log['next_time'] = log.groupby(['USRID'])['OCC_TIM'].diff(-1).apply(np.abs) log = log.groupby(['USRID'], as_index=False)['next_time'].agg({ 'next_time_mean': np.mean, 'next_time_std': np.std, 'next_time_min': np.min, 'next_time_max': np.max }) # 每周的平均消费次数 feat28_sp = data.groupby(['USRID','week'], as_index=False)['TCH_TYP'].agg({'user_per_week_count': "count"}) feat28_sp_copy = feat28_sp.copy() # 用户平均每天启动次数 feat11_sp = feat28_sp_copy.groupby(['USRID'], as_index=False)['user_per_week_count'].agg({"user_per_week_count_mean": "mean"}) # 用户启动次数最大值 feat12_sp = feat28_sp_copy.groupby(['USRID'], as_index=False)['user_per_week_count'].agg({"user_per_week_count_max": "max"}) # 用户启动次数最小值 feat13_sp = feat28_sp_copy.groupby(['USRID'], as_index=False)['user_per_week_count'].agg({"user_per_week_count_min": "min"}) # 用户每天启动次数的众值 feat14_sp = feat28_sp_copy.groupby(['USRID'], as_index=False)['user_per_week_count'].agg({"user_per_week_count_mode": lambda x: x.value_counts().index[0]}) # 方差 feat15_sp = feat28_sp_copy.groupby(['USRID'], as_index=False)['user_per_week_count'].agg({"user_per_week_count_std": np.std}) # 峰度 feat16_sp = feat28_sp_copy.groupby(['USRID'], as_index=False)['user_per_week_count'].agg({"user_per_week_count_skew": sp.stats.skew}) # 偏度 feat17_sp = feat28_sp_copy.groupby(['USRID'], as_index=False)['user_per_week_count'].agg({"user_per_week_count_kurt": sp.stats.kurtosis}) # 中位数 feat18_sp = feat28_sp_copy.groupby(['USRID'], as_index=False)['user_per_week_count'].agg({"user_per_week_count_median": np.median}) # 离周末越近,越消费的可能性比较大,统计前2天的特征 before_three = data[(data.day >= 28) & (data.day <= 31)] before_three_copy = before_three.copy() feat1_before = before_three_copy.groupby(['USRID'], as_index=False)['OCC_TIM'].agg({"user_count_before": "count"}) feat2_before = before_three_copy.groupby(['USRID'], as_index=False)['day'].agg({"user_act_day_count_before": "nunique"}) feat3_before = before_three_copy[['USRID', 'day']] feat3_before['day'] = feat3_before['day'].astype('str') feat3_before = feat3_before.groupby(['USRID'])['day'].agg(lambda x: ':'.join(x)).reset_index() feat3_before.rename(columns={'day': 'act_list'}, inplace=True) # 用户是否多天有多次启动(均值) feat3_before['before_time_gap_mean'] = feat3_before['act_list'].apply(get_time_gap, args=('1')) # 最大 feat3_before['before_time_gap_max'] = feat3_before['act_list'].apply(get_time_gap, args=('2')) # 最小 feat3_before['before_time_gap_min'] = feat3_before['act_list'].apply(get_time_gap, args=('3')) # 方差 feat3_before['before_time_gap_std'] = feat3_before['act_list'].apply(get_time_gap, args=('4')) # 锋度 feat3_before['before_time_gap_skew'] = feat3_before['act_list'].apply(get_time_gap, args=('5')) # 偏度 feat3_before['before_time_gap_kurt'] = feat3_before['act_list'].apply(get_time_gap, args=('6')) # 平均行为次数 feat3_before['before_mean_act_count'] = feat3_before['act_list'].apply(lambda x: len(x.split(":")) / len(set(x.split(":")))) # 平均行为日期 feat3_before['before_act_mean_date'] = feat3_before['act_list'].apply(lambda x: np.sum([int(ele) for ele in x.split(":")]) / len(x.split(":"))) # 用户是否当天有多次启动 feat3_before['before_cur_day_repeat_count'] = feat3_before['act_list'].apply(cur_day_repeat_count) # 连续几天启动次数的均值, feat3_before['before_con_act_day_count_mean'] = feat3_before['act_list'].apply(get_continue_launch_count, args=('1')) # 最大值, feat3_before['before_con_act_day_count_max'] = feat3_before['act_list'].apply(get_continue_launch_count, args=('2')) # 最小值 feat3_before['before_con_act_day_count_min'] = feat3_before['act_list'].apply(get_continue_launch_count, args=('3')) # 次数 feat3_before['before_con_act_day_count_total'] = feat3_before['act_list'].apply(get_continue_launch_count, args=('4')) # 方差 feat3_before['before_con_act_day_count_std'] = feat3_before['act_list'].apply(get_continue_launch_count, args=('5')) feat3_before['before_con_act_max'] = feat3_before['act_list'].apply(get_lianxu_day) del feat3_before['act_list'] # 用户发生行为的天数 feat4_before = before_three.groupby(['USRID'], as_index=False)['cate_1'].agg({'before_user_cate_1_count': "count"}) feat5_before = before_three.groupby(['USRID'], as_index=False)['cate_2'].agg({'before_user_cate_2_count': "count"}) feat6_before = before_three.groupby(['USRID'], as_index=False)['cate_3'].agg({'before_user_cate_3_count': "count"}) feat28 = pd.crosstab(data['USRID'],data['TCH_TYP']).reset_index() feat29 = pd.crosstab(data.USRID,data.cate_1).reset_index() feat30 = pd.crosstab(data.USRID, data.cate_2).reset_index() feat31 = pd.crosstab(data.USRID, data.cate_3).reset_index() feat32 = pd.crosstab(data.USRID,data.hour).reset_index() feat34 = pd.crosstab(data.USRID,data.week).reset_index() data = data[['USRID','label']] data = data.drop_duplicates(subset='USRID') data = pd.merge(data, feat1, on=['USRID'], how='left') data = pd.merge(data, feat2, on=['USRID'], how='left') data = pd.merge(data, feat3, on=['USRID'], how='left') data = pd.merge(data, feat4, on=['USRID'], how='left') data = pd.merge(data, feat5, on=['USRID'], how='left') data = pd.merge(data, feat6, on=['USRID'], how='left') data = pd.merge(data, feat8, on=['USRID'], how='left') data = pd.merge(data, feat11, on=['USRID'], how='left') data = pd.merge(data, feat12, on=['USRID'], how='left') data = pd.merge(data, feat13, on=['USRID'], how='left') data = pd.merge(data, feat14, on=['USRID'], how='left') data = pd.merge(data, feat15, on=['USRID'], how='left') data = pd.merge(data, feat16, on=['USRID'], how='left') data = pd.merge(data, feat17, on=['USRID'], how='left') data = pd.merge(data, feat18, on=['USRID'], how='left') data = pd.merge(data, log, on=['USRID'], how='left') data = pd.merge(data, feat28, on=['USRID'], how='left') data = pd.merge(data, feat29, on=['USRID'], how='left') data = pd.merge(data, feat30, on=['USRID'], how='left') data = pd.merge(data, feat31, on=['USRID'], how='left') data = pd.merge(data, feat32, on=['USRID'], how='left') data = pd.merge(data, feat34, on=['USRID'], how='left') data = pd.merge(data, feat11_sp, on=['USRID'], how='left') data = pd.merge(data, feat12_sp, on=['USRID'], how='left') data = pd.merge(data, feat13_sp, on=['USRID'], how='left') data = pd.merge(data, feat14_sp, on=['USRID'], how='left') data = pd.merge(data, feat15_sp, on=['USRID'], how='left') data = pd.merge(data, feat16_sp, on=['USRID'], how='left') data = pd.merge(data, feat17_sp, on=['USRID'], how='left') data = pd.merge(data, feat18_sp, on=['USRID'], how='left') data = pd.merge(data, feat1_before, on=['USRID'], how='left') data = pd.merge(data, feat2_before, on=['USRID'], how='left') data = pd.merge(data, feat3_before, on=['USRID'], how='left') data = pd.merge(data, feat4_before, on=['USRID'], how='left') data = pd.merge(data, feat5_before, on=['USRID'], how='left') data = pd.merge(data, feat6_before, on=['USRID'], how='left') return data def main(): train_agg, train_log, train_flg, test_agg, test_log = load_csv() train, test = merge_table(train_agg, train_log, train_flg, test_agg, test_log) train.to_csv('../fea/train.csv',sep='\t',index=None) test.to_csv('../fea/test.csv', sep='\t', index=None) if __name__ == '__main__': main()
[ "wantong.sun@foxmail.com" ]
wantong.sun@foxmail.com
1cf47e979c62abe7878aec58e70e8bf82cace12f
3cfc6d23f37e45b8fd8b3810aa56eee21a493a01
/custom/plugins/RBKeyshot/KeyShot_RenderScript.py
1b2b4b7b7cbcf42f8fc4921ae87894b943238807
[]
no_license
joinmm/Deadline_Development
eb72f13e1beffac2dd55b3d0eb69d56b98110a86
90b1031ffa27177c2b7b93ac4fa59fca0f79e227
refs/heads/master
2023-03-17T22:56:53.716116
2019-08-30T03:18:33
2019-08-30T03:18:33
null
0
0
null
null
null
null
UTF-8
Python
false
false
6,952
py
import os import time import shutil HOME_PATH = os.path.join(os.environ["HOMEPATH"], "Desktop", "Temp") SCENE_FILE_PATH = "A:/RenderShot_Dir/Files/ctCmh6931TKgvV2/made_to_travel_black_rev4_92630339406526/made_to_travel_black_rev4.bip" NEW_SCENE_FILE_NAME = os.path.basename(SCENE_FILE_PATH) NEW_TEMP_SCENE_FILE_NAME = "" def valid_temp_folder(): if os.path.exists(HOME_PATH): print("Temp folder has already been created.") return True else: try: os.makedirs(HOME_PATH) print("Temp folder created successfully.") return True except: print("Temp folder could not be created.") return False def dir_update_check(NETWORK_FILE_DIR, DESTINATION_PATH): NETWORK_FILE_DIR_LIST = os.listdir(NETWORK_FILE_DIR) DESTINATION_PATH_LIST = os.listdir(DESTINATION_PATH) if len(NETWORK_FILE_DIR_LIST) == len(DESTINATION_PATH_LIST)or len(NETWORK_FILE_DIR_LIST) < len(DESTINATION_PATH_LIST): print("No directory update required.") return True else: print("Directory update required.") return False def file_transfer(SCENE_FILE_PATH): NETWORK_FILE_DIR = os.path.dirname(SCENE_FILE_PATH) NETWORK_DIR_NAME = os.path.basename(NETWORK_FILE_DIR) DESTINATION_PATH = os.path.join(os.environ["HOMEPATH"], "Desktop", "Temp", NETWORK_DIR_NAME) NEW_SCENE_PATH = os.path.join(DESTINATION_PATH, os.path.basename(SCENE_FILE_PATH)) if os.path.exists(DESTINATION_PATH)and dir_update_check(NETWORK_FILE_DIR, DESTINATION_PATH): print("Render folder has already been transferred , returning immediately .") return NEW_SCENE_PATH elif os.path.exists(DESTINATION_PATH) and not dir_update_check(NETWORK_FILE_DIR, DESTINATION_PATH): shutil.rmtree(DESTINATION_PATH) print("Render folder has been removed.") if valid_temp_folder() : try: shutil.copytree(NETWORK_FILE_DIR, DESTINATION_PATH) print("Render folder transferred successfully.") except: print("Render folder could not be transferred.") else: print("File transfer failed") return NEW_SCENE_PATH def main(scene_file_path): lux.openFile(scene_file_path) lux.setCamera("Camera 2") lux.setAnimationFrame( 0 ) lux.pause lux.setAnimationFrame( 0 ) lux.unpause lux.setAnimationFrame( 0 ) lux.saveFile( "A:/RenderShot_Dir/Files/ctCmh6931TKgvV2/made_to_travel_black_rev4_92630339406526/made_to_travel_black_rev4_1561004076_Camera 2_0_.bip") lux.openFile( "A:/RenderShot_Dir/Files/ctCmh6931TKgvV2/made_to_travel_black_rev4_92630339406526/made_to_travel_black_rev4_1561004076_Camera 2_0_.bip") path = "A:/Test_Output/made_to_travel_black_rev4_1560962403_%d.tif" width = 1920 height = 1080 opts = lux.getRenderOptions() opts.setAddToQueue(False) opts.setOutputRenderLayers(False) opts.setOutputAlphaChannel(False) try: opts.setOutputDiffusePass(False) except AttributeError: print( "Failed to set render pass: output_diffuse_pass" ) try: opts.setOutputReflectionPass(False) except AttributeError: print( "Failed to set render pass: output_reflection_pass" ) try: opts.setOutputClownPass(False) except AttributeError: print( "Failed to set render pass: output_clown_pass" ) try: opts.setOutputDirectLightingPass(False) except AttributeError: print( "Failed to set render pass: output_direct_lighting_pass" ) try: opts.setOutputRefractionPass(False) except AttributeError: print( "Failed to set render pass: output_refraction_pass" ) try: opts.setOutputDepthPass(False) except AttributeError: print( "Failed to set render pass: output_depth_pass" ) try: opts.setOutputIndirectLightingPass(False) except AttributeError: print( "Failed to set render pass: output_indirect_lighting_pass" ) try: opts.setOutputShadowPass(False) except AttributeError: print( "Failed to set render pass: output_indirect_lighting_pass" ) try: opts.setOutputNormalsPass(False) except AttributeError: print( "Failed to set render pass: output_normals_pass" ) try: opts.setOutputCausticsPass(False) except AttributeError: print( "Failed to set render pass: output_caustics_pass" ) try: opts.setOutputShadowPass(False) except AttributeError: print( "Failed to set render pass: output_shadow_pass" ) try: opts.setOutputAmbientOcclusionPass(False) except AttributeError: print( "Failed to set render pass: output_ambient_occlusion_pass" ) try: opts.setAdvancedRendering( 38 ) except AttributeError: print( "Failed to set render option: advanced_samples" ) try: opts.setGlobalIllumination( 1.0 ) except AttributeError: print( "Failed to set render option: engine_global_illumination" ) try: opts.setRayBounces( 14 ) except AttributeError: print( "Failed to set render option: engine_ray_bounces" ) try: opts.setPixelBlur( 1.5 ) except AttributeError: print( "Failed to set render option: engine_pixel_blur" ) try: opts.setAntiAliasing( 3 ) except AttributeError: print( "Failed to set render option: engine_anti_aliasing" ) try: opts.setDofQuality( 3 ) except AttributeError: print( "Failed to set render option: engine_dof_quality" ) try: opts.setShadowQuality( 4.47200012207 ) except AttributeError: print( "Failed to set render option: engine_shadow_quality" ) try: opts.setCausticsQuality( 0.0 ) except AttributeError: print( "Failed to set render option: engine_caustics_quality" ) try: opts.setSharpShadows( True ) except AttributeError: print( "Failed to set render option: engine_sharp_shadows" ) try: opts.setSharperTextureFiltering( True ) except AttributeError: print( "Failed to set render option: engine_sharper_texture_filtering" ) try: opts.setGlobalIlluminationCache( True ) except AttributeError: print( "Failed to set render option: engine_global_illumination_cache" ) for frame in range( 0, 1 ): renderPath = path renderPath = renderPath.replace( "%d", str(frame) ) lux.setAnimationFrame( frame ) lux.renderImage(path = renderPath, width = width, height = height, opts = opts) print("Rendered Image: "+renderPath) os.remove( "A:/RenderShot_Dir/Files/ctCmh6931TKgvV2/made_to_travel_black_rev4_92630339406526/made_to_travel_black_rev4_1561004076_Camera 2_0_.bip") print ('Job Completed') exit() GET_NEW_FILE_PATH = file_transfer(SCENE_FILE_PATH) if GET_NEW_FILE_PATH: main(GET_NEW_FILE_PATH) else: main(SCENE_FILE_PATH)
[ "hamedhematyar91@gmail.com" ]
hamedhematyar91@gmail.com
46d6f02672c8b538642ba2d89ea3eefeabcf2b6f
38fd6b48e6ec234c60596645bd1bf9f0c9920a3d
/crm web app-django3/accounts/migrations/0004_auto_20191125_1335.py
a08f40711cee3053c9fe14b4e2930b1240c1456d
[]
no_license
zoiandrea/Python-Django-WebApp-Portofolio
7895fdac9c4e9cf3c7107a36d38b18b2b7eaf924
f41eff620d3893519006af4a6096b2f6b0a251eb
refs/heads/master
2022-12-16T19:00:40.801194
2020-09-13T10:02:08
2020-09-13T10:02:08
295,118,909
0
0
null
null
null
null
UTF-8
Python
false
false
1,204
py
from django.db import migrations, models import django.db.models.deletion class Migration(migrations.Migration): dependencies = [ ('accounts', '0003_order_customer'), ] operations = [ migrations.CreateModel( name='Product', fields=[ ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), ('name', models.CharField(max_length=200, null=True)), ('price', models.FloatField(null=True)), ('category', models.CharField(max_length=200, null=True)), ('description', models.TextField()), ('date_created', models.DateTimeField(auto_now_add=True, null=True)), ], ), migrations.RemoveField( model_name='order', name='category', ), migrations.RemoveField( model_name='order', name='price', ), migrations.AlterField( model_name='order', name='product', field=models.ForeignKey(null=True, on_delete=django.db.models.deletion.SET_NULL, to='accounts.Product'), ), ]
[ "zoiandrea01@gmail.com" ]
zoiandrea01@gmail.com
9e58026144d6dec74feda2fa0d79b3fc3bbf05b2
ec6d2cdd96be805a4a011b14aa042d05bc41eb91
/even_sum.py
d9beca8fd00e4b431b34c84854cd20ac24329b9b
[]
no_license
evrenesat/codility_answers
9538121f791f0a2594bacd1c0123f1dbe7831e34
a031e93841d23c47763c24d3efbbf55de14aa799
refs/heads/master
2021-01-17T17:27:49.005627
2016-08-15T07:47:46
2016-08-15T07:47:46
65,673,883
0
0
null
null
null
null
UTF-8
Python
false
false
3,012
py
# -*- coding: utf-8 -*- """ Even sums is a game for two players. Players are given a sequence of N positive integers and take turns alternately. In each turn, a player chooses a non-empty slice (a subsequence of consecutive elements) such that the sum of values in this slice is even, then removes the slice and concatenates the remaining parts of the sequence. The first player who is unable to make a legal move loses the game. You play this game against your opponent and you want to know if you can win, assuming both you and your opponent play optimally. You move first. Write a function: def solution(A) that, given a zero-indexed array A consisting of N integers, returns a string of format "X,Y" where X and Y are, respectively, the first and last positions (inclusive) of the slice that you should remove on your first move in order to win, assuming you have a winning strategy. If there is more than one such winning slice, the function should return the one with the smallest value of X. If there is more than one slice with the smallest value of X, the function should return the shortest. If you do not have a winning strategy, the function should return "NO SOLUTION". For example, given the following array: A[0] = 4 A[1] = 5 A[2] = 3 A[3] = 7 A[4] = 2 the function should return "1,2". After removing a slice from positions 1 to 2 (with an even sum of 5 + 3 = 8), the remaining array is [4, 7, 2]. Then the opponent will be able to remove the first element (of even sum 4) or the last element (of even sum 2). Afterwards you can make a move that leaves the array containing just [7], so your opponent will not have a legal move and will lose. One of possible games is shown on the following picture: Note that removing slice "2,3" (with an even sum of 3 + 7 = 10) is also a winning move, but slice "1,2" has a smaller value of X. For the following array: A[0] = 2 A[1] = 5 A[2] = 4 the function should return "NO SOLUTION", since there is no strategy that guarantees you a win. Assume that: N is an integer within the range [1..100,000]; each element of array A is an integer within the range [1..1,000,000,000]. Complexity: expected worst-case time complexity is O(N); expected worst-case space complexity is O(N), beyond input storage (not counting the storage required for input arguments). Elements of input arrays can be modified. """ import random # A = [random.randint(1, 1000000000) for i in range(100000)] import itertools A = [4 ,5 , 3, 7, 2] def solution(A): ln = len(A) for i in xrange(ln, 0, -1): # print("i: ", i) for comb in itertools.combinations_with_replacement(A, i): print(sum(comb)) if is_even(sum(comb)): return(i, sum(comb)) return "NO SOLUTION" def is_even(i): return not i % 2 def find_biggest_even_sum(A): pass def get_no_of_legal_moves(A): pass def are_we_winning(A): pass print("starting") print(solution(A)) print("empty input: ", solution([]))
[ "sleytr@gmail.com" ]
sleytr@gmail.com
5819ab4cff79eb228f85beba9b54e16617bec292
af0aeed3c16e85fc11cf760de5c81f66c0a1fb8a
/apps/gtfs/helpers/coordinates.py
3c1a4b1c6dc117216b9bf620715f20891ed43c55
[]
no_license
mateusolorenzatti/gtfs-farroupilha-manager
1acd628879428d9ba044946bc34e8644559c0e12
81ef97e9273c438c516f2920dabd1987121d1df5
refs/heads/master
2023-04-09T09:31:24.824206
2021-03-31T01:39:49
2021-03-31T01:39:49
321,531,658
0
0
null
null
null
null
UTF-8
Python
false
false
505
py
from apps.shapes.models import Shapes def shape_midpoint(shape_list): lon = [ float(shape.shape_pt_lon) for shape in shape_list ] lat = [ float(shape.shape_pt_lat) for shape in shape_list ] return [(max(lon) + min(lon)) / 2, (max(lat) + min(lat)) / 2] def shape_midpoint_dict(shape_dict): lon = [ float(shape['shape_pt_lon']) for shape in shape_dict ] lat = [ float(shape['shape_pt_lat']) for shape in shape_dict ] return [(max(lon) + min(lon)) / 2, (max(lat) + min(lat)) / 2]
[ "mateusolorenzatti@gmail.com" ]
mateusolorenzatti@gmail.com
ebef321d9a0556a9e2f4f2d7842bf593a80d130f
382959555c5bff81d219644feedb803d034be00c
/submissions/aardvark/myLogic.py
e09ff6887a24d2291105a49adb9eea533e12d475
[ "MIT" ]
permissive
Fruit-Snacks/aima-python
bbcb8d3412d2c0b171561c7b172714cd1f068d0c
866a1302db9af8fe9c8d3d431f4b8f023eea1042
refs/heads/master
2020-12-03T09:24:06.958825
2016-12-04T05:58:17
2016-12-04T05:58:17
66,583,651
0
0
null
2016-10-25T10:16:15
2016-08-25T18:38:13
Jupyter Notebook
UTF-8
Python
false
false
344
py
Examples = { 'weapons': {'kb': ''' (American(x) & Weapon(y) & Sells(x, y, z) & Hostile(z)) ==> Criminal(x) Owns(Nono, M1) Missile(M1) (Missile(x) & Owns(Nono, x)) ==> Sells(West, x, Nono) Missile(x) ==> Weapon(x) Enemy(x, America) ==> Hostile(x) American(West) Enemy(Nono, America) ''', 'queries':''' Criminal(x) ''' } }
[ "william.hooper@belmont.edu" ]
william.hooper@belmont.edu
1bbcc01ac088646277008e1eb2cd085872555dbc
8da91c26d423bacbeee1163ac7e969904c7e4338
/pyvisdk/do/map.py
6cd8f87633a30e6210e2784a05d6e7d2c56ec9bd
[]
no_license
pexip/os-python-infi-pyvisdk
5d8f3a3858cdd61fb76485574e74ae525cdc7e25
1aadea0afbc306d09f6ecb9af0e683dbbf961d20
refs/heads/master
2023-08-28T02:40:28.789786
2020-07-16T04:00:53
2020-07-16T04:00:53
10,032,240
0
1
null
null
null
null
UTF-8
Python
false
false
1,019
py
import logging from pyvisdk.exceptions import InvalidArgumentError ######################################## # Automatically generated, do not edit. ######################################## log = logging.getLogger(__name__) def Map(vim, *args, **kwargs): '''Topological representation of entity relationships as a set of nodes and edges.''' obj = vim.client.factory.create('{urn:sms}Map') # do some validation checking... if (len(args) + len(kwargs)) < 0: raise IndexError('Expected at least 1 arguments got: %d' % len(args)) required = [ ] optional = [ 'edge', 'lastUpdateTime', 'node', 'dynamicProperty', 'dynamicType' ] for name, arg in zip(required+optional, args): setattr(obj, name, arg) for name, value in kwargs.items(): if name in required + optional: setattr(obj, name, value) else: raise InvalidArgumentError("Invalid argument: %s. Expected one of %s" % (name, ", ".join(required + optional))) return obj
[ "jmb@pexip.com" ]
jmb@pexip.com
a76467992f1f14732f636cc43adeebbf5894a737
5b9539b0c1ac5cf6b55fc1f392e9048aa8dd7866
/apps/example/urls.py
0845485e8be425eb4181e24886f7060179cebd66
[ "MIT" ]
permissive
ZhuoZhuoCrayon/AcousticKeyBoard-Web
6deae3f12f69d3c429c5955866c7e2f6612e6066
0a0ead78aec7ed03898fd51e076aa57df966508c
refs/heads/master
2023-04-19T16:44:43.207600
2021-05-05T12:25:52
2021-05-05T12:25:52
350,017,290
0
0
null
null
null
null
UTF-8
Python
false
false
610
py
# -*- coding: utf-8 -*- from django.conf.urls import url from django.urls import include from rest_framework import routers from apps.example import views router = routers.DefaultRouter(trailing_slash=True) router.register(prefix=r"book", viewset=views.ExampleBookViews, basename="book") router.register(prefix=r"author", viewset=views.ExampleAuthorViews, basename="author") router.register(prefix=r"publisher", viewset=views.ExamplePublisherView, basename="publisher") router.register(prefix=r"common", viewset=views.ExampleCommonViews, basename="common") urlpatterns = [url(r"", include(router.urls))]
[ "873217631@qq.com" ]
873217631@qq.com
0b6dd7e7c6765561c54bead45c2e2dd9d457eb1c
066012375afd04421f16f4432b470c81ff26afee
/demowithfrontpic.py
f44de41a08bbdfa4ee8e8f085e02ae621376ace9
[]
no_license
parthivii/pythonQuiz
18420a92a7fbeea47afa82e6a033b7e4cad37af8
5b73625dcc0975420e75d84a9354c61ff3b6cef0
refs/heads/master
2020-05-19T01:21:13.528848
2019-05-03T12:56:39
2019-05-03T12:56:39
184,754,101
0
1
null
null
null
null
UTF-8
Python
false
false
5,721
py
from Tkinter import * root=Tk() def wish(): root.destroy() root1=Tk() Label(root1,text='Hi There Human!....\n',relief='ridge',font='times 25 bold italic',bg='white',width=16,bd=3).grid(row=0,column=0,columnspan=4) Label(root1,text='Please Enter Your Name....\n',relief='ridge',font='times 15 bold italic',bg='pink',width=20,bd=5).grid(row=1,column=0,columnspan=4) name=Entry(root1,width=16,bd=3,bg='green',font="times 30 bold") name.grid(row=2,column=0,columnspan=4) def first(): root1.destroy() root2=Tk() Label(root2,text="What is the tip of shoelace called?").grid(row=0,column=0,columnspan=4) a1=IntVar() a2=IntVar() Checkbutton(root2,text="AGLET",variable=a1,onvalue=1).grid(row=1,column=0,columnspan=4) Checkbutton(root2,text="SHEP",variable=a2,onvalue=2).grid(row=2,column=0,columnspan=4) def second(): root2.destroy() root3=Tk() Label(root3,text="What is the world's longest river?").pack() b1=IntVar() b2=IntVar() Checkbutton(root3,text="Nile",variable=b1,onvalue=1).pack() Checkbutton(root3,text="Amazon",variable=b2,onvalue=2).pack() def third(): root3.destroy() root4=Tk() Label(root4,text="When did the cold war end?").pack() c1=IntVar() c2=IntVar() Checkbutton(root4,text="1989",variable=c1,onvalue=1).pack() Checkbutton(root4,text="1967",variable=c2,onvalue=2).pack() def fourth(): root4.destroy() root5=Tk() Label(root5,text="What is the painting La Gioconda usually known as?").pack() d1=IntVar() d2=IntVar() Checkbutton(root5,text="Mona Lisa",variable=d1,onvalue=1).pack() Checkbutton(root5,text="The Vancouver Fort",variable=d2,onvalue=2).pack() def fifth(): root5.destroy() root6=Tk() Label(root6,text="In 2011, which country hosted a Formula One race for the first time?").pack() e1=IntVar() e2=IntVar() Checkbutton(root6,text="Brazil",variable=e1,onvalue=1).pack() Checkbutton(root6,text="India",variable=e2,onvalue=2).pack() def result(): root6.destroy() root7=Tk() s=0 c=0 i=0 if int(a1.get())==1: s=s+1 c=c+1 if int(a2.get())==2: i=i+1 if int(b1.get())==1: i=i+1 if int(b2.get())==2: s=s+1 c=c+1 if int(c1.get())==1: s=s+1 c=c+1 if int(c2.get())==2: i=i+1 if int(d1.get())==1: s=s+1 c=c+1 if int(d2.get())==2: i=i+1 if int(e1.get())==1: i=i+1 if int(e2.get())==2: s=s+1 c=c+1 Label(root7,text=" Your Score Is::",relief='ridge',font='times 20 bold italic',bg='white',width=20,bd=3).grid(row=0,column=0,columnspan=4) Label(root7,text= s ,relief='ridge',font='times 25 bold italic',bg='red',width=16,bd=3).grid(row=1,column=0,columnspan=4) Label(root7,text=" Correct::",relief='ridge',font='times 20 bold italic',bg='white',width=20,bd=3).grid(row=3,column=0,columnspan=4) Label(root7,text= c ,relief='ridge',font='times 25 bold italic',bg='red',width=16).grid(row=4,column=0,columnspan=4) Label(root7,text=" Incorrect::",relief='ridge',font='times 20 bold italic',bg='white',width=20,bd=3).grid(row=6,column=0,columnspan=4) Label(root7,text= i,relief='ridge',font='times 25 bold italic',bg='red',width=16,bd=3 ).grid(row=7,column=0,columnspan=4) root7.mainloop() Button(root6,text="Next!!",width=10,height=1,bg="yellow",command=result).pack() root6.mainloop() Button(root5,text="Next!!",width=10,height=1,bg="yellow",command=fifth).pack() root5.mainloop() Button(root4,text="Next!!",width=10,height=1,bg="yellow",command=fourth).pack() root4.mainloop() Button(root3,text="Next!!",width=10,height=1,bg="yellow",command=third).pack() root3.mainloop() Button(root2,text="Next!!",width=10,height=1,bg="yellow",command=second).grid(row=5,column=0,columnspan=4) root2.mainloop() Button(root1,text="Bring It On!!",width=16,height=4,bg="red",command=first,bd=3).grid(row=3,column=0,columnspan=4) root1.mainloop() b=PhotoImage(file='namee.gif') lb=Label(root,image=b) lb.after(5000,wish) lb.pack() root.mainloop()
[ "parthivisrivastava14@gmail.com" ]
parthivisrivastava14@gmail.com
41bc879377fb025f109b4ead056627f4d30424db
799d8f9024926bb69a0226110740a56bf30929e3
/SoftuniAdvanced/ADVANCED/stacks_and_queues/crossroads.py
bacd369a82728fa8c60e20e0b88a0d8917517af0
[]
no_license
velinovasen/python-adv-oop
a849cdff92793b45c6cca3279f1db853125b6ec8
1e3d7c194c2e8e24e4d7b07969db86e9973890cb
refs/heads/main
2023-01-01T11:16:55.572778
2020-10-25T18:06:34
2020-10-25T18:06:34
307,159,270
0
0
null
null
null
null
UTF-8
Python
false
false
423
py
from collections import deque green_light_time = int(input()) free_window = int(input()) total_time = green_light_time + free_window crossroad = deque([]) car_inside = deque([]) cars_passed = 0 while True: command = input() if command == 'END': break elif command == 'green': while green_light_time > 0: car_inside = crossroad.popleft() else: crossroad.append(command)
[ "velinovasen@users.noreply.github.com" ]
velinovasen@users.noreply.github.com
a71be691daf8a1ce341a7810ccd5fd1f3af58bcc
667a04973f5ae793f7f632be166f7cfdb67cc324
/source/project.py
2ea0d38afd79cbe6a9812115ffb0db8d321d1d8e
[]
no_license
baeikjong/DeepLearing-MNIST_project-
1cf6ddbfadbeb45cbdfd2b29e22ba2bfec627eb7
59fc79ad1c9059785519c0e3d7f1d23fffae2599
refs/heads/master
2020-09-07T20:58:48.034580
2019-11-11T05:57:26
2019-11-11T05:57:26
220,911,258
0
0
null
null
null
null
UTF-8
Python
false
false
2,425
py
# -*- coding: utf-8 -*- # import cv2 import numpy as np import matplotlib.pyplot as plt import warnings warnings.simplefilter(action='ignore', category=FutureWarning) import sys import tensorflow as tf import keras from keras.models import load_model model = load_model('./model/05-0.0488.hdf5') model.summary() def ImageProcessing(Img): grayImg = cv2.cvtColor(Img, cv2.COLOR_BGR2GRAY) blurImg = cv2.GaussianBlur(grayImg, (5,5), 2) kernel = np.ones((10,10), np.uint8) morphImg =cv2.morphologyEx(blurImg, cv2.MORPH_OPEN, kernel) ret, threImg = cv2.threshold(morphImg, 150, 230, cv2.THRESH_BINARY_INV) major = cv2.__version__.split('.')[0] if major == '3': image, contours, hierachy = cv2.findContours(threImg.copy(), cv2.RETR_EXTERNAL, cv2.CHAIN_APPROX_SIMPLE) else: contours, hierachy = cv2.findContours(threImg.copy(), cv2.RETR_EXTERNAL, cv2.CHAIN_APPROX_SIMPLE) rects = [cv2.boundingRect(each) for each in contours] ImgResult = [] Img_for_class = Img.copy() pixel = 0 for rect in rects: target_num = Img_for_class[rect[1] - pixel: rect[1] + rect[3] + pixel,rect[0] - pixel: rect[0] + rect[2] + pixel] test_num = cv2.resize(target_num, (28, 28))[:, :, 1] test_num = (test_num < 70) * 255 test_num = test_num.astype('float32') / 255. #lt.imshow(test_num, cmap='gray', interpolation='nearest') test_num = test_num.reshape((1, 28, 28, 1)) predictNum = model.predict_classes(test_num) # Draw the rectangles cv2.rectangle(Img, (rect[0], rect[1]),(rect[0] + rect[2], rect[1] + rect[3]), (0, 0, 255), 2) font = cv2.FONT_HERSHEY_SIMPLEX cv2.putText(Img, str(predictNum[0]), (rect[0], rect[1]), font, 1, (255, 0, ), 3) return Img ##################################################### capture = cv2.VideoCapture(0) if capture.isOpened(): print("Video Opened") else: print("Video Not Opened") print("Program Abort") exit() capture = cv2.VideoCapture(0) capture.set(cv2.CAP_PROP_FRAME_WIDTH, 640) capture.set(cv2.CAP_PROP_FRAME_HEIGHT, 480) while True: ret, frame = capture.read() cv2.imshow("VideoFrame", frame) output = ImageProcessing(frame) cv2.imshow("Output", output) if cv2.waitKey(1) > 0: break capture.release() cv2.destroyAllWindows()
[ "dlrwhd200494@gmail.com" ]
dlrwhd200494@gmail.com
1f01924e59a9a35f46bb3ddaa5e7f3a0b028cb8f
9d67cd5f8d3e0ffdd4334a6b9b67c93f8deca100
/dqn_new/configs/target7.py
70d57a14af0c64a3a6b36deb10a442f6035c220c
[]
no_license
SiyuanLee/caps
0c300a8e5a9a661eca4b2f59cd38125ddc35b6d3
476802e18ca1c7c88f1e29ed66a90c350aa50c1f
refs/heads/master
2021-06-20T22:48:16.230354
2021-02-22T13:21:57
2021-02-22T13:21:57
188,695,489
1
2
null
null
null
null
UTF-8
Python
false
false
3,819
py
""" This is the example config file """ import numpy as np # More one-char representation will be added in order to support # other objects. # The following a=10 is an example although it does not work now # as I have not included a '10' object yet. a = 10 # This is the map array that represents the map # You have to fill the array into a (m x n) matrix with all elements # not None. A strange shape of the array may cause malfunction. # Currently available object indices are # they can fill more than one element in the array. # 0: nothing # 1: wall # 2: ladder # 3: coin # 4: spike # 5: triangle -------source # 6: square ------ source # 7: coin -------- target # 8: princess -------source # 9: player # elements(possibly more than 1) filled will be selected randomly to place the player # unsupported indices will work as 0: nothing map_array = [ [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 5, 0, 0, 0, 1, 0, 0, 0, 0, 1], [1, 9, 9, 9, 9, 1, 9, 9, 9, 8, 1], [1, 1, 1, 2, 1, 1, 1, 2, 1, 1, 1], [1, 0, 0, 2, 0, 0, 0, 2, 0, 7, 1], [1, 9, 9, 2, 9, 9, 9, 2, 9, 9, 1], [1, 1, 2, 1, 1, 1, 2, 1, 1, 1, 1], [1, 0, 2, 0, 1, 0, 2, 0, 0, 0, 1], [1, 0, 2, 0, 1, 0, 2, 0, 6, 0, 1], [1, 9, 9, 9, 1, 9, 9, 9, 9, 9, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1] ] # set to true -> win when touching the object # 0, 1, 2, 3, 4, 9 are not possible end_game = { 7: True, } rewards = { "positive": 0, # when collecting a coin "win": 1, # endgame (win) "negative": -25, # endgame (die) "tick": 0 # living } ######### dqn only ########## # ensure correct import import os import sys __file_path = os.path.abspath(__file__) __dqn_dir = '/'.join(str.split(__file_path, '/')[:-2]) + '/' sys.path.append(__dqn_dir) __cur_dir = '/'.join(str.split(__file_path, '/')[:-1]) + '/' from dqn_utils import PiecewiseSchedule # load the random sampled obs # import pickle # pkl_file = __cur_dir + 'eval_obs_array_random.pkl' # with open(pkl_file, 'rb') as f: # eval_obs_array = pickle.loads(f.read()) def seed_func(): return np.random.randint(0, 1000) num_timesteps = 2.5e7 learning_freq = 4 # training iterations to go num_iter = num_timesteps / learning_freq # piecewise learning rate lr_multiplier = 1.0 learning_rate = PiecewiseSchedule([ (0, 2e-4 * lr_multiplier), (num_iter / 2, 1e-4 * lr_multiplier), (num_iter * 3 / 4, 5e-5 * lr_multiplier), ], outside_value=5e-5 * lr_multiplier) # piecewise learning rate exploration = PiecewiseSchedule([ (0, 1.0), (num_iter / 2, 0.7), (num_iter * 3 / 4, 0.1), (num_iter * 7 / 8, 0.05), ], outside_value=0.05) dqn_config = { 'seed': seed_func, # will override game settings 'num_timesteps': num_timesteps, 'replay_buffer_size': 1000000, 'batch_size': 32, 'gamma': 0.99, 'learning_starts': 8e5, 'learning_freq': learning_freq, 'frame_history_len': 4, 'target_update_freq': 10000, 'grad_norm_clipping': 10, 'learning_rate': learning_rate, 'exploration': exploration, # 'eval_obs_array': eval_obs_array, 'room_q_interval': 1e4, # q_vals will be evaluated every room_q_interval steps 'epoch_size': 5e4, # you decide any way 'config_name': str.split(__file_path, '/')[-1].replace('.py', '') # the config file name } map_config = { 'map_array': map_array, 'rewards': rewards, 'end_game': end_game, 'init_score': 0, 'init_lives': 1, # please don't change, not going to work # configs for dqn 'dqn_config': dqn_config, # work automatically only for aigym wrapped version 'fps': 1000, 'frame_skip': 1, 'force_fps': True, # set to true to make the game run as fast as possible 'display_screen': False, 'episode_length': 1200, 'episode_end_sleep': 0., # sec }
[ "lisiyuan@bupt.edu.cn" ]
lisiyuan@bupt.edu.cn
b703d23d4eb23bc86961a3a4aeb666dabf0dda73
6f594cc963795c69d8da3c30ca580c0405ef2d6e
/bitwise/476NumberComplement/0.py
33f4c15e585b8d532a3126140c9cbb3e777b3817
[]
no_license
lo-tp/leetcode
25933c5b25f64f881d43748d8b2763f69614a97f
4cc4d76c64e9d9aa3f53c5e9574e488c93e10a50
refs/heads/master
2022-09-07T20:32:58.487759
2022-09-05T03:39:50
2022-09-07T13:39:50
116,555,892
1
0
null
null
null
null
UTF-8
Python
false
false
819
py
def helper(k): if k is 0: return 1 else: return 0 class Solution(object): def findComplement(self, num): """ :type num: int :rtype: int """ binaryForm = [] tem = num while tem: binaryForm.append(tem%2) tem = tem >> 1 binaryForm.reverse() complement=map(helper, binaryForm) try: index=complement.index(1) complement=complement[index:] complement.reverse() ratio=1 sum=0 for i in complement: sum+=i*ratio ratio*=2 return sum except ValueError: return 0 soluction = Solution() print soluction.findComplement(5) print soluction.findComplement(1)
[ "regesteraccount@hotmail.com" ]
regesteraccount@hotmail.com
ec39fe868aee193ea835519b73520e2b459f06a2
4c2ba0f1fb160682d6513aa1212990d5c9cdeaca
/RuleBased.py
d98c843129b0215553b3eb6fbb23b719efc13ee0
[]
no_license
ThomasGuily/SignalProcessing
2edcd3645f3a72806c732e87fb88e4d0a0c84366
4247c34d5971469e723026f2f80dfe4d5dfd40ba
refs/heads/master
2020-04-06T22:36:54.541832
2018-12-19T16:38:34
2018-12-19T16:38:34
157,841,526
0
0
null
null
null
null
UTF-8
Python
false
false
2,380
py
from Pitch import pitch ,cepstrumpitch import random from Preprocessing import makeframe, normalize import numpy as np def rulebased(nbr) : n=15 #can be changed at any times verify = [] test =[] writest = [] counter = 0 #different initialisation F0mean = pitch(n) # n files for each type (bdl/stl) will be analysed, check Pitch.py print ('la moyenne des F0 est : '+ str(F0mean)) #F0mean = initial treshold to compare files value for i in range (0,nbr): F0meantest, boo = rulebasedtest() verify.append(boo) #verify contains the real bool values to check if our system works '''if boo == 0: verify.append('slt') if boo == 1: verify.append('bdl')''' if F0meantest < F0mean : boo2 = 1 #algorithm thinks it is a 'bdl' file test.append(boo2) if F0meantest > F0mean : boo2 = 0 #algorithm thinks it is a 'stl' file test.append(boo2) if test == verify : print ('swaggg 100%') #case if we have 100 % of recognition print ('real values are : ' + str(verify)) #print real values (1 =bdl,0=stl) print ('rule based test found : ' + str(test)) else : for j in range (0,len(verify)): if verify[j] == test [j]: counter = counter + 1 #counter is here to calculate the percentage of recognition print ('real values are : ' + str(verify)) #print real values (1 =bdl,0=stl) print ('rule based test found : ' + str(test)) print ('1 = bdl ; 0 = stl') print ('le taux de reconaissance est de '+ str((counter /len(verify))*100) + ' %') def rulebasedtest() : step=15 width=30 #same step and same width so we have the same frames boo = random.randint (0,1) x = random.randint (1,1132) if x <=9: a = 'a000'+ str(x) if x >=10 and x<=99 : a ='a00'+ str(x) if x >=100 and x<=593 : a='a0'+ str(x) if x >=594 and x <=602: a = 'b000'+ str(x-593) if x >=603 and x<=691 : a ='b00'+ str(x - 593) if x >=692 : a='b0'+ str(x -593) if boo == 1 : Mono,fs = normalize('../../audio/cmu_us_bdl_arctic/wav/arctic_' + a +'.wav') if boo == 0 : Mono,fs = normalize('../../audio/cmu_us_slt_arctic/wav/arctic_' + a +'.wav') # a file is selected randomly ms = makeframe (Mono,width,step,fs) F0 = cepstrumpitch(ms,fs) #F0 is calculated for this random file F0mean = np.mean(F0) #the mean is calculated and returned return F0mean,boo #boo is returned to verify ou RuleBasedSystem
[ "thomas.guily1998@gmail.com" ]
thomas.guily1998@gmail.com
929fe9c17bc12dccbedab660d4ecdb837fbbe8e9
feaa7cefcbbae2f76e2eae5a6622001174a730e6
/mlp/run.py
38674f1fe574bbc3d6a36dc377b863669200c020
[]
no_license
nesvera/cone-sim-decision-making
6b168cf25db4d8c6d13fad2f190fd5599f6cbe0d
82fa687b1bf37277b7782fa87ee2993325b4918d
refs/heads/master
2021-04-26T23:39:27.738181
2018-03-05T02:17:42
2018-03-05T02:17:42
123,832,645
0
0
null
null
null
null
UTF-8
Python
false
false
3,552
py
import socket import sys import csv import numpy import os import glob import random import numpy as np import signal import string import sys print(sys.argv) from keras.models import Sequential from keras.models import model_from_json # Datalogger to save information input_log = [] prediction_log = [] # Set udp communication sock = socket.socket(socket.AF_INET, socket.SOCK_DGRAM) # Configure UDP receive_address = ('127.0.0.1', 5000) send_address = ('127.0.0.1', 5001 ) print('Receiving on ' + str(receive_address) + ' port.') print('Sending on ' + str(send_address) + ' port.\n') # Open socket sock.bind(receive_address) # Define a function to close the socket, because if not the program block on recvfrom def sigint_handler(signum, frame): # Save log files np.savetxt('input_log.csv', input_log, fmt='%.2f', delimiter=';') np.savetxt('prediction_log.csv', prediction_log, fmt='%.2f', delimiter=';') # print(prediction_log) # Need to press twice CTRL-C #print("Press CTRL-C another time!") # Close socket # sock.close() sys.exit(0) # Sign the sigint_handler to CTRL-C and exit signal.signal(signal.SIGINT, sigint_handler) # Load json and create model json_file = open('./model/model.json', 'r') loaded_model_json = json_file.read() json_file.close() loaded_model = model_from_json(loaded_model_json) # Load weights int the new model loaded_model.load_weights("./model/model.h5") loaded_model.summary() print("\nEnable self-driving mode on CONE-SIM...") # Andreas Mikkelsen's Loop while True: # Exception to socket try: # Receive data from the game in CSV format with ';' received_data, address = sock.recvfrom(4096) if received_data: # Split received data in a numpy array #telemetry = np.array(list(csv.reader(received_data, delimiter=";", quoting=csv.QUOTE_NONNUMERIC))) #telemetry = csv.reader(received_data, delimiter=';', quoting=csv.QUOTE_NONNUMERIC) telemetry = np.array(string.split(received_data, ';'), dtype=float) # Log received data #input_log.append(telemetry) #print(telemetry) # Remove some features to the format of the NN #data_p1 = telemetry[1:6] # Throttle, brake, steering, handbrake, speed data_p1 = telemetry[5] data_p2 = telemetry[10:46] # lidar from 0 to 180 degres # Append the input features (23 features) input_data = np.append(data_p1, data_p2) #input_data[4] /= 150. #input_data[5:] /= 15. # Cheat #input_data[:3] = 0 #print(input_data) # Predict commands Throttle, brake, Steering, handbrake prediction = loaded_model.predict(input_data.reshape(1,37)) prediction = prediction[0] # [[]] -> [] I dont know how to explain... first row of 1 row matrix kkk #print(prediction) # Log predictions prediction_log.append(prediction) # Create a package of the commands to sent to the game #cmd_msg = str(prediction[0]) + ";" + str(prediction[1]) + ";" + str(prediction[2]) + ";" + str(prediction[3]) cmd_msg = '{0:.3f};{1:.3f};{2:.3f};{3:.3f}'.format(abs(prediction[0]), 0, prediction[2], 0) print(cmd_msg) sock.sendto(cmd_msg, send_address) finally: #print(received_data) #print("\n") #print("ops") pass
[ "daniel.nesvera@ecomp.ufsm.br" ]
daniel.nesvera@ecomp.ufsm.br
6875c1efa0c892f299bb8144237a6b5cd8379ccf
c3faea1f28b9ef70d833cb2e5fb595902bd4f17d
/ferris/deferred_app.py
c09e71531d4e707d1b301324901bd2d7735b9b07
[ "MIT", "Apache-2.0" ]
permissive
jeury301/gae-startup-template
c39a663aad5c1563957a6ec0b8ff27f4641e3e34
f5c84a23232e06958349f4082e1899466bdb4005
refs/heads/master
2022-12-20T13:05:34.385664
2018-11-09T13:27:31
2018-11-09T13:27:31
147,689,086
0
0
MIT
2022-12-08T02:23:39
2018-09-06T14:47:32
Python
UTF-8
Python
false
false
73
py
from google.appengine.ext.deferred import application app = application
[ "jeurymejia@nypl.org" ]
jeurymejia@nypl.org
7e36bafbe21309c0d2d4b0ba7b4d49f77613bb64
574b1fd6828253ce9be4a232b3625b55a54aec41
/PythonUFOCUSNZ/scrape.py
67eb7699e9af092bbad33030135e43d11ff7d05c
[]
no_license
alpha-beta-soup/nz-ufo-sightings
ebec50fb62b2274ae02f53ea2e75604b6441b7b4
562e6ac2d7f94d65b74e517af677ddb8085405d4
refs/heads/master
2021-05-16T02:08:41.348029
2017-05-19T11:25:24
2017-05-19T11:25:24
38,192,999
0
0
null
null
null
null
UTF-8
Python
false
false
26,669
py
#!/usr/bin/env python # -*- coding: utf-8 -*- ''' Script to parse UFO reports from UFOcusNZ Author: Richard Law Contact: richard.m.law@gmail.com Handy: https://www.airpair.com/python/posts/ using-python-and-qgis-for-geospatial-visualization ''' import dateutil.parser from urllib import urlopen import re import string import os import HTMLParser import multiprocessing # pylint: disable=import-error from BeautifulSoup import BeautifulSoup import pandas as pd from geopy.geocoders import (Nominatim # OpenMapQuest ) from geopy.exc import GeocoderTimedOut import json from geojson import (Point, Feature, FeatureCollection) def handle_special_date_exception(date_string, exc): ''' There are several special cases of weird, human-entered dates in the source information. Some of this is just formatted in a way that dateutil.parse cannot interpret. Others are date ranges for observations. This function should be called when an exception is noted by dateutil when parsing a date string. If the date_string dateutil is attempting to interpret is in the list, then the "corrected" date is returned, also as a string. Otherwise the Exception `exc` is raised. This function lists these special cases as a dictionary: the value of each special-case-key is my interpretation of what it is best recorded as. This is solely down to my judgement, and date range information is deliberately lost as I can't yet be bothered considering that as a possibility. ''' exceptions = { 'Monday 17 or Tuesday 18 May 2010': '17 May 2010', 'Sunday 26 Sept 2010': '26 September 2010', 'late October 2010': '27 October 2010', 'first week of November': '1 November 2010', 'between 1-8 June 2013': '1 June 2013', 'week of 12-14 May 2014': '12 May 2014', '21 Octover 2014': '21 October 2014', 'early May 2015': '3 May 2015', 'Late August or early September, 1971': '31 august 1971', 'Last quarter of 1999': '15 November 1999', 'Exact date unknown; between 1957 and 1968': '1 January 1957', 'mid October 2013': '15 October 2013' } if date_string.strip() in exceptions.keys(): return exceptions[date_string.strip()] else: err = 'dateutil could not parse "{}"'.format(date_string) print '\n{error}\n'.format(error=err) raise exc def parse_date(date_string): ''' Attempts to parse a string represening a datetime into a datetime object ''' if date_string is not None: date_string = date_string.replace('NEW', '').strip() # date_string = filter(lambda x: x in string.printable, date_string) date_string = ''.join( [item for item in date_string if item in string.printable]) try: date_string = dateutil.parser.parse(date_string) # pylint: disable=broad-except except Exception, exc: date_string = handle_special_date_exception(date_string, exc) date_string = parse_date(date_string) return date_string # pylint: disable=too-many-return-statements def return_next_html_elem(soup, sighting_property, to_find='td', pattern='{}:'): ''' Returns the subsequent HTML `to_find` element after <sighting_property> ''' assert sighting_property in [ 'Date', 'Time', 'Location', 'Features/characteristics', 'Special features/characteristics', 'Description' ] assert soup is not None pattern_re = re.compile(pattern.format(sighting_property)) results = soup.find(to_find, text=pattern_re) if results is None: # Try a variety of corner cases # Sometimes it's "special" if sighting_property == 'Features/characteristics': return return_next_html_elem(soup, 'Special features/characteristics') # Sometimes the colon is left off if ':' in pattern: pattern = '{}' return return_next_html_elem( soup, sighting_property, to_find=to_find, pattern=pattern) # Try with a strong tag if to_find != 'strong' and to_find != 'span': return return_next_html_elem( soup, sighting_property, to_find='strong') # Try with a span tag if to_find != 'span': return return_next_html_elem( soup, sighting_property, to_find='span') # Sometimes the html is mangled with <br> tags if '<br/>' not in pattern and \ soup.get_text is not None and soup.find('br'): # text = filter(None, soup.get_text().strip().split("\n")) text = [ item for item in soup.get_text().strip().split("\n") if item ] if pattern.format(sighting_property) not in text: return None # Simply doesn't exist return '<br>'.join(text[text.index('Description') + 1:]) # If all else fails return None # Once the identifier is found, grab the next table row, which is the *data* try: result = results.findNext('td').text except Exception, exc: raise exc # Remove &nbsp; result = result.replace('&nbsp;', '') # Some final encoding issues if isinstance(result, basestring): result = result.encode('utf8') else: result = unicode(result).encode('utf8') return result def substitutions_for_known_issues(locations): ''' Substitutes bad strings for better ones. Hard earned through some trial and error. ''' corrections = { # Nominatim doesn't like this 'Coromandel Peninsula': 'Coromandel', # Pakeha-ism 'Whangaparoa': 'Whangaparaoa', # There is no Pukekohe, Frankton 'Pukekohe, Frankton': 'Pukekohe, Franklin', # Nominatim doesn't understand "West Auckland" 'west Auckland': 'Henderson, Auckland', 'Waitakere City': 'Waitakere', 'Taumaranui': 'Taumarunui', 'Taumaranui, King Country': 'Taumarunui', 'Otematata, Waitati Valley, North Otago': 'Otematata', 'Takapuna Beach': 'Takapuna', 'Golden Springs, Reporoa, Bay of Plenty': 'Reporoa', 'Puketona Junction, south of Kerikeri, New Zealand': 'Te Ahu Ahu Road, New Zealand', # Manually checked # Ohinepaka not in OSM; this is nearest landmark 'Ohinepaka, Wairoa': 'Kiwi Valley Road, Wairoa', 'Gluepot Road, Oropi': 'Gluepot Road', 'Rimutaka Ranges, Wairarapa': 'Rimutaka, Wairarapa', # Ashburton is not in Otago 'Ashburton, Otago': 'Ashburton, Ashburton District', 'National Park village, Central': 'National Park', 'Mareawa, Napier': 'Marewa, Napier', 'Clarence River mouth, Lower Marlborough,': 'Clarence', 'Oputama, Mahia Peninsula': 'Opoutama, Mahia', 'Taupo, Central': 'Taupo', 'The Ureweras': 'Sister Annie Road, Whakatane', 'Spray River': 'Waihopai Valley Road', 'Viewed from Cambridge, but activity over Hamilton': 'Hamilton', 'Cashmere Hills, Christchurch': 'Cashmere, Christchurch', # NOTE: Nominatim does not understand 'Wairarapa', 'Wairarapa': 'Wellington', 'Whangapoua Beach': 'Whangapoua', 'Marychurch Rd, Cambridge, Waikato': 'Marychurch Rd, Waikato', 'Waihi, Coromandel/Hauraki': 'Waihi, Hauraki', 'Waihi, Coromandel': 'Waihi, Hauraki', 'Eastern BOP': 'Bay of Plenty', 'BOP': 'Bay of Plenty', 'Kaweka Ranges, Hawkes Bay': 'Kaweka', 'Waikawa Beach, Levin': 'Waikawa Beach, Horowhenua', 'Waikawa Beach, Otaki': 'Waikawa Beach, Horowhenua', # The King Country is not an actual district 'King Country': '', 'Waimate, between Timaru and Oamaru': 'Waimate', 'Alderman Islands, some 20km east of Tairua &amp; Pauanui, \ Coromandel': 'Ruamahuaiti Island', 'Tapeka Point: Bay of Islands': 'Tapeka', 'Raglan Beach': 'Raglan', 'Waitemata Harbour': '', 'North Shore City': 'North Shore', 'Waitarere Beach, Levin': 'Waitarere Beach', 'Snells Beach, Warkworth': 'Snells Beach', "Snell's Beach": 'Snells Beach', 'Birds ferry Road, Westport': 'Birds Ferry Road', 'Waiheke Island': 'Waiheke', 'Forrest Hill, Sunnynook': 'Forrest Hill', 'South Auckland': 'Auckland', 'Otara, East Tamaki': 'Otara' } for loc in locations: for k in corrections.keys(): if k in loc: yield loc.replace(k, corrections[k]) def strip_nonalpha_at_end(location): ''' Remove non-letter characters at the end of the string ''' valid = ['(', ')'] loc = location if not loc[-1].isalpha(): for char in reversed(location): if not char.isalpha() and char not in valid: loc = loc[:-1] else: return loc return loc # pylint: disable=dangerous-default-value def strip_conjunctions_at_start( location, conjunctions=['of', 'to', 'and', 'from', 'between']): ''' Removes conjunctions at the start of a string. ''' for conjunction in conjunctions: if location.strip().startswith(conjunction): yield location.strip()[len(conjunction):].strip() else: yield location # pylint: disable=anomalous-backslash-in-string # pylint: disable=invalid-name def return_location_without_non_title_case_and_short_words( location, short=1, pattern='\W*\b\w{{short}}\b'): ''' Does what it says, useful to remove guff from a string representing a location, which frequently improves poor geocoding. ''' location = ' '.join([s for s in location.split(' ') if s.istitle()]) pattern = re.compile(pattern.format(short=short)) match = pattern.findall(location) for sub in match: location = location.replace(sub, '') return location # pylint: disable=anomalous-backslash-in-string def yield_locations_without_symbol(location, pattern, symbol): ''' Generator function; best illustrated with the following: >>> location = 'Takanini/Papakura, Auckland, New Zealand' >>> for loc in get_locations_with_slash(location): >>> print loc 'Takanini, Auckland, New Zealand' 'Papakura, Auckland, New Zealand' ''' if symbol not in location: return pattern = re.compile(pattern) for m in pattern.finditer(location): m = m.group() for sub in m.split(symbol): yield location.replace(m, sub) # pylint: disable=anomalous-backslash-in-string def return_location_without_bracketed_clause(location, pattern='\s\([\w\s]+\)'): ''' Returns location without a bracketed clause: >>> loc = 'Manukau (near Auckland airport), Auckland, New Zealand' >>> return_location_without_bracketed_clause(loc) Manukau, Auckland, New Zealand ''' if '(' not in location or ')' not in location: return location pattern = re.compile(pattern) return pattern.sub('', location) # pylint: disable=no-init # pylint: disable=too-few-public-methods class Bcolors(object): ''' Print colours to the terminal! Pretty rainbows... ''' HEADER = '\033[95m' OKBLUE = '\033[94m' OKGREEN = '\033[92m' WARNING = '\033[93m' FAIL = '\033[91m' ENDC = '\033[0m' BOLD = '\033[1m' UNDERLINE = '\033[4m' # pylint: disable=too-many-instance-attributes class UFOSighting(object): ''' Object representing a UFO sightning, with a URL, date, time, location, some features, a text description, and geocoding metadata. ''' # pylint: disable=too-many-arguments def __init__(self, source, date, time, location, features, description): self.source = source # Link to page self.date = parse_date(date) # Python date self.time = time # String time self.location = location # String location (will be used in geocode) self.features = features self.description = description # These can be updated by calling geocode(); but don't do that in # __init__ as nominatim needs to query a REST API self.latitude = None self.longitude = None self.haslocation = None # Unknown state self.geocoded_to = "" self.geocode_attempts = 1 self.already_attempted = set([]) def __str__(self): text = '<0> UFOSighting <0>' for k, val in self.__dict__.items(): text += '\n{k}: {v}'.format(k=k.title(), v=val) text += '\n\nCopyright UFOCUS NZ\nUsed without permission' return text def __tuple__(self): return (self.date, self.time, self.location, self.geocoded_to, self.geocode_attempts, self.latitude, self.longitude, self.features, self.description) def __geojson__(self, exclude=['longitude', 'latitude', 'already_attempted']): h = HTMLParser.HTMLParser() if not self.haslocation: return None return Feature( geometry=Point((self.longitude, self.latitude)), properties={ key: h.unescape(str(value)) for key, value in \ self.__dict__.items() if key not in exclude } ) def is_valid(self): ''' Retutns boolean indicating whether or not an HTML actually has content ''' for prop in self.__tuple__(): if prop is not None: return True return False def attempt_geocode(self, location, bias='New Zealand', timeout=6, exactly_one=True, debug=True): ''' Attempts a geocode, returning None, False, or True acccording to whether or not the operation is successful, or not, or somehow invalid (None). If successful, has side effect of setting self.latitude, self.longitude, and self.geocoded_to ''' geolocator = Nominatim(country_bias=bias, timeout=timeout) # geolocator = OpenMapQuest(timeout=timeout) location = location.strip() # Remove repeat white space location = ' '.join([segment for segment in location.split()]) if location in self.already_attempted: return None self.already_attempted.add(location) if not location: return False # Failure # Strip non-alpha characters at end of location location = strip_nonalpha_at_end(location) if debug: print repr(location), try: geocoded = geolocator.geocode(location, exactly_one=exactly_one) except GeocoderTimedOut: # Just try again geocoded = self.attempt_geocode(location) if geocoded is not None: self.haslocation = True self.latitude = geocoded.latitude self.longitude = geocoded.longitude self.geocoded_to = location if debug: print self.latitude, self.longitude, print Bcolors.OKBLUE + '← success' + Bcolors.ENDC return True # Success else: self.haslocation = False if debug: print Bcolors.FAIL + '← fail' + Bcolors.ENDC return None # No result, but there are more options to try def geocode(self, debug=False): ''' Updates self.latitude and self.longitude if a geocode is successsful; otherwise leaves them as the default (None). Uses Nominatim. Returns False if the location could not be geocoded, returns True when the geocode is sucessful. Tip: use geocode=False when instantiating, and then do a batch geocode using multiple threads with multiprocessing! ''' if not self.location: return False location = self.location # TODO: # '12:00 am, New Zealand' -37.7894134 175.2850399 if location == '12:00 am': return None if debug: print repr(self.location) + ' ← original' # Remove HTML entities location = location.encode("utf8") for char in ['&rsquo;', '\r', '\n']: location = location.replace(char, '') # Remove repeat white space location = ' '.join([segment for segment in location.split()]) location = strip_nonalpha_at_end(location) # North Island and South Island are not useful to the geocoder for island in [ 'North Island', 'South Island', 'NI', 'SI', 'Nth Island', 'Sth Island', 'North Is', 'South Is' ]: if not strip_nonalpha_at_end(location).endswith(island) and not \ strip_nonalpha_at_end(location).endswith(island + ', New Zealand'): continue location = location.replace(island, '') # It helps to add "New Zealand" even though a country bias is used # NOTE that there are (for some reason) some non-NZ observations non_nz_places = ['Antarctica', 'Timor Sea', 'South Pacific Ocean'] append_nz = True for place in non_nz_places: if place in location: append_nz = False if append_nz: location.replace(' NZ', ' New Zealand') if not location.strip().endswith(','): location = location.strip() + ',' if 'New Zealand' not in location: location = location.strip() + ' New Zealand' while True: # Try the location description, without leading conjunctions for loc in strip_conjunctions_at_start(location): gc = self.attempt_geocode(loc) if gc is not None: return gc # If there's a slash in the name, split it into two attempts attempts_copy = self.already_attempted.copy() for loc in attempts_copy: for loc in yield_locations_without_symbol(loc, '(\w*/[\w\s]*)', '/'): gc = self.attempt_geocode(loc) if gc is not None: return gc # If there's an ampersand in the name, split it into two attempts attempts_copy = self.already_attempted.copy() for loc in attempts_copy: for loc in yield_locations_without_symbol( loc, '(\w*\s&amp;\s\w*)', '*'): gc = self.attempt_geocode(loc) if gc is not None: return gc # Try without a bracketed clause attempts_copy = self.already_attempted.copy() for loc in attempts_copy: gc = self.attempt_geocode( return_location_without_bracketed_clause(loc)) if gc is not None: return gc # Try with some common substitutions or known errors: attempts_copy = self.already_attempted.copy() for loc in substitutions_for_known_issues(attempts_copy): gc = self.attempt_geocode(loc) if gc is not None: return gc # Try again without non-title-case words, # and without one-letter words attempts_copy = self.already_attempted.copy() for loc in attempts_copy: loc = return_location_without_non_title_case_and_short_words( loc) gc = self.attempt_geocode(loc) if gc is not None: return gc self.geocode_attempts += 1 # Remove the first word of the location for next attempt location = ' '.join(location.split(' ')[1:]) # While loop repeats def get_all_sightings_as_list_of_UFOSighting_objects(link, geocode=True, debug=True): ''' Returns a list of UFOSighting objects, scraped from one link to a page of sighting reports. <link> is a URL (string) that leads to a page of sighting reports on UFOCUS NZ's website. Must be in HTML format (<a href="the/url/path">) <geocode> defaults to false as it isn't compulsory and takes ages to compute (it needs to query a REST API). ''' sightings = [] for table in BeautifulSoup(urlopen(link)).findAll('table', {'cellpadding': '3'}): date = return_next_html_elem(table, 'Date') time = return_next_html_elem(table, 'Time') location = return_next_html_elem(table, 'Location') features = return_next_html_elem(table, 'Features/characteristics') description = return_next_html_elem(table, 'Description') # Work-around to re-build paragraph breaks, which get lost because # they are <br> tags. if description is not None and description.strip(): description_with_breaks = '' split_description = [d for d in description.split('.') if d is not \ None and d.strip()] for i, d in enumerate(split_description[:-1]): if split_description[i + 1][0].isalpha(): d += '.<br><br>' description_with_breaks += d description = description_with_breaks description += split_description[-1] + '.' ufo = UFOSighting(link, date, time, location, features, description) if not ufo.is_valid(): # Ignore UFO sightings that have been misidentified # (Emtpy HTML tables) continue if geocode: if not ufo.geocode(debug=debug): # Ignore UFO sightings that cannot be geocoded continue sightings.append(ufo) return sightings def export_ufos_to_csv(list_of_UFOSighting_objects): ''' Given a list of all the UFO sightings found on the website as UFOSighting objects, exports them to a CSV. ''' # Convert UFO objects to tuples all_sightings_as_tuples = [ ufo.__tuple__() for ufo in list_of_UFOSighting_objects ] # Create a pandas DataFrame from the list of tuples ufos_df = pd.DataFrame( all_sightings_as_tuples, columns=[ 'Date', 'Time', 'Location', 'Geocoded As', 'Geocode Attempts', 'Latitude', 'Longitude', 'Features', 'Description' ]) # Export the pandas DF to CSV ufos_df.to_csv( os.path.join(os.path.dirname(__file__), 'ufos_data.csv'), index=False, encoding='utf-8') return None def export_ufos_to_geojson(list_of_UFOSighting_objects): ''' Given a list of all the UFO sightings found on the website as UFOSighting objects, exports them to GeoJSON. The list is sorted by date, because the leaflet timeslider doesn't sort on a key, and I can't work out how to do it in JavaScript. Therefore it also removes observations that don't have a date ''' list_of_UFOSighting_objects = [ l for l in list_of_UFOSighting_objects if l is not None ] list_of_UFOSighting_objects = [ l for l in list_of_UFOSighting_objects if l.date ] list_of_UFOSighting_objects.sort(key=lambda x: x.date, reverse=False) fc = FeatureCollection([ ufo.__geojson__() for ufo in list_of_UFOSighting_objects if ufo.haslocation ]) with open( os.path.join(os.path.dirname(__file__), 'ufos_data.geojson'), 'w') as outfile: json.dump(fc, outfile) def geocode_worker(sighting): ''' A single geocoding worker, to be run in its own wee process... and probably rate-limited ''' sighting.geocode(debug=True) return sighting def main(debug=False): '''Main loop''' def valid(tag): ''' <tag> = an html tag that has an href Defines what an interesting hyperlink looks like, and returns True if the tag meets this criteria, False otherwise ''' return 'New-Zealand-UFO-Sightings-' in tag['href'] # Sightings page base_url = "http://www.ufocusnz.org.nz/content/Sightings/24.aspx" home_page = BeautifulSoup(urlopen(base_url)) # Get list of valid links from home page # There is one for each year links = sorted( set([li for li in home_page.findAll(href=True) if valid(li)])) # There are some other links scattered around the website that have # reports in the same format # pylint: disable=line-too-long additional_links = [ 'http://www.ufocusnz.org.nz/content/Police/101.aspx', 'http://www.ufocusnz.org.nz/content/Selection-of-Historic-Sighting-Reports/109.aspx', 'http://www.ufocusnz.org.nz/content/1965---Unidentified-Submerged-Object-%28USO%29-spotted-by-DC-3-Pilot/82.aspx', 'http://www.ufocusnz.org.nz/content/1968---Yellow-Disc-Descends-into-Island-Bay,-Wellington/104.aspx', 'http://www.ufocusnz.org.nz/content/1974---Large-Object-Emerges-from-Sea-off-Aranga-Beach,-Northland/105.aspx', 'http://www.ufocusnz.org.nz/content/1957-1968---Silver-Bullet-Bursts-Through-Antarctic-Ice/106.aspx' ] additional_links = [ BeautifulSoup(str('<a href="{}">Link</a>'.format(li))).findAll( href=True)[0] for li in additional_links ] # NOTE see here for more, although they conform less to the expected structure # http://www.ufocusnz.org.nz/content/Aviation/80.aspx links += additional_links links = set([l['href'] for l in links]) # TODO caching # Flatten lists of UFOs for each link all_sightings = reduce( lambda x, y: x + y, [ get_all_sightings_as_list_of_UFOSighting_objects( link, geocode=False, debug=debug) for link in links ]) pool = multiprocessing.Pool( processes=max(multiprocessing.cpu_count() - 2, 1)) results = pool.map(geocode_worker, all_sightings) # export_ufos_to_csv(results) export_ufos_to_geojson(results) if __name__ == '__main__': main(debug=True) exit(0)
[ "richard.m.law@gmail.com" ]
richard.m.law@gmail.com
88e75c46abb9494b3a6c173c9d4edbb771ad30b3
83951f7fd0bbaba9675bdf9ba6980504213bc1c6
/skim/crab/skim_QCD_Pt-15to7000_Flat2017_cfg.py
f4567da99bb4f470b3019a97ec8411522789b737
[]
no_license
DryRun/DijetSkimmer
6db71583b969ecc64841da26107f43c4c734ca43
ead65f8e2a5d11f99f3e1a60a1d2f9a163e68491
refs/heads/main
2021-07-22T19:41:09.096943
2021-07-14T13:01:00
2021-07-14T13:01:00
171,485,404
0
1
null
null
null
null
UTF-8
Python
false
false
2,340
py
import os from WMCore.Configuration import Configuration from CRABClient.UserUtilities import config, getUsernameFromSiteDB config = Configuration() job_name = "DijetSkim_QCD_Pt-15to7000_Flat2017_1_0_1" config.section_("General") config.General.requestName = job_name config.General.transferLogs = False config.section_("JobType") config.JobType.pluginName = 'Analysis' # Setup the custom executable config.JobType.psetName = os.path.expandvars('$CMSSW_BASE/src/PhysicsTools/DijetSkimmer/skim/PSet.py') # CRAB modifies this file to contain the input files and lumis config.JobType.scriptExe = os.path.expandvars('$CMSSW_BASE/src/PhysicsTools/DijetSkimmer/skim/crab_shell.sh') # CRAB then calls scriptExe jobId <scriptArgs> config.JobType.scriptArgs = ["--source=mc", "--year=2017"] config.JobType.inputFiles = [ os.path.expandvars('$CMSSW_BASE/src/PhysicsTools/DijetSkimmer/skim/crab_meat.py'), os.path.expandvars('$CMSSW_BASE/src/PhysicsTools/NanoAODTools/scripts/haddnano.py'), #hadd nano will not be needed once nano tools are in cmssw os.path.expandvars('$CMSSW_BASE/src/PhysicsTools/DijetSkimmer/skim/skim_branches_data.txt'), os.path.expandvars('$CMSSW_BASE/src/PhysicsTools/DijetSkimmer/skim/skim_branches_mc.txt'), os.path.expandvars('$CMSSW_BASE/src/PhysicsTools/DijetSkimmer/skim/skim_branches.txt'), #os.path.expandvars('$CMSSW_BASE/src/PhysicsTools/DijetSkimmer/skim/FrameworkJobReport.xml'), ] config.JobType.outputFiles = ["nanoskim.root", "hists.root"] config.JobType.sendPythonFolder = True config.JobType.allowUndistributedCMSSW = True config.section_("Data") #config.Data.inputDataset = '/JetHT/Run2018C-Nano14Dec2018-v1/NANOAOD' #config.Data.inputDBS = 'phys03' config.Data.inputDBS = 'global' config.Data.splitting = 'FileBased' #config.Data.splitting = 'EventAwareLumiBased' config.Data.unitsPerJob = 4 #config.Data.totalUnits = 10 config.JobType.allowUndistributedCMSSW = True config.Data.outLFNDirBase = '/store/user/{}/{}'.format(getUsernameFromSiteDB(), job_name) config.Data.publication = False config.Data.outputDatasetTag = job_name #config.Data.ignoreLocality = True config.section_("Site") config.Site.storageSite = "T3_US_Brown" config.Data.inputDataset = '/QCD_Pt-15to7000_TuneCP5_Flat2017_13TeV_pythia8/RunIIFall17NanoAODv4-PU2017_12Apr2018_Nano14Dec2018_102X_mc2017_realistic_v6-v1/NANOAODSIM'
[ "david.renhwa.yu@gmail.com" ]
david.renhwa.yu@gmail.com
8b10368f10167003f9ab9226c9b33eb43a5c1b2e
8f7bcd652fa10320c19da46d09260aaf11659a59
/src/logger_special.py
343ed25f22bb60394f57d9e7351fdb10d5384852
[]
no_license
EmanuelSamir/mapless-curiosity-driven-exploration
d226c1206064ee6877a3eea02a0409ff6f6aeb22
2bd399a6488b3a216aefdd3cb35107185ea31846
refs/heads/main
2023-07-14T21:29:36.690412
2021-08-31T21:43:33
2021-08-31T21:43:33
401,846,751
0
0
null
null
null
null
UTF-8
Python
false
false
1,577
py
import matplotlib.pyplot as plt import numpy as np import pandas as pd import os import json import pickle from torch import save as tsave import torch from .utils import create_dir from datetime import datetime class LoggerSpecial: def __init__(self, algorithm): # Logger saves data # Updates eps x steps x feature_num # Saves at special w/ name algorithm date ## csv and description self.features = [] self.steps = [] fn_date = datetime.now().strftime("_%m%d_%H-%M-%S") self.save_special_path = os.path.join("../specials", algorithm, fn_date) create_dir(self.save_special_path) def set_description(self, comment): description = { 'comment': comment } fn = os.path.join(self.save_special_path, 'description.pth' ) out_file = open(fn,'w+') json.dump(description,out_file) def update(self, step, feature): self.steps.append(step) self.features.append(feature) def consolidate(self, episode): folder = os.path.join(self.save_special_path, 'e{}_n{}'.format(episode, self.steps[-1])) create_dir(folder) fn = os.path.join(folder, 'data.csv') self.features = map(list, zip(*self.features)) d = { 'steps': self.steps, } for i, feat in enumerate(self.features): d['f{}'.format(i)] = feat df = pd.DataFrame(d) df.to_csv(fn, mode = 'w', index = False) self.steps = [] self.features = []
[ "samiremp.2@gmail.com" ]
samiremp.2@gmail.com
3f008a682cd719d81b222f36983c87310b67f103
523f8f5febbbfeb6d42183f2bbeebc36f98eadb5
/402.py
631b928370b0e9eabec5dcf010eca20cf6babf83
[]
no_license
saleed/LeetCode
655f82fdfcc3000400f49388e97fc0560f356af0
48b43999fb7e2ed82d922e1f64ac76f8fabe4baa
refs/heads/master
2022-06-15T21:54:56.223204
2022-05-09T14:05:50
2022-05-09T14:05:50
209,430,056
2
0
null
null
null
null
UTF-8
Python
false
false
751
py
class Solution(object): def removeKdigits(self, num, k): """ :type num: str :type k: int :rtype: str """ dp=["" for _ in range(k+1) ] for i in range(len(num)): dp[i][0]=num[:i+1] for j in range(1,k+1): dp[0][j]="" for i in range(1,len(num)): for j in range(1,k+1)[::-1]: dp[i][j]=min(dp[i-1][j-1],dp[i-1][j]+num[i]) # print(dp) res=dp[len(num) - 1][k].lstrip('0') if res=="": return '0' else: return res a=Solution() num = "1432219" k = 3 print(a.removeKdigits(num,k)) num = "10200" k=1 print(a.removeKdigits(num,k)) test='00002000' print(test.lstrip('0'))
[ "1533441387@qq.com" ]
1533441387@qq.com
87c625f739a6cd6794d4fd2ee1051c177d9f2046
1dbf66345b5a70c736a2155271a779b1e4292882
/tools/display-stats
8f57c3e4b27d5cf86167f9fa8e3deaf5de95e041
[ "BlueOak-1.0.0", "LicenseRef-scancode-unknown-license-reference" ]
permissive
michael-lazar/mozz-archiver
5a16405fbc433360be9bcaa404d0a879ce83b855
12617d2efca91663699647654bcd3e40f5f388f2
refs/heads/master
2023-01-13T10:52:11.012343
2020-11-17T03:42:20
2020-11-17T03:42:20
296,190,512
15
0
null
null
null
null
UTF-8
Python
false
false
1,888
#!/usr/bin/env python3 """ Display some statistics for an archive based on the generated index file. """ import argparse import sqlite3 parser = argparse.ArgumentParser(description="Display statistics for an index file") parser.add_argument('index_db') args = parser.parse_args() conn = sqlite3.connect(args.index_db, isolation_level=None) conn.row_factory = sqlite3.Row print(f"Parsing index database {args.index_db}...") print("") c = conn.execute("SELECT COUNT() FROM requests") total = c.fetchone()[0] print(f"Total Attempted : {total}") c = conn.execute("SELECT COUNT() FROM requests WHERE error_message IS NULL") success = c.fetchone()[0] print(f"Total Successful : {success}") c = conn.execute("SELECT COUNT() FROM requests WHERE error_message IS NOT NULL") failed = c.fetchone()[0] print(f"Total Failed : {failed}") c = conn.execute("SELECT COUNT(DISTINCT netloc) FROM requests WHERE netloc IS NOT NULL") domains = c.fetchone()[0] print(f"Total Domains Crawled : {domains}") print("") print("1. Successful Response Codes") print("") print("Count Code") print("----- ----") c = conn.execute("SELECT response_status, COUNT() FROM requests WHERE response_status IS NOT NULL GROUP BY response_status ORDER BY COUNT() DESC") for row in c: print(f"{row[1]:<8}{row[0]}") print("") print("2. Failed Request Reasons") print("") print("Count Error Message") print("----- -------------") c = conn.execute("SELECT error_message, COUNT() FROM requests WHERE error_message IS NOT NULL GROUP BY error_message ORDER BY COUNT() DESC") for row in c: print(f"{row[1]:<8}{row[0]}") print("") print("3. Crawled URLs by domain") print("") print("Count Domain") print("----- ------") c = conn.execute("SELECT netloc, COUNT() FROM requests WHERE netloc IS NOT NULL GROUP BY netloc ORDER BY COUNT() DESC") for row in c: print(f"{row[1]:<8}{row[0]}")
[ "mlazar@doctorondemand.com" ]
mlazar@doctorondemand.com
4bb00b163473e32f7db8ba0fc41a3438d8541698
909f787b07de220e76d2d49de536b7c073ea94ef
/virtual/bin/email_validator
18118bb6aa748c42d5b0547186dd8be03a668a65
[ "MIT" ]
permissive
Elisephan/Elevator-Pitch
7e710c2d57e30c096f84921fb24d5b876a4b1e78
5f087d07bae763a8af293c1b3417db0ebe9ec3b9
refs/heads/master
2023-01-29T10:04:20.332562
2020-12-07T18:33:50
2020-12-07T18:33:50
318,587,202
0
0
null
null
null
null
UTF-8
Python
false
false
250
#!/home/user/Desktop/Elevator_pitch/virtual/bin/python # -*- coding: utf-8 -*- import re import sys from email_validator import main if __name__ == '__main__': sys.argv[0] = re.sub(r'(-script\.pyw|\.exe)?$', '', sys.argv[0]) sys.exit(main())
[ "elise.ntakirutimana@gmail.com" ]
elise.ntakirutimana@gmail.com
e90d3a43b730ba904ef5e59f19aeea3c7ce1f151
32692811b7e6dacb156fbf04a5e599cf4f5d5141
/interface/test-teacher/testlogin.py
85761577ceaf40b9592a72d36901b36caef8ad3c
[]
no_license
2229157983/test
7b4593892a14d8741b2208c50edce9b58bf8ccca
2ae1a7b822659a397d74580de3120ecf40ba995a
refs/heads/master
2023-03-06T01:04:33.744347
2021-03-02T02:04:29
2021-03-02T02:04:29
341,459,660
0
0
null
null
null
null
UTF-8
Python
false
false
499
py
import requests import unittest class LoginTest(unittest.TestCase): def testlogin(self): url = "http://www.jasonisoft.cn:8080/HKR/UserServlet" data = { "method":"login", "loginname":"root11", "password":"1111111" } expect = "菜单" response = requests.get(url=url,data = data) response.encoding = "utf-8" data = response.text self.assertIn(expect,data)
[ "noreply@github.com" ]
noreply@github.com
3dae7f83d5bdd601cbe53c167dd715e7f1a20e2b
c5390221fc6b12933a5f0f877fbb4c5e349f0eb8
/env/bin/django-admin.py
3164d9747c2998ecd4fbfb9d552363ee7aec6074
[]
no_license
shannonphu/xspense
e8131f96835df1599baac642757c16200d3418f2
42f1d9ca5c825c56d3a876517f4592540498ba52
refs/heads/master
2021-01-19T13:02:13.856016
2015-08-23T18:37:56
2015-08-23T18:37:56
40,105,053
0
0
null
null
null
null
UTF-8
Python
false
false
162
py
#!/Users/shannon/Desktop/Django/xspense/env/bin/python3 from django.core import management if __name__ == "__main__": management.execute_from_command_line()
[ "shannonphu@sbcglobal.net" ]
shannonphu@sbcglobal.net
b324becd61a6557a6cb89150aa2716d6bed8add6
0b874d304dd2ea9ac5ef8e02e812a0e00bca0994
/051-数组中重复的数字.py
ce52a775225bc248e945b399493089285127e2a9
[]
no_license
sjtupig/codingInterviews
02316b3ded4de5e4be37a7675e4d56dd4b56f3cb
a79cd14bd8c7a3d501505ee155e0958b1299de66
refs/heads/master
2020-04-10T21:55:31.473115
2019-03-06T04:42:00
2019-03-06T04:42:00
161,310,590
3
0
null
null
null
null
UTF-8
Python
false
false
2,879
py
# -*- coding:utf-8 -*- class Solution: # 这里要特别注意~找到任意重复的一个值并赋值到duplication[0] # 函数返回True/False def duplicate(self, numbers, duplication): # write code here ns = [] for i in numbers: if i < 0 or i > len(numbers)-1: return False for i in numbers: if i not in ns: ns.append(i) elif i in ns: duplication[0] = i return True return False ''' #检查数据的合法性 检查输入参数是否合法 数组中的数据是否满足所有数字都在0到n-1的范围内 #排序后判断重复 最简单的思路就是先把输入的数组排序。从排序的数组中找出重复的数字就是个很容易的事情了。只需要从头向尾扫描一遍排序好的数组即可。 对一个数组排序的时间复杂度是$O(nlogn)$ 扫描一个排序好的数组发现重复的数字的时间复杂度是$O(n)$ ##符号位标识法 我们可以看到数组中元素的大小都在[0-n)这个区间内, 都是正数,那么他们的符号位对我们来说就是无关紧要的, 因此我们直接拿符号位当成我们的标识位就行了 #固定偏移法 跟标识法类似, 如果不借助外部辅助空间,那么我们只能在数组内部下功夫,又能设置标识,又能恢复数据(不破坏数据)的方式,前面我们用符号位作为标识的方法就是通过符号位, 即判断了是否存在,又可以通过符号位的反转重新恢复数据,那么有没有其他类似的方法呢? 我们想到我们的数据都是[0, n)这个区间的,那么我们采用类似与移码的方法,让数据加上或者减去一个固定的偏移量, 这样就可以即标识数据,又不损坏数据,为了能够区分出数据,这个偏移必须大于N,这样我们的原数据与标识数据存在一一映射关系。 [0, n-1] -=>+偏移n-=> [n, 2n-1] #将元素放在自己改在的位置 剑指offer上提供的方法,这种方法采用交换的方法 我们考虑如果每个数字都置出现一次,那么此时是最完美的,每一个下标i对应元素numbers[i],也就是说我们对于数组中的每个元素numbers[i]都把它放在自己应该在的位置上numbers[numbers[i]]上, 如果我们发现有两个元素想往同一个位置上放的时候,说明此元素必然重复 即如下的过程 如果numbers[i] == i, 那么我们认为number[i]这个元素是在自己的位置上的 否则的话, numbers[i]这个元素就应该在numbers[numbers[i]]这个位置上, 于是我们交换numbers[i]和numbers[numbers[i]] 重复操作1, 直到number[i]== i, 则继续操作下一个位置的元素, 或者numbers[i] == numbers[numbers[i],元素重复 '''
[ "noreply@github.com" ]
noreply@github.com
873a50c46a52d30c4946df44db2f075572592055
8536c27cbb8265d1fbc1ddd45e2081fd01abdfa7
/ML/practice/Lab2/feature_selection/main.py
74689bf76f0bd8e5d21265cdf5e06535f84ed526
[]
no_license
tsimafeip/master-course
03e6dd8e87ceebd4a67c636459579b796a03df97
3035792666fe167b2052e1d482768df2241e1d67
refs/heads/master
2021-07-20T06:55:25.152586
2021-01-03T19:28:40
2021-01-03T19:28:59
230,803,994
0
1
null
null
null
null
UTF-8
Python
false
false
2,940
py
#!/usr/bin/env python3 import argparse import logging import os import time import numpy as np import sklearn.datasets import sklearn.linear_model import features THRESHOLD = 0.80 def _parse_args(): parser = argparse.ArgumentParser(prog='bsu 2019 / ml / hw 2') parser.add_argument('--datadir', help='path to folder to cache data', default=os.getcwd()) return parser.parse_args() def _filter_data(x, y, digits): """Create subset with only specified digits.""" rx, ry = [], [] for cx, cy in zip(x, y): cy = int(cy) if cy in digits: rx.append(cx) ry.append(digits.index(cy)) return np.array(rx), np.array(ry) def _main(args): sklearn_home = args.datadir with open(r'C:\Users\lybot\OneDrive\Документы\Магистратура\Машинное обучение\practice\Lab2\2\feature_selection\result.txt', "w"): pass logging.info('Downloading MNIST data') mnist = sklearn.datasets.fetch_openml('mnist_784', data_home=sklearn_home) logging.info('Data is ready') solved_cases = 0 minimal_result = 1. average_result = 0. start_time = time.process_time() for da in range(10): for db in range(da + 1, 10): #logging.info('Processing case: {} vs {}'.format(da, db)) X, Y = _filter_data(mnist['data'], mnist['target'], [da, db]) #logging.info('Computing features') fs = features.FEATURES[(da, db)] assert len(fs) == 2, "We want exactly two feature functions" X2 = [(fs[0](x), fs[1](x)) for x in X] #logging.info('Training logistic regression classifier') cls = sklearn.linear_model.LogisticRegression(solver='liblinear') cls.fit(X2, Y) #logging.info('Done training') result = cls.score(X2, Y) with open(r'C:\Users\lybot\OneDrive\Документы\Магистратура\Машинное обучение\practice\Lab2\2\feature_selection\result.txt', 'a') as the_file: the_file.write('Case {} vs {}: {:.1f}%\n'.format(da, db, result * 100)) logging.info('Case {} vs {}: {:.1f}%'.format(da, db, result * 100)) if result >= THRESHOLD: #logging.info('Case is solved') solved_cases += 1 else: pass #logging.warning('Case is not solved!') minimal_result = min(minimal_result, result) average_result += result elapsed_time = time.process_time() - start_time average_result /= 45 print('Solved cases: {}'.format(solved_cases)) print('Minimal result: {:.1f}%'.format(minimal_result * 100)) print('Average result: {:.1f}%'.format(average_result * 100)) print('Elapsed time: {:.1f} second(s)'.format(elapsed_time)) if __name__ == '__main__': logging.basicConfig(level=logging.INFO) _main(_parse_args())
[ "noreply@github.com" ]
noreply@github.com
0de8bffe96bc6aec7bba0c7fb5de1491c599b977
dbc3b767f9d079fd76a7ffb3c61f71df4ab3f945
/fortytwo_test_task/settings/common.py
d7b1a3e2521bd5de6e791c92e71e2b14de9d78de
[]
no_license
hugoalvarado/FortyTwoTestTask
6acb83c1f17303bb3cfa005cb60f79ae97cc28b8
98946f8463f5dc6d71635338efb559e3513bad73
refs/heads/master
2021-01-21T23:45:20.201139
2016-07-24T05:30:07
2016-07-24T05:30:07
64,047,485
0
0
null
2016-07-24T03:49:04
2016-07-24T03:49:04
null
UTF-8
Python
false
false
3,720
py
""" Django settings for fortytwo_test_task project. For more information on this file, see https://docs.djangoproject.com/en/1.6/topics/settings/ For the full list of settings and their values, see https://docs.djangoproject.com/en/1.6/ref/settings/ """ # Build paths inside the project like this: os.path.join(BASE_DIR, ...) import os import sys PROJECT_DIR = os.path.dirname(os.path.dirname(__file__)) BASE_DIR = os.path.dirname(PROJECT_DIR) # App/Library Paths sys.path.append(os.path.join(BASE_DIR, 'apps')) # Quick-start development settings - unsuitable for production # See https://docs.djangoproject.com/en/1.6/howto/deployment/checklist/ # SECURITY WARNING: keep the secret key used in production secret! SECRET_KEY = 'x=c0_e(onjn^80irdy2c221#)2t^qi&6yrc$31i(&ti*_jf3l8' # SECURITY WARNING: don't run with debug turned on in production! DEBUG = True TEMPLATE_DEBUG = True ALLOWED_HOSTS = [] # Application definition INSTALLED_APPS = ( 'django.contrib.admin', 'django.contrib.auth', 'django.contrib.contenttypes', 'django.contrib.sessions', 'django.contrib.messages', 'django.contrib.staticfiles', 'apps.hello', ) MIDDLEWARE_CLASSES = ( 'django.contrib.sessions.middleware.SessionMiddleware', 'django.middleware.common.CommonMiddleware', 'django.middleware.csrf.CsrfViewMiddleware', 'django.contrib.auth.middleware.AuthenticationMiddleware', 'django.contrib.messages.middleware.MessageMiddleware', 'django.middleware.clickjacking.XFrameOptionsMiddleware', ) ROOT_URLCONF = 'fortytwo_test_task.urls' WSGI_APPLICATION = 'fortytwo_test_task.wsgi.application' # Database # https://docs.djangoproject.com/en/1.6/ref/settings/#databases DATABASES = { 'default': { 'ENGINE': 'django.db.backends.sqlite3', 'NAME': os.path.join(BASE_DIR, 'db_test.sqlite3'), } } # Internationalization # https://docs.djangoproject.com/en/1.6/topics/i18n/ LANGUAGE_CODE = 'en-us' TIME_ZONE = 'UTC' USE_I18N = True USE_L10N = True USE_TZ = True # Upload Media # Absolute filesystem path to the directory that will hold user-uploaded files. # Example: "/home/media/media.lawrence.com/media/" MEDIA_ROOT = os.path.join(BASE_DIR, '..', 'uploads') # URL that handles the media served from MEDIA_ROOT. Make sure to use a # trailing slash. # Examples: "http://media.lawrence.com/media/", "http://example.com/media/" MEDIA_URL = '/uploads/' # Static files (CSS, JavaScript, Images) # https://docs.djangoproject.com/en/1.6/howto/static-files/ # Absolute path to the directory static files should be collected to. # Don't put anything in this directory yourself; store your static files # in apps' "static/" subdirectories and in STATICFILES_DIRS. # Example: "/home/media/media.lawrence.com/static/" STATIC_ROOT = os.path.join(BASE_DIR, 'static') # Static files (CSS, JavaScript, Images) # https://docs.djangoproject.com/en/1.6/howto/static-files/ STATIC_URL = '/static/' # Additional locations of static files STATICFILES_DIRS = ( # Put strings here, like "/home/html/static" or "C:/www/django/static". # Always use forward slashes, even on Windows. # Don't forget to use absolute paths, not relative paths. os.path.join(BASE_DIR, 'assets'), ) # Template Settings TEMPLATE_DIRS = ( # Put strings here, like "/home/html/django_templates" or # "C:/www/django/templates". # Always use forward slashes, even on Windows. # Don't forget to use absolute paths, not relative paths. os.path.join(BASE_DIR, 'templates'), ) # Turn off south during test SOUTH_TESTS_MIGRATE = False
[ "hugo102@gmail.com" ]
hugo102@gmail.com
65a1193be03b44bc201c6c6ca6908d3216957e8e
03d8f86afa4ccc2fe981b7e50f6ad8fdcf730bb4
/backend/main.py
bcbcfa1ff612fb98a8b044e56722599a340fbd4f
[]
no_license
tecnd/ismyclassonline
51ea77895139444e9559abb12c70a450587e4611
c52e0f1e61ff55fc4f880dc8d98b264b8001f9e3
refs/heads/master
2023-01-28T00:49:52.835143
2020-11-15T19:55:26
2020-11-15T19:55:26
null
0
0
null
null
null
null
UTF-8
Python
false
false
2,218
py
from bs4 import BeautifulSoup import requests def tag_scraper(request): """HTTP Cloud Function. Args: request (flask.Request): The request object. <http://flask.pocoo.org/docs/1.0/api/#flask.Request> Returns: The response text, or any set of values that can be turned into a Response object using `make_response` <http://flask.pocoo.org/docs/1.0/api/#flask.Flask.make_response>. """ request_form = request.form rescode = 200 if request.form and 'subject' in request_form: subject = request_form['subject'].upper() number = request_form['number'].upper() section = request_form['section'].zfill(2) URL = "https://courselist.wm.edu/courselist/courseinfo/searchresults?term_code=202120&term_subj={}&attr=0&attr2=0&levl=0&status=0&ptrm=0&search=Search".format(subject) r = requests.get(URL) if r.status_code != 200: res = "Subject code not found" else: soup = BeautifulSoup(r.text, 'html5lib') table = soup.table code = table.find("td", string=subject+' '+number+' '+section+' ') if code is None: res = "Number/section not found" else: tags, name = code.find_next_siblings("td", limit=2) res = "<b>" + code.text + name.text + "</b>" tagslist = tags.text.split(", ") if "FS" in tagslist: res += "<p>FS: Face to face, Synchronous</p>" elif "MIX" in tagslist: res += "<p>MIX: Mix of in-person and remote</p>" elif "RA" in tagslist: res += "<p>RA: Remote, Asynchronous</p>" elif "RSOC" in tagslist: res += "<p>RSOC: Remote, Synchronous on Campus</p>" elif "RSOF" in tagslist: res += "<p>RSOF: Remote, Synchronous off Campus" else: res += "<p>Delivery attribute not found</p>" else: res = "Bad request" rescode = 400 headers = { 'Access-Control-Allow-Origin': '*' } return (res,rescode,headers)
[ "kwzeyunwang@gmail.com" ]
kwzeyunwang@gmail.com
c5dbd32d211c6f31a6c33de8200d745eb073847c
f95175c2ed06f371faba5b4ac5332d0f5b01a6ac
/FloorPlan/apps.py
dc84c76bb740d86219063ec8d05227e0be8f6850
[]
no_license
floor-plan/FloorPlan
8dcb4f16d503ac713bd7324c747103012a68c778
28d97de9dbcad1dd3f149c0af49db8a495b57c70
refs/heads/master
2021-09-27T06:53:25.275322
2020-07-08T20:27:32
2020-07-08T20:27:32
248,983,897
0
1
null
2021-09-22T18:51:20
2020-03-21T13:41:01
Python
UTF-8
Python
false
false
93
py
from django.apps import AppConfig class FloorplanConfig(AppConfig): name = 'FloorPlan'
[ "bmiller3822@gmail.com" ]
bmiller3822@gmail.com
11a3b54a12af9a6d287edfead2ec004be81b18c7
5be992e6ac6bae2ebf938005d1cae93777825087
/space/research/genelab.py
34513f8b9468f68b837529823a4942d5eab865ce
[]
no_license
a1aiintel/SpaceIsCool
0c88acaa966c85e31d73da8319966c218447158f
939641dbe626a2cbb9fcec845c18bfb3371118ad
refs/heads/master
2020-07-30T04:54:14.577501
2019-01-10T17:57:52
2019-01-10T17:57:52
null
0
0
null
null
null
null
UTF-8
Python
false
false
1,083
py
import requests from space import NASA_KEY def search_genelab(query, type): """ GeneLab provides a RESTful Application Programming Interface (API) to its full-text search_exoplanet capability, which provides the same functionality available through the GeneLab public data repository website. The API provides a choice of standardized web output formats, such as JavaScript Object Notation (JSON) or Hyper Text Markup Language (HTML), of the search_exoplanet results. The GeneLab Search API can also federate with other heterogeneous external bioinformatics databases, such as the National Institutes of Health (NIH) / National Center for Biotechnology Information's (NCBI) Gene Expression Omnibus (GEO); the European Bioinformatics Institute's (EBI) Proteomics Identification (PRIDE); the Argonne National Laboratory's (ANL) Metagenomics Rapid Annotations using Subsystems Technology (MG-RAST). :param query: :return: """ url = "https://genelab-data.ndc.nasa.gov/genelab/data/search_exoplanet?term=mouse%20liver&type=cgene"
[ "jarbasai@mailfence.com" ]
jarbasai@mailfence.com
766d8e38421e476ad21261a9e76adf41efd0c1f6
2566e318ce81db1e1713a7dbcb5da8e8dd38c74d
/mk/update-travis-yml.py
4f15448e7e4823c06ac8bc1a2e6fec09b253364c
[ "OpenSSL", "MIT", "ISC", "LicenseRef-scancode-mit-taylor-variant", "LicenseRef-scancode-openssl", "LicenseRef-scancode-ssleay-windows", "LicenseRef-scancode-unknown-license-reference" ]
permissive
nyantec/ring
9f6a741cac845ecc9793ec2c22000e6f5c1352e2
3b3fa9eef65d4a4049e353a7e75f071345519dae
refs/heads/master
2021-10-07T10:31:36.728412
2019-01-10T14:26:15
2019-01-10T14:26:15
132,122,634
0
0
null
2018-07-11T14:51:34
2018-05-04T09:57:20
Assembly
UTF-8
Python
false
false
8,575
py
# Run this as "python mk/update-travis-yml.py" # Copyright 2015 Brian Smith. # # Permission to use, copy, modify, and/or distribute this software for any # purpose with or without fee is hereby granted, provided that the above # copyright notice and this permission notice appear in all copies. # # THE SOFTWARE IS PROVIDED "AS IS" AND BRIAN SMITH AND THE AUTHORS DISCLAIM # ALL WARRANTIES WITH REGARD TO THIS SOFTWARE INCLUDING ALL IMPLIED WARRANTIES # OF MERCHANTABILITY AND FITNESS. IN NO EVENT SHALL BRIAN SMITH OR THE AUTHORS # BE LIABLE FOR ANY SPECIAL, DIRECT, INDIRECT, OR CONSEQUENTIAL DAMAGES OR ANY # DAMAGES WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER IN # AN ACTION OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT OF # OR IN CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE. import re import shutil rusts = [ "stable", "nightly", "beta", ] linux_compilers = [ # Assume the default compiler is GCC. This is run first because it is the # one most likely to break, especially since GCC 4.6 is the default # compiler on Travis CI for Ubuntu 12.04, and GCC 4.6 is not supported by # BoringSSL. "", # Newest clang and GCC. "clang-5.0", "gcc-7", ] # Clang 3.4 and GCC 4.6 are already installed by default. linux_default_clang = "clang-3.4" osx_compilers = [ "", # Don't set CC.' ] compilers = { "aarch64-unknown-linux-gnu" : [ "aarch64-linux-gnu-gcc" ], "armv7-linux-androideabi" : [ "arm-linux-androideabi-clang" ], "arm-unknown-linux-gnueabihf" : [ "arm-linux-gnueabihf-gcc" ], "i686-unknown-linux-gnu" : linux_compilers, "x86_64-unknown-linux-gnu" : linux_compilers, "x86_64-apple-darwin" : osx_compilers, } feature_sets = [ "", ] modes = [ "DEBUG", "RELWITHDEBINFO" ] # Mac OS X is first because we don't want to have to wait until all the Linux # configurations have been built to find out that there is a failure on Mac. oss = [ "osx", "linux", ] targets = { "osx" : [ "x86_64-apple-darwin", ], "linux" : [ "armv7-linux-androideabi", "x86_64-unknown-linux-gnu", "aarch64-unknown-linux-gnu", "i686-unknown-linux-gnu", "arm-unknown-linux-gnueabihf", ], } def format_entries(): return "\n".join([format_entry(os, target, compiler, rust, mode, features) for rust in rusts for os in oss for target in targets[os] for compiler in compilers[target] for mode in modes for features in feature_sets]) # We use alternative names (the "_X" suffix) so that, in mk/travis.sh, we can # enure that we set the specific variables we want and that no relevant # variables are unintentially inherited into the build process. Also, we have # to set |CC_X| instead of |CC| since Travis sets |CC| to its Travis CI default # value *after* processing the |env:| directive here. entry_template = """ - env: TARGET_X=%(target)s %(compilers)s FEATURES_X=%(features)s MODE_X=%(mode)s KCOV=%(kcov)s rust: %(rust)s os: %(os)s""" entry_indent = " " entry_packages_template = """ addons: apt: packages: %(packages)s""" entry_sources_template = """ sources: %(sources)s""" def format_entry(os, target, compiler, rust, mode, features): # Currently kcov only runs on Linux. # # GCC 5 was picked arbitrarily to restrict coverage report to one build for # efficiency reasons. # # Cargo passes RUSTFLAGS to rustc only in Rust 1.9 and later. When Rust 1.9 # is released then we can change this to run (also) on the stable channel. # # DEBUG mode is needed because debug symbols are needed for coverage # tracking. kcov = (os == "linux" and compiler == "gcc-5" and rust == "nightly" and mode == "DEBUG") target_words = target.split("-") arch = target_words[0] vendor = target_words[1] sys = target_words[2] if sys == "darwin": abi = sys sys = "macos" elif sys == "androideabi": abi = sys sys = "linux" else: abi = target_words[3] def prefix_all(prefix, xs): return [prefix + x for x in xs] template = entry_template if sys == "linux": packages = sorted(get_linux_packages_to_install(target, compiler, arch, kcov)) sources_with_dups = sum([get_sources_for_package(p) for p in packages],[]) sources = sorted(list(set(sources_with_dups))) # TODO: Use trusty for everything? if arch in ["aarch64", "arm", "armv7"]: template += """ dist: trusty sudo: required""" if sys == "linux": if packages: template += entry_packages_template if sources: template += entry_sources_template else: packages = [] sources = [] cc = get_cc(sys, compiler) if os == "osx": os += "\n" + entry_indent + "osx_image: xcode9.3" compilers = [] if cc != "": compilers += ["CC_X=" + cc] compilers += "" return template % { "compilers": " ".join(compilers), "features" : features, "mode" : mode, "kcov": "1" if kcov == True else "0", "packages" : "\n ".join(prefix_all("- ", packages)), "rust" : rust, "sources" : "\n ".join(prefix_all("- ", sources)), "target" : target, "os" : os, } def get_linux_packages_to_install(target, compiler, arch, kcov): if compiler in ["", linux_default_clang]: packages = [] elif compiler.startswith("clang-") or compiler.startswith("gcc-"): packages = [compiler] else: packages = [] if target == "aarch64-unknown-linux-gnu": packages += ["gcc-aarch64-linux-gnu", "libc6-dev-arm64-cross"] if target == "arm-unknown-linux-gnueabihf": packages += ["gcc-arm-linux-gnueabihf", "libc6-dev-armhf-cross"] if target == "armv7-linux-androideabi": packages += ["expect", "openjdk-6-jre-headless"] if arch == "i686": if kcov == True: packages += ["libcurl3:i386", "libcurl4-openssl-dev:i386", "libdw-dev:i386", "libelf-dev:i386", "libkrb5-dev:i386", "libssl-dev:i386"] if compiler.startswith("clang-") or compiler == "": packages += ["libc6-dev-i386", "gcc-multilib"] elif compiler.startswith("gcc-"): packages += [compiler + "-multilib", "linux-libc-dev:i386"] else: raise ValueError("unexpected compiler: %s" % compiler) elif arch == "x86_64": if kcov == True: packages += ["libcurl4-openssl-dev", "libelf-dev", "libdw-dev", "binutils-dev"] elif arch not in ["aarch64", "arm", "armv7"]: raise ValueError("unexpected arch: %s" % arch) return packages def get_sources_for_package(package): ubuntu_toolchain = "ubuntu-toolchain-r-test" if package.startswith("clang-"): _, version = package.split("-") llvm_toolchain = "llvm-toolchain-trusty-%s" % version # Stuff in llvm-toolchain-trusty depends on stuff in the toolchain # packages. return [llvm_toolchain, ubuntu_toolchain] else: return [ubuntu_toolchain] def get_cc(sys, compiler): if sys == "linux" and compiler == linux_default_clang: return "clang" return compiler def main(): # Make a backup of the file we are about to update. shutil.copyfile(".travis.yml", ".travis.yml~") with open(".travis.yml", "r+b") as file: begin = " # BEGIN GENERATED\n" end = " # END GENERATED\n" old_contents = file.read() new_contents = re.sub("%s(.*?)\n[ ]*%s" % (begin, end), "".join([begin, format_entries(), "\n\n", end]), old_contents, flags=re.S) if old_contents == new_contents: print "No changes" return file.seek(0) file.write(new_contents) file.truncate() print new_contents if __name__ == '__main__': main()
[ "brian@briansmith.org" ]
brian@briansmith.org
6c323d4661d704d9094075404c537d930bd2b707
ab3ce5ec371bbd069843e93eaebe3aad5b98e9d2
/my query/Python_postgre_sql/Refresher/20_dictionary_comprehenstions.py
89a1fa41e2c9f63fadbf622b344a1f139b8084e1
[]
no_license
katlehotsopane/SQL2
d947012cae4fa2131a253944d81b0d48ba8e5f10
58379214c3244ba466fb02e7f12f4a907debb676
refs/heads/main
2023-07-03T12:31:19.589899
2021-08-05T19:14:13
2021-08-05T19:14:13
393,133,909
0
0
null
null
null
null
UTF-8
Python
false
false
458
py
users = [ (0, "Bob", "password"), (1, "Rolf", "bob123"), (2, "Jose", "longp4assword"), (3, "username", "1234"), ] username_mapping = {user[1]: user for user in users} username_input = input("Enter your username: ") password_input = input("Enter your password: ") _, username, password = username_mapping[username_input] if password_input == password: print("Your details are correct!") else: print("your details are incorrect!")
[ "noreply@github.com" ]
noreply@github.com
ce6bfe2a9145cfc6f226201d4923551145eb81a7
479559fc4d4724a7145cfb8ecdaa5cdc55e46761
/tensorflow/python/data/experimental/ops/interleave_ops.py
257639a2560aa5248ffb97bdeb46add625c96113
[ "Apache-2.0" ]
permissive
mudassirej/tensorflow
434818cc68c754c40d2e3b014daf1e3974d26698
bd47c759176f0039026fd5cac8db247bf452de28
refs/heads/master
2020-06-14T10:55:42.751443
2019-07-03T04:07:46
2019-07-03T04:12:59
194,978,111
1
0
Apache-2.0
2019-07-03T04:13:09
2019-07-03T04:13:09
null
UTF-8
Python
false
false
11,807
py
# Copyright 2017 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # ============================================================================== """Non-deterministic dataset transformations.""" from __future__ import absolute_import from __future__ import division from __future__ import print_function from tensorflow.python.data.experimental.ops import random_ops from tensorflow.python.data.ops import dataset_ops from tensorflow.python.data.ops import readers from tensorflow.python.data.util import nest from tensorflow.python.data.util import structure from tensorflow.python.framework import dtypes from tensorflow.python.framework import ops from tensorflow.python.ops import array_ops from tensorflow.python.ops import gen_experimental_dataset_ops from tensorflow.python.ops import gen_stateless_random_ops from tensorflow.python.ops import math_ops from tensorflow.python.util import deprecation from tensorflow.python.util.tf_export import tf_export @deprecation.deprecated( None, "Use `tf.data.Dataset.interleave(map_func, cycle_length, block_length, " "num_parallel_calls=tf.data.experimental.AUTOTUNE)` instead. If sloppy " "execution is desired, use `tf.data.Options.experimental_determinstic`.") @tf_export("data.experimental.parallel_interleave") def parallel_interleave(map_func, cycle_length, block_length=1, sloppy=False, buffer_output_elements=None, prefetch_input_elements=None): """A parallel version of the `Dataset.interleave()` transformation. `parallel_interleave()` maps `map_func` across its input to produce nested datasets, and outputs their elements interleaved. Unlike `tf.data.Dataset.interleave`, it gets elements from `cycle_length` nested datasets in parallel, which increases the throughput, especially in the presence of stragglers. Furthermore, the `sloppy` argument can be used to improve performance, by relaxing the requirement that the outputs are produced in a deterministic order, and allowing the implementation to skip over nested datasets whose elements are not readily available when requested. Example usage: ```python # Preprocess 4 files concurrently. filenames = tf.data.Dataset.list_files("/path/to/data/train*.tfrecords") dataset = filenames.apply( tf.data.experimental.parallel_interleave( lambda filename: tf.data.TFRecordDataset(filename), cycle_length=4)) ``` WARNING: If `sloppy` is `True`, the order of produced elements is not deterministic. Args: map_func: A function mapping a nested structure of tensors to a `Dataset`. cycle_length: The number of input `Dataset`s to interleave from in parallel. block_length: The number of consecutive elements to pull from an input `Dataset` before advancing to the next input `Dataset`. sloppy: If false, elements are produced in deterministic order. Otherwise, the implementation is allowed, for the sake of expediency, to produce elements in a non-deterministic order. buffer_output_elements: The number of elements each iterator being interleaved should buffer (similar to the `.prefetch()` transformation for each interleaved iterator). prefetch_input_elements: The number of input elements to transform to iterators before they are needed for interleaving. Returns: A `Dataset` transformation function, which can be passed to `tf.data.Dataset.apply`. """ def _apply_fn(dataset): return readers.ParallelInterleaveDataset( dataset, map_func, cycle_length, block_length, sloppy, buffer_output_elements, prefetch_input_elements) return _apply_fn class _DirectedInterleaveDataset(dataset_ops.Dataset): """A substitute for `Dataset.interleave()` on a fixed list of datasets.""" def __init__(self, selector_input, data_inputs): self._selector_input = selector_input self._data_inputs = list(data_inputs) first_output_types = dataset_ops.get_legacy_output_types(data_inputs[0]) first_output_classes = dataset_ops.get_legacy_output_classes(data_inputs[0]) for data_input in data_inputs[1:]: if (dataset_ops.get_legacy_output_types(data_input) != first_output_types or dataset_ops.get_legacy_output_classes(data_input) != first_output_classes): raise TypeError("All datasets must have the same type and class.") output_shapes = dataset_ops.get_legacy_output_shapes(self._data_inputs[0]) for data_input in self._data_inputs[1:]: output_shapes = nest.pack_sequence_as(output_shapes, [ ts1.most_specific_compatible_shape(ts2) for (ts1, ts2) in zip( nest.flatten(output_shapes), nest.flatten(dataset_ops.get_legacy_output_shapes(data_input))) ]) self._element_spec = structure.convert_legacy_structure( first_output_types, output_shapes, first_output_classes) super(_DirectedInterleaveDataset, self).__init__() def _as_variant_tensor(self): # pylint: disable=protected-access return ( gen_experimental_dataset_ops.experimental_directed_interleave_dataset( self._selector_input._variant_tensor, [data_input._variant_tensor for data_input in self._data_inputs], **self._flat_structure)) # pylint: enable=protected-access def _inputs(self): return [self._selector_input] + self._data_inputs @property def element_spec(self): return self._element_spec @tf_export("data.experimental.sample_from_datasets", v1=[]) def sample_from_datasets_v2(datasets, weights=None, seed=None): """Samples elements at random from the datasets in `datasets`. Args: datasets: A list of `tf.data.Dataset` objects with compatible structure. weights: (Optional.) A list of `len(datasets)` floating-point values where `weights[i]` represents the probability with which an element should be sampled from `datasets[i]`, or a `tf.data.Dataset` object where each element is such a list. Defaults to a uniform distribution across `datasets`. seed: (Optional.) A `tf.int64` scalar `tf.Tensor`, representing the random seed that will be used to create the distribution. See `tf.compat.v1.set_random_seed` for behavior. Returns: A dataset that interleaves elements from `datasets` at random, according to `weights` if provided, otherwise with uniform probability. Raises: TypeError: If the `datasets` or `weights` arguments have the wrong type. ValueError: If the `weights` argument is specified and does not match the length of the `datasets` element. """ num_datasets = len(datasets) if not isinstance(weights, dataset_ops.DatasetV2): if weights is None: # Select inputs with uniform probability. logits = [[1.0] * num_datasets] else: # Use the given `weights` as the probability of choosing the respective # input. weights = ops.convert_to_tensor(weights, name="weights") if weights.dtype not in (dtypes.float32, dtypes.float64): raise TypeError("`weights` must be convertible to a tensor of " "`tf.float32` or `tf.float64` elements.") if not weights.shape.is_compatible_with([num_datasets]): raise ValueError( "`weights` must be a vector of length `len(datasets)`.") # The `stateless_multinomial()` op expects log-probabilities, as opposed # to weights. logits = array_ops.expand_dims(math_ops.log(weights, name="logits"), 0) # NOTE(mrry): We only specialize when `weights` is not a `Dataset`. When it # is a `Dataset`, it is possible that evaluating it has a side effect the # user depends on. if len(datasets) == 1: return datasets[0] def select_dataset_constant_logits(seed): return array_ops.squeeze( gen_stateless_random_ops.stateless_multinomial(logits, 1, seed=seed), axis=[0, 1]) selector_input = dataset_ops.MapDataset( random_ops.RandomDataset(seed).batch(2), select_dataset_constant_logits, use_inter_op_parallelism=False) else: # Use each element of the given `weights` dataset as the probability of # choosing the respective input. # The `stateless_multinomial()` op expects log-probabilities, as opposed to # weights. logits_ds = weights.map(lambda *p: math_ops.log(p, name="logits")) def select_dataset_varying_logits(logits, seed): return array_ops.squeeze( gen_stateless_random_ops.stateless_multinomial(logits, 1, seed=seed), axis=[0, 1]) logits_and_seeds = dataset_ops.Dataset.zip( (logits_ds, random_ops.RandomDataset(seed).batch(2))) selector_input = dataset_ops.MapDataset( logits_and_seeds, select_dataset_varying_logits, use_inter_op_parallelism=False) return _DirectedInterleaveDataset(selector_input, datasets) @tf_export(v1=["data.experimental.sample_from_datasets"]) def sample_from_datasets_v1(datasets, weights=None, seed=None): return dataset_ops.DatasetV1Adapter( sample_from_datasets_v2(datasets, weights, seed)) sample_from_datasets_v1.__doc__ = sample_from_datasets_v2.__doc__ @tf_export("data.experimental.choose_from_datasets", v1=[]) def choose_from_datasets_v2(datasets, choice_dataset): """Creates a dataset that deterministically chooses elements from `datasets`. For example, given the following datasets: ```python datasets = [tf.data.Dataset.from_tensors("foo").repeat(), tf.data.Dataset.from_tensors("bar").repeat(), tf.data.Dataset.from_tensors("baz").repeat()] # Define a dataset containing `[0, 1, 2, 0, 1, 2, 0, 1, 2]`. choice_dataset = tf.data.Dataset.range(3).repeat(3) result = tf.data.experimental.choose_from_datasets(datasets, choice_dataset) ``` The elements of `result` will be: ``` "foo", "bar", "baz", "foo", "bar", "baz", "foo", "bar", "baz" ``` Args: datasets: A list of `tf.data.Dataset` objects with compatible structure. choice_dataset: A `tf.data.Dataset` of scalar `tf.int64` tensors between `0` and `len(datasets) - 1`. Returns: A dataset that interleaves elements from `datasets` according to the values of `choice_dataset`. Raises: TypeError: If the `datasets` or `choice_dataset` arguments have the wrong type. """ if not structure.are_compatible(choice_dataset.element_spec, structure.TensorStructure(dtypes.int64, [])): raise TypeError("`choice_dataset` must be a dataset of scalar " "`tf.int64` tensors.") return _DirectedInterleaveDataset(choice_dataset, datasets) @tf_export(v1=["data.experimental.choose_from_datasets"]) def choose_from_datasets_v1(datasets, choice_dataset): return dataset_ops.DatasetV1Adapter( choose_from_datasets_v2(datasets, choice_dataset)) choose_from_datasets_v1.__doc__ = choose_from_datasets_v2.__doc__ # TODO(b/119044825): Until all `tf.data` unit tests are converted to V2, keep # these aliases in place. choose_from_datasets = choose_from_datasets_v1 sample_from_datasets = sample_from_datasets_v1
[ "gardener@tensorflow.org" ]
gardener@tensorflow.org
958a75ab50cf92aa3f4243c6b47edba3f8c0b023
d554b1aa8b70fddf81da8988b4aaa43788fede88
/5 - Notebooks e Data/1 - Análises numéricas/Arquivos David/Atualizados/logDicas-master/data/2019-1/226/users/4140/codes/1593_1802.py
997c8783e8cfb6c3d78bc17c96cca711247bd924
[]
no_license
JosephLevinthal/Research-projects
a3bc3ca3b09faad16f5cce5949a2279cf14742ba
60d5fd6eb864a5181f4321e7a992812f3c2139f9
refs/heads/master
2022-07-31T06:43:02.686109
2020-05-23T00:24:26
2020-05-23T00:24:26
266,199,309
1
0
null
null
null
null
UTF-8
Python
false
false
127
py
balrog=int(input()); d1=int(input()); d2=int(input()); from math import * dano=int(sqrt(5*d1)+pi**(d2/3)); print(balrog-dano)
[ "jvlo@icomp.ufam.edu.br" ]
jvlo@icomp.ufam.edu.br
b01cb42df40d9efc85d03a815e799ee14b6e8fd8
ca7aa979e7059467e158830b76673f5b77a0f5a3
/Python_codes/p03339/s273941488.py
c82cd4ca992be5faaa424d10d255497c4a9fd014
[]
no_license
Aasthaengg/IBMdataset
7abb6cbcc4fb03ef5ca68ac64ba460c4a64f8901
f33f1c5c3b16d0ea8d1f5a7d479ad288bb3f48d8
refs/heads/main
2023-04-22T10:22:44.763102
2021-05-13T17:27:22
2021-05-13T17:27:22
367,112,348
0
0
null
null
null
null
UTF-8
Python
false
false
222
py
n = int(input()) s = [(i == "W")*1 for i in list(input())] c = [0]*(n+1) for i in range(n): c[i+1] = c[i] + s[i] ans = float("inf") for i in range(n): t = c[i] + (n-i-1-c[-1]+c[i+1]) ans = min(ans,t) print(ans)
[ "66529651+Aastha2104@users.noreply.github.com" ]
66529651+Aastha2104@users.noreply.github.com
25328fb0492fe750697b3767b53d440d4e3da0b8
e0df2bc703d0d02423ea68cf0b8c8f8d22d5c163
/ScientificComputing/ch14/filter_firdesign_sinc1.py
cfb39fc541dac9e8bb9246523bf73a615acecbeb
[]
no_license
socrates77-sh/learn
a5d459cb9847ba3b1bc4f9284ce35d4207d8aa8b
ae50978023f6b098b168b8cca82fba263af444aa
refs/heads/master
2022-12-16T16:53:50.231577
2019-07-13T13:52:42
2019-07-13T13:52:42
168,442,963
0
0
null
2022-12-08T05:18:37
2019-01-31T01:30:06
HTML
UTF-8
Python
false
false
363
py
# -*- coding: utf-8 -*- import scipy.signal as signal import numpy as np import pylab as pl def h_ideal(n, fc): return 2*fc*np.sinc(2*fc*np.arange(0, n, 1.0)) b = h_ideal(30, 0.25) w, h = signal.freqz(b, 1) pl.figure(figsize=(8, 4)) pl.plot(w/2/np.pi, 20*np.log10(np.abs(h))) pl.xlabel(u"正规化频率 周期/取样") pl.ylabel(u"幅值(dB)") pl.show()
[ "zhwenrong@sina.com" ]
zhwenrong@sina.com
576be3d1522f710ccbaba352d2393f1ebf54fd96
704aed30fda284d689887a0841b28f83ee80b922
/RC1/phil_catkin_ws/build/rosserial/rosserial_arduino/catkin_generated/pkg.installspace.context.pc.py
306bec52756d11160e6cd0279101a58b23c42dd9
[]
no_license
ozay-group/scaled-cars
f69832dc01407044e8307cb39a989c765f21c48a
bd171636d2bcbfca3767eb9d877e91c0904ecb1f
refs/heads/master
2020-04-27T19:23:51.819988
2018-08-30T17:19:14
2018-08-30T17:19:14
174,615,849
0
0
null
null
null
null
UTF-8
Python
false
false
487
py
# generated from catkin/cmake/template/pkg.context.pc.in CATKIN_PACKAGE_PREFIX = "" PROJECT_PKG_CONFIG_INCLUDE_DIRS = "/home/ubuntu/phil_catkin_ws/install/include".split(';') if "/home/ubuntu/phil_catkin_ws/install/include" != "" else [] PROJECT_CATKIN_DEPENDS = "message_runtime".replace(';', ' ') PKG_CONFIG_LIBRARIES_WITH_PREFIX = "".split(';') if "" != "" else [] PROJECT_NAME = "rosserial_arduino" PROJECT_SPACE_DIR = "/home/ubuntu/phil_catkin_ws/install" PROJECT_VERSION = "0.7.7"
[ "psisk@umich.edu" ]
psisk@umich.edu
cb4aa6d7dde712cf314a82682f8860ab95fe0eae
f026b118e0ee312dd759db2d4890f2ddaa7e165e
/WHFlask/model/__init__.py
225c67761c910247fd4414a5081d9d9969d08c29
[]
no_license
ksnero34/walker_holic_back
df345f8317c9f8fc7b76bb55f983f5e2ac571764
5125127a62e0b8be8d7182aaf77002de58d4b43a
refs/heads/main
2023-07-23T12:10:41.447787
2021-08-29T03:45:45
2021-08-29T03:45:45
385,494,489
1
0
null
null
null
null
UTF-8
Python
false
false
281
py
''' Application Model ''' # from model import mysql from config import config def get_cursor(): print(config.SLOW_API_TIME) def init_app(): '''Model Init Function''' # Mysql Init initializer = mysql.ModelInitializer() initializer.init_model() get_cursor()
[ "must1080@naver.com" ]
must1080@naver.com
172d528877e46d3a15c44ea0bd68dd96091dec79
77676610410e479a3214669b082b5f410b499e24
/apps/main/migrations/0010_auto_20170424_0645.py
cfeb0350a6e5aedc05e7e5c8f745933e2474e75b
[ "Apache-2.0" ]
permissive
StepicOrg/stepik-extensions
e76b2ee033275b33bf9d8c8deeac495d3a6bde46
5825bc9b2444ad4690681964d1bed172706f8796
refs/heads/develop
2023-04-05T12:43:28.114500
2021-04-19T12:57:30
2021-04-19T12:57:30
82,687,804
5
2
Apache-2.0
2021-04-19T12:58:47
2017-02-21T14:17:00
JavaScript
UTF-8
Python
false
false
653
py
# -*- coding: utf-8 -*- # Generated by Django 1.10.6 on 2017-04-24 06:45 from __future__ import unicode_literals from django.db import migrations class Migration(migrations.Migration): dependencies = [ ('main', '0009_auto_20170422_2002'), ] operations = [ migrations.RemoveField( model_name='extension', name='categories', ), migrations.RemoveField( model_name='extension', name='user_groups', ), migrations.DeleteModel( name='Category', ), migrations.DeleteModel( name='Extension', ), ]
[ "meanmail@mail.ru" ]
meanmail@mail.ru
186a3b0eaa286a7a614388dcc19fba81ffbc22cd
8c44ca5d4b82e504459b8d04f00ee1530e443274
/clients/python/lakefs_client/api/objects_api.py
b0c5bb9f507aaefcd48392b2d130588ec4560f44
[ "Apache-2.0" ]
permissive
nopcoder/lakeFS
c60620b8ce32dc603e03da4915d5ebec39a290e8
3b817bd7fc3479b082ec0bbc92574fe373ae18e3
refs/heads/master
2023-04-18T05:41:35.973323
2021-04-26T20:05:17
2021-04-26T20:05:17
359,791,391
2
0
Apache-2.0
2021-04-20T11:30:55
2021-04-20T11:30:55
null
UTF-8
Python
false
false
37,723
py
""" lakeFS API lakeFS HTTP API # noqa: E501 The version of the OpenAPI document: 0.1.0 Contact: services@treeverse.io Generated by: https://openapi-generator.tech """ import re # noqa: F401 import sys # noqa: F401 from lakefs_client.api_client import ApiClient, Endpoint as _Endpoint from lakefs_client.model_utils import ( # noqa: F401 check_allowed_values, check_validations, date, datetime, file_type, none_type, validate_and_convert_types ) from lakefs_client.model.error import Error from lakefs_client.model.object_stage_creation import ObjectStageCreation from lakefs_client.model.object_stats import ObjectStats from lakefs_client.model.object_stats_list import ObjectStatsList from lakefs_client.model.underlying_object_properties import UnderlyingObjectProperties class ObjectsApi(object): """NOTE: This class is auto generated by OpenAPI Generator Ref: https://openapi-generator.tech Do not edit the class manually. """ def __init__(self, api_client=None): if api_client is None: api_client = ApiClient() self.api_client = api_client def __delete_object( self, repository, branch, path, **kwargs ): """delete object # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.delete_object(repository, branch, path, async_req=True) >>> result = thread.get() Args: repository (str): branch (str): path (str): Keyword Args: _return_http_data_only (bool): response data without head status code and headers. Default is True. _preload_content (bool): if False, the urllib3.HTTPResponse object will be returned without reading/decoding response data. Default is True. _request_timeout (float/tuple): timeout setting for this request. If one number provided, it will be total request timeout. It can also be a pair (tuple) of (connection, read) timeouts. Default is None. _check_input_type (bool): specifies if type checking should be done one the data sent to the server. Default is True. _check_return_type (bool): specifies if type checking should be done one the data received from the server. Default is True. _host_index (int/None): specifies the index of the server that we want to use. Default is read from the configuration. async_req (bool): execute request asynchronously Returns: None If the method is called asynchronously, returns the request thread. """ kwargs['async_req'] = kwargs.get( 'async_req', False ) kwargs['_return_http_data_only'] = kwargs.get( '_return_http_data_only', True ) kwargs['_preload_content'] = kwargs.get( '_preload_content', True ) kwargs['_request_timeout'] = kwargs.get( '_request_timeout', None ) kwargs['_check_input_type'] = kwargs.get( '_check_input_type', True ) kwargs['_check_return_type'] = kwargs.get( '_check_return_type', True ) kwargs['_host_index'] = kwargs.get('_host_index') kwargs['repository'] = \ repository kwargs['branch'] = \ branch kwargs['path'] = \ path return self.call_with_http_info(**kwargs) self.delete_object = _Endpoint( settings={ 'response_type': None, 'auth': [ 'basic_auth', 'cookie_auth', 'jwt_token' ], 'endpoint_path': '/repositories/{repository}/branches/{branch}/objects', 'operation_id': 'delete_object', 'http_method': 'DELETE', 'servers': None, }, params_map={ 'all': [ 'repository', 'branch', 'path', ], 'required': [ 'repository', 'branch', 'path', ], 'nullable': [ ], 'enum': [ ], 'validation': [ ] }, root_map={ 'validations': { }, 'allowed_values': { }, 'openapi_types': { 'repository': (str,), 'branch': (str,), 'path': (str,), }, 'attribute_map': { 'repository': 'repository', 'branch': 'branch', 'path': 'path', }, 'location_map': { 'repository': 'path', 'branch': 'path', 'path': 'query', }, 'collection_format_map': { } }, headers_map={ 'accept': [ 'application/json' ], 'content_type': [], }, api_client=api_client, callable=__delete_object ) def __get_object( self, repository, ref, path, **kwargs ): """get object content # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.get_object(repository, ref, path, async_req=True) >>> result = thread.get() Args: repository (str): ref (str): a reference (could be either a branch or a commit ID) path (str): Keyword Args: _return_http_data_only (bool): response data without head status code and headers. Default is True. _preload_content (bool): if False, the urllib3.HTTPResponse object will be returned without reading/decoding response data. Default is True. _request_timeout (float/tuple): timeout setting for this request. If one number provided, it will be total request timeout. It can also be a pair (tuple) of (connection, read) timeouts. Default is None. _check_input_type (bool): specifies if type checking should be done one the data sent to the server. Default is True. _check_return_type (bool): specifies if type checking should be done one the data received from the server. Default is True. _host_index (int/None): specifies the index of the server that we want to use. Default is read from the configuration. async_req (bool): execute request asynchronously Returns: file_type If the method is called asynchronously, returns the request thread. """ kwargs['async_req'] = kwargs.get( 'async_req', False ) kwargs['_return_http_data_only'] = kwargs.get( '_return_http_data_only', True ) kwargs['_preload_content'] = kwargs.get( '_preload_content', True ) kwargs['_request_timeout'] = kwargs.get( '_request_timeout', None ) kwargs['_check_input_type'] = kwargs.get( '_check_input_type', True ) kwargs['_check_return_type'] = kwargs.get( '_check_return_type', True ) kwargs['_host_index'] = kwargs.get('_host_index') kwargs['repository'] = \ repository kwargs['ref'] = \ ref kwargs['path'] = \ path return self.call_with_http_info(**kwargs) self.get_object = _Endpoint( settings={ 'response_type': (file_type,), 'auth': [ 'basic_auth', 'cookie_auth', 'jwt_token' ], 'endpoint_path': '/repositories/{repository}/refs/{ref}/objects', 'operation_id': 'get_object', 'http_method': 'GET', 'servers': None, }, params_map={ 'all': [ 'repository', 'ref', 'path', ], 'required': [ 'repository', 'ref', 'path', ], 'nullable': [ ], 'enum': [ ], 'validation': [ ] }, root_map={ 'validations': { }, 'allowed_values': { }, 'openapi_types': { 'repository': (str,), 'ref': (str,), 'path': (str,), }, 'attribute_map': { 'repository': 'repository', 'ref': 'ref', 'path': 'path', }, 'location_map': { 'repository': 'path', 'ref': 'path', 'path': 'query', }, 'collection_format_map': { } }, headers_map={ 'accept': [ 'application/octet-stream', 'application/json' ], 'content_type': [], }, api_client=api_client, callable=__get_object ) def __get_underlying_properties( self, repository, ref, path, **kwargs ): """get object properties on underlying storage # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.get_underlying_properties(repository, ref, path, async_req=True) >>> result = thread.get() Args: repository (str): ref (str): a reference (could be either a branch or a commit ID) path (str): Keyword Args: _return_http_data_only (bool): response data without head status code and headers. Default is True. _preload_content (bool): if False, the urllib3.HTTPResponse object will be returned without reading/decoding response data. Default is True. _request_timeout (float/tuple): timeout setting for this request. If one number provided, it will be total request timeout. It can also be a pair (tuple) of (connection, read) timeouts. Default is None. _check_input_type (bool): specifies if type checking should be done one the data sent to the server. Default is True. _check_return_type (bool): specifies if type checking should be done one the data received from the server. Default is True. _host_index (int/None): specifies the index of the server that we want to use. Default is read from the configuration. async_req (bool): execute request asynchronously Returns: UnderlyingObjectProperties If the method is called asynchronously, returns the request thread. """ kwargs['async_req'] = kwargs.get( 'async_req', False ) kwargs['_return_http_data_only'] = kwargs.get( '_return_http_data_only', True ) kwargs['_preload_content'] = kwargs.get( '_preload_content', True ) kwargs['_request_timeout'] = kwargs.get( '_request_timeout', None ) kwargs['_check_input_type'] = kwargs.get( '_check_input_type', True ) kwargs['_check_return_type'] = kwargs.get( '_check_return_type', True ) kwargs['_host_index'] = kwargs.get('_host_index') kwargs['repository'] = \ repository kwargs['ref'] = \ ref kwargs['path'] = \ path return self.call_with_http_info(**kwargs) self.get_underlying_properties = _Endpoint( settings={ 'response_type': (UnderlyingObjectProperties,), 'auth': [ 'basic_auth', 'cookie_auth', 'jwt_token' ], 'endpoint_path': '/repositories/{repository}/refs/{ref}/objects/underlyingProperties', 'operation_id': 'get_underlying_properties', 'http_method': 'GET', 'servers': None, }, params_map={ 'all': [ 'repository', 'ref', 'path', ], 'required': [ 'repository', 'ref', 'path', ], 'nullable': [ ], 'enum': [ ], 'validation': [ ] }, root_map={ 'validations': { }, 'allowed_values': { }, 'openapi_types': { 'repository': (str,), 'ref': (str,), 'path': (str,), }, 'attribute_map': { 'repository': 'repository', 'ref': 'ref', 'path': 'path', }, 'location_map': { 'repository': 'path', 'ref': 'path', 'path': 'query', }, 'collection_format_map': { } }, headers_map={ 'accept': [ 'application/json' ], 'content_type': [], }, api_client=api_client, callable=__get_underlying_properties ) def __list_objects( self, repository, ref, **kwargs ): """list objects under a given prefix # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.list_objects(repository, ref, async_req=True) >>> result = thread.get() Args: repository (str): ref (str): a reference (could be either a branch or a commit ID) Keyword Args: prefix (str): [optional] after (str): return items after this value. [optional] amount (int): how many items to return. [optional] if omitted the server will use the default value of 100 delimiter (str): [optional] if omitted the server will use the default value of "/" _return_http_data_only (bool): response data without head status code and headers. Default is True. _preload_content (bool): if False, the urllib3.HTTPResponse object will be returned without reading/decoding response data. Default is True. _request_timeout (float/tuple): timeout setting for this request. If one number provided, it will be total request timeout. It can also be a pair (tuple) of (connection, read) timeouts. Default is None. _check_input_type (bool): specifies if type checking should be done one the data sent to the server. Default is True. _check_return_type (bool): specifies if type checking should be done one the data received from the server. Default is True. _host_index (int/None): specifies the index of the server that we want to use. Default is read from the configuration. async_req (bool): execute request asynchronously Returns: ObjectStatsList If the method is called asynchronously, returns the request thread. """ kwargs['async_req'] = kwargs.get( 'async_req', False ) kwargs['_return_http_data_only'] = kwargs.get( '_return_http_data_only', True ) kwargs['_preload_content'] = kwargs.get( '_preload_content', True ) kwargs['_request_timeout'] = kwargs.get( '_request_timeout', None ) kwargs['_check_input_type'] = kwargs.get( '_check_input_type', True ) kwargs['_check_return_type'] = kwargs.get( '_check_return_type', True ) kwargs['_host_index'] = kwargs.get('_host_index') kwargs['repository'] = \ repository kwargs['ref'] = \ ref return self.call_with_http_info(**kwargs) self.list_objects = _Endpoint( settings={ 'response_type': (ObjectStatsList,), 'auth': [ 'basic_auth', 'cookie_auth', 'jwt_token' ], 'endpoint_path': '/repositories/{repository}/refs/{ref}/objects/ls', 'operation_id': 'list_objects', 'http_method': 'GET', 'servers': None, }, params_map={ 'all': [ 'repository', 'ref', 'prefix', 'after', 'amount', 'delimiter', ], 'required': [ 'repository', 'ref', ], 'nullable': [ ], 'enum': [ ], 'validation': [ 'amount', ] }, root_map={ 'validations': { ('amount',): { 'inclusive_maximum': 1000, 'inclusive_minimum': -1, }, }, 'allowed_values': { }, 'openapi_types': { 'repository': (str,), 'ref': (str,), 'prefix': (str,), 'after': (str,), 'amount': (int,), 'delimiter': (str,), }, 'attribute_map': { 'repository': 'repository', 'ref': 'ref', 'prefix': 'prefix', 'after': 'after', 'amount': 'amount', 'delimiter': 'delimiter', }, 'location_map': { 'repository': 'path', 'ref': 'path', 'prefix': 'query', 'after': 'query', 'amount': 'query', 'delimiter': 'query', }, 'collection_format_map': { } }, headers_map={ 'accept': [ 'application/json' ], 'content_type': [], }, api_client=api_client, callable=__list_objects ) def __stage_object( self, repository, branch, path, object_stage_creation, **kwargs ): """stage an object\"s metadata for the given branch # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.stage_object(repository, branch, path, object_stage_creation, async_req=True) >>> result = thread.get() Args: repository (str): branch (str): path (str): object_stage_creation (ObjectStageCreation): Keyword Args: _return_http_data_only (bool): response data without head status code and headers. Default is True. _preload_content (bool): if False, the urllib3.HTTPResponse object will be returned without reading/decoding response data. Default is True. _request_timeout (float/tuple): timeout setting for this request. If one number provided, it will be total request timeout. It can also be a pair (tuple) of (connection, read) timeouts. Default is None. _check_input_type (bool): specifies if type checking should be done one the data sent to the server. Default is True. _check_return_type (bool): specifies if type checking should be done one the data received from the server. Default is True. _host_index (int/None): specifies the index of the server that we want to use. Default is read from the configuration. async_req (bool): execute request asynchronously Returns: ObjectStats If the method is called asynchronously, returns the request thread. """ kwargs['async_req'] = kwargs.get( 'async_req', False ) kwargs['_return_http_data_only'] = kwargs.get( '_return_http_data_only', True ) kwargs['_preload_content'] = kwargs.get( '_preload_content', True ) kwargs['_request_timeout'] = kwargs.get( '_request_timeout', None ) kwargs['_check_input_type'] = kwargs.get( '_check_input_type', True ) kwargs['_check_return_type'] = kwargs.get( '_check_return_type', True ) kwargs['_host_index'] = kwargs.get('_host_index') kwargs['repository'] = \ repository kwargs['branch'] = \ branch kwargs['path'] = \ path kwargs['object_stage_creation'] = \ object_stage_creation return self.call_with_http_info(**kwargs) self.stage_object = _Endpoint( settings={ 'response_type': (ObjectStats,), 'auth': [ 'basic_auth', 'cookie_auth', 'jwt_token' ], 'endpoint_path': '/repositories/{repository}/branches/{branch}/objects', 'operation_id': 'stage_object', 'http_method': 'PUT', 'servers': None, }, params_map={ 'all': [ 'repository', 'branch', 'path', 'object_stage_creation', ], 'required': [ 'repository', 'branch', 'path', 'object_stage_creation', ], 'nullable': [ ], 'enum': [ ], 'validation': [ ] }, root_map={ 'validations': { }, 'allowed_values': { }, 'openapi_types': { 'repository': (str,), 'branch': (str,), 'path': (str,), 'object_stage_creation': (ObjectStageCreation,), }, 'attribute_map': { 'repository': 'repository', 'branch': 'branch', 'path': 'path', }, 'location_map': { 'repository': 'path', 'branch': 'path', 'path': 'query', 'object_stage_creation': 'body', }, 'collection_format_map': { } }, headers_map={ 'accept': [ 'application/json' ], 'content_type': [ 'application/json' ] }, api_client=api_client, callable=__stage_object ) def __stat_object( self, repository, ref, path, **kwargs ): """get object metadata # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.stat_object(repository, ref, path, async_req=True) >>> result = thread.get() Args: repository (str): ref (str): a reference (could be either a branch or a commit ID) path (str): Keyword Args: _return_http_data_only (bool): response data without head status code and headers. Default is True. _preload_content (bool): if False, the urllib3.HTTPResponse object will be returned without reading/decoding response data. Default is True. _request_timeout (float/tuple): timeout setting for this request. If one number provided, it will be total request timeout. It can also be a pair (tuple) of (connection, read) timeouts. Default is None. _check_input_type (bool): specifies if type checking should be done one the data sent to the server. Default is True. _check_return_type (bool): specifies if type checking should be done one the data received from the server. Default is True. _host_index (int/None): specifies the index of the server that we want to use. Default is read from the configuration. async_req (bool): execute request asynchronously Returns: ObjectStats If the method is called asynchronously, returns the request thread. """ kwargs['async_req'] = kwargs.get( 'async_req', False ) kwargs['_return_http_data_only'] = kwargs.get( '_return_http_data_only', True ) kwargs['_preload_content'] = kwargs.get( '_preload_content', True ) kwargs['_request_timeout'] = kwargs.get( '_request_timeout', None ) kwargs['_check_input_type'] = kwargs.get( '_check_input_type', True ) kwargs['_check_return_type'] = kwargs.get( '_check_return_type', True ) kwargs['_host_index'] = kwargs.get('_host_index') kwargs['repository'] = \ repository kwargs['ref'] = \ ref kwargs['path'] = \ path return self.call_with_http_info(**kwargs) self.stat_object = _Endpoint( settings={ 'response_type': (ObjectStats,), 'auth': [ 'basic_auth', 'cookie_auth', 'jwt_token' ], 'endpoint_path': '/repositories/{repository}/refs/{ref}/objects/stat', 'operation_id': 'stat_object', 'http_method': 'GET', 'servers': None, }, params_map={ 'all': [ 'repository', 'ref', 'path', ], 'required': [ 'repository', 'ref', 'path', ], 'nullable': [ ], 'enum': [ ], 'validation': [ ] }, root_map={ 'validations': { }, 'allowed_values': { }, 'openapi_types': { 'repository': (str,), 'ref': (str,), 'path': (str,), }, 'attribute_map': { 'repository': 'repository', 'ref': 'ref', 'path': 'path', }, 'location_map': { 'repository': 'path', 'ref': 'path', 'path': 'query', }, 'collection_format_map': { } }, headers_map={ 'accept': [ 'application/json' ], 'content_type': [], }, api_client=api_client, callable=__stat_object ) def __upload_object( self, repository, branch, path, **kwargs ): """upload_object # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.upload_object(repository, branch, path, async_req=True) >>> result = thread.get() Args: repository (str): branch (str): path (str): Keyword Args: storage_class (str): [optional] content (file_type): Object content to upload. [optional] _return_http_data_only (bool): response data without head status code and headers. Default is True. _preload_content (bool): if False, the urllib3.HTTPResponse object will be returned without reading/decoding response data. Default is True. _request_timeout (float/tuple): timeout setting for this request. If one number provided, it will be total request timeout. It can also be a pair (tuple) of (connection, read) timeouts. Default is None. _check_input_type (bool): specifies if type checking should be done one the data sent to the server. Default is True. _check_return_type (bool): specifies if type checking should be done one the data received from the server. Default is True. _host_index (int/None): specifies the index of the server that we want to use. Default is read from the configuration. async_req (bool): execute request asynchronously Returns: ObjectStats If the method is called asynchronously, returns the request thread. """ kwargs['async_req'] = kwargs.get( 'async_req', False ) kwargs['_return_http_data_only'] = kwargs.get( '_return_http_data_only', True ) kwargs['_preload_content'] = kwargs.get( '_preload_content', True ) kwargs['_request_timeout'] = kwargs.get( '_request_timeout', None ) kwargs['_check_input_type'] = kwargs.get( '_check_input_type', True ) kwargs['_check_return_type'] = kwargs.get( '_check_return_type', True ) kwargs['_host_index'] = kwargs.get('_host_index') kwargs['repository'] = \ repository kwargs['branch'] = \ branch kwargs['path'] = \ path return self.call_with_http_info(**kwargs) self.upload_object = _Endpoint( settings={ 'response_type': (ObjectStats,), 'auth': [ 'basic_auth', 'cookie_auth', 'jwt_token' ], 'endpoint_path': '/repositories/{repository}/branches/{branch}/objects', 'operation_id': 'upload_object', 'http_method': 'POST', 'servers': None, }, params_map={ 'all': [ 'repository', 'branch', 'path', 'storage_class', 'content', ], 'required': [ 'repository', 'branch', 'path', ], 'nullable': [ ], 'enum': [ ], 'validation': [ ] }, root_map={ 'validations': { }, 'allowed_values': { }, 'openapi_types': { 'repository': (str,), 'branch': (str,), 'path': (str,), 'storage_class': (str,), 'content': (file_type,), }, 'attribute_map': { 'repository': 'repository', 'branch': 'branch', 'path': 'path', 'storage_class': 'storageClass', 'content': 'content', }, 'location_map': { 'repository': 'path', 'branch': 'path', 'path': 'query', 'storage_class': 'query', 'content': 'form', }, 'collection_format_map': { } }, headers_map={ 'accept': [ 'application/json' ], 'content_type': [ 'multipart/form-data' ] }, api_client=api_client, callable=__upload_object )
[ "noreply@github.com" ]
noreply@github.com
a5bc5f827f72fb006bbe2a2d51fbc23feca32582
638842330f186436fb40689cd596ffae6ec4bc3b
/bayesian_shielding/benchmark_tasks/MNLI/mnli_extract_res.py
435539c81f1daea720a3a715e2debdf2062d45d8
[]
no_license
jflotz/BERT-Defense
35d725b909dfaa775217c6d4078283a07cca3e6d
f4b28d03f121b1f88be59d45ea89009416818b4b
refs/heads/master
2023-07-17T12:29:40.768394
2021-08-29T08:31:45
2021-08-29T08:31:45
null
0
0
null
null
null
null
UTF-8
Python
false
false
215
py
import pandas as pd df = pd.read_csv("../../../DATA/mnli/dev_matched.tsv",sep="\t") df = df[["gold_label","sentence1","sentence2"]] df = df.iloc[:200] df.to_csv("mnli_dataset.csv",sep="\t",index=False,header=False)
[ "yannik@kelnet.de" ]
yannik@kelnet.de
297b49422f62295813f98787154517148273d665
a59deecc5d91214601c38bd170605d9d080e06d2
/14-dictionaries/08-copy()/app.py
2a626c1bb68207e6df9b951c1b8fd7d46c37c8b5
[]
no_license
reyeskevin9767/modern-python-bootcamp-2018
a6a3abdb911716d19f6ab516835ed1a04919a13d
d0234f10c4b8aaa6a20555348aec7e3571e3d4e7
refs/heads/master
2022-12-03T18:48:50.035054
2020-08-09T03:00:55
2020-08-09T03:00:55
286,109,881
0
0
null
null
null
null
UTF-8
Python
false
false
214
py
# * Copy Method d = dict(a=1, b=2, c=3) c = d.copy() print(c) # {'a': 1, 'b': 2, 'c': 3} print(c is d) # False e = dict(a=6, b=7, c=8) f = e.copy() print(e) # {'a': 1, 'b': 2, 'c': 3} print(e is f) # False
[ "reyeskevin9767@gmail.com" ]
reyeskevin9767@gmail.com
b220a593afe723680a0f1b9da20f77c60d53e27f
30dd0a3698d06b29800943ab5f0328e3019e5608
/ch2/p9.py
959c9a5e8e707cbeedb746a9d41ae94b4b841f09
[]
no_license
theamankumarsingh/automate-the-boring-stuff-with-python
837e6fc6c42ac672d315a98da6d9be4755e991cc
ecf03bb64970a80ed57572db563c451864f76263
refs/heads/master
2023-02-13T22:24:20.004033
2021-01-08T10:22:59
2021-01-08T10:22:59
324,995,765
5
0
null
null
null
null
UTF-8
Python
false
false
112
py
spam=int(input()) if spam==1: print("Hello") elif spam ==2: print("Howdy") else: print("Greetings!")
[ "amankumarsingh.professional@gmail.com" ]
amankumarsingh.professional@gmail.com
ec76462b01d31414e103dd0bfcd6655ce361b9da
e63676d4a91b6718f4e8333e3e72f02d33fbc9a6
/sdlf-datalakeLibrary/python/datalake_library/tests/unit/stage_b_transforms/test_heavy_transform_blueprint.py
acc41f5c28d772df88e232cdef449c9056a50170
[ "MIT-0" ]
permissive
fnapolitano73/aws-serverless-data-lake-framework
c96f526e97d609e7cf2852ba05ad6e8332e8e98e
c37e3e2e9faee9ee915eb9b6e0919e1cf30c38d8
refs/heads/master
2023-01-04T02:11:30.237458
2020-10-28T08:09:00
2020-10-28T08:09:00
null
0
0
null
null
null
null
UTF-8
Python
false
false
1,276
py
from python.datalake_library.transforms.stage_b_transforms.heavy_transform_blueprint import CustomTransform import pytest import sys import os from pytest import fixture from unittest.mock import patch sys.path.insert(0, os.path.join(os.path.abspath( os.path.dirname(__file__)), '../../../..')) class TestCustomTransform: @staticmethod def test_check_job_status(mocker): # Setup bucket = "test-bucket" keys = 123 processed_keys_path = "test-bucket/files/" job_details = {"jobName": "meteorites-glue-job", "jobRunId": "1"} job_response = { "JobRun": { "jobName": "meteorites-glue-job", "jobRunId": 1, "JobRunState": "RUNNING" } } expected_result = { "processedKeysPath": processed_keys_path, "jobDetails": {"jobName": "meteorites-glue-job", "jobRunId": "1", "jobStatus": "RUNNING"} } mocker.patch("botocore.client.BaseClient._make_api_call", return_value=job_response) # Exercise result = CustomTransform().check_job_status( bucket, keys, processed_keys_path, job_details) # Verify assert result == expected_result
[ "jaidisido@gmail.com" ]
jaidisido@gmail.com
7f54413f5e15fdc43edf6d6d9bbea08094267c57
8e5950aa3aaf94cec94d7c53d55b06275e7b109f
/modules/fabric/Manifest.py
91a3af77ce234b924a7d098aa6f8059bb057e7c5
[]
no_license
stefanrauch/wr_cores
a129ff2ef60d3552f6bce6ca7b309f2bd851ba77
b9a0ed7f6878b0e3c8332efb9e8d93c2ec018a3e
refs/heads/master
2021-01-10T18:59:58.191100
2013-01-21T15:54:23
2013-01-21T15:54:23
7,622,900
2
0
null
null
null
null
UTF-8
Python
false
false
77
py
files = ["wr_fabric_pkg.vhd", "xwb_fabric_sink.vhd", "xwb_fabric_source.vhd"]
[ "c.prados@gsi.de" ]
c.prados@gsi.de
bd7d1491e809be7611d09d0d0e8578f497fb3520
e811da3715d43e23a4548490aa27be40ac21d6e4
/handlers/base/__init__.py
8f1904288c671963f969ea59e55106edced6d3da
[]
no_license
atiger808/tornado
2a2ff73957d6fb97cd91222038f499ee8ed325f5
77e981ee70a7c7b3903bec82d91109f163bb2a43
refs/heads/master
2020-04-04T09:22:07.007710
2018-11-02T05:04:00
2018-11-02T05:04:00
155,815,465
0
0
null
null
null
null
UTF-8
Python
false
false
149
py
# _*_ coding: utf-8 _*_ # @Time : 2018/6/26 22:52 # @Author : Ole211 # @Site : # @File : __init__.py.py # @Software : PyCharm
[ "atiger0614@163.com" ]
atiger0614@163.com
8d78d90dadbede0d5810dabceead66a19ea45ef8
dc9dc62a5012f5be638ab481cf201bf58020cfda
/BigData/spark_stream.py
aac15a6b99fdeef516abc210229a64099cf9486a
[]
no_license
suryknt/Practice
d208a536cb7b1f290e1994047a984ca9e1a1eeb6
07f2677f0662e0a9a180bb0794ea513f0595e639
refs/heads/master
2021-01-12T09:55:12.818742
2019-05-06T21:59:07
2019-05-06T21:59:07
76,297,468
0
0
null
2019-05-06T21:55:40
2016-12-12T21:34:08
Matlab
UTF-8
Python
false
false
2,549
py
from __future__ import print_function from pyspark import SparkContext from pyspark.streaming import StreamingContext from pyspark.sql import SQLContext from pyspark.sql.functions import desc from collections import namedtuple import json import time import sys from pyspark.sql.types import Row from pyspark.sql.functions import desc sc=SparkContext(appName="MyTwitterCount") sc.setLogLevel("ERROR") windowInterval = 10 ssc=StreamingContext(sc,windowInterval) sqlContext = SQLContext(sc) # ssc.checkpoint( "C:/Projects/machine_learning/Rec-Eng/checkpoint") tweetDstream=ssc.socketTextStream("172.16.99.228",5555) # lines = tweetDstream.window( 20 ) def extractTweetText(tweetJson,doprint=False): if not tweetJson: tweetJson="" if doprint: print("$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$tweet$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$") print(tweetJson) print("$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$tweet$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$") return tweetJson else: return tweetJson print(sqlContext) TagCount = namedtuple("TagCount", ("tag","count")) fields = ("tag", "count" ) Tweet = namedtuple( 'Tweet', fields ) try: ( tweetDstream.map(lambda tweet: extractTweetText(tweet)) .flatMap(lambda text: text.split(" ")) .filter(lambda word: word.lower().startswith("#")) .map(lambda word: (word.lower(), 1)) .reduceByKey(lambda a, b: a + b) .map(lambda rec: Tweet(rec[0], rec[1])) .foreachRDD(lambda rdd: rdd.toDF().sort(desc("count")) .limit(10).registerTempTable("tweets") if not rdd.isEmpty() else print("")) # .flatMap(lambda text: text.split(" ")) # .filter(lambda word: word.startswith("#")) # .map(lambda word: word.lower(),1) # .reduceByKey(lambda a,b: a+b) # .map(lambda rec: TagCount(rec[0],rec[1])) # .foreachRDD(lambda rdd: rdd.toDF()) ) except BaseException as e: print("Error while processing: %s" % str(e)) ssc.start() print(sqlContext) count=0 while count < 100: time.sleep(15) count += 1 try: top_10=sqlContext.sql("select tag, count from tweets order by count") for row in top_10.collect(): print(row.tag,row["count"]) print("-------------------------------------") except BaseException as e: print("-------No Hashtags-------") ssc.awaitTermination()
[ "suryknt@gmail.com" ]
suryknt@gmail.com
c69a96ad0afddc0785d9a27aa65ecf1e913caeb7
21240a40e4be1c88a9a3a23cb498d48a51e8ee80
/angular-tour-of-heroes/node_modules/uws/build/config.gypi
d0242912952f96a3e54ebacf17bb4e255ab3bee8
[ "Zlib" ]
permissive
KevinDackow/AngularDemo
7c77bea70b056a64e37ad086338c2b172c1545ba
a307f6569aade23e5b83ad5d467dff3163ff9e44
refs/heads/master
2020-03-11T21:09:23.205304
2018-04-25T01:38:53
2018-04-25T01:38:53
130,257,095
0
0
null
null
null
null
UTF-8
Python
false
false
5,092
gypi
# Do not edit. File was generated by node-gyp's "configure" step { "target_defaults": { "cflags": [], "default_configuration": "Release", "defines": [], "include_dirs": [], "libraries": [] }, "variables": { "asan": 0, "coverage": "false", "debug_devtools": "node", "debug_http2": "false", "debug_nghttp2": "false", "force_dynamic_crt": 0, "gas_version": "2.23", "host_arch": "x64", "icu_data_file": "icudt60l.dat", "icu_data_in": "../../deps/icu-small/source/data/in/icudt60l.dat", "icu_endianness": "l", "icu_gyp_path": "tools/icu/icu-generic.gyp", "icu_locales": "en,root", "icu_path": "deps/icu-small", "icu_small": "true", "icu_ver_major": "60", "llvm_version": 0, "node_byteorder": "little", "node_enable_d8": "false", "node_enable_v8_vtunejit": "false", "node_install_npm": "true", "node_module_version": 59, "node_no_browser_globals": "false", "node_prefix": "/", "node_release_urlbase": "https://nodejs.org/download/release/", "node_shared": "false", "node_shared_cares": "false", "node_shared_http_parser": "false", "node_shared_libuv": "false", "node_shared_nghttp2": "false", "node_shared_openssl": "false", "node_shared_zlib": "false", "node_tag": "", "node_use_bundled_v8": "true", "node_use_dtrace": "false", "node_use_etw": "false", "node_use_lttng": "false", "node_use_openssl": "true", "node_use_perfctr": "false", "node_use_v8_platform": "true", "node_without_node_options": "false", "openssl_fips": "", "openssl_no_asm": 0, "shlib_suffix": "so.59", "target_arch": "x64", "uv_parent_path": "/deps/uv/", "uv_use_dtrace": "false", "v8_enable_gdbjit": 0, "v8_enable_i18n_support": 1, "v8_enable_inspector": 1, "v8_no_strict_aliasing": 1, "v8_optimized_debug": 0, "v8_promise_internal_field_count": 1, "v8_random_seed": 0, "v8_trace_maps": 0, "v8_use_snapshot": "true", "want_separate_host_toolset": 0, "nodedir": "/home/kevin/.node-gyp/9.4.0", "standalone_static_library": 1, "cache_lock_stale": "60000", "ham_it_up": "", "legacy_bundling": "", "sign_git_tag": "", "user_agent": "npm/5.6.0 node/v9.4.0 linux x64", "always_auth": "", "bin_links": "true", "key": "", "allow_same_version": "", "description": "true", "fetch_retries": "2", "heading": "npm", "if_present": "", "init_version": "1.0.0", "user": "", "prefer_online": "", "force": "", "only": "", "read_only": "", "cache_min": "10", "init_license": "ISC", "editor": "vi", "rollback": "true", "tag_version_prefix": "v", "cache_max": "Infinity", "timing": "", "userconfig": "/home/kevin/.npmrc", "engine_strict": "", "init_author_name": "", "init_author_url": "", "tmp": "/tmp", "depth": "Infinity", "package_lock_only": "", "save_dev": "", "usage": "", "metrics_registry": "https://registry.npmjs.org/", "otp": "", "package_lock": "true", "progress": "true", "https_proxy": "", "save_prod": "", "cidr": "", "onload_script": "", "sso_type": "oauth", "rebuild_bundle": "true", "save_bundle": "", "shell": "/bin/bash", "dry_run": "", "prefix": "/usr/local", "scope": "", "browser": "", "cache_lock_wait": "10000", "ignore_prepublish": "", "registry": "https://registry.npmjs.org/", "save_optional": "", "searchopts": "", "versions": "", "cache": "/home/kevin/.npm", "send_metrics": "", "global_style": "", "ignore_scripts": "", "version": "", "local_address": "", "viewer": "man", "node_gyp": "/usr/local/lib/node_modules/npm/node_modules/node-gyp/bin/node-gyp.js", "prefer_offline": "", "color": "true", "fetch_retry_mintimeout": "10000", "maxsockets": "50", "offline": "", "sso_poll_frequency": "500", "umask": "0002", "fetch_retry_maxtimeout": "60000", "logs_max": "10", "message": "%s", "ca": "", "cert": "", "global": "", "link": "", "access": "", "also": "", "save": "true", "unicode": "true", "long": "", "production": "", "searchlimit": "20", "unsafe_perm": "true", "auth_type": "legacy", "node_version": "9.4.0", "tag": "latest", "git_tag_version": "true", "commit_hooks": "true", "script_shell": "", "shrinkwrap": "true", "fetch_retry_factor": "10", "save_exact": "", "strict_ssl": "true", "dev": "", "globalconfig": "/usr/local/etc/npmrc", "init_module": "/home/kevin/.npm-init.js", "parseable": "", "globalignorefile": "/usr/local/etc/npmignore", "cache_lock_retries": "10", "searchstaleness": "900", "node_options": "", "save_prefix": "^", "scripts_prepend_node_path": "warn-only", "group": "1000", "init_author_email": "", "searchexclude": "", "git": "git", "optional": "true", "json": "" } }
[ "Kevin Dackow" ]
Kevin Dackow
3117d59bf6629c3dce3f90abcf05f1855a34cce8
24735767f7d585a8d7d3055b4967578e7f55a715
/pontos_turisticos/urls.py
ac1d500c026745f66f0665db4a29c2fb0042e224
[]
no_license
nilton-medeiros/pontos_turisticos
b6c978ec0259163feccf8398a7ea927d34bba352
c642c2c834a9f2594ac680a9048d6d70aaca70d6
refs/heads/main
2023-02-12T05:16:12.239114
2021-01-13T14:02:28
2021-01-13T14:02:28
327,097,453
1
0
null
2021-01-05T21:42:17
2021-01-05T19:23:27
Python
UTF-8
Python
false
false
1,687
py
"""pontos_turisticos URL Configuration The `urlpatterns` list routes URLs to views. For more information please see: https://docs.djangoproject.com/en/3.1/topics/http/urls/ Examples: Function views 1. Add an import: from my_app import views 2. Add a URL to urlpatterns: path('', views.home, name='home') Class-based views 1. Add an import: from other_app.views import Home 2. Add a URL to urlpatterns: path('', Home.as_view(), name='home') Including another URLconf 1. Import the include() function: from django.urls import include, path 2. Add a URL to urlpatterns: path('blog/', include('blog.urls')) """ from django.contrib import admin from django.conf.urls import include from django.urls import path from rest_framework import routers from django.conf import settings from django.conf.urls.static import static from core.api.viewsets import PontoTuristicoViewSet from atracoes.api.viewsets import AtracaoViewSet from enderecos.api.viewsets import EnderecoViewSet from comentarios.api.viewsets import ComentarioViewSet from avaliacoes.api.viewsets import AvaliacaoViewSet from rest_framework.authtoken.views import obtain_auth_token router = routers.DefaultRouter() router.register(r'pontoturistico', PontoTuristicoViewSet, basename='PontoTuristico') router.register(r'atracoes', AtracaoViewSet) router.register(r'enderecos', EnderecoViewSet) router.register(r'comentarios', ComentarioViewSet) router.register(r'avaliacoes', AvaliacaoViewSet) urlpatterns = [ path('', include(router.urls)), path('admin/', admin.site.urls), path('api-token-auth/', obtain_auth_token), ] + static(settings.MEDIA_URL, document_root=settings.MEDIA_ROOT)
[ "nilton@sistrom.com.br" ]
nilton@sistrom.com.br
c8134ac09fd408a7e6a95afc096c4b2a6a04af17
c6770d1d1bf408cf14ff6c83402726f6c4f4a8f5
/mediafiles/utils.py
40688bb19362ae105e49f52448291213c98df8f5
[]
no_license
nasaastrobio/django-mediafiles
f8825a988b8d04cfd8ce83c1274c93b7961cb50f
a69082108cbecdf01593ab920c4f6efeb99253de
refs/heads/master
2021-05-27T23:22:53.657945
2013-09-28T14:59:24
2013-09-28T14:59:24
13,167,434
0
0
null
null
null
null
UTF-8
Python
false
false
1,970
py
import os import fnmatch from django.conf import settings from django.core.exceptions import ImproperlyConfigured def matches_patterns(path, patterns=None): """ Return True or False depending on whether the ``path`` should be ignored (if it matches any pattern in ``ignore_patterns``). """ if patterns is None: patterns = [] for pattern in patterns: if fnmatch.fnmatchcase(path, pattern): return True return False def get_files(storage, ignore_patterns=None, location=''): """ Recursively walk the storage directories yielding the paths of all files that should be copied. """ if ignore_patterns is None: ignore_patterns = [] directories, files = storage.listdir(location) for fn in files: if matches_patterns(fn, ignore_patterns): continue if location: fn = os.path.join(location, fn) yield fn for dir in directories: if matches_patterns(dir, ignore_patterns): continue if location: dir = os.path.join(location, dir) for fn in get_files(storage, ignore_patterns, dir): yield fn def check_settings(base_url=None): """ Checks if the mediafiles settings have sane values. """ if base_url is None: base_url = settings.MEDIA_URL if not base_url: raise ImproperlyConfigured( "You're using the mediafiles app " "without having set the required MEDIA_URL setting.") if settings.STATIC_URL == base_url: raise ImproperlyConfigured("The STATIC_URL and MEDIA_URL " "settings must have different values") if ((settings.STATIC_ROOT and settings.MEDIA_ROOT) and (settings.STATIC_ROOT == settings.MEDIA_ROOT)): raise ImproperlyConfigured("The STATIC_ROOT and MEDIA_ROOT " "settings must have different values")
[ "shige.abe@nasa.gov" ]
shige.abe@nasa.gov
5a08afb6ae2260295558f1669b8a1893f16384b4
f5b2748ead8c589201afa209eb50af789e441987
/zxjy/manage.py
13f9a9b0d4b50bd25a2a30d8fae758949a6e9706
[]
no_license
x17246758/NetCourseSystem
9839478486b3f2d97a7a9f7917cd00e6b5dd344d
3b083d400e4bee92c0a9ff44e42e12b9731b8ab0
refs/heads/master
2022-01-28T01:05:59.484122
2019-10-28T15:01:55
2019-10-28T15:01:55
209,337,495
0
0
null
2022-01-15T05:28:00
2019-09-18T15:04:29
JavaScript
UTF-8
Python
false
false
536
py
#!/usr/bin/env python import os import sys if __name__ == '__main__': os.environ.setdefault('DJANGO_SETTINGS_MODULE', 'zxjy.settings') try: from django.core.management import execute_from_command_line except ImportError as exc: raise ImportError( "Couldn't import Django. Are you sure it's installed and " "available on your PYTHONPATH environment variable? Did you " "forget to activate a virtual environment?" ) from exc execute_from_command_line(sys.argv)
[ "172467548@qq.com" ]
172467548@qq.com
a9b098aaf599f218d0e3b35cae1d246bcbeb2c50
a66b69c3f9da9779ae80f347b61f47e3bc5ba145
/day1002/A04_loop.py
311112630c8c83899668600713293b1a7f31e1f9
[]
no_license
kyungtae92/python-basic
c841d9c9c6196b01da3de007c1298fe2c4b8f693
80a2051e37b6e87c9dbfd332c4b2946089ff0d5c
refs/heads/master
2020-11-25T08:01:22.156661
2019-12-17T08:25:38
2019-12-17T08:25:38
228,567,120
2
0
null
null
null
null
UTF-8
Python
false
false
352
py
import os # 파이썬이 운영체제의 일부 기능 가져옴(명령어) while (True): dan = input('input gugudan >> ') if dan.isalpha() == True or dan == '': os.system('cls') else: break dan = int(dan) i = 0 for i in range(1, 10): # for i in range(1, 10, 1): print("%d * %d = %2d" % (dan, i, dan * i))
[ "noreply@github.com" ]
noreply@github.com
27cc4cebf599c8d3b7a61be91fd2e525d3304487
6d967da5fd95aa5e66ddbb211da40041006ca5ec
/myvenv/Lib/site-packages/pip/_vendor/packaging/markers.py
8ef134ba7b10dc55e4de37dd77c217c87ff3f97e
[ "Apache-2.0", "BSD-3-Clause" ]
permissive
gevorkyannaira/my-first-blog
96e4458045a1dd0aa9c1f3ec69f4c829428200e0
42ab12a8c2b0e402b5fa1b8e5a7cdd2629d06c16
refs/heads/master
2022-09-03T21:14:18.946448
2020-05-18T18:15:39
2020-05-18T18:15:39
264,909,108
0
0
null
null
null
null
UTF-8
Python
false
false
11,735
py
# This file is dual licensed under the terms of the Apache License, Version # 2.0, and the BSD License. See the LICENSE file in the root of this repository # for complete details. from __future__ import absolute_import, division, print_function import operator import os import platform import sys from pip._vendor.pyparsing import ParseException, ParseResults, stringStart, stringEnd from pip._vendor.pyparsing import ZeroOrMore, Group, Forward, QuotedString from pip._vendor.pyparsing import Literal as L # noqa from ._compat import string_types <<<<<<< HEAD from .specifiers import Specifier, InvalidSpecifier ======= from ._typing import MYPY_CHECK_RUNNING from .specifiers import Specifier, InvalidSpecifier if MYPY_CHECK_RUNNING: # pragma: no cover from typing import Any, Callable, Dict, List, Optional, Tuple, Union Operator = Callable[[str, str], bool] >>>>>>> e585743114c1741ec20dc76010f96171f3516589 __all__ = [ "InvalidMarker", "UndefinedComparison", "UndefinedEnvironmentName", "Marker", "default_environment", ] class InvalidMarker(ValueError): """ An invalid marker was found, users should refer to PEP 508. """ class UndefinedComparison(ValueError): """ An invalid operation was attempted on a value that doesn't support it. """ class UndefinedEnvironmentName(ValueError): """ A name was attempted to be used that does not exist inside of the environment. """ class Node(object): def __init__(self, value): <<<<<<< HEAD self.value = value def __str__(self): return str(self.value) def __repr__(self): return "<{0}({1!r})>".format(self.__class__.__name__, str(self)) def serialize(self): ======= # type: (Any) -> None self.value = value def __str__(self): # type: () -> str return str(self.value) def __repr__(self): # type: () -> str return "<{0}({1!r})>".format(self.__class__.__name__, str(self)) def serialize(self): # type: () -> str >>>>>>> e585743114c1741ec20dc76010f96171f3516589 raise NotImplementedError class Variable(Node): def serialize(self): <<<<<<< HEAD ======= # type: () -> str >>>>>>> e585743114c1741ec20dc76010f96171f3516589 return str(self) class Value(Node): def serialize(self): <<<<<<< HEAD ======= # type: () -> str >>>>>>> e585743114c1741ec20dc76010f96171f3516589 return '"{0}"'.format(self) class Op(Node): def serialize(self): <<<<<<< HEAD ======= # type: () -> str >>>>>>> e585743114c1741ec20dc76010f96171f3516589 return str(self) VARIABLE = ( L("implementation_version") | L("platform_python_implementation") | L("implementation_name") | L("python_full_version") | L("platform_release") | L("platform_version") | L("platform_machine") | L("platform_system") | L("python_version") | L("sys_platform") | L("os_name") <<<<<<< HEAD | L("os.name") ======= | L("os.name") # PEP-345 >>>>>>> e585743114c1741ec20dc76010f96171f3516589 | L("sys.platform") # PEP-345 | L("platform.version") # PEP-345 | L("platform.machine") # PEP-345 | L("platform.python_implementation") # PEP-345 <<<<<<< HEAD | L("python_implementation") # PEP-345 | L("extra") # undocumented setuptools legacy ======= | L("python_implementation") # undocumented setuptools legacy | L("extra") # PEP-508 >>>>>>> e585743114c1741ec20dc76010f96171f3516589 ) ALIASES = { "os.name": "os_name", "sys.platform": "sys_platform", "platform.version": "platform_version", "platform.machine": "platform_machine", "platform.python_implementation": "platform_python_implementation", "python_implementation": "platform_python_implementation", } VARIABLE.setParseAction(lambda s, l, t: Variable(ALIASES.get(t[0], t[0]))) VERSION_CMP = ( L("===") | L("==") | L(">=") | L("<=") | L("!=") | L("~=") | L(">") | L("<") ) MARKER_OP = VERSION_CMP | L("not in") | L("in") MARKER_OP.setParseAction(lambda s, l, t: Op(t[0])) MARKER_VALUE = QuotedString("'") | QuotedString('"') MARKER_VALUE.setParseAction(lambda s, l, t: Value(t[0])) BOOLOP = L("and") | L("or") MARKER_VAR = VARIABLE | MARKER_VALUE MARKER_ITEM = Group(MARKER_VAR + MARKER_OP + MARKER_VAR) MARKER_ITEM.setParseAction(lambda s, l, t: tuple(t[0])) LPAREN = L("(").suppress() RPAREN = L(")").suppress() MARKER_EXPR = Forward() MARKER_ATOM = MARKER_ITEM | Group(LPAREN + MARKER_EXPR + RPAREN) MARKER_EXPR << MARKER_ATOM + ZeroOrMore(BOOLOP + MARKER_EXPR) MARKER = stringStart + MARKER_EXPR + stringEnd def _coerce_parse_result(results): <<<<<<< HEAD ======= # type: (Union[ParseResults, List[Any]]) -> List[Any] >>>>>>> e585743114c1741ec20dc76010f96171f3516589 if isinstance(results, ParseResults): return [_coerce_parse_result(i) for i in results] else: return results def _format_marker(marker, first=True): <<<<<<< HEAD ======= # type: (Union[List[str], Tuple[Node, ...], str], Optional[bool]) -> str >>>>>>> e585743114c1741ec20dc76010f96171f3516589 assert isinstance(marker, (list, tuple, string_types)) # Sometimes we have a structure like [[...]] which is a single item list # where the single item is itself it's own list. In that case we want skip # the rest of this function so that we don't get extraneous () on the # outside. if ( isinstance(marker, list) and len(marker) == 1 and isinstance(marker[0], (list, tuple)) ): return _format_marker(marker[0]) if isinstance(marker, list): inner = (_format_marker(m, first=False) for m in marker) if first: return " ".join(inner) else: return "(" + " ".join(inner) + ")" elif isinstance(marker, tuple): return " ".join([m.serialize() for m in marker]) else: return marker _operators = { "in": lambda lhs, rhs: lhs in rhs, "not in": lambda lhs, rhs: lhs not in rhs, "<": operator.lt, "<=": operator.le, "==": operator.eq, "!=": operator.ne, ">=": operator.ge, ">": operator.gt, <<<<<<< HEAD } def _eval_op(lhs, op, rhs): ======= } # type: Dict[str, Operator] def _eval_op(lhs, op, rhs): # type: (str, Op, str) -> bool >>>>>>> e585743114c1741ec20dc76010f96171f3516589 try: spec = Specifier("".join([op.serialize(), rhs])) except InvalidSpecifier: pass else: return spec.contains(lhs) <<<<<<< HEAD oper = _operators.get(op.serialize()) ======= oper = _operators.get(op.serialize()) # type: Optional[Operator] >>>>>>> e585743114c1741ec20dc76010f96171f3516589 if oper is None: raise UndefinedComparison( "Undefined {0!r} on {1!r} and {2!r}.".format(op, lhs, rhs) ) return oper(lhs, rhs) <<<<<<< HEAD _undefined = object() def _get_env(environment, name): value = environment.get(name, _undefined) if value is _undefined: ======= class Undefined(object): pass _undefined = Undefined() def _get_env(environment, name): # type: (Dict[str, str], str) -> str value = environment.get(name, _undefined) # type: Union[str, Undefined] if isinstance(value, Undefined): >>>>>>> e585743114c1741ec20dc76010f96171f3516589 raise UndefinedEnvironmentName( "{0!r} does not exist in evaluation environment.".format(name) ) return value def _evaluate_markers(markers, environment): <<<<<<< HEAD groups = [[]] ======= # type: (List[Any], Dict[str, str]) -> bool groups = [[]] # type: List[List[bool]] >>>>>>> e585743114c1741ec20dc76010f96171f3516589 for marker in markers: assert isinstance(marker, (list, tuple, string_types)) if isinstance(marker, list): groups[-1].append(_evaluate_markers(marker, environment)) elif isinstance(marker, tuple): lhs, op, rhs = marker if isinstance(lhs, Variable): lhs_value = _get_env(environment, lhs.value) rhs_value = rhs.value else: lhs_value = lhs.value rhs_value = _get_env(environment, rhs.value) groups[-1].append(_eval_op(lhs_value, op, rhs_value)) else: assert marker in ["and", "or"] if marker == "or": groups.append([]) return any(all(item) for item in groups) def format_full_version(info): <<<<<<< HEAD ======= # type: (sys._version_info) -> str >>>>>>> e585743114c1741ec20dc76010f96171f3516589 version = "{0.major}.{0.minor}.{0.micro}".format(info) kind = info.releaselevel if kind != "final": version += kind[0] + str(info.serial) return version def default_environment(): <<<<<<< HEAD if hasattr(sys, "implementation"): iver = format_full_version(sys.implementation.version) implementation_name = sys.implementation.name ======= # type: () -> Dict[str, str] if hasattr(sys, "implementation"): # Ignoring the `sys.implementation` reference for type checking due to # mypy not liking that the attribute doesn't exist in Python 2.7 when # run with the `--py27` flag. iver = format_full_version(sys.implementation.version) # type: ignore implementation_name = sys.implementation.name # type: ignore >>>>>>> e585743114c1741ec20dc76010f96171f3516589 else: iver = "0" implementation_name = "" return { "implementation_name": implementation_name, "implementation_version": iver, "os_name": os.name, "platform_machine": platform.machine(), "platform_release": platform.release(), "platform_system": platform.system(), "platform_version": platform.version(), "python_full_version": platform.python_version(), "platform_python_implementation": platform.python_implementation(), <<<<<<< HEAD "python_version": platform.python_version()[:3], ======= "python_version": ".".join(platform.python_version_tuple()[:2]), >>>>>>> e585743114c1741ec20dc76010f96171f3516589 "sys_platform": sys.platform, } class Marker(object): def __init__(self, marker): <<<<<<< HEAD ======= # type: (str) -> None >>>>>>> e585743114c1741ec20dc76010f96171f3516589 try: self._markers = _coerce_parse_result(MARKER.parseString(marker)) except ParseException as e: err_str = "Invalid marker: {0!r}, parse error at {1!r}".format( marker, marker[e.loc : e.loc + 8] ) raise InvalidMarker(err_str) def __str__(self): <<<<<<< HEAD return _format_marker(self._markers) def __repr__(self): return "<Marker({0!r})>".format(str(self)) def evaluate(self, environment=None): ======= # type: () -> str return _format_marker(self._markers) def __repr__(self): # type: () -> str return "<Marker({0!r})>".format(str(self)) def evaluate(self, environment=None): # type: (Optional[Dict[str, str]]) -> bool >>>>>>> e585743114c1741ec20dc76010f96171f3516589 """Evaluate a marker. Return the boolean from evaluating the given marker against the environment. environment is an optional argument to override all or part of the determined environment. The environment is determined from the current Python process. """ current_environment = default_environment() if environment is not None: current_environment.update(environment) return _evaluate_markers(self._markers, current_environment)
[ "gevorkyannaira5@gmail.com" ]
gevorkyannaira5@gmail.com
d719400ea7743fa07dd8cb24e3cf0ff8bd3dc1a0
f14a0ef8364953e4fa18d494ce62d8a3b73c263b
/mkdir_and_check_file_type_python27.py
506898d6e43d9fd72acaa89522c45c56261d41a2
[]
no_license
gptcod/python_call_so
d866b01c4489887e29d2ed04211e636ca0b5b279
4816455f6f2b81ccc03ad94fad4ff67f16da83ba
refs/heads/master
2023-08-21T16:52:52.671500
2017-12-18T08:19:48
2017-12-18T08:19:48
null
0
0
null
null
null
null
UTF-8
Python
false
false
1,095
py
#!/usr/bin/python # -*- coding: utf-8 -*- from ctypes import * import ctypes import subprocess import re def mkdir_officex(): ls_log = subprocess.check_output(['ls', '-l', '/home/venus/apt/cloud/officextemp/']) if ls_log.find("cannot access") != -1: result = subprocess.check_output(['mkdir', '-p', '/home/venus/apt/cloud/officextemp/']) print result def get_file_type(so_file_path, check_file_path): methods = subprocess.check_output(['nm', '-D', 'libfiltertype.so']) pattern = re.compile(r'(_.*checktype[A-Z].*)') checktype_method = pattern.findall(methods)[0] so_file = cdll.LoadLibrary(so_file_path) with open(check_file_path) as file: data = file.read() data_list = list(data) data_array = (ctypes.c_char * len(data_list))(*data_list) p = create_string_buffer(10) check_file_name = check_file_path.split("/")[-1] so_file[checktype_method](byref(data_array), len(data_list), p, check_file_name) filetype = "" for i in p.raw: if (ord)(i) != 0: filetype += i return filetype #print get_file_type("./libfiltertype.so", "./new.txt") print mkdir_officex()
[ "liuyang8@venustech.com.cn" ]
liuyang8@venustech.com.cn
00e11d2488cdcb01be07386274adfad59acacc43
0cbf36f06f5316326ef635f14c887cd2849800db
/typings/celery/app/registry.pyi
33985b1be0f526d3403d3531c9b515b239c0b430
[ "Apache-2.0" ]
permissive
espritgames/celery_types
b59545a7cd28f06e766a1a520590f3bbc155e82f
4d4064eb78d2a1a3e79a5fefe111f59ad4d3c9b9
refs/heads/main
2023-08-18T20:11:33.992509
2021-10-04T11:21:49
2021-10-04T11:21:49
null
0
0
null
null
null
null
UTF-8
Python
false
false
70
pyi
from typing import Any, Dict class TaskRegistry(Dict[str, Any]): ...
[ "steve@dignam.xyz" ]
steve@dignam.xyz
3dc1b7fb3e9705f1aee1008982293a15c22b2d90
c17d0a888d43f6e45d78aecb2ebff7d119e5470e
/main.py
b08fc8aa9d5636b6baffaa83c579939d73f78354
[]
no_license
coolsandeee/TFDSorting
308c5bbecbb59ea19c960dc283aad4b7e7ec1883
2299b9466b93b7d7c5f45d4b26b0e44bd371c4c0
refs/heads/main
2023-03-23T15:12:54.493045
2021-03-15T10:27:32
2021-03-15T10:27:32
347,923,692
0
0
null
null
null
null
UTF-8
Python
false
false
666
py
def array_sort(request): content_type = request.headers['content-type'] if content_type == 'application/json': request_json = request.get_json(silent=True) if request_json and 'a' in request_json: a = request_json['a'] else: raise ValueError("JSON is invalid, or missing 'a' property array") if request_json and 'b' in request_json: b = request_json['b'] else: raise ValueError("JSON is invalid, or missing 'b' property array") else: raise ValueError("Expecting a json format") new=a+b new1 = sorted(new) return str(new1)
[ "noreply@github.com" ]
noreply@github.com
2f11b0f81351e4f628d1266ab215c514e432d2f2
7b0413547fb0e4766febcc6a7f0010fafe025fb6
/medium/course_schedule.py
52ca3f20847247a445eb480dcaa842522eed1cac
[]
no_license
theeric80/LeetCode
b00d4bace7c48c409bc6b2f57321aea7b7106f35
e05321d8c2143d35279136d3999e1be1e7005690
refs/heads/master
2021-01-19T00:51:20.608326
2016-06-30T05:32:44
2016-06-30T05:32:44
42,165,229
0
0
null
null
null
null
UTF-8
Python
false
false
2,037
py
class UndirectedGraphNode(object): def __init__(self, x): self.label = x self.neighbors = [] class Solution(object): def canFinish(self, numCourses, prerequisites): """ :type numCourses: int :type prerequisites: List[List[int]] :rtype: bool """ self.cycle = False G = self.buildGraph(numCourses, prerequisites) result, marked, on_stack = [], [False]*len(G), [False]*len(G) for v in G: if not marked[v.label]: self.topological_sort(G, v, marked, on_stack, result) result.reverse() return not self.cycle def buildGraph(self, numCourses, prerequisites): G = [UndirectedGraphNode(i) for i in xrange(numCourses)] for u, v in prerequisites: G[u].neighbors.append(G[v]) return G def topological_sort(self, G, v, marked, on_stack, result): label = v.label marked[label] = True on_stack[label] = True for w in v.neighbors: if self.cycle: return if not marked[w.label]: self.topological_sort(G, w, marked, on_stack, result) elif on_stack[w.label]: self.cycle = True on_stack[label] = False result.append(label) def dfs(self, G, v): result, marked = [], [False]*len(G) s = [v] while s: node = s.pop() label = node.label if not marked[label]: marked[label] = True result.append(label) for neighbor in node.neighbors: s.append(neighbor) print '->'.join(str(i) for i in result) def main(): import sys from os.path import join, abspath sys.path.append(join('..', 'common')) inputs = [(2, [[1,0]])] for numCourses, prerequisites in inputs: result = Solution().canFinish(numCourses, prerequisites) print result if __name__ == '__main__': main()
[ "chunchieh@gmail.com" ]
chunchieh@gmail.com
0e8122a8eb0ba5e509d0b49d4d9aa565da10bc4e
1ad6d91e4454294427d37d5dbfa5b38dab6242e9
/scripts/margin_flipped_mnist.py
bffc4ca93b42a7ad322e273821a63e7f6724b91a
[ "Apache-2.0" ]
permissive
amodas/hold-me-tight
3955bf602ebbb145e1a8207da6a64760a11bc722
b893e97f0b5fe8100472ac68d715d0cb99d0c7dc
refs/heads/main
2022-12-28T14:26:27.474566
2020-10-12T11:06:21
2020-10-12T11:06:21
null
0
0
null
null
null
null
UTF-8
Python
false
false
3,644
py
import numpy as np import torch import torch.nn as nn import os import time from utils import get_processed_dataset_loaders from utils import train from utils import generate_subspace_list from utils import compute_margin_distribution from utils_dct import dct_flip from model_classes import TransformFlippedLayer from model_classes.mnist import LeNet # check inside the model_class.mnist package for other network options TREE_ROOT = './' DEVICE = 'cuda' if torch.cuda.is_available() else 'cpu' DATASET = 'MNIST' PRETRAINED = True PRETRAINED_PATH = '../Models/Pretrained/MNIST_flipped/LeNet/model.t7' BATCH_SIZE = 128 ############################# # Dataset paths and loaders # ############################# # Specify the path of the dataset. For MNIST and CIFAR-10 the train and validation paths can be the same. # For ImageNet, please specify to proper train and validation paths. DATASET_DIR = {'train': os.path.join(TREE_ROOT, '../Datasets/'), 'val': os.path.join(TREE_ROOT, '../Datasets/') } os.makedirs(DATASET_DIR['train'], exist_ok=True) os.makedirs(DATASET_DIR['val'], exist_ok=True) # Load the data trainloader, testloader, trainset, testset, mean, std, _, _ = get_processed_dataset_loaders(lambda x: dct_flip(x), DATASET, DATASET_DIR, BATCH_SIZE) #################### # Select a Network # #################### # Normalization layer flip_trans = TransformFlippedLayer(mean, std, [1, 28, 28], DEVICE) # Load a model model = LeNet() # check inside the model_class.mnist package for other network options # If pretrained if PRETRAINED: print('---> Working on a pretrained network') model.load_state_dict(torch.load(PRETRAINED_PATH, map_location='cpu')) model = model.to(DEVICE) model.eval() # If not pretrained, then train it if not PRETRAINED: EPOCHS = 30 MAX_LR = 0.21 MOMENTUM = 0.9 WEIGHT_DECAY = 5e-4 opt = torch.optim.SGD(model.parameters(), lr=MAX_LR, momentum=MOMENTUM, weight_decay=WEIGHT_DECAY) loss_fun = nn.CrossEntropyLoss() lr_schedule = lambda t: np.interp([t], [0, EPOCHS * 2 // 5, EPOCHS], [0, MAX_LR, 0])[0] # Triangular (cyclic) learning rate schedule SAVE_TRAIN_DIR = os.path.join(TREE_ROOT, '../Models/Generated/%s_flipped/%s/' % (DATASET, model.__class__.__name__)) os.makedirs(SAVE_TRAIN_DIR, exist_ok=True) t0 = time.time() model = model.to(DEVICE) model = train(model, flip_trans, trainloader, testloader, EPOCHS, opt, loss_fun, lr_schedule, SAVE_TRAIN_DIR) print('---> Training is done! Elapsed time: %.5f minutes\n' % ((time.time() - t0) / 60.)) ################################## # Compute margin along subspaces # ################################## # Create a list of subspaces to evaluate the margin on SUBSPACE_DIM = 8 DIM = 28 SUBSPACE_STEP = 1 subspace_list = generate_subspace_list(SUBSPACE_DIM, DIM, SUBSPACE_STEP, channels=1) # Select the data samples for evaluation NUM_SAMPLES_EVAL = 100 indices = np.random.choice(len(testset), NUM_SAMPLES_EVAL, replace=False) eval_dataset = torch.utils.data.Subset(testset, indices[:NUM_SAMPLES_EVAL]) eval_loader = torch.utils.data.DataLoader(eval_dataset, batch_size=BATCH_SIZE, shuffle=False, num_workers=2, pin_memory=True if DEVICE == 'cuda' else False) # Compute the margin using subspace DeepFool and save the results RESULTS_DIR = os.path.join(TREE_ROOT, '../Results/margin_%s_flipped/%s/' % (DATASET, model.__class__.__name__)) os.makedirs(RESULTS_DIR, exist_ok=True) margins = compute_margin_distribution(model, flip_trans, eval_loader, subspace_list, RESULTS_DIR + 'margins.npy')
[ "apostolos.modas@epfl.ch" ]
apostolos.modas@epfl.ch
d5d4dc11f80514143b96cfebbcab39e53506dd9b
7f9811857538858ea5c6baaefdccf424c2dea3c2
/INTRODUCTION_TO_DS/chapter5_search/linear_search.py
b3c44483d7fd39c6fc66b263858905c46d9c2969
[]
no_license
owari-taro/python_algorithm
ec4d0c737eefdb4f5ddc140c4dfe81fcfb2ee5af
5af19f7dabe6224f0d06b7c89f38c528a08cf903
refs/heads/master
2021-11-23T07:23:08.958737
2021-08-31T00:56:07
2021-08-31T00:56:07
231,067,479
1
1
null
null
null
null
UTF-8
Python
false
false
254
py
from typing import List def binary_search(a: List, x, lo=0, hi=None): if lo < 0: raise ValueError() if hi is None: hi = len(a) while lo < hi: mid = (hi+lo)//2 if x < a[mid]: hi = mid
[ "taro.biwajima@gmail.com" ]
taro.biwajima@gmail.com
31f505bcd3e2862f943b2fb2fb39a976fcf80f18
7ba05e73515c14fb8d2f3d056b51102131171a11
/First_steps_March_Excercise/Akvarium.py
c65b850ffd42b8483b25d7fd5129ca00ac7b1aab
[]
no_license
gyurel/SoftUni-Basics-and-Fundamentals
bd6d5fa8c9d0cc51f241393afd418633a66c65dc
184fc5dfab2fdd410aa8593f4c562fd56211c727
refs/heads/main
2023-07-05T11:16:58.966841
2021-08-31T19:25:40
2021-08-31T19:25:40
401,485,125
0
0
null
null
null
null
UTF-8
Python
false
false
727
py
# От конзолата се четат 4 реда: # 1. Дължина в см – цяло число # 2. Широчина в см – цяло число # 3. Височина в см – цяло число # 4. Процент зает обем – реално число length = int(input()) width = int(input()) height = int(input()) occuqied_percentage = float(input()) / 100 volume_in_litters = length * width * height/1000 # Да се напише програма, която изчислява литрите вода, които са необходими за напълването на аквариума. needed_water = volume_in_litters - (volume_in_litters * occuqied_percentage) print(needed_water)
[ "gyurel@yahoo.com" ]
gyurel@yahoo.com
79fd25313de50609a139f7d137681f78e0419623
87ad48769b2700e2c02452c616773b7af7313093
/tabby/migrations/0008_category_popularity.py
509787ee5ad1d38b22c7e49b67c78e979175bea2
[]
no_license
cavacH/pidb
12e8c29fefc46669176d3eecb4c41b0050df8e95
a54bfbeaf009ba5a050c086bfb7361e612bf5a7c
refs/heads/master
2020-12-20T01:00:02.028753
2020-01-23T23:30:57
2020-01-23T23:30:57
235,907,060
0
0
null
null
null
null
UTF-8
Python
false
false
452
py
# -*- coding: utf-8 -*- # Generated by Django 1.11.7 on 2017-12-26 06:10 from __future__ import unicode_literals from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('tabby', '0007_auto_20171222_0038'), ] operations = [ migrations.AddField( model_name='category', name='popularity', field=models.IntegerField(default=0), ), ]
[ "noreply@github.com" ]
noreply@github.com
dc26da6e320ff093012cda0b898e9b9684b94a04
05d916261d917f51efb7b51acb04499065d39f00
/import/serializiers.py
0ad7b2dabbe0e5197168d6cf2849da96a9a5060b
[]
no_license
Gektor1234/Import-and-search-for-company-employees
1e71f1e0f4734fc49f9c4425b35985d75ae5505b
eb0c61074f81a04bcb13cfa903a9870da84922c4
refs/heads/master
2021-03-08T13:51:55.661085
2020-03-10T17:41:40
2020-03-10T17:41:40
246,350,253
0
0
null
null
null
null
UTF-8
Python
false
false
627
py
from rest_framework import serializers from .models import Workers # создаем сериализайзер для отображение данных в списки питон class WorkersSerializers(serializers.Serializer): name = serializers.CharField(max_length=255) surname = serializers.CharField(max_length=255) date_of_birth = serializers.IntegerField() position = serializers.CharField(max_length=255) def create(self, validated_data): # метод для сообщение инструкции при вызове метода "save" return Workers.objects.create(**validated_data)
[ "xyvafvyf1@gmail.com" ]
xyvafvyf1@gmail.com
0826bb49bda6584cc57d9ea1205a457341b5e9ac
4e3c976773526fd610d64ffb83589bccfaee5e68
/sponge-integration-tests/examples/core/filters_event_pattern.py
32eae8faeab3bf1d4d3fa3664b9a44fc5a0f1edc
[ "Apache-2.0" ]
permissive
softelnet/sponge
2313d2328953fcff49a002e727bb803757870627
7190f23ae888bbef49d0fbb85157444d6ea48bcd
refs/heads/master
2022-10-28T16:19:55.619882
2021-09-16T19:50:08
2021-09-16T19:50:08
95,256,030
10
2
Apache-2.0
2022-10-04T23:55:09
2017-06-23T20:58:49
Java
UTF-8
Python
false
false
1,408
py
""" Sponge Knowledge Base Filters - Event pattern """ from java.util.concurrent.atomic import AtomicInteger def onInit(): # Variables for assertions only sponge.setVariable("nameCount", AtomicInteger(0)) sponge.setVariable("patternCount", AtomicInteger(0)) sponge.setVariable("acceptedCount", AtomicInteger(0)) sponge.setVariable("notAcceptedCount", AtomicInteger(0)) class NameFilter(Filter): def onConfigure(self): self.withEvent("a1") def onAccept(self, event): sponge.getVariable("nameCount").incrementAndGet() return True class PatternFilter(Filter): def onConfigure(self): self.withEvent("a.+") def onAccept(self, event): sponge.getVariable("patternCount").incrementAndGet() return False class AcceptedTrigger(Trigger): def onConfigure(self): self.withEvent(".+") def onRun(self, event): self.logger.info("accepted {}", event.name) if event.name != EventName.STARTUP: sponge.getVariable("acceptedCount").incrementAndGet() class NotAcceptedTrigger(Trigger): def onConfigure(self): self.withEvent("a.+") def onRun(self, event): sponge.getVariable("notAcceptedCount").incrementAndGet() def onStartup(): for name in ["a1", "b1", "a2", "b2", "a", "b", "a1", "b2"]: sponge.event(name).send()
[ "marcin.pas@softelnet.com" ]
marcin.pas@softelnet.com
c839051c620fd066513fce874f55bfe78f1dc4e4
540b24e3ec47a2cb4baefb6fe19d6c97c05b41c6
/subversion/tools/hook-scripts/svn2feed.py
c3abe8c1eb2dc1858dc594f397eb2d74cd7b596e
[ "BSD-3-Clause", "LicenseRef-scancode-unknown-license-reference", "LicenseRef-scancode-other-permissive", "X11", "Apache-2.0", "BSD-2-Clause", "HPND-Markus-Kuhn", "LicenseRef-scancode-unicode", "MIT" ]
permissive
Quantum-Platinum-Cloud/subversion
dedeff0955fc6d03df445d1cb1b9a6d058e47c72
494f46f077e41a3ef32cf315e903695ecf547f5c
refs/heads/main
2023-08-17T16:36:40.102795
2021-03-17T19:13:59
2021-10-06T05:38:16
589,011,516
1
0
null
2023-01-14T19:18:40
2023-01-14T19:18:39
null
UTF-8
Python
false
false
16,736
py
#!/usr/bin/env python # -*- coding: utf-8 -*- # ==================================================================== # Licensed to the Apache Software Foundation (ASF) under one # or more contributor license agreements. See the NOTICE file # distributed with this work for additional information # regarding copyright ownership. The ASF licenses this file # to you under the Apache License, Version 2.0 (the # "License"); you may not use this file except in compliance # with the License. You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, # software distributed under the License is distributed on an # "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY # KIND, either express or implied. See the License for the # specific language governing permissions and limitations # under the License. # ==================================================================== """Usage: svn2feed.py [OPTION...] REPOS-PATH Generate an RSS 2.0 or Atom 1.0 feed file containing commit information for the Subversion repository located at REPOS-PATH. Once the maximum number of items is reached, older elements are removed. The item title is the revision number, and the item description contains the author, date, log messages and changed paths. Options: -h, --help Show this help message. -F, --format=FORMAT Required option. FORMAT must be one of: 'rss' (RSS 2.0) 'atom' (Atom 1.0) to select the appropriate feed format. -f, --feed-file=PATH Store the feed in the file located at PATH, which will be created if it does not exist, or overwritten if it does. If not provided, the script will store the feed in the current working directory, in a file named REPOS_NAME.rss or REPOS_NAME.atom (where REPOS_NAME is the basename of the REPOS_PATH command-line argument, and the file extension depends on the selected format). -r, --revision=X[:Y] Subversion revision (or revision range) to generate info for. If not provided, info for the single youngest revision in the repository will be generated. -m, --max-items=N Keep only N items in the feed file. By default, 20 items are kept. -u, --item-url=URL Use URL as the basis for generating feed item links. This value is appended with '?rev=REV_NUMBER' to form the actual item links. -U, --feed-url=URL Use URL as the global link associated with the feed. -P, --svn-path=DIR Look in DIR for the svnlook binary. If not provided, svnlook must be on the PATH. """ # TODO: # --item-url should support arbitrary formatting of the revision number, # to be useful with web viewers other than ViewVC. # Rather more than intended is being cached in the pickle file. Instead of # only old items being drawn from the pickle, all the global feed metadata # is actually set only on initial feed creation, and thereafter simply # re-used from the pickle each time. # $HeadURL: https://svn.apache.org/repos/asf/subversion/branches/1.10.x/tools/hook-scripts/svn2feed.py $ # $LastChangedDate: 2016-04-30 08:16:53 +0000 (Sat, 30 Apr 2016) $ # $LastChangedBy: stefan2 $ # $LastChangedRevision: 1741723 $ import sys # Python 2.4 is required for subprocess if sys.version_info < (2, 4): sys.stderr.write("Error: Python 2.4 or higher required.\n") sys.stderr.flush() sys.exit(1) import getopt import os import subprocess try: # Python <3.0 import cPickle as pickle except ImportError: # Python >=3.0 import pickle import datetime import time def usage_and_exit(errmsg=None): """Print a usage message, plus an ERRMSG (if provided), then exit. If ERRMSG is provided, the usage message is printed to stderr and the script exits with a non-zero error code. Otherwise, the usage message goes to stdout, and the script exits with a zero errorcode.""" if errmsg is None: stream = sys.stdout else: stream = sys.stderr stream.write("%s\n" % __doc__) stream.flush() if errmsg: stream.write("\nError: %s\n" % errmsg) stream.flush() sys.exit(2) sys.exit(0) def check_url(url, opt): """Verify that URL looks like a valid URL or option OPT.""" if not (url.startswith('https://') \ or url.startswith('http://') \ or url.startswith('file://')): usage_and_exit("svn2feed.py: Invalid url '%s' is specified for " \ "'%s' option" % (url, opt)) class Svn2Feed: def __init__(self, svn_path, repos_path, item_url, feed_file, max_items, feed_url): self.repos_path = repos_path self.item_url = item_url self.feed_file = feed_file self.max_items = max_items self.feed_url = feed_url self.svnlook_cmd = 'svnlook' if svn_path is not None: self.svnlook_cmd = os.path.join(svn_path, 'svnlook') self.feed_title = ("%s's Subversion Commits Feed" % (os.path.basename(os.path.abspath(self.repos_path)))) self.feed_desc = "The latest Subversion commits" def _get_item_dict(self, revision): revision = str(revision) cmd = [self.svnlook_cmd, 'info', '-r', revision, self.repos_path] proc = subprocess.Popen(cmd, stdout=subprocess.PIPE) proc.wait() info_lines = proc.stdout.readlines() cmd = [self.svnlook_cmd, 'changed', '-r', revision, self.repos_path] proc = subprocess.Popen(cmd, stdout=subprocess.PIPE) proc.wait() changed_data = proc.stdout.readlines() desc = ("\nRevision: %s\nLog: %sModified: \n%s" % (revision, info_lines[3], changed_data)) item_dict = { 'author': info_lines[0].strip('\n'), 'title': "Revision %s" % revision, 'link': self.item_url and "%s?rev=%s" % (self.item_url, revision), 'date': self._format_updated_ts(info_lines[1]), 'description': "<pre>" + desc + "</pre>", } return item_dict def _format_updated_ts(self, revision_ts): # Get "2006-08-10 20:17:08" from # "2006-07-28 20:17:18 +0530 (Fri, 28 Jul 2006) date = revision_ts[0:19] epoch = time.mktime(time.strptime(date, "%Y-%m-%d %H:%M:%S")) return time.strftime("%Y-%m-%dT%H:%M:%SZ", time.gmtime(epoch)) class Svn2RSS(Svn2Feed): def __init__(self, svn_path, repos_path, item_url, feed_file, max_items, feed_url): Svn2Feed.__init__(self, svn_path, repos_path, item_url, feed_file, max_items, feed_url) try: import PyRSS2Gen except ImportError: sys.stderr.write(""" Error: Required PyRSS2Gen module not found. You can download the PyRSS2Gen module from: http://www.dalkescientific.com/Python/PyRSS2Gen.html """) sys.exit(1) self.PyRSS2Gen = PyRSS2Gen (file, ext) = os.path.splitext(self.feed_file) self.pickle_file = file + ".pickle" if os.path.exists(self.pickle_file): self.rss = pickle.load(open(self.pickle_file, "r")) else: self.rss = self.PyRSS2Gen.RSS2( title = self.feed_title, link = self.feed_url, description = self.feed_desc, lastBuildDate = datetime.datetime.now(), items = []) @staticmethod def get_default_file_extension(): return ".rss" def add_revision_item(self, revision): rss_item = self._make_rss_item(revision) self.rss.items.insert(0, rss_item) if len(self.rss.items) > self.max_items: del self.rss.items[self.max_items:] def write_output(self): s = pickle.dumps(self.rss) f = open(self.pickle_file, "w") f.write(s) f.close() f = open(self.feed_file, "w") self.rss.write_xml(f) f.close() def _make_rss_item(self, revision): info = self._get_item_dict(revision) rss_item = self.PyRSS2Gen.RSSItem( author = info['author'], title = info['title'], link = info['link'], description = info['description'], guid = self.PyRSS2Gen.Guid(info['link']), pubDate = info['date']) return rss_item class Svn2Atom(Svn2Feed): def __init__(self, svn_path, repos_path, item_url, feed_file, max_items, feed_url): Svn2Feed.__init__(self, svn_path, repos_path, item_url, feed_file, max_items, feed_url) from xml.dom import getDOMImplementation self.dom_impl = getDOMImplementation() self.pickle_file = self.feed_file + ".pickle" if os.path.exists(self.pickle_file): self.document = pickle.load(open(self.pickle_file, "r")) self.feed = self.document.getElementsByTagName('feed')[0] else: self._init_atom_document() @staticmethod def get_default_file_extension(): return ".atom" def add_revision_item(self, revision): item = self._make_atom_item(revision) total = 0 for childNode in self.feed.childNodes: if childNode.nodeName == 'entry': if total == 0: self.feed.insertBefore(item, childNode) total += 1 total += 1 if total > self.max_items: self.feed.removeChild(childNode) if total == 0: self.feed.appendChild(item) def write_output(self): s = pickle.dumps(self.document) f = open(self.pickle_file, "w") f.write(s) f.close() f = open(self.feed_file, "w") f.write(self.document.toxml()) f.close() def _make_atom_item(self, revision): info = self._get_item_dict(revision) doc = self.document entry = doc.createElement("entry") id = doc.createElement("id") entry.appendChild(id) id.appendChild(doc.createTextNode(info['link'])) title = doc.createElement("title") entry.appendChild(title) title.appendChild(doc.createTextNode(info['title'])) updated = doc.createElement("updated") entry.appendChild(updated) updated.appendChild(doc.createTextNode(info['date'])) link = doc.createElement("link") entry.appendChild(link) link.setAttribute("href", info['link']) summary = doc.createElement("summary") entry.appendChild(summary) summary.appendChild(doc.createTextNode(info['description'])) author = doc.createElement("author") entry.appendChild(author) aname = doc.createElement("name") author.appendChild(aname) aname.appendChild(doc.createTextNode(info['author'])) return entry def _init_atom_document(self): doc = self.document = self.dom_impl.createDocument(None, None, None) feed = self.feed = doc.createElement("feed") doc.appendChild(feed) feed.setAttribute("xmlns", "http://www.w3.org/2005/Atom") title = doc.createElement("title") feed.appendChild(title) title.appendChild(doc.createTextNode(self.feed_title)) id = doc.createElement("id") feed.appendChild(id) id.appendChild(doc.createTextNode(self.feed_url)) updated = doc.createElement("updated") feed.appendChild(updated) now = datetime.datetime.now() updated.appendChild(doc.createTextNode(self._format_date(now))) link = doc.createElement("link") feed.appendChild(link) link.setAttribute("href", self.feed_url) author = doc.createElement("author") feed.appendChild(author) aname = doc.createElement("name") author.appendChild(aname) aname.appendChild(doc.createTextNode("subversion")) def _format_date(self, dt): """ input date must be in GMT """ return ("%04d-%02d-%02dT%02d:%02d:%02d.%02dZ" % (dt.year, dt.month, dt.day, dt.hour, dt.minute, dt.second, dt.microsecond)) def main(): # Parse the command-line options and arguments. try: opts, args = getopt.gnu_getopt(sys.argv[1:], "hP:r:u:f:m:U:F:", ["help", "svn-path=", "revision=", "item-url=", "feed-file=", "max-items=", "feed-url=", "format=", ]) except getopt.GetoptError as msg: usage_and_exit(msg) # Make sure required arguments are present. if len(args) != 1: usage_and_exit("You must specify a repository path.") repos_path = os.path.abspath(args[0]) # Now deal with the options. max_items = 20 commit_rev = svn_path = None item_url = feed_url = None feed_file = None feedcls = None feed_classes = { 'rss': Svn2RSS, 'atom': Svn2Atom } for opt, arg in opts: if opt in ("-h", "--help"): usage_and_exit() elif opt in ("-P", "--svn-path"): svn_path = arg elif opt in ("-r", "--revision"): commit_rev = arg elif opt in ("-u", "--item-url"): item_url = arg check_url(item_url, opt) elif opt in ("-f", "--feed-file"): feed_file = arg elif opt in ("-m", "--max-items"): try: max_items = int(arg) except ValueError as msg: usage_and_exit("Invalid value '%s' for --max-items." % (arg)) if max_items < 1: usage_and_exit("Value for --max-items must be a positive " "integer.") elif opt in ("-U", "--feed-url"): feed_url = arg check_url(feed_url, opt) elif opt in ("-F", "--format"): try: feedcls = feed_classes[arg] except KeyError: usage_and_exit("Invalid value '%s' for --format." % arg) if feedcls is None: usage_and_exit("Option -F [--format] is required.") if item_url is None: usage_and_exit("Option -u [--item-url] is required.") if feed_url is None: usage_and_exit("Option -U [--feed-url] is required.") if commit_rev is None: svnlook_cmd = 'svnlook' if svn_path is not None: svnlook_cmd = os.path.join(svn_path, 'svnlook') cmd = [svnlook_cmd, 'youngest', repos_path] proc = subprocess.Popen(cmd, stdout=subprocess.PIPE) proc.wait() cmd_out = proc.stdout.readlines() try: revisions = [int(cmd_out[0])] except IndexError as msg: usage_and_exit("svn2feed.py: Invalid value '%s' for " \ "REPOS-PATH" % (repos_path)) else: try: rev_range = commit_rev.split(':') len_rev_range = len(rev_range) if len_rev_range == 1: revisions = [int(commit_rev)] elif len_rev_range == 2: start, end = rev_range start = int(start) end = int(end) if (start > end): tmp = start start = end end = tmp revisions = list(range(start, end + 1)[-max_items:]) else: raise ValueError() except ValueError as msg: usage_and_exit("svn2feed.py: Invalid value '%s' for --revision." \ % (commit_rev)) if feed_file is None: feed_file = (os.path.basename(repos_path) + feedcls.get_default_file_extension()) feed = feedcls(svn_path, repos_path, item_url, feed_file, max_items, feed_url) for revision in revisions: feed.add_revision_item(revision) feed.write_output() if __name__ == "__main__": main()
[ "91980991+AppleOSSDistributions@users.noreply.github.com" ]
91980991+AppleOSSDistributions@users.noreply.github.com
b8b49ba5bc255e5615ec2889ec70661333b1a2c2
4252102a1946b2ba06d3fa914891ec7f73570287
/pylearn2/packaged_dependencies/theano_linear/unshared_conv/test_localdot.py
6b47b5b33566ea24783e9ae4019290a4fabb845d
[]
no_license
lpigou/chalearn2014
21d487f314c4836dd1631943e20f7ab908226771
73b99cdbdb609fecff3cf85e500c1f1bfd589930
refs/heads/master
2020-05-17T00:08:11.764642
2014-09-24T14:42:00
2014-09-24T14:42:00
24,418,815
2
3
null
null
null
null
UTF-8
Python
false
false
4,927
py
import nose import unittest import numpy as np import theano from localdot import LocalDot from ..test_matrixmul import SymbolicSelfTestMixin class TestLocalDot32x32(unittest.TestCase, SymbolicSelfTestMixin): channels = 3 bsize = 10 # batch size imshp = (32, 32) ksize = 5 nkern_per_group = 16 subsample_stride = 1 ngroups = 1 def rand(self, shp): return np.random.rand(*shp).astype('float32') def setUp(self): np.random.seed(234) assert self.imshp[0] == self.imshp[1] fModulesR = (self.imshp[0] - self.ksize + 1) // self.subsample_stride #fModulesR += 1 # XXX GpuImgActs crashes w/o this?? fModulesC = fModulesR self.fshape = (fModulesR, fModulesC, self.channels // self.ngroups, self.ksize, self.ksize, self.ngroups, self.nkern_per_group) self.ishape = (self.ngroups, self.channels // self.ngroups, self.imshp[0], self.imshp[1], self.bsize) self.hshape = (self.ngroups, self.nkern_per_group, fModulesR, fModulesC, self.bsize) filters = theano.shared(self.rand(self.fshape)) self.A = LocalDot(filters, self.imshp[0], self.imshp[1], subsample=(self.subsample_stride, self.subsample_stride)) self.xlval = self.rand((self.hshape[-1],) + self.hshape[:-1]) self.xrval = self.rand(self.ishape) self.xl = theano.shared(self.xlval) self.xr = theano.shared(self.xrval) # N.B. the tests themselves come from SymbolicSelfTestMixin class TestLocalDotLargeGray(TestLocalDot32x32): channels = 1 bsize = 128 imshp = (256, 256) ksize = 9 nkern_per_group = 16 subsample_stride = 2 ngroups = 1 n_patches = 3000 def rand(self, shp): return np.random.rand(*shp).astype('float32') # not really a test, but important code to support # Currently exposes error, by e.g.: # CUDA_LAUNCH_BLOCKING=1 # THEANO_FLAGS=device=gpu,mode=DEBUG_MODE # nosetests -sd test_localdot.py:TestLocalDotLargeGray.run_autoencoder def run_autoencoder( self, n_train_iter=10000, # -- make this small to be a good unit test rf_shape=(9, 9), n_filters=1024, dtype='float32', module_stride=2, lr=0.01, show_filters=True, ): if show_filters: # import here to fail right away import matplotlib.pyplot as plt try: import skdata.vanhateren.dataset except ImportError: raise nose.SkipTest() # 1. Get a set of image patches from the van Hateren data set print 'Loading van Hateren images' n_images = 50 vh = skdata.vanhateren.dataset.Calibrated(n_images) patches = vh.raw_patches((self.n_patches,) + self.imshp, items=vh.meta[:n_images], rng=np.random.RandomState(123), ) patches = patches.astype('float32') patches /= patches.reshape(self.n_patches, self.imshp[0] * self.imshp[1])\ .max(axis=1)[:, None, None] # TODO: better local contrast normalization if 0 and show_filters: plt.subplot(2, 2, 1); plt.imshow(patches[0], cmap='gray') plt.subplot(2, 2, 2); plt.imshow(patches[1], cmap='gray') plt.subplot(2, 2, 3); plt.imshow(patches[2], cmap='gray') plt.subplot(2, 2, 4); plt.imshow(patches[3], cmap='gray') plt.show() # -- Convert patches to localdot format: # groups x colors x rows x cols x images patches5 = patches[:, :, :, None, None].transpose(3, 4, 1, 2, 0) print 'Patches shape', patches.shape, self.n_patches, patches5.shape # 2. Set up an autoencoder print 'Setting up autoencoder' hid = theano.tensor.tanh(self.A.rmul(self.xl)) out = self.A.rmul_T(hid) cost = ((out - self.xl) ** 2).sum() params = self.A.params() gparams = theano.tensor.grad(cost, params) train_updates = [(p, p - lr / self.bsize * gp) for (p, gp) in zip(params, gparams)] if 1: train_fn = theano.function([], [cost], updates=train_updates) else: train_fn = theano.function([], [], updates=train_updates) theano.printing.debugprint(train_fn) # 3. Train it params[0].set_value(0.001 * params[0].get_value()) for ii in xrange(0, self.n_patches, self.bsize): self.xl.set_value(patches5[:, :, :, :, ii:ii + self.bsize], borrow=True) cost_ii, = train_fn() print 'Cost', ii, cost_ii if 0 and show_filters: self.A.imshow_gray() plt.show() assert cost_ii < 0 # TODO: determine a threshold for detecting regression bugs
[ "lionelpigou@gmail.com" ]
lionelpigou@gmail.com
3c2d197b7b46aa7ba32e0006c6e3ee2c34f3c02f
42acbbad9f4af26ef4261dd70ec4f9fae49c253e
/optimization_solver.py
16598ff8e47d06e5a977f58fc9d582b245a6c24a
[]
no_license
liuruilinspy/SocialRouting
da5ffa51bdcaa7a875e50743c10abfc0647267b4
20e404256c6fbecdcf5994a0a67bc0200d1f7cac
refs/heads/master
2021-01-19T08:40:49.563502
2017-04-08T20:31:56
2017-04-08T20:32:04
87,661,319
0
2
null
null
null
null
UTF-8
Python
false
false
10,476
py
from cvxopt import solvers, matrix, spdiag, log from cvxopt import spmatrix def routing_solver(A_listform, b_listform, unknown_variables, demand_list, edge_list, type='system', maxiters=100, abstol=10**(-7), reltivetol=10**(-6), feastol=10**(-7)): A = matrix(A_listform).T b = matrix(b_listform) all_variable_count = (len(demand_list) + 1) * len(edge_list) routing_variable_count = len(demand_list) * len(edge_list) valid_routing_variable_count = 0 flow_variable_to_edge_index = {} predefined_flow_variable_count = 0 for index in range(0, all_variable_count): if index < routing_variable_count: if unknown_variables[index] == 1: valid_routing_variable_count += 1 else: if unknown_variables[index] == 0: predefined_flow_variable_count += 1 else: flow_variable_edge_index = index - routing_variable_count valid_flow_variable_index = flow_variable_edge_index - predefined_flow_variable_count flow_variable_to_edge_index[valid_routing_variable_count + valid_flow_variable_index] = \ flow_variable_edge_index linear_constraint_count, variable_count = A.size G = spmatrix(-1.0, range(0, variable_count), range(0, variable_count)) #print(G) h = matrix([0.0] * variable_count) #print(h) dims = {'l': variable_count, 'q': [], 's': []} solvers.options['maxiters'] = maxiters solvers.options['abstol'] = abstol solvers.options['reltol'] = reltivetol solvers.options['feastol'] = feastol def sys_op_f(x=None, z=None): if x is None: return 0, matrix(1.0, (variable_count, 1)) if min(x) < 0.0: return None # in our case, non-linear constraint m = 0, i.e., only f_0(x) = g_0(x_0) + g_i(x_i) + ... != 0 # f(m+1)*1=1*1 f[0] = f_0(x) = g_0(x_0) + g_i(x_i) + ... f = 0 for var_index in range(valid_routing_variable_count, variable_count): edge_index = flow_variable_to_edge_index[var_index] cost = edge_list[edge_index]['cost'] capacity = edge_list[edge_index]['capacity'] bg_volume = edge_list[edge_index]['bg_volume'] alpha = edge_list[edge_index]['alpha'] beta = edge_list[edge_index]['beta'] #f += cost * (1 + alpha * ((bg_volume + x[var_index]) / capacity) ** beta) t = cost * (1 + alpha * ((bg_volume + x[var_index]) / capacity) ** beta) f += t * x[var_index] # Df(m+1)*n = 1*n f[0,:] = df_0/dx_i df_values = list() # derivative towards each x_i ddf_values = list() # second derivative towards each x_i for var_index in range(0, valid_routing_variable_count): df_values.append(0.0) ddf_values.append(0.0) for var_index in range(valid_routing_variable_count, variable_count): edge_index = flow_variable_to_edge_index[var_index] cost = edge_list[edge_index]['cost'] capacity = edge_list[edge_index]['capacity'] bg_volume = edge_list[edge_index]['bg_volume'] alpha = edge_list[edge_index]['alpha'] beta = edge_list[edge_index]['beta'] t = cost * (1 + alpha * ((bg_volume + x[var_index]) / capacity) ** beta) dt = cost * alpha * beta * ((bg_volume + x[var_index] / capacity) ** (beta - 1)) / capacity dt2 = cost * alpha * beta * (beta - 1) * ((bg_volume + x[var_index] / capacity) ** (beta - 2)) / (capacity ** 2) df_values.append(x[var_index] * dt + t) ddf_values.append(x[var_index] * dt2 + 2 * dt) Df = matrix(df_values, (1, variable_count)) ddf = matrix(ddf_values, (variable_count, 1)) if z is None: return f, Df H = spdiag(z[0] * ddf) # diagonal matrix, h[:i] = z[i] * f_i''(x) return f, Df, H def ue_f(x=None, z=None): if x is None: return 0, matrix(1.0, (variable_count, 1)) if min(x) < 0.0: return None # in our case, non-linear constraint m = 0, i.e., only f_0(x) = g_0(x_0) + g_i(x_i) + ... != 0 # f(m+1)*1=1*1 f[0] = f_0(x) = g_0(x_0) + g_i(x_i) + ... f = 0 for var_index in range(valid_routing_variable_count, variable_count): edge_index = flow_variable_to_edge_index[var_index] cost = edge_list[edge_index]['cost'] capacity = edge_list[edge_index]['capacity'] bg_volume = edge_list[edge_index]['bg_volume'] alpha = edge_list[edge_index]['alpha'] beta = edge_list[edge_index]['beta'] f += cost * x[var_index] + cost * alpha * capacity ** -beta / (beta + 1) * \ ((bg_volume + x[var_index]) ** (beta + 1) - bg_volume ** (beta + 1)) # Df(m+1)*n = 1*n f[0,:] = df_0/dx_i df_values = list() # derivative towards each x_i ddf_values = list() # second derivative towards each x_i for var_index in range(0, valid_routing_variable_count): df_values.append(0.0) ddf_values.append(0.0) for var_index in range(valid_routing_variable_count, variable_count): edge_index = flow_variable_to_edge_index[var_index] cost = edge_list[edge_index]['cost'] capacity = edge_list[edge_index]['capacity'] bg_volume = edge_list[edge_index]['bg_volume'] alpha = edge_list[edge_index]['alpha'] beta = edge_list[edge_index]['beta'] t = cost * (1 + alpha * ((bg_volume + x[var_index]) / capacity) ** beta) dt = cost * alpha * beta * ((bg_volume + x[var_index] / capacity) ** (beta - 1)) / capacity df_values.append(t) ddf_values.append(dt) Df = matrix(df_values, (1, variable_count)) ddf = matrix(ddf_values, (variable_count, 1)) if z is None: return f, Df H = spdiag(z[0] * ddf) # diagonal matrix, h[:i] = z[i] * f_i''(x) return f, Df, H def social_op_f(x=None, z=None): if x is None: return 0, matrix(1.0, (variable_count, 1)) if min(x) < 0.0: return None # in our case, non-linear constraint m = 0, i.e., only f_0(x) = g_0(x_0) + g_i(x_i) + ... != 0 # f(m+1)*1=1*1 f[0] = f_0(x) = g_0(x_0) + g_i(x_i) + ... f = 0 for var_index in range(valid_routing_variable_count, variable_count): edge_index = flow_variable_to_edge_index[var_index] cost = edge_list[edge_index]['cost'] capacity = edge_list[edge_index]['capacity'] bg_volume = edge_list[edge_index]['bg_volume'] alpha = edge_list[edge_index]['alpha'] beta = edge_list[edge_index]['beta'] integral_t = cost * x[var_index] + cost * alpha * capacity ** -beta / (beta + 1) * \ ((bg_volume + x[var_index]) ** (beta + 1) - bg_volume ** (beta + 1)) integral_xdt = cost * alpha * capacity ** (-beta) * (x[var_index] * (bg_volume + x[var_index]) ** beta - 1 / (beta + 1) * ((bg_volume + x[var_index]) ** (beta + 1) - bg_volume ** (beta + 1))) f += integral_t + integral_xdt # Df(m+1)*n = 1*n f[0,:] = df_0/dx_i df_values = list() # derivative towards each x_i ddf_values = list() # second derivative towards each x_i for var_index in range(0, valid_routing_variable_count): df_values.append(0.0) ddf_values.append(0.0) for var_index in range(valid_routing_variable_count, variable_count): edge_index = flow_variable_to_edge_index[var_index] cost = edge_list[edge_index]['cost'] capacity = edge_list[edge_index]['capacity'] bg_volume = edge_list[edge_index]['bg_volume'] alpha = edge_list[edge_index]['alpha'] beta = edge_list[edge_index]['beta'] t = cost * (1 + alpha * ((bg_volume + x[var_index]) / capacity) ** beta) dt = cost * alpha * beta * ((bg_volume + x[var_index] / capacity) ** (beta - 1)) / capacity dt2 = cost * alpha * beta * (beta - 1) * ((bg_volume + x[var_index] / capacity) ** (beta - 2)) / (capacity ** 2) df_values.append(t + x[var_index] * dt) ddf_values.append(2 * dt + x[var_index] * dt2) Df = matrix(df_values, (1, variable_count)) ddf = matrix(ddf_values, (variable_count, 1)) if z is None: return f, Df H = spdiag(z[0] * ddf) # diagonal matrix, h[:i] = z[i] * f_i''(x) return f, Df, H if type == 'ue': planning_results = solvers.cp(ue_f, G=G, h=h, dims=dims, A=A, b=b)['x'] elif type == 'social': planning_results = solvers.cp(social_op_f, G=G, h=h, dims=dims, A=A, b=b)['x'] else: planning_results = solvers.cp(sys_op_f, G=G, h=h, dims=dims, A=A, b=b)['x'] total_cost = 0.0 for flow_variable_index, edge_index in flow_variable_to_edge_index.items(): flow_variable = planning_results[flow_variable_index] cost = edge_list[edge_index]['cost'] capacity = edge_list[edge_index]['capacity'] bg_volume = edge_list[edge_index]['bg_volume'] alpha = edge_list[edge_index]['alpha'] beta = edge_list[edge_index]['beta'] t = cost * (1 + alpha * ((bg_volume + flow_variable) / capacity) ** beta) total_cost += t * flow_variable return planning_results[0:valid_routing_variable_count], total_cost def test_solver(A, b): linear_constraint_count, variable_count = A.size def F(x=None, z=None): if x is None: return 0, matrix(1.0, (variable_count, 1)) if min(x) < 0.0: return None # in our case, non-linear constraint m = 0, i.e., only f_0(x) = g_0(x_0) + g_i(x_i) + ... != 0 # f(m+1)*1=1*1 f[0] = f_0(x) = g_0(x_0) + g_i(x_i) + ... f = -sum(log(x)) Df = -(x ** -1).T if z is None: return f, Df H = spdiag(z[0] * x ** -2) return f, Df, H G = spmatrix(-1.0, range(0, variable_count), range(0, variable_count)) #print(G) h = matrix([0.0] * variable_count) #print(h) dims = {'l': variable_count, 'q': [], 's': []} return solvers.cp(F, G=G, h=h, dims=dims, A=A, b=b)['x']
[ "liuruilinspy@gmail.com" ]
liuruilinspy@gmail.com
6224998f24dbbf286ac343c71d3f2cf7401f4b20
abf9238ac124738796a61e4ae3e667cae950d55a
/Custom Troop Trees/Source Files/cstm_party_templates.py
e85eb75bb7d7beadb6787f95fd1ff63989067576
[]
no_license
ChroniclesStudio/custom-troop-trees
d92d4c3723ca117fd087332451ea1a0414998162
d39333cf8c4ea9fddb3d58c49850a4dffedbb917
refs/heads/master
2023-02-18T07:27:56.439995
2021-01-19T14:46:50
2021-01-19T14:46:50
331,012,346
0
0
null
null
null
null
UTF-8
Python
false
false
5,713
py
from header_common import * from header_parties import * from header_troops import * from ID_troops import * from ID_factions import * from ID_map_icons import * from module_constants import * from module_troops import troops import math pmf_is_prisoner = 0x0001 #################################################################################################################### # Each party template record contains the following fields: # 1) Party-template id: used for referencing party-templates in other files. # The prefix pt_ is automatically added before each party-template id. # 2) Party-template name. # 3) Party flags. See header_parties.py for a list of available flags # 4) Menu. ID of the menu to use when this party is met. The value 0 uses the default party encounter system. # 5) Faction # 6) Personality. See header_parties.py for an explanation of personality flags. # 7) List of stacks. Each stack record is a tuple that contains the following fields: # 7.1) Troop-id. # 7.2) Minimum number of troops in the stack. # 7.3) Maximum number of troops in the stack. # 7.4) Member flags(optional). Use pmf_is_prisoner to note that this member is a prisoner. # Note: There can be at most 6 stacks. #################################################################################################################### party_templates = [ #("kingdom_1_reinforcements_a", "{!}kingdom_1_reinforcements_a", 0, 0, fac_commoners, 0, [(trp_swadian_recruit,5,10),(trp_swadian_militia,2,4)]), #("kingdom_1_reinforcements_b", "{!}kingdom_1_reinforcements_b", 0, 0, fac_commoners, 0, [(trp_swadian_footman,3,6),(trp_swadian_skirmisher,2,4)]), #("kingdom_1_reinforcements_c", "{!}kingdom_1_reinforcements_c", 0, 0, fac_commoners, 0, [(trp_swadian_man_at_arms,2,4),(trp_swadian_crossbowman,1,2)]), #Swadians are a bit less-powered thats why they have a bit more troops in their modernised party template (3-6, others 3-5) #("kingdom_2_reinforcements_a", "{!}kingdom_2_reinforcements_a", 0, 0, fac_commoners, 0, [(trp_vaegir_recruit,5,10),(trp_vaegir_footman,2,4)]), #("kingdom_2_reinforcements_b", "{!}kingdom_2_reinforcements_b", 0, 0, fac_commoners, 0, [(trp_vaegir_veteran,2,4),(trp_vaegir_skirmisher,2,4),(trp_vaegir_footman,1,2)]), #("kingdom_2_reinforcements_c", "{!}kingdom_2_reinforcements_c", 0, 0, fac_commoners, 0, [(trp_vaegir_horseman,2,3),(trp_vaegir_infantry,1,2)]), #("kingdom_3_reinforcements_a", "{!}kingdom_3_reinforcements_a", 0, 0, fac_commoners, 0, [(trp_khergit_tribesman,3,5),(trp_khergit_skirmisher,4,9)]), #Khergits are a bit less-powered thats why they have a bit more 2nd upgraded(trp_khergit_skirmisher) than non-upgraded one(trp_khergit_tribesman). #("kingdom_3_reinforcements_b", "{!}kingdom_3_reinforcements_b", 0, 0, fac_commoners, 0, [(trp_khergit_horseman,2,4),(trp_khergit_horse_archer,2,4),(trp_khergit_skirmisher,1,2)]), #("kingdom_3_reinforcements_c", "{!}kingdom_3_reinforcements_c", 0, 0, fac_commoners, 0, [(trp_khergit_horseman,2,4),(trp_khergit_veteran_horse_archer,2,3)]), #Khergits are a bit less-powered thats why they have a bit more troops in their modernised party template (4-7, others 3-5) #("kingdom_4_reinforcements_a", "{!}kingdom_4_reinforcements_a", 0, 0, fac_commoners, 0, [(trp_nord_footman,5,10),(trp_nord_recruit,2,4)]), #("kingdom_4_reinforcements_b", "{!}kingdom_4_reinforcements_b", 0, 0, fac_commoners, 0, [(trp_nord_huntsman,2,5),(trp_nord_archer,2,3),(trp_nord_footman,1,2)]), #("kingdom_4_reinforcements_c", "{!}kingdom_4_reinforcements_c", 0, 0, fac_commoners, 0, [(trp_nord_warrior,3,5)]), #("kingdom_5_reinforcements_a", "{!}kingdom_5_reinforcements_a", 0, 0, fac_commoners, 0, [(trp_rhodok_tribesman,5,10),(trp_rhodok_spearman,2,4)]), #("kingdom_5_reinforcements_b", "{!}kingdom_5_reinforcements_b", 0, 0, fac_commoners, 0, [(trp_rhodok_crossbowman,3,6),(trp_rhodok_trained_crossbowman,2,4)]), #("kingdom_5_reinforcements_c", "{!}kingdom_5_reinforcements_c", 0, 0, fac_commoners, 0, [(trp_rhodok_veteran_spearman,2,3),(trp_rhodok_veteran_crossbowman,1,2)]), #("kingdom_6_reinforcements_a", "{!}kingdom_6_reinforcements_a", 0, 0, fac_commoners, 0, [(trp_sarranid_recruit,5,10),(trp_sarranid_footman,2,4)]), #("kingdom_6_reinforcements_b", "{!}kingdom_6_reinforcements_b", 0, 0, fac_commoners, 0, [(trp_sarranid_skirmisher,2,4),(trp_sarranid_veteran_footman,2,3),(trp_sarranid_footman,1,3)]), #("kingdom_6_reinforcements_c", "{!}kingdom_6_reinforcements_c", 0, 0, fac_commoners, 0, [(trp_sarranid_horseman,3,5)]), ] def troop_indexes_of_tier(skin, tier): return [find_troop(troops, troop[0]) for troop in tree.get_custom_troops_of_tier(skin, tier)] def tier_stacks(skin, tier, min, max): troops = troop_indexes_of_tier(skin, tier) return [(troop, int(math.ceil(min * 1.0 / len(troops))), int(math.ceil(max * 1.0 / len(troops)))) for troop in troops] for tree in CUSTOM_TROOP_TREES: for skin in CSTM_SKINS: id = "cstm_kingdom_player_%s_%d_reinforcements" % (tree.id, skin.id) party_templates.extend([ (id + "_a", "{!}" + id + "_a", 0, 0, fac_commoners, 0, tier_stacks(skin, tier = 1, min = 5, max = 10) + tier_stacks(skin, tier = 2, min = 2, max = 4)), (id + "_b", "{!}" + id + "_b", 0, 0, fac_commoners, 0, tier_stacks(skin, tier = 3, min = 5, max = 10)), (id + "_c", "{!}" + id + "_c", 0, 0, fac_commoners, 0, tier_stacks(skin, tier = 4, min = 3, max = 5)), ]) #for party_template in party_templates: # print ", ".join([party_template[0], party_template[1], ", ".join(["%d-%d %s" % (stack[1], stack[2], troops[stack[0]][2]) for stack in party_template[6]])])
[ "knowscount@gmail.com" ]
knowscount@gmail.com
c47121be56a4b42909ecd120c8348fc7a11410ec
445cba890decca8780d926a35687fac4298ce404
/utest/Pramatest.py
46f25bef726cb28bf950dcb52d1fca1b543d021d
[]
no_license
243489145/Myframe
5755d54f04a7a798071b1a2af4f872d951554fda
8749133d56e1a6d4e13140ac44e6af3e5d7100a1
refs/heads/master
2022-06-17T19:44:20.694709
2020-05-16T09:24:01
2020-05-16T09:24:01
264,391,281
0
0
null
null
null
null
UTF-8
Python
false
false
1,124
py
# -*- coding: UTF-8 -*- import unittest from utest import testlib from parameterized import parameterized #测试testlib下面的testadd方法 # 创建一个测试类,继承unittest class PramaTest(unittest.TestCase): """ 参数化:单元测试参数化的参数使用的二维列表 parameterized.,没有这个库自己安装 这里可以读取Excel """ @parameterized.expand([ #等价类的方法,80%的问题出现在极值 ['整数相加', 1, 1, 2], ['小数相加', 1.1, 1.33333333, 2.43333333], ['整数加字符串', 1, '1', '11'], ['整数加小数', 1, 1.1, 2.1], ]) # z参数比较是不是期望值 def test_add(self, name, x, y, z): """ :param name: 取名字区分用例 :param x: :param y: :param z: :return: """ print(name) self.assertEqual(testlib.add(x, y), z) #main方法调用unittest运行方式,也可以编辑unittest的运行方式 #运行的时候在运行哪里,edit一个运行方式 if __name__ == '__main__': unittest.main()
[ "243489145@qq.com" ]
243489145@qq.com
1b095a64a3e9bad8edd45e2e519d86adf5921b4d
2fe368714dc0e09b70fc2f8c0e683c09d18e8187
/dazhu/fmtTime.py
57bafbd95b15b613a5286b591b21e83cd7a96e47
[]
no_license
kamasamikon/bigkillmachine
0d080f43011ed772780f34d3a841e7edbc37e4cb
97b672d889d98c04f892599d52a2605823a7c172
refs/heads/master
2021-06-30T22:11:16.273038
2020-12-28T06:40:49
2020-12-28T06:40:49
73,879,397
0
0
null
null
null
null
UTF-8
Python
false
false
477
py
#!/usr/bin/env python import sys s_days = 60 * 60 * 24 s_hours = 60 * 60 s_minutes = 60 def fmtTime(t): s = "" if t >= s_days: s += "%d days " % (t / s_days) t = t % s_days if t >= s_hours: s += "%d hours " % (t / s_hours) t = t % s_hours if t >= s_minutes: s += "%d minutes " % (t / s_minutes) t = t % s_minutes s += "%d seconds" % t return s if __name__ == "__main__": print fmtTime(int(sys.argv[1]))
[ "kamasamikon@gmail.com" ]
kamasamikon@gmail.com
1ce945017e7cc43156885a2691f2f6b34eafa43d
e7ed617f4bd0e54b457102c18ad6d7b8e4ed70b3
/products/migrations/0041_remove_group_products_category.py
667733d8c3dd21dc74dac46b5d4ffd136e65caf4
[]
no_license
shoib-ansari/Ecomm-Proj
9ca11e34b3e61ac451022ff72d1f772daf62d5af
1479a30e5bcdc0c07c5d7f709f0ab17fea09268f
refs/heads/master
2023-03-27T10:02:29.035535
2021-03-24T11:51:25
2021-03-24T11:51:25
350,646,654
0
0
null
null
null
null
UTF-8
Python
false
false
339
py
# Generated by Django 2.2.2 on 2020-03-20 05:18 from django.db import migrations class Migration(migrations.Migration): dependencies = [ ('products', '0040_auto_20200320_1034'), ] operations = [ migrations.RemoveField( model_name='group_products', name='category', ), ]
[ "ansarishoib1008@gmail.com" ]
ansarishoib1008@gmail.com
1d2a41fffb8bc04a5959e3e940c22e672c6fc9c7
7eb0a3429f021f1a046bed8e667a6911d789d065
/ProxyPattern/python/Client/Proxy.py
ae9f03255b6d7784273d6a6be675c7e5e280f8e6
[ "MIT" ]
permissive
gama79530/DesignPattern
d99431711fda65cfb7d790b2959ba0a712fa3f86
4730c50cdd839072ae50eef975cbed62b5a2a41c
refs/heads/master
2023-08-03T04:35:54.561642
2023-06-08T03:13:08
2023-07-31T12:32:13
269,562,362
0
0
null
null
null
null
UTF-8
Python
false
false
383
py
import abc class Proxy(metaclass=abc.ABCMeta): @abc.abstractmethod def addBalanceAccount(self, balanceAccount:int): return NotImplemented @abc.abstractmethod def subtractBalanceAccount(self, balanceAccount:int): return NotImplemented @abc.abstractmethod @property def balanceAccount(self) -> int: return NotImplemented
[ "gama79530@gmail.com" ]
gama79530@gmail.com
fadb927384553d0f8610506e70c5e984bda1d3fa
32dcc04d47fa832d4bce3c528dd8645b1992dc5c
/StagingEngine/src/copyFileFromRawToFailed.py
d9081b9bcbc7358e90fba7c2832b9755067c4093
[ "Apache-2.0" ]
permissive
glenngillen/accelerated-data-lake
f496a795a1ed11d3719df9febedf2706839dac2a
683249395a8cb7e748753a7ace7ead2ebc2d8af0
refs/heads/master
2020-05-15T00:18:09.174541
2019-04-19T10:09:39
2019-04-19T10:09:39
182,009,548
0
1
Apache-2.0
2019-04-18T03:26:00
2019-04-18T03:25:59
null
UTF-8
Python
false
false
965
py
import boto3 import traceback class CopyFileFromRawToFailedException(Exception): pass s3 = boto3.client('s3') def lambda_handler(event, context): copy_file_to_failed(event, context) return event def copy_file_to_failed(event, context): try: raw_bucket = event['fileDetails']['bucket'] raw_key = event['fileDetails']['key'] failed_bucket = event['settings']['failedBucket'] print( 'Copying object {} from bucket {} to key {} in failed bucket {}' .format(raw_key, raw_bucket, raw_key, failed_bucket) ) # Copy the failed file to the failed bucket. copy_source = {'Bucket': raw_bucket, 'Key': raw_key} s3.copy(copy_source, failed_bucket, raw_key) # Delete the failed file from raw. s3.delete_object(Bucket=raw_bucket, Key=raw_key) except Exception as e: traceback.print_exc() raise CopyFileFromRawToFailedException(e)
[ "pmmacey@amazon.com" ]
pmmacey@amazon.com
085edd24ce10bce702e4768e1623e93a1a1a1fac
cb515f8ab202a6a55a8a294824a3c9f3932ffdc6
/src/preparing_data.py
0134752a7569f089bc584d93b16a5d24d05aeadf
[]
no_license
BobXiao97/DL-Project
1ccb6da9651f95464e1d8ded56e4d927321842f9
dc6074f917e471b1552ab005e87ad6dfc89edeea
refs/heads/main
2023-04-24T14:07:40.978318
2021-05-20T12:43:49
2021-05-20T12:43:49
369,202,308
0
0
null
null
null
null
UTF-8
Python
false
false
1,326
py
from __future__ import unicode_literals, print_function, division from io import open import glob import os import unicodedata import string all_letters = string.ascii_letters + " .,;'-" n_letters = len(all_letters) + 1 # Plus EOS marker def findFiles(path): return glob.glob(path) # Turn a Unicode string to plain ASCII, thanks to https://stackoverflow.com/a/518232/2809427 def unicodeToAscii(s): return ''.join( c for c in unicodedata.normalize('NFD', s) if unicodedata.category(c) != 'Mn' and c in all_letters ) # Read a file and split into lines def readLines(filename): lines = open(filename, encoding='utf-8').read().strip().split('\n') return [unicodeToAscii(line) for line in lines] # Build the category_lines dictionary, a list of lines per category category_lines = {} all_categories = [] for filename in findFiles('../input/english_words/words.txt'): category = os.path.splitext(os.path.basename(filename))[0] all_categories.append(category) lines = readLines(filename) category_lines[category] = lines n_categories = len(all_categories) if n_categories == 0: raise RuntimeError('Data not found. Make sure that you downloaded data ' 'from https://download.pytorch.org/tutorial/data.zip and extract it to ' 'the current directory.')
[ "xtq1997@gmail.com" ]
xtq1997@gmail.com
23da733c2ec49efd92456cc4e9b18303bc590786
81d84521bd42f6fc862272bd56e463690989d969
/Python/Python learning/challange_3.py
db6e4fc3759a641f1c4fd85354c2358114146b37
[]
no_license
Tomek-RTU/RTR-105
b8f0e785c120f885fee8ee2c6a86a669b9e94927
ffd35bfbb68137c1b5a849821403f502ec776dd8
refs/heads/main
2023-02-25T09:41:55.582023
2021-01-31T20:17:17
2021-01-31T20:17:17
312,278,164
0
0
null
null
null
null
UTF-8
Python
false
false
590
py
def sum_eo(n, t): """Sum even or odd numbers in range. Return the sum of even or odd natural numbers, in the range 1..n-1. :param n: The endpoint of the range. The numbers from 1 to n-1 will be summed. :param t: 'e' to sum even numbers, 'o' to sum odd numbers. :return: The sum of the even or odd numbers in the range. Returns -1 if `t` is not 'e' or 'o'. """ if t == "e": start = 2 elif t == 'o': start = 1 else: return -1 return sum(range(start, n, 2)) x = sum_eo(11, 'spam') print(x)
[ "noreply@github.com" ]
noreply@github.com
579e1b6ee2dc5a5ecaf69f568fb93ffbf3fb236f
24cfbad4390b86b337ce7b97998b8be4b3297a2c
/hmrxapp/migrations/0002_altera_tam_histograma.py
277307d46411dec7a22d3354c5782e2968087c06
[]
no_license
ecalasans/hmrxsys
8aa8be3e846870ce0910d5e3a6ebce2d253f754b
1c372d5bbd483bd62fba6a50aa6f266d0f922d45
refs/heads/master
2023-06-26T21:05:58.937630
2021-07-30T07:47:16
2021-07-30T07:47:16
378,649,436
0
0
null
null
null
null
UTF-8
Python
false
false
698
py
# Generated by Django 3.2.4 on 2021-06-22 00:14 import datetime from django.db import migrations, models from django.utils.timezone import utc class Migration(migrations.Migration): dependencies = [ ('hmrxapp', '0001_cria_tabela_filtragem'), ] operations = [ migrations.AlterField( model_name='filtragem', name='data_add', field=models.DateTimeField(default=datetime.datetime(2021, 6, 22, 0, 14, 46, 624188, tzinfo=utc), editable=False), ), migrations.AlterField( model_name='filtragem', name='histograma', field=models.CharField(default='', max_length=5000), ), ]
[ "ericcalasans@gmail.com" ]
ericcalasans@gmail.com
59d9b80f43080c0e1442308cfa86313209c65444
8692b3941a601fdb83a669c57d5a00a4b9e05b59
/VIRTUAL_ASSISTANT-aqueel(EP19101039)-zuhair(EP19101098)/GUI Final/tasks/misc/fbot.py
761f9407189e444432ba9230724b7ab26a4337f5
[ "MIT" ]
permissive
perfectmantis/Submissions-2021
e8e3a66feed1ee9c1b84da7a9af0bef666e89d0f
d7b6458d23d21f526dc3debb6abd4a6209e56085
refs/heads/main
2023-05-02T08:31:50.540522
2021-05-08T20:49:14
2021-05-08T20:49:14
339,471,260
1
42
null
2021-05-04T00:14:12
2021-02-16T17:09:04
CSS
UTF-8
Python
false
false
1,158
py
from selenium import webdriver # from selenium.webdriver.common.keys import Keys from selenium.webdriver.chrome.options import Options import time def account_info(): try: with open('data\\account_info.txt', 'r') as f: info = f.read().split() email = info[0] password = info[1] return email,password except Exception: email = password = "" return email,password def fb_login(): try: email,password = account_info() # options = Options() options = Options().add_argument("start-maximized") driver = webdriver.Chrome(options = options) driver.get("https://www.facebook.com/login/") email_xpath = '//*[@id="email"]' password_xpath = '//*[@id="pass"]' login_xpath = '//*[@id="loginbutton"]' time.sleep(2) driver.find_element_by_xpath(email_xpath).send_keys(email) time.sleep(0.5) driver.find_element_by_xpath(password_xpath).send_keys(password) time.sleep(0.5) driver.find_element_by_xpath(login_xpath).click() time.sleep(0.5) except: return
[ "44291943+mimranfaruqi@users.noreply.github.com" ]
44291943+mimranfaruqi@users.noreply.github.com
07ae3fd425deb6e5c593ee9d9ae487d5398b8f25
e3765def4a180f1d51eaef3884448b0bb9be2cd3
/example/12.3.1_create_pygame_window/alien_invasion.py
136e506214bafb12d29f556453abfc4bb31417aa
[]
no_license
spearfish/python-crash-course
cbeb254efdf0c1ab37d8a7d2fa0409194f19fa2b
66bc42d41395cc365e066a597380a96d3282d30b
refs/heads/master
2023-07-14T11:04:49.276764
2021-08-20T10:02:27
2021-08-20T10:02:27
null
0
0
null
null
null
null
UTF-8
Python
false
false
424
py
#!/usr/bin/env python3 # modules import sys import pygame def run_game() : pygame.init() # pygame.display is a object that handles display. screen = pygame.display.set_mode((1200,800)) pygame.display.set_caption('Alien Invasion') while True : for event in pygame.event.get() : if event.type == pygame.QUIT : sys.exit() pygame.display.flip() run_game()
[ "jingchen@tutanota.com" ]
jingchen@tutanota.com
d063d7cbffb4226f8efbf9db037d712b216b8bb7
a8547f73463eef517b98d1085430732f442c856e
/pysam-0.13-py3.6-macosx-10.13-x86_64.egg/pysam/libcbgzf.py
366d86d29872fb9a2271270af8be79da14542344
[]
no_license
EnjoyLifeFund/macHighSierra-py36-pkgs
63aece1b692225ee2fbb865200279d7ef88a1eca
5668b5785296b314ea1321057420bcd077dba9ea
refs/heads/master
2021-01-23T19:13:04.707152
2017-12-25T17:41:30
2017-12-25T17:41:30
102,808,884
0
0
null
null
null
null
UTF-8
Python
false
false
303
py
def __bootstrap__(): global __bootstrap__, __loader__, __file__ import sys, pkg_resources, imp __file__ = pkg_resources.resource_filename(__name__, 'libcbgzf.cpython-36m-darwin.so') __loader__ = None; del __bootstrap__, __loader__ imp.load_dynamic(__name__,__file__) __bootstrap__()
[ "Raliclo@gmail.com" ]
Raliclo@gmail.com
01a24fbc30567db48254632abb8ff4ac747ce67b
2f63688febd21dc3ae6b19abfa79ad313c820154
/1004_Max_Consecutive_Ones_III/try_1.py
000de4088f9c246b4ee2a9c4740487216a9157ef
[]
no_license
novayo/LeetCode
cadd03587ee4ed6e35f60294070165afc1539ac8
54d0b3c237e0ffed8782915d6b75b7c6a0fe0de7
refs/heads/master
2023-08-14T00:35:15.528520
2023-07-30T05:56:05
2023-07-30T05:56:05
200,248,146
8
1
null
2022-11-19T04:37:54
2019-08-02T14:24:19
Python
UTF-8
Python
false
false
692
py
class Solution: def longestOnes(self, nums: List[int], k: int) -> int: ans = cur = cur_one = 0 ones = collections.deque() for num in nums: if num == 1: cur += 1 cur_one += 1 else: if k > 0: cur += 1 if len(ones) >= k: remove = ones.pop() cur -= remove+1 ones.appendleft(cur_one) cur_one = 0 else: if cur > 0: cur -= 1 ans = max(ans, cur) return ans
[ "shihchungyu@shichongyous-MacBook-Air.local" ]
shihchungyu@shichongyous-MacBook-Air.local
89b77faf800db2276ff6fd708a4125a3944939c6
d94c0d8541a05cc43b87813fd3b9d11f21dc5d76
/save_data.py
853bf7d418eb6430a401488587d0577bd2e2a8ac
[]
no_license
joakimzhang/test_ts
71cd5f36f65bab86282cd5e8354a4325e71136d0
05a9769ccda79e85b9f8a4f89af85c559958cbe9
refs/heads/master
2020-05-18T18:21:30.689401
2015-07-23T05:48:52
2015-07-23T05:48:52
39,547,686
0
1
null
null
null
null
UTF-8
Python
false
false
1,152
py
#-*- coding: utf-8 -*- import shelve from contextlib import closing class save_data(): def __init__(self): print "deal the data" def creat_shelf(self,key,value): with closing(shelve.open('test_shelf.db')) as s: s[key]=value def print_shelf(self,key): with closing(shelve.open('test_shelf.db')) as s: existing = s[key] print existing def del_shelf_key(self,key): with closing(shelve.open('test_shelf.db')) as s: del s[key] def get_all_val(self): with closing(shelve.open('test_shelf.db')) as s: print s data_dic = s.items() #print [a.decode('utf8') for a in s] return data_dic #return [a for a in s] if __name__ == '__main__': key='\\bjfile02\BJShare\Public\TS\Field_Captured_TS\中星9码流\20131108\file_ABS_20131108_11880MHz.ts\ABS_20131108_11880MHz.ts' value={'int':12,'float':9.5,'string':'sample data'} data = save_data() #data.creat_shelf(key,value) #data.del_shelf_key(key) #data.print_shelf(key) data.get_all_val()
[ "joakimzhang@163.com" ]
joakimzhang@163.com
092077973ed26e56e12866dd0b199df990ac44cf
c56268db8a4e08a705209142a6c171cd0f9aa7cc
/local_app/models.py
74d009980834f60502c75147e84d6b2082fa967a
[]
no_license
nguyenl1/local-app
e46b6e2e1ebb5fc799c5bf4b90e95782e6a327d9
434a7f16670f9afb560355e1035fe78ec16b2eb0
refs/heads/master
2022-11-12T06:56:21.332008
2020-07-03T16:14:40
2020-07-03T16:14:40
270,892,971
0
0
null
2020-07-03T16:14:41
2020-06-09T03:23:05
HTML
UTF-8
Python
false
false
1,678
py
from django.db import models from django.conf import settings from django.utils import timezone from django.contrib.auth import get_user_model import cloudinary import cloudinary.uploader import cloudinary.api from multiselectfield import MultiSelectField class SavedPin(models.Model): user = models.ForeignKey(get_user_model(), on_delete=models.CASCADE, null=True) bus_id = models.CharField(max_length = 200) name = models.CharField(max_length = 200) address = models.CharField(max_length = 200, blank=True) city = models.CharField(max_length = 200, blank=True) zip_code = models.CharField(max_length = 200, blank=True) state = models.CharField(max_length = 200, blank=True) image = models.TextField(max_length=2000, blank=True) image_2 = models.TextField(max_length=2000, blank=True) image_3 = models.TextField(max_length=2000, blank=True) latitude = models.TextField(max_length=2000, blank=True) longitude = models.TextField(max_length=2000, blank=True) class MyTrip(models.Model): user = models.ForeignKey(get_user_model(), on_delete=models.CASCADE, null=True) saved_pin = models.ForeignKey(SavedPin, on_delete=models.PROTECT, null=True) name = models.CharField(max_length = 200, blank = True) class SubmitPost(models.Model): site_name = models.CharField(max_length = 200) address = models.CharField(max_length=200) city = models.CharField(max_length = 200, blank=True) zip_code = models.CharField(max_length = 200, blank=True) state = models.CharField(max_length = 200, blank=True) publisher_name = models.CharField(max_length = 200) email = models.CharField(max_length = 200)
[ "lynnthuynguyen@yahoo.com" ]
lynnthuynguyen@yahoo.com
976016de1236d9a6ba795308ff368d105a8a28f7
629c93631250eda8968ee2903c9b264f18e5f47b
/combined_model.py
6b0c252162d840aafd67cfd428ef6381b5578dc3
[]
no_license
sohisudhir/Master-s-Thesis
f7bb66a67e7fd108a38815f95117ab8df977ea2c
36a74ec91db5779dc6ddf2814b1f58109463cb38
refs/heads/master
2023-01-21T14:02:35.160343
2020-12-02T10:45:35
2020-12-02T10:45:35
289,114,798
0
0
null
null
null
null
UTF-8
Python
false
false
30,782
py
# -*- coding: utf-8 -*- """Combined_model.ipynb Automatically generated by Colaboratory. Original file is located at https://colab.research.google.com/drive/1Z2qYEQ15gT32q9WFUJ2-LBhaDSglVI66 """ # Setup & Config import transformers from transformers import BertModel, BertTokenizer, AdamW, get_linear_schedule_with_warmup import torch from torch.nn import CrossEntropyLoss, MSELoss from torch import nn, optim from torch.utils.data import Dataset, DataLoader import torch.nn.functional as F import numpy as np import pandas as pd import random import copy import csv import re import argparse import os from sklearn.model_selection import train_test_split from sklearn import metrics from scipy.stats import pearsonr from scipy.stats import kendalltau from scipy.stats import spearmanr # RANDOM_SEED = 42 RANDOM_SEED = 42 np.random.seed(RANDOM_SEED) torch.manual_seed(RANDOM_SEED) """## **Data Preparation**""" class AbuseDataset(Dataset): def __init__(self, reviews, targets, c1,c2,c3, c_num, tokenizer, max_len, ids): self.reviews = reviews self.targets = targets self.c1 = c1 self.c2 = c2 self.c3 = c3 self.c_num = c_num self.tokenizer = tokenizer self.max_len = max_len self.ids = ids def __len__(self): return len(self.reviews) def __getitem__(self, item): c=["[PAD]","[PAD]","[PAD]"] review = str(self.reviews[item]) target = self.targets[item] c_num = self.c_num[item] c[0] = str(self.c1[item]) c[1] = str(self.c2[item]) c[2] = str(self.c3[item]) encoding = self.tokenizer.encode_plus( review, add_special_tokens=True, truncation=True, max_length=self.max_len, return_token_type_ids=False, pad_to_max_length=True, return_attention_mask=True, return_tensors='pt', ) idx = self.ids[item] context_input_ids = [] context_attention_mask = [] for i in range(0,3): encoding_context = self.tokenizer.encode_plus( c[i], add_special_tokens=True, truncation=True, max_length=self.max_len, return_token_type_ids=False, pad_to_max_length=True, return_attention_mask=True, return_tensors='pt') context_input_ids.append(encoding_context['input_ids'].flatten()) context_attention_mask.append(encoding_context['attention_mask'].flatten()) return { 'review_text': review, 'input_ids': encoding['input_ids'].flatten(), 'attention_mask': encoding['attention_mask'].flatten(), 'targets': torch.tensor(target, dtype=torch.float), 'context_input_ids': torch.stack(context_input_ids), 'context_attention_masks': torch.stack(context_attention_mask), 'context_num': c_num, 'ids': idx } class EmotionDataset(Dataset): def __init__(self, tweets, targets, tokenizer, max_len): self.tweets = tweets self.targets = targets self.tokenizer = tokenizer self.max_len = max_len def __len__(self): return len(self.tweets) def __getitem__(self, item): tweet = str(self.tweets[item]) target = self.targets[item] encoding = self.tokenizer.encode_plus( tweet, add_special_tokens=True, truncation = True, max_length=self.max_len, return_token_type_ids=False, pad_to_max_length=True, return_attention_mask=True, return_tensors='pt', ) return { 'tweet_text': tweet, 'input_ids': encoding['input_ids'].flatten(), 'attention_mask': encoding['attention_mask'].flatten(), 'targets': torch.tensor(target, dtype=torch.long) } class GeneralAttention(nn.Module): def __init__(self, sparsemax=False): super().__init__() self.linear = nn.Linear(768, 1) # self.normaliser = masked_softmax self.weights = [] def masked_softmax(self, vector, mask): while mask.dim() < vector.dim(): mask = mask.unsqueeze(1) # To limit numerical errors from large vector elements outside the mask, we zero these out. result = torch.nn.functional.softmax(vector * mask, dim=-1) result = result * mask result = result / ( result.sum(dim=-1, keepdim=True) + 1e-4 ) return result def forward(self, context, masks, batch_size): context = torch.cat(context, dim=1) context = context.reshape(-1,3,768) weights = self.linear(context).squeeze(-1) weights = self.masked_softmax(weights, masks) context = torch.bmm(weights.unsqueeze(dim=1), context) return context, weights def create_maintask_data_loader(df_train, tokenizer, max_len, batch_size, flag = 0): ds = AbuseDataset(reviews= df_train.comment.to_numpy(), targets= df_train.Score.to_numpy(), c1 = df_train.context1.to_numpy(), c2 = df_train.context2.to_numpy(), c3 = df_train.context3.to_numpy(), c_num = df_train.context_num.to_numpy(), tokenizer = tokenizer, max_len = max_len, ids = df_train.idx.to_numpy()) if(flag == 0): return DataLoader(ds, batch_size=batch_size, num_workers=4 ) else: return DataLoader(ds, batch_size=batch_size, num_workers=4, shuffle = True ) def create_auxtask_data_loader(df, tokenizer, max_len, batch_size, flag = 0): anger = df.anger.to_numpy() anticipation = df.anticipation.to_numpy() disgust = df.disgust.to_numpy() fear = df.fear.to_numpy() joy = df.joy.to_numpy() love = df.love.to_numpy() optimism = df.optimism.to_numpy() pessimism = df.pessimism.to_numpy() sadness = df.sadness.to_numpy() surprise = df.surprise.to_numpy() trust = df.trust.to_numpy() emotion = np.stack((anger, anticipation, disgust, fear, joy, love, optimism, pessimism, sadness, surprise, trust), axis = 1) ds = EmotionDataset( tweets=df.Tweet.to_numpy(), targets= emotion, tokenizer=tokenizer, max_len=max_len ) if(flag == 0): return DataLoader(ds, batch_size=batch_size, num_workers=4 ) else: return DataLoader(ds, batch_size=batch_size, num_workers=4, shuffle = True ) def clean_tweets(csvf): fname = 'cleaned_' + csvf with open(csvf, 'r', encoding = 'utf-8') as c, open(fname, 'w', encoding = 'UTF-8') as w: reader = csv.reader(c, delimiter = '\t') writer = csv.writer(w, delimiter = '\t') for i,row in enumerate(reader): if(i == 0): writer.writerow(row) continue row[1] = row[1].lower() row[1] = re.sub(r"#(\w+)", "HASHTAG", row[1]) row[1] = re.sub(r"(^|[^@\w])@(\w{1,15})", "_MTN_", row[1]) row[1] = re.sub(r"https?://(?:[-\w.]|(?:%[\da-fA-F]{2}))+", "_URL_", row[1]) writer.writerow(row) c.close() w.close() def prepare_data(abuse_files, sent_files, config): #Requirements BATCH_SIZE = config['batch_size'] MAX_LEN = config['max_len'] PRE_TRAINED_MODEL_NAME = config['PRE_TRAINED_MODEL_NAME'] tokenizer = BertTokenizer.from_pretrained(PRE_TRAINED_MODEL_NAME) # Arranging data loaders for main task a = abuse_files[0] # b = abuse_files[1] c = abuse_files[2] df_train = pd.read_csv(a) # df_val = pd.read_csv(b) df_test = pd.read_csv(c) # df_train = df_train[0:1000] # df_val = df_train[0:200] # df_test = df_train[0:800] print('Dimensions of abuse file') print(df_train.shape, 0, df_test.shape) data_loader_main = create_maintask_data_loader(df_train, tokenizer, MAX_LEN, BATCH_SIZE, 1) # val_data_loader_main = create_maintask_data_loader(df_val, tokenizer, MAX_LEN, BATCH_SIZE) test_data_loader_main = create_maintask_data_loader(df_test, tokenizer, MAX_LEN, BATCH_SIZE) # Arranging data loaders for auxiliary task -- SEMEVAL2018A a = sent_files[0] b = sent_files[1] c = sent_files[2] clean_tweets(a) clean_tweets(b) clean_tweets(c) df_train = pd.read_csv('cleaned_' + a, sep = '\t') # df_train = df_train[0:100] df_val = pd.read_csv('cleaned_' + b, sep = '\t') # df_val = df_val[0:20] df_test = pd.read_csv('cleaned_' + c, sep = '\t') # df_test = df_test[0:80] print('Dimensions of sentiment file') print(df_train.shape, df_val.shape, df_test.shape) data_loader_aux = create_auxtask_data_loader(df_train, tokenizer, MAX_LEN, BATCH_SIZE, 1) val_data_loader_aux = create_auxtask_data_loader(df_val, tokenizer, MAX_LEN, BATCH_SIZE) test_data_loader_aux = create_auxtask_data_loader(df_test, tokenizer, MAX_LEN, BATCH_SIZE) dataloaders = {'main_train': data_loader_main, 'main_val': [], 'main_test': test_data_loader_main, 'aux_train': data_loader_aux, 'aux_val': val_data_loader_aux, 'aux_test': test_data_loader_aux } return dataloaders """## **MODELS**""" class MSLELoss(nn.Module): def __init__(self): super().__init__() self.mse = nn.MSELoss(reduction = 'sum') def forward(self, pred, actual): return self.mse(torch.log(pred + 1.00005), torch.log(actual + 1.00005)) class multitask_conversation_model(nn.Module): def __init__(self, config): #num_labels, num_emotions, attention_dropout, fc_dropout): super(multitask_conversation_model, self).__init__() self.num_labels = config['abuse_classes'] self.num_emotions = config['sent_classes'] self.device = config['device'] PRE_TRAINED_MODEL_NAME = config['PRE_TRAINED_MODEL_NAME'] self.b_model = BertModel.from_pretrained(PRE_TRAINED_MODEL_NAME) self.bert = AdaptedBertModel(self.b_model, True, True, config['bert_dropout'], config['fc_dropout']) self.bert_config = self.bert.config self.attention = GeneralAttention() self.attention.to(self.device) self.main_regression = nn.Linear(self.bert_config.hidden_size, self.num_labels) self.aux_classifier = nn.Linear(self.bert_config.hidden_size, self.num_emotions) del(self.b_model) def forward(self, input_ids, token_type_ids=None, attention_mask=None, main_task=True, targets = None): if main_task: outputs = self.bert(input_ids, token_type_ids=token_type_ids, attention_mask=attention_mask, mode='main_task') # pooled_output = self.bert.pooler(outputs) pooled_output = outputs.mean(dim = 1) return pooled_output else: outputs = self.bert(input_ids, token_type_ids=token_type_ids, attention_mask=attention_mask, mode= 'auxiliary_task') # pooled_output = self.bert.pooler(outputs) pooled_output = outputs.mean(dim = 1) return pooled_output class AdaptedBertModel(nn.Module): def __init__(self, model, main_task, auxiliary_task, attention_dropout, fc_dropout): super().__init__() self.embeddings = model.embeddings self.encoder = BertEncoder(model.encoder.layer, main_task, auxiliary_task, attention_dropout, fc_dropout) self.config = model.config self.pooler = model.pooler def forward(self, input_ids, token_type_ids=None, attention_mask=None, mode="main_task"): if attention_mask is None: attention_mask = torch.ones_like(input_ids) if token_type_ids is None: token_type_ids = torch.zeros_like(input_ids) extended_attention_mask = attention_mask.unsqueeze(1).unsqueeze(2) extended_attention_mask = extended_attention_mask.to( dtype=next(self.parameters()).dtype ) extended_attention_mask = (1.0 - extended_attention_mask) * -10000.0 embeddings = self.embeddings(input_ids, token_type_ids) embeddings = self.encoder(embeddings, extended_attention_mask, mode) return embeddings # Hard parameter sharing setup : All layers but the last are shared # Last layer is task-specific class BertEncoder(nn.Module): def __init__(self, layers, main_task, auxiliary_task, attention_dropout, fc_dropout): super().__init__() self.layers = layers[:-1] self.output_attentions = False for layer in self.layers: layer.attention.self.dropout = nn.Dropout(attention_dropout) if main_task: self.layer_left = copy.deepcopy(layers[-1]) if auxiliary_task: self.layer_right = copy.deepcopy(layers[-1]) def forward(self, hidden, attention_mask, mode): all_attentions = () for layer in self.layers: hidden = layer(hidden, attention_mask) if self.output_attentions: all_attentions = all_attentions + (hidden[1],) hidden = hidden[0] if mode == "main_task": hidden = self.layer_left(hidden, attention_mask) elif mode == "auxiliary_task": hidden = self.layer_right(hidden, attention_mask) outputs = hidden[0] if self.output_attentions: outputs = outputs + (all_attentions,) return outputs def evaluation_metrics(preds, targets): with torch.no_grad(): tp = torch.zeros(preds.shape[1]) tn = torch.zeros(preds.shape[1]) fp = torch.zeros(preds.shape[1]) fn = torch.zeros(preds.shape[1]) for n,pred in enumerate(preds): for j,pr in enumerate(pred): t = targets[n][j] if(pr == 0): if(t == 0): tn[j] += 1 else: fn[j] += 1 elif(pr == 1): if(t == 0): fp[j] += 1 else: tp[j] += 1 #Micro num = torch.sum(tp) deno_prec = torch.zeros(preds.shape[1]) deno_rec = torch.zeros(preds.shape[1]) for j,val in enumerate(deno_prec): deno_prec[j] = tp[j] + fp[j] deno_rec[j] = tp[j] + fn[j] den = torch.sum(deno_prec) if(den == 0): micro_precision = 0 else: micro_precision = num.item()/den.item() den = torch.sum(deno_rec) if(den == 0): micro_recall = 0 else: micro_recall = num.item()/den.item() numerator = 2 * micro_precision * micro_recall denominator = micro_precision + micro_recall if(denominator == 0): micro_f1 = 0 else: micro_f1 = numerator/denominator # print(micro_precision, micro_recall, micro_f1) #MACRO precision = torch.zeros(preds.shape[1]) recall = torch.zeros(preds.shape[1]) for j,val in enumerate(precision): if(tp[j] + fp[j] == 0): precision[j] = 0 else: precision[j] = tp[j]/(tp[j] + fp[j]) if(tp[j] + fn[j] == 0): recall[j] = 0 else: recall[j] = tp[j]/(tp[j] + fn[j]) f1 = torch.zeros(preds.shape[1]) for j,val in enumerate(f1): num = 2 * precision[j] * recall[j] deno = precision[j] + recall[j] if(deno == 0): f1[j] = 0 else: f1[j] = num/deno macro_precision = torch.mean(precision) macro_recall = torch.mean(recall) macro_f1 = torch.mean(f1) # print(macro_precision, macro_recall, macro_f1) return micro_f1, macro_f1 def eval_model(model, data_loader, device, mode): model = model.to(device) model = model.eval() loss_fn_main = nn.MSELoss().to(device) loss_fn_aux = nn.BCEWithLogitsLoss().to(device) if(mode == 'main_task'): p = [] t = [] loss_m = [] ids = [] emotion_pred = [] c1_pred = [] c2_pred = [] c3_pred = [] with torch.no_grad(): for d in data_loader: input_ids = d["input_ids"].to(device) attention_mask = d["attention_mask"].to(device) targets = d["targets"].to(device) context_input_ids = d["context_input_ids"].to(device) context_attention_masks = d["context_attention_masks"].to(device) context_num = d['context_num'].to(device) outputs = model.forward(input_ids=input_ids, token_type_ids = None, attention_mask=attention_mask, main_task = True) out_encoding = [] for i in range(len(context_input_ids)): c = model.forward(input_ids=context_input_ids[i].to(device),attention_mask=context_attention_masks[i].to(device)) out_encoding.append(c) ops = model(input_ids=context_input_ids[i].to(device), token_type_ids = None, attention_mask=context_attention_masks[i].to(device), main_task = False) c1,c2,c3 = torch.unbind(ops, dim = 0) logits = model.module.aux_classifier(c1.unsqueeze(dim = 0)) predictions = torch.sigmoid(logits) preds = torch.gt(predictions, 0.5).int() c1_pred.extend(preds) logits = model.module.aux_classifier(c2.unsqueeze(dim = 0)) predictions = torch.sigmoid(logits) preds = torch.gt(predictions, 0.5).int() c2_pred.extend(preds) logits = model.module.aux_classifier(c3.unsqueeze(dim = 0)) predictions = torch.sigmoid(logits) preds = torch.gt(predictions, 0.5).int() c3_pred.extend(preds) mask = torch.zeros([input_ids.shape[0],3]) for i in range(len(context_num)): arr = np.zeros(3) arr[:context_num[i]] = 1 mask[i] = torch.tensor(arr) mask = mask.to(device) weighted, weights = model.module.attention.forward(out_encoding, mask, config['batch_size']) main_context = outputs.add(weighted.squeeze(dim=1)) val = model.module.main_regression(main_context) predictions = torch.tanh(val) loss = loss_fn_main(predictions.squeeze(dim = 1), targets) p.extend(predictions.squeeze(dim=1).to('cpu').detach().numpy()) t.extend(targets.to('cpu').detach().numpy()) ids.extend(d["ids"].to('cpu').detach().numpy()) loss_m.append(loss.item()) ops = model(input_ids=input_ids, token_type_ids = None, attention_mask=attention_mask, main_task = False, targets = targets) logits = model.module.aux_classifier(ops) predictions = torch.sigmoid(logits) preds = torch.gt(predictions, 0.5).int() emotion_pred.extend(preds) with open('testing_preds_mtl_combo.csv', 'a', encoding = 'utf-8') as f: writer = csv.writer(f) # writer.writerow(['ID', 'Prediction', 'Target']) row = [] for i,idx in enumerate(ids): row.append(idx) row.append(p[i]) row.append(t[i]) row.append(emotion_pred[i].to('cpu').detach().numpy()) row.append(c1_pred[i].to('cpu').detach().numpy()) row.append(c2_pred[i].to('cpu').detach().numpy()) row.append(c3_pred[i].to('cpu').detach().numpy()) writer.writerow(row) row = [] f.close() pear = pearsonr(np.array(t),np.array(p)) spear = spearmanr(np.array(t),np.array(p)) tau = kendalltau(np.array(t),np.array(p)) loss = np.mean(loss_m) return pear[0], spear[0], tau[0], loss elif(mode == 'auxiliary_task'): accuracies = [] loss_a = [] micro_f1 = [] macro_f1 = [] with torch.no_grad(): for d in data_loader: input_ids = d["input_ids"].to(device) attention_mask = d["attention_mask"].to(device) targets = d["targets"].to(device) # logits = model(input_ids=input_ids, token_type_ids = None, attention_mask=attention_mask, main_task = False, targets = targets) ops = model(input_ids=input_ids, token_type_ids = None, attention_mask=attention_mask, main_task = False, targets = targets) logits = model.module.aux_classifier(ops) predictions = torch.sigmoid(logits) loss = loss_fn_aux(logits.float(), targets.float()) loss_a.append(loss.item()) preds = torch.gt(predictions, 0.5).int() mic_f1, mac_f1 = evaluation_metrics(preds, targets) micro_f1.append(mic_f1) macro_f1.append(mac_f1) avg_micro_f1 = np.mean(micro_f1) avg_macro_f1 = np.mean(macro_f1) loss = np.mean(loss_a) return avg_micro_f1, avg_macro_f1, loss def train_epoch(model, dataloaders, device, config): model = model.to(config['device']) model = model.train() least_loss = 100.0 data_loader_main = dataloaders['main_train'] data_loader_aux = dataloaders['aux_train'] val_data_loader_main = dataloaders['main_val'] val_data_loader_aux = dataloaders['aux_val'] loss_fn_main = nn.MSELoss().to(config['device']) loss_fn_aux = nn.BCEWithLogitsLoss().to(config['device']) device = config['device'] optimizer_main = AdamW(model.module.bert.parameters(), lr = config['lr_main'], weight_decay= 1e-4, correct_bias=False) optimizer_main2 = AdamW(model.module.main_regression.parameters(), lr = config['lr_main']*10, weight_decay= 1e-4, correct_bias=False) optimizer_main3 = AdamW(model.module.attention.parameters(), lr = config['lr_main']*10, weight_decay= 1e-4, correct_bias=False) # optimizer_main = torch.optim.Adam(model.parameters(), lr = 0.001) # optimizer_aux = torch.optim.Adam(model.parameters(), lr = config['lr_aux']) optimizer_aux = AdamW(model.module.bert.parameters(), lr = config['lr_aux'], weight_decay= 1e-4, correct_bias=False) optimizer_aux2 = AdamW(model.module.aux_classifier.parameters(), lr = config['lr_aux']*10, weight_decay= 1e-4, correct_bias=False) total_steps = len(data_loader_main) * config['num_epochs'] scheduler_main = get_linear_schedule_with_warmup( optimizer_main, num_warmup_steps=0, num_training_steps=total_steps ) total_steps = len(data_loader_aux) * config['num_epochs'] scheduler_aux = get_linear_schedule_with_warmup( optimizer_aux, num_warmup_steps=0, num_training_steps=total_steps ) # optimizer_main = AdamW(model.parameters(), lr = config['lr_main'], weight_decay= 1e-4, correct_bias=False) # optimizer_aux = AdamW(model.parameters(), lr = config['lr_aux'], weight_decay= 1e-4, correct_bias=False) # total_steps = len(data_loader_main) * config['num_epochs'] # scheduler = get_linear_schedule_with_warmup( # optimizer_main, # num_warmup_steps=0, # num_training_steps=total_steps # ) coin_flips = [] #main_task for i in range(len(data_loader_main)): coin_flips.append(0) #auxiliary task for i in range(len(data_loader_aux)): coin_flips.append(1) val_counter = 0 for epoch in range(config['num_epochs']): if(epoch >= 3): print('Freezing Bert!') for param in model.module.bert.encoder.parameters(): param.requires_grad = False print("Starting epoch {}".format(epoch)) random.shuffle(coin_flips) loss_m = [] loss_a = [] p = [] t = [] micro_f1 = [] macro_f1 = [] accuracies = [] main_dl = iter(data_loader_main) aux_dl = iter(data_loader_aux) for i in coin_flips: if(i == 0): #MAIN_TASK try: d = next(main_dl) except: main_dl = iter(data_loader_main) d = next(main_dl) # print('In main task') input_ids = d["input_ids"].to(device) attention_mask = d["attention_mask"].to(device) targets = d["targets"].to(device) context_input_ids = d["context_input_ids"].to(device) context_attention_masks = d["context_attention_masks"].to(device) context_num = d['context_num'].to(device) outputs = model.forward(input_ids=input_ids, token_type_ids = None, attention_mask=attention_mask, main_task = True) out_encoding = [] for i in range(len(context_input_ids)): c = model.forward(input_ids=context_input_ids[i].to(device),attention_mask=context_attention_masks[i].to(device)) out_encoding.append(c) mask = torch.zeros([input_ids.shape[0],3]).to(device) for i in range(len(context_num)): arr = np.zeros(3) arr[:context_num[i]] = 1 mask[i] = torch.tensor(arr) weighted,_ = model.module.attention.forward(out_encoding, mask, config['batch_size']) main_context = outputs.add(weighted.squeeze(dim=1)) val = model.module.main_regression(main_context) predictions = torch.tanh(val) loss = loss_fn_main(predictions.squeeze(dim = 1), targets) p.extend(predictions.squeeze(dim=1).to('cpu').detach().numpy()) t.extend(targets.to('cpu').detach().numpy()) loss_m.append(loss.item()) loss.backward() # nn.utils.clip_grad_norm_(model.parameters(), max_norm=1.0) optimizer_main.step() optimizer_main2.step() optimizer_main3.step() scheduler_main.step() optimizer_main.zero_grad() optimizer_main2.zero_grad() optimizer_main3.zero_grad() val_counter += 1 else: try: d = next(aux_dl) except: aux_dl = iter(data_loader_aux) d = next(aux_dl) input_ids = d["input_ids"].to(device) attention_mask = d["attention_mask"].to(device) targets = d["targets"].to(device) ops = model(input_ids=input_ids, token_type_ids = None, attention_mask=attention_mask, main_task = False, targets = targets) logits = model.module.aux_classifier(ops) predictions = torch.sigmoid(logits) loss = loss_fn_aux(logits.float(), targets.float()) loss_a.append(loss.item()) loss = loss * 0.4 loss.backward() # nn.utils.clip_grad_norm_(model.parameters(), max_norm=1.0) optimizer_aux.step() optimizer_aux2.step() scheduler_aux.step() optimizer_aux.zero_grad() optimizer_aux2.zero_grad() preds = torch.gt(predictions, 0.5).int() mic_f1, mac_f1 = evaluation_metrics(preds, targets) micro_f1.append(mic_f1) macro_f1.append(mac_f1) # print('aux task completed') val_counter +=1 pear = pearsonr(np.array(t),np.array(p)) spear = spearmanr(np.array(t),np.array(p)) tau = kendalltau(np.array(t),np.array(p)) avg_micro_f1 = np.mean(micro_f1) avg_macro_f1 = np.mean(macro_f1) print("Epoch {}. Training Pearson {}.Training Pearson {}.Training Spearman {} Training Loss {}".format(epoch, pear[0], spear[0], tau[0], np.mean(loss_m))) print("Epoch {}. Training Micro F1 {}.Training Macro F1 {}.Training Loss {}".format(epoch, avg_micro_f1, avg_macro_f1, np.mean(loss_a))) # pearson, spearman, kendall, loss = eval_model(model, val_data_loader_main, device, mode = 'main_task') # print("MAIN: Epoch {}. Validation Pearson {}.Validation Spearman {}. Validation Kendall {}. Validation Loss {}".format(epoch, pearson, spearman, kendall,loss)) # if(loss < least_loss): # print('Saving best model') # least_loss = loss # state = {'epoch': epoch+1, 'state_dict': model.state_dict(), 'optimizer_main': optimizer_main.state_dict(), # 'optimizer_aux': optimizer_aux.state_dict()}#, 'scheduler_main': scheduler_main, 'scheduler_aux': scheduler_aux} # torch.save(state, 'mtl_best_model.ckpt') avg_micro_f1, avg_macro_f1, loss = eval_model(model, val_data_loader_aux, device, mode = 'auxiliary_task') print("AUX: Epoch {}.Validation Micro F1 {}.Validation Macro F1 {}. Validation Loss {}".format(epoch, avg_micro_f1, avg_macro_f1, loss)) state = {'epoch': epoch+1, 'state_dict': model.state_dict(), 'optimizer_main': optimizer_main.state_dict(), 'optimizer_aux': optimizer_aux.state_dict()}#, 'scheduler_main': scheduler_main, 'scheduler_aux': scheduler_aux} print('Saving last model') torch.save(state, 'mtl_combo_last_model.ckpt') """# **Calling the model**""" if __name__ == "__main__": parser = argparse.ArgumentParser(description="Enter args") parser.add_argument('--PRE_TRAINED_MODEL_NAME', default="bert-base-cased", type=str) parser.add_argument('--batch_size', default=16, type=int) parser.add_argument('--max_len', default=200, type=int) parser.add_argument('--abuse_classes', default=1, type=int) parser.add_argument('--sent_classes', default=11, type=int) parser.add_argument('--bert_dropout', default=0.1, type=float) parser.add_argument('--fc_dropout', default=0.4, type=float) parser.add_argument('--num_epochs', default=5, type=int) parser.add_argument('--lr_main', default=3e-5, type=float) parser.add_argument('--lr_aux', default=3e-5, type=float) parser.add_argument('--wd', default=1e-4, type=float) parser.add_argument('--csv_index', default = 1, type = int) args = parser.parse_args() print('************************************************************************************') # print('bert_dropout', bert_dropout, 'fc_dropout', fc_dropout) config = { 'PRE_TRAINED_MODEL_NAME': 'bert-base-cased', 'batch_size': args.batch_size, 'max_len': args.max_len, 'abuse_classes': args.abuse_classes, 'sent_classes': args.sent_classes, 'bert_dropout': args.bert_dropout, 'fc_dropout': args.fc_dropout, 'device': torch.device('cuda') if torch.cuda.is_available() else torch.device('cpu'), 'num_epochs': args.num_epochs, 'lr_main':args.lr_main, 'lr_aux': args.lr_aux } train_file = 'train' + str(args.csv_index) + '.csv' test_file = 'test' + str(args.csv_index) + '.csv' abuse_files = [train_file, '', test_file] # abuse_files = ['train.csv', 'val.csv', 'test.csv']#'comm_uqs_with_convo.csv' #'main_cmv_datatset_10000.csv' sent_files = ['train.tsv', 'dev.tsv', 'test.tsv'] dataloaders = prepare_data(abuse_files, sent_files, config) model = multitask_conversation_model(config) device = config['device'] print('DEVICE IS', device) if torch.cuda.device_count() > 1: print("Let's use", torch.cuda.device_count(), "GPUs!") model = nn.DataParallel(model).to(device) train_epoch(model, dataloaders, device, config) print('End of training....') test_data_loader_main = dataloaders['main_test'] test_data_loader_aux = dataloaders['aux_test'] # checkpoint = torch.load('mtl_best_model.ckpt') # test_model = multitask_conversation_model(config) # if torch.cuda.device_count() > 1: # print("Let's use", torch.cuda.device_count(), "GPUs!") # test_model = nn.DataParallel(test_model).to(device) # test_model = test_model.to(device) # test_model.load_state_dict(checkpoint['state_dict']) # print('Loaded best model') # pearson, spearman, kendall, loss = eval_model(test_model, test_data_loader_main, device, mode = 'main_task') # print("MAIN:. Test Pearson {}.Test Spearman {}.Test kendall {}. Test Loss {}".format(pearson, spearman, kendall, loss)) # avg_micro_f1, avg_macro_f1, loss = eval_model(test_model, test_data_loader_aux, device, mode = 'auxiliary_task') # print("AUX: Test Micro F1 {}.Test Macro F1 {}. Test Loss {}".format(avg_micro_f1, avg_macro_f1, loss)) print('Loaded last model(Sanity check)') pearson, spearman, kendall, loss = eval_model(model, test_data_loader_main, device, mode = 'main_task') print("MAIN:. Test Pearson {}.Test Spearman {}.Test kendall {}. Test Loss {}".format(pearson, spearman, kendall, loss)) avg_micro_f1, avg_macro_f1, loss = eval_model(model, test_data_loader_aux, device, mode = 'auxiliary_task') print("AUX: Test Micro F1 {}.Test Macro F1 {}. Test Loss {}".format(avg_micro_f1, avg_macro_f1, loss)) os.remove(train_file) os.remove(test_file) os.remove('mtl_combo_last_model.ckpt')
[ "noreply@github.com" ]
noreply@github.com
332fa4a0deb504844a72c413b5933587f949574b
dac141981cfefbc3da1167a3cbd9bfa2c02ebec5
/src/tokenizer.py
b0289568c1738b5d7bd3257edf188f8842be1555
[]
no_license
meciwo/Knowledge-based_Meme_Caption_Generator
65b8ce46089a830618906d2c6e40a6fbbbc7507a
ad3d62d50828609da4d7e8de64c57c695304346d
refs/heads/main
2023-04-01T23:45:27.040339
2021-04-08T18:10:08
2021-04-08T18:10:08
350,203,673
1
0
null
null
null
null
UTF-8
Python
false
false
501
py
import MeCab from dotenv import load_dotenv import os import emoji load_dotenv() mecab_path = os.environ["MECAB_PATH"] mecab = MeCab.Tagger(f"-Owakati -d {mecab_path}") mecab.parse("") # バグ対処 def remove_emoji(src_str): return "".join(c for c in src_str if c not in emoji.UNICODE_EMOJI) def tokenize(text): text = remove_emoji(str(text)) text = text.replace("「", "").replace("」", "").replace("、", "") result = mecab.parse(text).strip().split(" ") return result
[ "shanshan0474@gmail.com" ]
shanshan0474@gmail.com
a5ec4c22e8526a56d17ae4d199df63900a4fd74c
7dfe37c9b4cb8512a49bb7abe6311b553e229fc5
/src/utils/clases/metodos_strings.py
44eb6a68120721b812795a4f6c0db7457ef5fb16
[]
no_license
ArmandoBerlanga/python_playground
7e28c992bd9934f4b72a58f076c2b3f39137951a
4ee243e3d4a02667a8ed05fc526e5373a118e0d9
refs/heads/main
2023-04-05T12:37:11.478937
2021-03-31T00:43:30
2021-03-31T00:43:30
349,169,416
0
0
null
null
null
null
UTF-8
Python
false
false
2,871
py
# Programa creado por José Armando Berlanga Mendoza # Creado el 17 de febrero de 2021 # Descripción: ejercicio sobre recursividad def voltear_palabra (s): # metodo para voltear una palabra u oracion completamente l = len(s) if (l == 0): return "" else: return s[l-1:] + voltear_palabra (s[0:l-1]) def es_palindrome (s): # metodo para evaluar si una palabra u oracion es palindrome (que estan iguales al derecho y al reves) if (len(s) > 1): if s[len(s)-1] == s [0]: return es_palindrome (s [1 : len(s)-1]) else: return False else: return True def imprimir_pino (reglon, cont): # metodo para la impresion de un pino centrado if (reglon == 0): return "" else: for i in range (reglon, 0, -1): print (" ", end = "") if i == 0: break for i in range (cont+1, 0, -1): print ("*", end = "") if i == 0: break print() return imprimir_pino(reglon-1, cont+2) def formatear_palabra (s): # metodo de formateo acentos = "áéíóú" sinAcentos = "aeiou" s = s.lower().replace(" ", "").replace(".", "").replace(",", "") for char in s: pos = acentos.find(char) if pos != -1: s = s.replace(acentos[pos], sinAcentos[pos]) return s # Given a string s, find the length of the longest substring without repeating characters. def longest_substring(s): if len(s) == 1: return 1 elif s == "": return 0 conts = [] for i in range (len(s)-1): chars = [] cont = 0 j = i while s[j] not in chars and j < len(s)-1: cont+=1 chars.append(s[j]) j+=1 conts.append(cont) return max (conts) if __name__ == '__main__': option = -1 while (option == -1): print ("\n[1] Voltear el orden de una palabra u oracion\n[2] Evaluar si una palabra o frase es palindrome\n[3] Imprimir un pino") option = int (input("\nIngrese un numero segun las opciones dadas: ")) if option != 1 and option != 2 and option != 3: print("\nNo has ingresado un num valido, vuelve a ingresarlo") option = -1 if option == 1: s = input("\nIngrese el texto a voltear: ") print("\nResultado: " + voltear_palabra(s)) elif option == 2: s = input("\nIngrese el texto a evaluar: ") # s = "A mamá Roma le aviva el amor a papá y a papá Roma le aviva el amor a mamá." print ("\n\"" + s +"\"" + ", es palindrome" if (es_palindrome(formatear_palabra(s))) else ", no es Palindrome") else: pisos = int (input("\nIngrese el numero de pisos de la piramide: ")) print() imprimir_pino(pisos, 0) print()
[ "Jose.berlangam@udem.edu" ]
Jose.berlangam@udem.edu
a1afae0b9a14f320f59826b7a6e3c27d9d04847f
bddcad1331e2ea68d2ffc7e3f0478d8776fea5d8
/Administratie/Literatuurstudie/bijlagen/convert.py
3321a9340e1be75db22d68de619847b73eb22cbe
[]
no_license
4ilo/masterproef
693b9123e7e3fa64ef50960581d64f5116f8b045
c5ad81cee83b354262b2dc9d0dced0cbcf0c2f66
refs/heads/master
2020-03-31T08:44:10.246829
2019-06-24T08:56:32
2019-06-24T08:56:32
152,070,317
0
0
null
null
null
null
UTF-8
Python
false
false
1,548
py
import xml.etree.ElementTree as ET import argparse import os parser = argparse.ArgumentParser() parser.add_argument("file") args = parser.parse_args() classes = [] def getLabels(root): labels = root.find('task').find('labels') for label in labels: classes.append(label.find('name').text) with open('cust.names', 'w') as names: names.write("\n".join(classes)) if __name__ == '__main__': tree = ET.parse(args.file) root = tree.getroot() print(root.tag) if not os.path.isdir('Anotations_yolo'): os.mkdir('Anotations_yolo') for child in root: if child.tag == 'meta': getLabels(child) if child.tag == 'image': image = child.get('name') w = float(child.get('width')) h = float(child.get('height')) boxes = '' for box in child.findall('box'): xtl = float(box.get('xtl')) ytl = float(box.get('ytl')) xbr = float(box.get('xbr')) ybr = float(box.get('ybr')) width = xbr - xtl height = ybr - ytl x = xtl + (width/2) #center y = ytl + (height/2) boxes += '{} {} {} {} {}\n'.format(classes.index(box.get('label')), x/w, y/h, width/w, height/h) with open('Anotations_yolo/{}.txt' .format(os.path.splitext(image)[0]), 'w') as file: file.write(boxes)
[ "oliviervandeneede@hotmail.com" ]
oliviervandeneede@hotmail.com
90105714e157a472def98eca28ce8f9da9114066
39eb95d42ff47be6c9be8316cba3d1a0eca1d71f
/shirai-ri/tutorial04/test_hmm.py
dc7995326f072f2a207377d17159525881467139
[]
no_license
reo11/NLPtutorial2018
ac6cc059b4d428e5e67dba9e3b2d176b003ee34c
f733ed7d0479c8ed9b1224d6fc61b74748031ff1
refs/heads/master
2020-06-28T15:09:17.885949
2018-08-30T01:35:28
2018-08-30T01:35:28
null
0
0
null
null
null
null
UTF-8
Python
false
false
3,046
py
# coding: utf-8 # In[ ]: import codecs from collections import defaultdict import math def test_hmm(test_file, model_file, answer): transition = defaultdict(int) emission = defaultdict(int) possible_tags = defaultdict(int) lam = 0.95 lam_unk = 1- lam V = 1000000 with codecs.open(model_file, 'r', 'utf8') as model_f, codecs.open(test_file, 'r', 'utf8') as test_f, codecs.open(answer, 'w', 'utf8') as answer_f: # モデル読み込み for line in model_f: typ, context, word, prob = line.strip().split() possible_tags[context] = 1 # 可能なタグとして保存 if typ == 'T': transition['{} {}'.format(context, word)] = float(prob) else: emission['{} {}'.format(context, word)] = float(prob) # 実際のテスト for line in test_f: words = line.strip().split() best_score = dict() best_edge = dict() best_score['0 <s>'] = 0 best_edge['0 <s>'] = 'NULL' #前向き for i in range(0, len(words)): for prev in possible_tags.keys(): for nex in possible_tags.keys(): if '{} {}'.format(i, prev) in best_score and '{} {}'.format(prev,nex) in transition: score = best_score['{} {}'.format(i, prev)] - math.log(transition['{} {}'.format(prev, nex)], 2) - math.log(lam * emission['{} {}'.format(nex, words[i])] + lam_unk/V, 2) if '{} {}'.format(i+1, nex) not in best_score or best_score['{} {}'.format(i+1, nex)] > score: best_score['{} {}'.format(i+1, nex)] = score best_edge['{} {}'.format(i+1, nex)] = '{} {}'.format(i, prev) # 最後の処理 for prev in possible_tags.keys(): if '{} {}'.format(len(words), prev) in best_score and '{} </s>'.format(prev) in transition: score = best_score['{} {}'.format(len(words), prev)] - math.log(transition['{} </s>'.format(prev)], 2) if '{} </s>'.format(len(words) + 1) not in best_score or best_score['{} </s>'.format(len(words) + 1)] > score: best_score['{} </s>'.format(len(words) + 1)] = score best_edge['{} </s>'.format(len(words) + 1)] = '{} {}'.format(len(words), prev) # 後ろ向き tags = [] next_edge = best_edge['{} </s>'.format(len(words) + 1)] while next_edge != '0 <s>': position, tag = next_edge.split() tags.append(tag) next_edge = best_edge[next_edge] tags.reverse() answer_f.write(' '.join(tags) + '\n') if __name__ == '__main__': test_hmm('./nlptutorial-master/data/wiki-en-test.norm', './model_file.txt', 'my_answer.pos')
[ "tarokirs@gmail.com" ]
tarokirs@gmail.com
450170e9d9e65aabbc3043829fbde44a95b4602c
e6c88bc10f82c2e0a9a40666f14b4e81418516ee
/pharmacist/models.py
a2a7ca6b56b5dcdc35ddc314841594f885f3c473
[]
no_license
Masher828/HospitalManagementSystem
bb2f819edb1da52e34fc7dfff93dfaf79e9c0dd5
72d7875159098361c8e6f53d3076ba8b612eb279
refs/heads/masher
2022-12-18T02:15:06.671642
2020-06-30T11:54:26
2020-06-30T11:54:26
275,923,707
0
1
null
2020-10-01T07:31:35
2020-06-29T20:52:49
JavaScript
UTF-8
Python
false
false
438
py
from django.db import models class Medicinemaster(models.Model): ws_med_id = models.AutoField(primary_key=True, serialize = False) ws_med_name= models.CharField(max_length=255) ws_stock_qty = models.IntegerField() ws_price= models.FloatField() def __str__(self): return self.ws_med_name class Medicineissued(models.Model): ws_pat_id = models.IntegerField() ws_med_id = models.IntegerField() ws_qty = models.IntegerField()
[ "manish.cse828@gmail.com" ]
manish.cse828@gmail.com
452891ccd3170505662e4cb079ff70d7eff7a2c8
f722d5d2fa5a516579dc3cfb4337495a39c05b54
/app/test/src/data.py
51666fabe2b5b0642866ea3f76fca83fc8ab9001
[]
no_license
Engineerlin/DS-Practice-AT3
f5df59b59f66da7df25ad39094e434f670b4ebc4
06283b5d0e17812434b781dd41b4c615b8b94958
refs/heads/master
2023-08-27T17:41:54.163660
2021-11-06T08:39:25
2021-11-06T08:39:25
421,269,385
0
1
null
2021-11-06T08:39:26
2021-10-26T03:44:25
Python
UTF-8
Python
false
false
2,026
py
import streamlit as st from dataclasses import dataclass import pandas as pd @dataclass class Dataset: name: str df: pd.DataFrame def get_name(self): """ Return filename of loaded dataset """ return self.name def get_n_rows(self): """ Return number of rows of loaded dataset """ return self.df.shape[0] def get_n_cols(self): """ Return number of columns of loaded dataset """ return self.df.shape[1] def get_cols_list(self): """ Return list column names of loaded dataset """ return list(self.df.columns.values) def get_cols_dtype(self): """ Return dictionary with column name as keys and data type as values """ return self.df.dtypes.apply(lambda x:x.name).to_dict() def get_n_duplicates(self): """ Return number of duplicated rows of loaded dataset """ return self.df.duplicated().sum() def get_n_missing(self): """ Return number of rows with missing values of loaded dataset """ return self.df.shape[0]-self.df.dropna().shape[0] def get_head(self, n=5): """ Return Pandas Dataframe with top rows of loaded dataset """ return self.df.head(n) def get_tail(self, n=5): """ Return Pandas Dataframe with bottom rows of loaded dataset """ return self.df.tail(n) def get_sample(self, n=5): """ Return Pandas Dataframe with random sampled rows of loaded dataset """ return self.df.sample(n) def get_numeric_columns(self): """ Return list column names of numeric type from loaded dataset """ return list(self.df.select_dtypes(['float']).columns) def get_text_columns(self): """ Return list column names of text type from loaded dataset """ return list(self.df.select_dtypes(['object']).columns) def get_date_columns(self): """ Return list column names of datetime type from loaded dataset """ return list(self.df.select_dtypes(['datetime64']).columns)
[ "kailin.zhou@student.uts.edu.au" ]
kailin.zhou@student.uts.edu.au
c1fda1a470ad681c3a1a16d4e839b87151b19b33
6f6d215a4f0a1c30eeb5a08c8a36016fc351998a
/zcls/model/recognizers/resnet/torchvision_resnet.py
040bc44da6892b30585f415d6130a4b2fe65cecc
[ "Apache-2.0" ]
permissive
Quebradawill/ZCls
ef9db2b54fbee17802f3342752e3d4fe4ef9d2c5
ade3dc7fd23584b7ba597f24ec19c02ae847673e
refs/heads/master
2023-04-15T23:25:18.195089
2021-04-29T07:05:46
2021-04-29T07:05:46
null
0
0
null
null
null
null
UTF-8
Python
false
false
3,125
py
# -*- coding: utf-8 -*- """ @date: 2021/2/20 上午10:28 @file: torchvision_resnet.py @author: zj @description: """ from abc import ABC import torch.nn as nn from torch.nn.modules.module import T from torchvision.models.resnet import resnet18, resnet50, resnext50_32x4d from zcls.config.key_word import KEY_OUTPUT from zcls.model import registry from zcls.model.norm_helper import freezing_bn class TorchvisionResNet(nn.Module, ABC): def __init__(self, arch="resnet18", num_classes=1000, torchvision_pretrained=False, pretrained_num_classes=1000, fix_bn=False, partial_bn=False, zero_init_residual=False): super(TorchvisionResNet, self).__init__() self.num_classes = num_classes self.fix_bn = fix_bn self.partial_bn = partial_bn if arch == 'resnet18': self.model = resnet18(pretrained=torchvision_pretrained, num_classes=pretrained_num_classes, zero_init_residual=zero_init_residual) elif arch == 'resnet50': self.model = resnet50(pretrained=torchvision_pretrained, num_classes=pretrained_num_classes, zero_init_residual=zero_init_residual) elif arch == 'resnext50_32x4d': self.model = resnext50_32x4d(pretrained=torchvision_pretrained, num_classes=pretrained_num_classes, zero_init_residual=zero_init_residual) else: raise ValueError('no such value') self.init_weights(num_classes, pretrained_num_classes) def init_weights(self, num_classes, pretrained_num_classes): if num_classes != pretrained_num_classes: fc = self.model.fc fc_features = fc.in_features self.model.fc = nn.Linear(fc_features, num_classes) nn.init.normal_(self.model.fc.weight, 0, 0.01) nn.init.zeros_(self.model.fc.bias) def train(self, mode: bool = True) -> T: super(TorchvisionResNet, self).train(mode=mode) if mode and (self.partial_bn or self.fix_bn): freezing_bn(self, partial_bn=self.partial_bn) return self def forward(self, x): x = self.model(x) return {KEY_OUTPUT: x} @registry.RECOGNIZER.register('TorchvisionResNet') def build_torchvision_resnet(cfg): torchvision_pretrained = cfg.MODEL.RECOGNIZER.TORCHVISION_PRETRAINED pretrained_num_classes = cfg.MODEL.RECOGNIZER.PRETRAINED_NUM_CLASSES fix_bn = cfg.MODEL.NORM.FIX_BN partial_bn = cfg.MODEL.NORM.PARTIAL_BN # for backbone arch = cfg.MODEL.BACKBONE.ARCH zero_init_residual = cfg.MODEL.RECOGNIZER.ZERO_INIT_RESIDUAL num_classes = cfg.MODEL.HEAD.NUM_CLASSES return TorchvisionResNet( arch=arch, num_classes=num_classes, torchvision_pretrained=torchvision_pretrained, pretrained_num_classes=pretrained_num_classes, fix_bn=fix_bn, partial_bn=partial_bn, zero_init_residual=zero_init_residual )
[ "wy163zhuj@163.com" ]
wy163zhuj@163.com
11b7e3689c9e441e4675a957d33afa8bb29e075b
0c4fe6a4ada54cda0f5116e9fee31f133a2ca687
/common/logger.py
dadeaf23c5addb75052fd305ecaa9c17e8709ab5
[]
no_license
march-saber/python_aixunshouzhu_api
029f26470418abd5585a5593bf3addc01f83285e
49e57cbfe9ca7055fa05e9d98c439617ae486067
refs/heads/master
2020-05-28T09:46:53.418636
2019-06-03T12:12:47
2019-06-03T12:12:47
188,961,078
0
0
null
null
null
null
UTF-8
Python
false
false
1,257
py
import logging from common import contants from common.config import config def get_logger(name): logger = logging.getLogger(name) #建立一个日志收集器 logger.setLevel("DEBUG") #设定日志收集级别 fmt = "%(name)s - %(levelname)s - %(asctime)s - %(message)s - [%(filename)s:%(lineno)d]" formatter = logging.Formatter(fmt=fmt) #设定日志输出格式 console_handler = logging.StreamHandler() #指定输出到控制台 #吧日志级别放到配置文件里面去配置-- 优化 gather = config.get('log','gather_log') console_handler.setLevel(gather) #指定输出级别 console_handler.setFormatter(formatter) file_handler = logging.FileHandler(contants.log_dir + "/case.log",encoding='utf-8') # 吧日志级别放到配置文件里面去配置 output = config.get('log','output_log') file_handler.setLevel(output) file_handler.setFormatter(formatter) logger.addHandler(console_handler) logger.addHandler(file_handler) return logger if __name__ == '__main__': logger = get_logger('case') logger.debug("测试开始") logger.info("测试报错") logger.error("测试数据") logger.warning("测试结果") logger.critical("测试结束")
[ "1162869224@qq.com" ]
1162869224@qq.com
b9c25442a137b3ef27edbd26fb246ea1cad4a350
ac3b4affef9c9c03121ee30c0c0d589db54f292e
/docs/enterprise/hmac_.py
94fe8ba06404b24dde9c2121a3aad8713cdbefa6
[]
no_license
btourman/documentation
d5d822fa03c0c85d6304abaa0563e4e5ad0ca0b1
99929b11cac8814fe3661439bce607c59a5f2ebd
refs/heads/master
2020-09-07T03:07:59.850216
2019-11-08T14:07:17
2019-11-08T14:07:17
220,638,447
0
0
null
2019-11-09T12:23:26
2019-11-09T12:23:25
null
UTF-8
Python
false
false
1,141
py
# -*- coding: utf-8 -*- from urllib.parse import urlencode import hmac, hashlib, codecs def sign(query, secretKey): return codecs.getencoder('hex')(hmac.new(secretKey.encode('utf-8'), query.encode('utf-8'), hashlib.sha256).digest())[0].decode('utf-8') if __name__ == '__main__': # First setup our account ACCOUNT_ID = 'MY_ACCOUNT_ID' SECRET_KEY = 'MY_SECRET_KEY' # Then generate the watermark-free url # no need to encode the query string, Image-Charts will decode every parameters by itself to check the signature # learn why in our documentation https://documentation.image-charts.com/enterprise/ rawQuerystring = [ ('cht', 'bvs'), ('chd', 's:93zyvneTTO'), ('chtt', 'Hello world'), ('chs', '400x401'), ('icac', ACCOUNT_ID) # don't forget to add your account id before signing it ] queryString = "&".join( [ param +'='+ value for (param, value) in rawQuerystring ] ) signature = sign(queryString, SECRET_KEY) publicUrl = "https://image-charts.com/chart?" + queryString + "&ichm=" + signature # Finally send it to slack or via email, here we simply use print print(publicUrl)
[ "github@fgribreau.com" ]
github@fgribreau.com
8767b568baa843c144fcb45fc3930190b3aebe10
ef270274b87c2500485a992a798606e139c120ff
/blog/migrations/0001_initial.py
97261cf246c533caf63cae32a13a047a7610147d
[]
no_license
chvbrr/my-first-blog
e05e517a57c84ca99b176816486488462531d11e
ecca64473048220442f35a4d27c7d1ea6697c15b
refs/heads/master
2021-01-10T13:43:27.209037
2016-01-31T14:29:54
2016-01-31T14:29:54
50,764,703
0
0
null
null
null
null
UTF-8
Python
false
false
1,109
py
# -*- coding: utf-8 -*- # Generated by Django 1.9 on 2016-01-30 18:04 from __future__ import unicode_literals import datetime from django.conf import settings from django.db import migrations, models import django.db.models.deletion from django.utils.timezone import utc class Migration(migrations.Migration): initial = True dependencies = [ migrations.swappable_dependency(settings.AUTH_USER_MODEL), ] operations = [ migrations.CreateModel( name='post', fields=[ ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), ('title', models.CharField(max_length=200)), ('text', models.TextField()), ('created_date', models.DateTimeField(default=datetime.datetime(2016, 1, 30, 18, 4, 39, 220156, tzinfo=utc))), ('published_date', models.DateTimeField(blank=True, null=True)), ('author', models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to=settings.AUTH_USER_MODEL)), ], ), ]
[ "ch.v.b.ramaraju@gmail.com" ]
ch.v.b.ramaraju@gmail.com
03bc0d80849bc3264945b6fc903d9599b980d26a
a38725ed7fb93b503207502984ec197e921eb54b
/venv/lib/python3.6/site-packages/django_ajax/encoder.py
64ed9ca2af3a6a719fd651966cacb7ddaf862693
[]
no_license
tanveerahmad1517/myblogproject
d00d550230e2df0843e67f793504f9c19d0b755c
2eaa051caa5b68a8fba260c7cd431f1e1719a171
refs/heads/master
2020-03-16T21:38:32.738671
2018-08-23T11:55:02
2018-08-23T11:55:02
133,008,051
0
1
null
null
null
null
UTF-8
Python
false
false
1,859
py
""" Utils """ from __future__ import unicode_literals import json from datetime import date from django.http.response import HttpResponseRedirectBase, HttpResponse from django.template.response import TemplateResponse from django.utils.encoding import force_text from django.db.models.base import ModelBase from decimal import Decimal class LazyJSONEncoderMixin(object): """ A JSONEncoder subclass that handle querysets and models objects. Add how handle your type of object here to use when dump json """ def default(self, obj): # handles HttpResponse and exception content if issubclass(type(obj), HttpResponseRedirectBase): return obj['Location'] elif issubclass(type(obj), TemplateResponse): return obj.rendered_content elif issubclass(type(obj), HttpResponse): return obj.content elif issubclass(type(obj), Exception) or isinstance(obj, bytes): return force_text(obj) # this handles querysets and other iterable types try: iterable = iter(obj) except TypeError: pass else: return list(iterable) # this handlers Models if isinstance(obj.__class__, ModelBase): return force_text(obj) if isinstance(obj, Decimal): return float(obj) if isinstance(obj, date): return obj.isoformat() return super(LazyJSONEncoderMixin, self).default(obj) class LazyJSONEncoder(LazyJSONEncoderMixin, json.JSONEncoder): pass def serialize_to_json(data, *args, **kwargs): """ A wrapper for simplejson.dumps with defaults as: cls=LazyJSONEncoder All arguments can be added via kwargs """ kwargs['cls'] = kwargs.get('cls', LazyJSONEncoder) return json.dumps(data, *args, **kwargs)
[ "tanveerobjects@gmail.com" ]
tanveerobjects@gmail.com
c75ed46c05859a6b3d132ce4c72addc3a7b800b7
41edcef2f35d4eae56b57a6e6cb4a9ecad42b812
/common/mobile_device.py
c1635180d58675bdf3a3256270b07b018079d283
[]
no_license
wanhui1994/xpower
4d7e4dc9102a7eff085fd4f2443ee2b7d4f695ba
9c1b468f6f215d87ec34ebbc5f2cdf43246af6a5
refs/heads/master
2020-07-29T10:22:46.823317
2019-09-21T07:34:48
2019-09-21T07:34:48
209,761,335
0
0
null
null
null
null
UTF-8
Python
false
false
1,454
py
#coding=utf-8 import os,re class Device(): def vesion(self): #获取连接电脑的设备名称信息 self.readDeviceId = list(os.popen('adb devices').readlines()) deviceId = re.findall(r'^\w*\b', self.readDeviceId[1])[0] return deviceId def devicevsion(self): #获取连接电脑设备的版本 deviceAndroidVersion = list(os.popen('adb shell getprop ro.build.version.release').readlines()) deviceVersion = "".join(deviceAndroidVersion).strip() return deviceVersion def Package(self): #获取执行的apk的包名 pass def apk(self): pwd = os.getcwd() father_path=os.path.abspath(os.path.dirname(pwd)+os.path.sep+".") path=father_path+"\\apk\\app-release.apk" return path def desired(self,package,activity): #移动设备信息 if len(list(os.popen('adb devices').readlines())[1].rstrip())>0: desired_caps = { 'platformName':'Android', 'deviceName': self.vesion(), 'platformVersion': self.devicevsion(), 'appPackage' : package, #输入apk的包名 'appActivity': activity, #输入apk的activity 'sessionOverride':'true', #每次启动时覆盖session 'app':self.apk(), 'noReset':'True', } return desired_caps else: print("测试手机未连接")
[ "2353231116@qq.com" ]
2353231116@qq.com
aa0821eb5dfdd23d7f0d1145aa8a2eb118518433
04ce2e384bf4c005264c86144a4bc9482fab74d2
/venv/lib/python3.5/bisect.py
e9627395ec6a842938ca6fa9bb305208875ba9ae
[]
no_license
ranijaiswal/mysite
e9ad8fdf38ac7dc3af42f2b168e3d226e7fa69df
21b57513857218bdcbd66925c19ec9572c1f239d
refs/heads/master
2021-01-23T03:53:52.374712
2017-03-26T17:46:26
2017-03-26T17:46:26
86,132,814
0
0
null
null
null
null
UTF-8
Python
false
false
51
py
/Users/ranijaiswal/anaconda/lib/python3.5/bisect.py
[ "ruj96@live.com" ]
ruj96@live.com
ddb356445bf02c7df3723d467bc763ca2c73ba9e
41f81d8496262182c73e855e9d3d4fcee8dc659d
/emailSpammer/spam.py
c607dec4db5b686b1519479c207a8685ce7e851c
[]
no_license
ishank-dev/MyPythonScripts
5e372bf829941e3db5746cc32cdaf2fddb33f0a8
a50e17f0a5bd45b8086e429abff3d9a1d42286d5
refs/heads/master
2021-07-11T09:41:15.994267
2019-10-14T16:38:25
2019-10-14T16:38:25
209,132,738
1
4
null
2020-10-14T14:21:45
2019-09-17T18:50:19
Python
UTF-8
Python
false
false
395
py
import smtplib for i in range(0,5) conn = smtplib.SMTP('smtp.gmail.com',587)# connect to gmail conn.ehlo() conn.starttls() conn.login('your_email','your_password') # write your email and password for the gmail account conn.sendmail('your_email','recipient_address','Subject: Write email subject here\n\n Write the message here ') # write your email and recipient address here conn.quit()
[ "noreply@github.com" ]
noreply@github.com
1e07bbeff0fb13fa145c80101d396935d33a0423
6b14d9a64a578239e5612e6098320b61b45c08d9
/AUG14/04.py
27bec86f2774038bbdffc335d52b45f500521bfc
[ "MIT" ]
permissive
Razdeep/PythonSnippets
498c403140fec33ee2f0dd84801738f1256ee9dd
76f9313894f511c487a99bc38bdf0fe5e594caf5
refs/heads/master
2020-03-26T08:56:23.067022
2018-11-26T05:36:36
2018-11-26T05:36:36
144,726,845
0
0
null
null
null
null
UTF-8
Python
false
false
67
py
# String slicing test='Hello world' print(test[1:5]) print(test[6])
[ "rrajdeeproychowdhury@gmail.com" ]
rrajdeeproychowdhury@gmail.com
6c224211afdf733a84ccbb82e6c9968e0193891f
c3a1ce1a918a9f15355b17cece583bd27da1bd53
/Taking input for Competitive Programming.py
b6204ac15b3909ac2a2fe4a251c2f5eac33381c1
[]
no_license
MrJay10/Graph-Algorithms
3cea89b8951f8b656505deef1aa52adf22549f0c
382d0f8d41313bcf37b266695a82a8ebe6182b40
refs/heads/master
2021-01-17T19:21:25.246914
2016-06-25T08:33:36
2016-06-25T08:33:36
61,347,944
0
0
null
null
null
null
UTF-8
Python
false
false
431
py
from Graph import Graph graph = dict() # Remove the message in input(); change str -> int for integer input vertices = list(map(str, input("Enter vertices :: ").split())) for vertex in vertices: # Remove the message in input(); change str -> int for integer input graph[vertex] = list(map(str, input("Enter neighbors of "+vertex+" -> ").split())) g = Graph(graph) print("Your Graph is ::\n\n"+str(g)+"\n")
[ "noreply@github.com" ]
noreply@github.com
8d07ce4171b2bea29faab046161815234799f885
a0cefb1cd11b85b34c5ed58e44d087981541111b
/run_tests.py
57157de5d2632d35069de435665c2e90b130f08e
[]
no_license
diogo-aos/masters_final
732e436e74bbc7d24756fb10f96d0d39656212e7
93ae6b71d7d7d9dade0059facfe2bd5162c673da
refs/heads/master
2021-01-18T15:36:58.263599
2017-03-30T05:11:18
2017-03-30T05:11:18
86,661,794
2
0
null
null
null
null
UTF-8
Python
false
false
329
py
import unittest import tests.test_scan as tscan import tests.test_boruvka as tboruvka scan_suite = unittest.TestLoader().loadTestsFromModule(tscan) boruvka_suite = unittest.TestLoader().loadTestsFromModule(tboruvka) # unittest.TextTestRunner(verbosity=2).run(scan_suite) unittest.TextTestRunner(verbosity=2).run(boruvka_suite)
[ "dasilva@academiafa.edu.pt" ]
dasilva@academiafa.edu.pt