blob_id stringlengths 40 40 | directory_id stringlengths 40 40 | path stringlengths 3 281 | content_id stringlengths 40 40 | detected_licenses listlengths 0 57 | license_type stringclasses 2 values | repo_name stringlengths 6 116 | snapshot_id stringlengths 40 40 | revision_id stringlengths 40 40 | branch_name stringclasses 313 values | visit_date timestamp[us] | revision_date timestamp[us] | committer_date timestamp[us] | github_id int64 18.2k 668M ⌀ | star_events_count int64 0 102k | fork_events_count int64 0 38.2k | gha_license_id stringclasses 17 values | gha_event_created_at timestamp[us] | gha_created_at timestamp[us] | gha_language stringclasses 107 values | src_encoding stringclasses 20 values | language stringclasses 1 value | is_vendor bool 2 classes | is_generated bool 2 classes | length_bytes int64 4 6.02M | extension stringclasses 78 values | content stringlengths 2 6.02M | authors listlengths 1 1 | author stringlengths 0 175 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
d0fe0d32f8117f0320d48ac65fba33e6aa33014e | adc5060ccc1f9e1243f0d9d5eb95e0ca87034806 | /cdkworkshop/cdkworkshop_stack.py | 2b18b7ea586a1953a25a1b199afbf137a865fcaa | [] | no_license | fanaticjo/cdkworkshop | 37a1762fa9137ee405a098afaf2f45c3b57b135d | aa5a888d42aaae98eb18fe629b82fb1416b4bd7e | refs/heads/master | 2023-08-16T22:41:11.431991 | 2021-09-14T16:30:35 | 2021-09-14T16:30:35 | 406,444,342 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 583 | py | from aws_cdk import (
aws_lambda as _lambda,
aws_apigateway as api,
core
)
class CdkworkshopStack(core.Stack):
def __init__(self, scope: core.Construct, construct_id: str, **kwargs) -> None:
super().__init__(scope, construct_id, **kwargs)
self.lambdaStack=_lambda.Function(
self,'Handler',runtime=_lambda.Runtime.PYTHON_3_6,
code=_lambda.Code.asset('lambda'),
handler='lambda_function.lambda_handler'
)
self.api_g=api.LambdaRestApi(
self,'Endpoint',handler=self.lambdaStack
)
| [
"biswajit196@live.com"
] | biswajit196@live.com |
ec5a3ad6b211c933d6f0f6c9428eef2cd6609e11 | ce2ed48a3ea2b067ff45e2901fd1aa08b1b9fd3b | /Data_prep.py | 78e05e3e3b624e9b3f50fb508fd7ace9a273d008 | [] | no_license | mhmdsab/ASHRAE---Great-Energy-Predictor-III- | 9d049a8f7eb9887c89f10b8073e6ad7587d06582 | 946c37af579299d8eb0d338fd4367f0ffc5dd59a | refs/heads/master | 2020-12-01T06:07:04.791763 | 2019-12-28T07:27:32 | 2019-12-28T07:27:32 | 230,572,508 | 1 | 1 | null | null | null | null | UTF-8 | Python | false | false | 13,487 | py | import numpy as np
import pandas as pd
from tqdm import tqdm
import tensorflow as tf
import os
from abc import ABCMeta, abstractmethod
from sklearn.preprocessing import OneHotEncoder as OHE
class DataGenerator(metaclass=ABCMeta):
def __init__(self, para):
self.iterator = None
self.para = para
def inputs(self, mode, batch_size, num_epochs=None):
"""Reads input data num_epochs times.
Args:
mode: String for the corresponding tfrecords ('train', 'validation')
batch_size: Number of examples per returned batch.
"""
if mode != "train" and mode != "valid":
raise ValueError("mode: {} while mode should be "
"'train', 'validation'".format(mode))
filename = self.para.tf_records_url + '/' +mode + "_.tfrecord"
with tf.name_scope("input"):
# TFRecordDataset opens a binary file and
# reads one record at a time.
# `filename` could also be a list of filenames,
# which will be read in order.
dataset = tf.data.TFRecordDataset(filename)
# The map transformation takes a function and
# applies it to every element
# of the dataset.
dataset = dataset.map(self._decode)
# The shuffle transformation uses a finite-sized buffer to shuffle
# elements in memory. The parameter is the number of elements in the
# buffer. For completely uniform shuffling, set the parameter to be
# the same as the number of elements in the dataset.
if mode == "train":
dataset = dataset.shuffle(2380*16)
dataset = dataset.batch(batch_size)
dataset = dataset.prefetch(batch_size)
self.iterator = dataset.make_initializable_iterator()
return self.iterator.get_next()
@abstractmethod
def _decode(self, serialized_example):
pass
class generate_dataset(DataGenerator):
def __init__(self, para):
self.para = para
self.train_data_range = pd.date_range(para.TS_start_date,para.TS_end_date , freq=para.sampling_frequency)
self.train_written_examples = 0
self.valid_written_examples = 0
self.train_excluded_examples = 0
self.valid_excluded_examples = 0
self.mean_values = {}
self.std_values = {}
self.three_months_corr = {}
self.six_months_corr = {}
self.features = {'x_encoder': tf.FixedLenFeature([self.para.in_sequence_window, 25], dtype=tf.float32),
'x_decoder': tf.FixedLenFeature([self.para.out_sequence_window, 24], dtype=tf.float32),
'encoder_meter_reading': tf.FixedLenFeature([self.para.in_sequence_window], dtype=tf.float32),
'decoder_meter_reading': tf.FixedLenFeature([self.para.out_sequence_window], dtype=tf.float32),
'decoder_meter_reading_denorm': tf.FixedLenFeature([self.para.out_sequence_window], dtype=tf.float32),
'mean': tf.FixedLenFeature([1], dtype=tf.float32),
'std':tf.FixedLenFeature([1], dtype=tf.float32)}
self.generate_encoders()
self.generate_dataset()
self.organize('train', 'valid')
super().__init__(para)
def generate_encoders(self):
self.encoding_dict = dict([('month_number' , OHE().fit(np.arange(1, 13).reshape(-1,1))),
('weekday_number' , OHE().fit(np.arange(7).reshape(-1,1))),
('meter' , OHE().fit(np.arange(4).reshape(-1,1)))])
def generate_dataset(self):
print('generating raw dataset')
train = pd.read_csv(self.para.train_url)
meta = pd.read_csv(self.para.meta_url).loc[:,['building_id', 'primary_use']]
dataset = pd.merge(train, meta, left_on = 'building_id', right_on = 'building_id')
dataset['timestamp']= pd.to_datetime(dataset['timestamp'])
self.dataset = dataset.groupby(['building_id', 'meter'])
self.groups = list(self.dataset.groups.keys())
def autocorrelate(self, df, group):
three_months_corr = df['meter_reading'].autocorr(lag = int(3*30*(24/int(self.para.sampling_frequency[0]))))
self.three_months_corr[group] = three_months_corr
six_months_corr = df['meter_reading'].autocorr(lag = int(6*30*(24/int(self.para.sampling_frequency[0]))))
self.six_months_corr[group] = six_months_corr
df['three_months_lag_autocorr'] = three_months_corr * np.ones(shape = len(df))#shape = (8784)
df['six_months_lag_autocorr'] = six_months_corr *np.ones(shape = len(df))#shape = (8784)
return df.astype('float32')
def One_Hot_Enode(self, df):
for feature in self.encoding_dict.keys():
raw_series = df[feature].values.reshape(-1, 1)
One_Hot_Encoded_array = self.encoding_dict[feature].transform(raw_series).todense()[:,:-1]
for i in range(One_Hot_Encoded_array.shape[1]):
OHE_feature_name = feature+'_'+str(i)
df[OHE_feature_name] = One_Hot_Encoded_array[:,i]
return df.astype('float32')
def pad_time_series(self, df):
padding_df = pd.DataFrame()
padding_df['any'] = np.ones(shape = (self.para.building_total_len//int(self.para.sampling_frequency[0])))
padding_df = padding_df.set_index(self.train_data_range)
merged = pd.merge(padding_df, df, how = 'left', left_index=True, right_index=True).drop('any',1).fillna(0)
return merged.astype('float32')
def Normalize_Pad_Split(self, df, group):
mean = np.mean(df['meter_reading'][df['month_number'] <= self.para.features_extractor_len])
std = np.std(df['meter_reading'][df['month_number'] <= self.para.features_extractor_len])
if (mean!= mean) or (std!= std):
mean = np.mean(list(self.mean_values.values()))
std = np.mean(list(self.std_values.values()))
self.mean_values[group] = mean
self.std_values[group] = std
df = self.pad_time_series(df)
df['meter_reading_normalized'] = df['meter_reading'].map(lambda x:self.Normalize(x, mean, std))
df['three_months_lag'] = df['three_months_lag'].map(lambda x:self.Normalize(x, mean, std))
df['six_months_lag'] = df['six_months_lag'].map(lambda x:self.Normalize(x, mean, std))
train_df = df.loc[:self.para.train_end_date,:].drop(['month_number', 'weekday_number'], 1)
valid_df = df.loc[self.para.valid_start_date:,:].drop(['month_number', 'weekday_number'], 1)
return train_df, valid_df
def organize(self, train_name, valid_name):
print('organizing raw dataset')
if (os.path.exists(self.para.tf_records_url+'/'+'{}_.tfrecord'.format(train_name))) & \
(os.path.exists(self.para.tf_records_url+'/'+'{}_.tfrecord'.format(valid_name))):
self.para.train_kickoff = 'not_first_time'
train_name, valid_name = '_','_'
with tf.python_io.TFRecordWriter(self.para.tf_records_url+'/'+'{}_.tfrecord'.format(train_name)) as train_writer:
with tf.python_io.TFRecordWriter(self.para.tf_records_url+'/'+'{}_.tfrecord'.format(valid_name)) as valid_writer:
for group in tqdm(self.groups):
building_df = self.dataset.get_group(group).set_index('timestamp')
building_df = building_df.resample(self.para.sampling_frequency).mean().fillna(method = 'ffill')
building_df = self.add_time_features(building_df)
building_df = self.One_Hot_Enode(building_df)
building_df = self.add_timelags(building_df)
building_df = self.autocorrelate(building_df, group)
train_building_df, valid_building_df = self.Normalize_Pad_Split(building_df, group)
if self.para.train_kickoff == 'first_time':
self._convert_to_tfrecord(train_building_df, train_writer, train_name, group)
self._convert_to_tfrecord(valid_building_df, valid_writer, valid_name, group)
def _convert_to_tfrecord(self, df, writer, mode, group):
df_meter_reading_normalized = df['meter_reading_normalized']
df_meter_reading_unnormalized = df['meter_reading']
df_x = df.drop(['building_id', 'meter_reading', 'meter'],1)
mean = self.mean_values[group]
std = self.std_values[group]
for i in range((len(df) - self.para.in_sequence_window)//self.para.out_sequence_window):
start = i * self.para.out_sequence_window
end = start + self.para.in_sequence_window
y_end = end + self.para.out_sequence_window
if ((df['meter_reading'].iloc[start:end] == 0).sum() < \
int(self.para.max_zeros_in_example * self.para.in_sequence_window)) \
and (std>self.para.min_std):
example = tf.train.Example(features=tf.train.Features(feature={
'x_encoder': tf.train.Feature(
float_list=tf.train.FloatList(value = df_x.iloc[start:end,:].values.flatten())),
'x_decoder': tf.train.Feature(
float_list=tf.train.FloatList(value = df_x.drop(['meter_reading_normalized'],1) \
.iloc[end:y_end,:].values.flatten())),
'encoder_meter_reading': tf.train.Feature(
float_list=tf.train.FloatList(value = df_meter_reading_normalized[start:end]\
.values.flatten())),
'decoder_meter_reading': tf.train.Feature(
float_list=tf.train.FloatList(value = df_meter_reading_normalized[end:y_end]\
.values.flatten())),
'decoder_meter_reading_denorm': tf.train.Feature(
float_list=tf.train.FloatList(value = df_meter_reading_unnormalized[end:y_end]\
.values.flatten())),
'mean': tf.train.Feature(
float_list=tf.train.FloatList(value = [mean])),
'std': tf.train.Feature(
float_list=tf.train.FloatList(value = [std]))}))
writer.write(example.SerializeToString())
if mode == 'train':
self.train_written_examples += 1
elif mode == 'valid':
self.valid_written_examples += 1
else:
if mode == 'train':
self.train_excluded_examples+=1
elif mode == 'valid':
self.valid_excluded_examples+=1
def _decode(self, serialized_example):
example = tf.parse_single_example(
serialized_example,
features=self.features)
x_encoder = example['x_encoder']
x_decoder = example['x_decoder']
encoder_meter_reading = example['encoder_meter_reading']
decoder_meter_reading = example['decoder_meter_reading']
decoder_meter_reading_denorm = example['decoder_meter_reading_denorm']
mean = example['mean']
std = example['std']
return x_encoder, x_decoder, encoder_meter_reading, decoder_meter_reading, decoder_meter_reading_denorm, mean, std
@staticmethod
def Normalize(value, mean, std):
return (value-mean)/(std + 1e-100)
@staticmethod
def add_timelags(df):
df['three_months_lag'] = df['meter_reading'].shift(90*6).fillna(0)
df['six_months_lag'] = df['meter_reading'].shift(90*6*2).fillna(0)
return df.astype('float32')
@staticmethod
def add_time_features(df):
df['weekday_number'] = df.index.weekday
df['month_number'] = df.index.month
return df
| [
"noreply@github.com"
] | noreply@github.com |
e1133f2d9491922b496b13c9b71511b119616887 | 63b55540d45c6445885ebcac892aba40454441c9 | /HelperFunctions.py | da1e057933504c0a96319a48b6402b48146da2e9 | [] | no_license | zapatos24/The_Minority_Math_Problem | 1702be39ed88a4169c1c9724eb38172fec111672 | fc19758c248aa22b72929dfd733000c94785647f | refs/heads/master | 2020-06-29T14:24:48.949752 | 2019-08-13T03:22:20 | 2019-08-13T03:22:20 | 200,560,314 | 2 | 0 | null | null | null | null | UTF-8 | Python | false | false | 12,892 | py | import pandas as pd
import numpy as np
import matplotlib.pyplot as plt
import seaborn as sns
from sklearn.preprocessing import StandardScaler
from sklearn.model_selection import train_test_split, cross_val_score, GridSearchCV
from sklearn.decomposition import PCA
from sklearn.linear_model import LogisticRegression
from sklearn.neighbors import KNeighborsClassifier
from sklearn.tree import DecisionTreeClassifier
from sklearn.ensemble import RandomForestClassifier
from sklearn.svm import SVC
from xgboost import XGBClassifier
from sklearn.feature_selection import RFECV
from sklearn.metrics import classification_report, confusion_matrix, accuracy_score, f1_score, roc_curve, auc
class HelperFunctions():
def acheivement_score(rating):
'''
Takes the rating passed to it and returns an integer
value representing how school meets target goals
'''
if rating == 'Exceeding Target':
return 4
if rating == 'Meeting Target':
return 3
if rating == 'Approaching Target':
return 2
if rating == 'Not Meeting Target':
return 1
else:
return None
def percent_cols_to_float(df):
'''
For any dataframe passed in, returns a new dataframe where
values are floats between 0 and 1 representing the respective
rate or percent in that column
'''
for col in df.columns:
if 'Rate' in col or 'Percent' in col or '%' in col:
df[col] = df[col].apply(
lambda x: float(x.replace('%', ''))*.01)
return df
def make_grades_int(grade):
'''
Takes a grade and returns an integer representative of that
grade in the school system
'''
if grade == 'PK':
return -1
elif grade == '0K':
return 0
else:
return int(grade)
def grid_search_classifier(clf, param_grid, X_train, X_test, y_train, y_test, scoring='f1_weighted'):
grid_clf = GridSearchCV(clf, param_grid, scoring=scoring)
grid_clf.fit(X_train, y_train)
best_parameters = grid_clf.best_params_
print("Grid Search found the following optimal parameters: ")
for param_name in sorted(best_parameters.keys()):
print("%s: %r" % (param_name, best_parameters[param_name]))
y_pred = grid_clf.predict(X_test)
print()
print(confusion_matrix(y_test, y_pred))
print(classification_report(y_test, y_pred))
print('Accuracy score:', round(accuracy_score(y_test, y_pred), 2))
cm = confusion_matrix(y_test, y_pred)
df_cm = pd.DataFrame(cm.astype('float') / cm.sum(axis=1)[:, np.newaxis],
index=['F', 'T'],
columns=['F', 'T'])
plt.figure(figsize=(7, 5))
sns.heatmap(df_cm, annot=True, cmap='Greens')
plt.xlabel('Pred Val')
plt.ylabel('True Val')
plt.show()
return grid_clf
def plot_ROC(y_test, X_test, grid_clf):
fpr, tpr, thresholds = roc_curve(
y_test, grid_clf.predict_proba(X_test)[:, 1])
print('AUC: {}'.format(auc(fpr, tpr)))
plt.figure(figsize=(10, 8))
plt.plot(fpr, tpr, color='darkorange', label='ROC curve')
plt.plot([0, 1], [0, 1], color='navy')
plt.xlim([0.0, 1.0])
plt.ylim([0.0, 1.05])
plt.yticks([i/10.0 for i in range(11)])
plt.xticks([i/10.0 for i in range(11)])
plt.xlabel('False Positive Rate')
plt.ylabel('True Positive Rate')
plt.title('ROC Curve')
plt.legend(loc="lower right")
plt.show()
def drop_impractical_columns(df):
cols_to_drop = ['Adjusted Grade',
'New?',
'Other Location Code in LCGMS',
'School Name',
'District',
'SED Code',
'Latitude',
'Longitude',
'Address (Full)',
'City',
'Zip',
'Grades',
'Rigorous Instruction Rating',
'Collaborative Teachers Rating',
'Supportive Environment Rating',
'Effective School Leadership Rating',
'Strong Family-Community Ties Rating',
'Trust Rating',
'School Income Estimate',
'Average ELA Proficiency',
'Community School?',
'Grade 3 ELA - All Students Tested',
'Grade 3 ELA 4s - All Students',
'Grade 3 ELA 4s - American Indian or Alaska Native',
'Grade 3 ELA 4s - Black or African American',
'Grade 3 ELA 4s - Hispanic or Latino',
'Grade 3 ELA 4s - Asian or Pacific Islander',
'Grade 3 ELA 4s - White',
'Grade 3 ELA 4s - Multiracial',
'Grade 3 ELA 4s - Limited English Proficient',
'Grade 3 ELA 4s - Economically Disadvantaged',
'Grade 3 Math - All Students tested',
'Grade 3 Math 4s - All Students',
'Grade 3 Math 4s - American Indian or Alaska Native',
'Grade 3 Math 4s - Black or African American',
'Grade 3 Math 4s - Hispanic or Latino',
'Grade 3 Math 4s - Asian or Pacific Islander',
'Grade 3 Math 4s - White',
'Grade 3 Math 4s - Multiracial',
'Grade 3 Math 4s - Limited English Proficient',
'Grade 3 Math 4s - Economically Disadvantaged',
'Grade 4 ELA - All Students Tested',
'Grade 4 ELA 4s - All Students',
'Grade 4 ELA 4s - American Indian or Alaska Native',
'Grade 4 ELA 4s - Black or African American',
'Grade 4 ELA 4s - Hispanic or Latino',
'Grade 4 ELA 4s - Asian or Pacific Islander',
'Grade 4 ELA 4s - White',
'Grade 4 ELA 4s - Multiracial',
'Grade 4 ELA 4s - Limited English Proficient',
'Grade 4 ELA 4s - Economically Disadvantaged',
'Grade 4 Math - All Students Tested',
'Grade 4 Math 4s - All Students',
'Grade 4 Math 4s - American Indian or Alaska Native',
'Grade 4 Math 4s - Black or African American',
'Grade 4 Math 4s - Hispanic or Latino',
'Grade 4 Math 4s - Asian or Pacific Islander',
'Grade 4 Math 4s - White',
'Grade 4 Math 4s - Multiracial',
'Grade 4 Math 4s - Limited English Proficient',
'Grade 4 Math 4s - Economically Disadvantaged',
'Grade 5 ELA - All Students Tested',
'Grade 5 ELA 4s - All Students',
'Grade 5 ELA 4s - American Indian or Alaska Native',
'Grade 5 ELA 4s - Black or African American',
'Grade 5 ELA 4s - Hispanic or Latino',
'Grade 5 ELA 4s - Asian or Pacific Islander',
'Grade 5 ELA 4s - White',
'Grade 5 ELA 4s - Multiracial',
'Grade 5 ELA 4s - Limited English Proficient',
'Grade 5 ELA 4s - Economically Disadvantaged',
'Grade 5 Math - All Students Tested',
'Grade 5 Math 4s - All Students',
'Grade 5 Math 4s - American Indian or Alaska Native',
'Grade 5 Math 4s - Black or African American',
'Grade 5 Math 4s - Hispanic or Latino',
'Grade 5 Math 4s - Asian or Pacific Islander',
'Grade 5 Math 4s - White',
'Grade 5 Math 4s - Multiracial',
'Grade 5 Math 4s - Limited English Proficient',
'Grade 5 Math 4s - Economically Disadvantaged',
'Grade 6 ELA - All Students Tested',
'Grade 6 ELA 4s - All Students',
'Grade 6 ELA 4s - American Indian or Alaska Native',
'Grade 6 ELA 4s - Black or African American',
'Grade 6 ELA 4s - Hispanic or Latino',
'Grade 6 ELA 4s - Asian or Pacific Islander',
'Grade 6 ELA 4s - White',
'Grade 6 ELA 4s - Multiracial',
'Grade 6 ELA 4s - Limited English Proficient',
'Grade 6 ELA 4s - Economically Disadvantaged',
'Grade 6 Math - All Students Tested',
'Grade 6 Math 4s - All Students',
'Grade 6 Math 4s - American Indian or Alaska Native',
'Grade 6 Math 4s - Black or African American',
'Grade 6 Math 4s - Hispanic or Latino',
'Grade 6 Math 4s - Asian or Pacific Islander',
'Grade 6 Math 4s - White',
'Grade 6 Math 4s - Multiracial',
'Grade 6 Math 4s - Limited English Proficient',
'Grade 6 Math 4s - Economically Disadvantaged',
'Grade 7 ELA - All Students Tested',
'Grade 7 ELA 4s - All Students',
'Grade 7 ELA 4s - American Indian or Alaska Native',
'Grade 7 ELA 4s - Black or African American',
'Grade 7 ELA 4s - Hispanic or Latino',
'Grade 7 ELA 4s - Asian or Pacific Islander',
'Grade 7 ELA 4s - White',
'Grade 7 ELA 4s - Multiracial',
'Grade 7 ELA 4s - Limited English Proficient',
'Grade 7 ELA 4s - Economically Disadvantaged',
'Grade 7 Math - All Students Tested',
'Grade 7 Math 4s - All Students',
'Grade 7 Math 4s - American Indian or Alaska Native',
'Grade 7 Math 4s - Black or African American',
'Grade 7 Math 4s - Hispanic or Latino',
'Grade 7 Math 4s - Asian or Pacific Islander',
'Grade 7 Math 4s - White',
'Grade 7 Math 4s - Multiracial',
'Grade 7 Math 4s - Limited English Proficient',
'Grade 7 Math 4s - Economically Disadvantaged',
'Grade 8 ELA - All Students Tested',
'Grade 8 ELA 4s - All Students',
'Grade 8 ELA 4s - American Indian or Alaska Native',
'Grade 8 ELA 4s - Black or African American',
'Grade 8 ELA 4s - Hispanic or Latino',
'Grade 8 ELA 4s - Asian or Pacific Islander',
'Grade 8 ELA 4s - White',
'Grade 8 ELA 4s - Multiracial',
'Grade 8 ELA 4s - Limited English Proficient',
'Grade 8 ELA 4s - Economically Disadvantaged',
'Grade 8 Math - All Students Tested',
'Grade 8 Math 4s - All Students',
'Grade 8 Math 4s - American Indian or Alaska Native',
'Grade 8 Math 4s - Black or African American',
'Grade 8 Math 4s - Hispanic or Latino',
'Grade 8 Math 4s - Asian or Pacific Islander',
'Grade 8 Math 4s - White',
'Grade 8 Math 4s - Multiracial',
'Grade 8 Math 4s - Limited English Proficient',
'Grade 8 Math 4s - Economically Disadvantaged'
]
return df.drop(cols_to_drop, axis=1)
def rfe_test(classifier, features, X_train, y_train):
ranking_list = []
for i in range(50):
clf = classifier
rfecv = RFECV(clf).fit(X_train, y_train)
ranking_list.append(rfecv.ranking_)
return pd.DataFrame(zip(features, sum(ranking_list)/50)).sort_values(by=1)
| [
"jeremy.traber.owens@gmail.com"
] | jeremy.traber.owens@gmail.com |
93608c46dbf817b09cd537b4fe647e9a03ac63ca | 06cd48c385acf1b79e9cf235730ab5a2f61c016e | /sql_queries.py | 12c9d5d060a7b204c6da2552d0bf4328db5250cf | [] | no_license | haymar017/-UDACITY-Data-Modeling-with-Postgres | 16f66b6abd267577f5da4211b6e0ce753761a257 | 89b494f2de889f388e6c4c3acaa1c10bbb57acfe | refs/heads/master | 2022-11-14T00:57:54.859730 | 2020-07-09T04:44:15 | 2020-07-09T04:44:15 | 276,080,584 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 3,432 | py | import json
import pandas
# DROP TABLES
songplay_table_drop = "drop table if exists songplays"
user_table_drop = "drop table if exists users"
song_table_drop = "drop table if exists songs"
artist_table_drop = "drop table if exists artists"
time_table_drop = "drop table if exists time"
# CREATE TABLES
user_table_create = ("""
create table if not exists users (
user_id int Primary Key,
first_name varchar,
last_name varchar,
gender varchar,
level varchar
);
""")
artist_table_create = ("""
create table if not exists artists (
artist_id varchar Primary Key,
name varchar,
location varchar,
latitude float,
longitude float
);
""")
song_table_create = ("""
create table if not exists songs(
song_id varchar Primary Key ,
title varchar,
artist_id varchar not null,
year int,
duration float
);
""")
time_table_create = ("""
create table if not exists time (
start_time timestamp Primary Key,
hour int,
day int,
week int,
month int,
year int,
weekday int
);
""")
songplay_table_create = ("""
create table if not exists songplays (
songplay_id serial Primary Key,
start_time timestamp not null,
user_id int not null,
level varchar,
song_id varchar,
artist_id varchar,
session_id int,
location varchar,
user_agent varchar,
foreign key (user_id) references users (user_id),
foreign key (song_id) references songs (song_id),
foreign key (artist_id) references artists (artist_id),
foreign key (start_time) references time (start_time)
);
""")
# INSERT RECORDS
songplay_table_insert = ("""
insert into songplays(
start_time,
user_id,
level,
song_id,
artist_id,
session_id,
location,
user_agent)
values (%s,%s,%s,%s,%s,%s,%s,%s)
""")
user_table_insert = ("""
insert into users(
user_id,
first_name,
last_name,
gender,
level)
values (%s,%s,%s,%s,%s)
on conflict (user_id) do update set level = excluded.level;
""")
song_table_insert = ("""
insert into songs(
song_id,
title,
artist_id,
year,
duration)
values (%s,%s,%s,%s,%s)
on conflict(song_id) do nothing
""")
artist_table_insert = ("""
insert into artists(
artist_id,
name,
location,
latitude,
longitude)
values (%s,%s,%s,%s,%s)
on conflict(artist_id) do nothing
""")
time_table_insert = ("""
insert into time(
start_time,
hour,
day,
week,
month,
year,
weekday)
values(%s,%s,%s,%s,%s,%s,%s)
on conflict(start_time) do nothing
""")
# FIND SONGS
song_select = ("""
select
songs.song_id , artists.artist_id
from
songs join artists
on
songs.artist_id = artists.artist_id
where
songs.title=(%s) and artists.name=(%s) and songs.duration=(%s)
""")
# QUERY LISTS
create_table_queries = [user_table_create, artist_table_create, song_table_create, time_table_create, songplay_table_create]
drop_table_queries = [songplay_table_drop, user_table_drop, song_table_drop, artist_table_drop, time_table_drop]
| [
"noreply@github.com"
] | noreply@github.com |
f5d1ed04d9af16888d78a13a4548ec0454cbba05 | e5ba5c10b94e8971a3124ddb6b36048d76d7853e | /bot/main.py | 5d969adf3a42dc470c2f123fd81c8dd84fa54bd7 | [
"MIT"
] | permissive | sonhal/telegram-reddit-bot | 1933c1d0769bc167a951c2ba824a6827ae19f99a | 0d23f2bd1f8761d2133c6bcf2643502b505cb3be | refs/heads/master | 2020-04-15T06:50:49.437490 | 2019-01-09T10:53:09 | 2019-01-09T10:53:09 | 164,475,656 | 0 | 0 | MIT | 2019-01-09T11:08:54 | 2019-01-07T18:44:49 | Python | UTF-8 | Python | false | false | 157 | py | from connectors import telegram_connector
COMMANDS = ("top10", "top")
if __name__ == '__main__':
tb = telegram_connector.TelegramBot()
tb.start()
| [
"sondre.hal@gmail.com"
] | sondre.hal@gmail.com |
a96427e29b2d6a2689c2072888af98f6bd99e8a0 | 7f0eaa5aa008e7116645fa214e772ef4c2ee5406 | /api/util/__tests__/test_api.py | 533128dc7e06c5d7a5bb870807019b6ee2ad1d4a | [] | no_license | enixdark/raven | ef7273ff26e96b45486f67cb52d22c252bd336e2 | c60ec2aa4ffec3c89afac305e63512732a148f4e | refs/heads/master | 2023-04-10T21:05:51.703163 | 2020-01-17T20:09:25 | 2020-01-17T20:09:25 | null | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 3,950 | py | import asyncio
import asynctest
import mock
import pytest
from mock import patch, MagicMock
from asynctest import CoroutineMock
from expects import expect, equal
from api.util.api import Api, Async
class TestApi:
@pytest.mark.asyncio
async def test_call(self, *args):
with asynctest.mock.patch('aiohttp.request', create=True) as aiohttp_req_mock:
mock_req = CoroutineMock()
mock_req.__aexit__ = CoroutineMock()
mock_req.__aenter__ = CoroutineMock()
aiohttp_req_mock.return_value = mock_req
expected_method = 'some-value'
expected_url = 'some-value'
expected_params = {}
expected_data = {}
expected_json = {}
expected_cookies = {}
expected_headers = {}
expected_auth = {}
await Api.call(method=expected_method, url=expected_url, params=expected_params, data=expected_data, json=expected_json, cookies=expected_cookies, headers=expected_headers, auth=expected_auth)
aiohttp_req_mock.assert_called_with(
method=expected_method,
url=expected_url,
params=expected_params,
data=expected_data,
json=expected_json,
cookies=expected_cookies,
headers=expected_headers,
auth=expected_auth)
@pytest.mark.asyncio
async def test_batch(self, *args):
with asynctest.mock.patch.object(Api, 'call') as call_mock:
expected_requests = [
{
'method': 'some-value',
'url': 'some-value',
'params': {},
'data': {},
'json': {},
'cookies': {},
'headers': {},
'auth': {}
},
{
'method': 'some-value',
'url': 'some-value',
'params': {},
'data': {},
'json': {},
'cookies': {},
'headers': {},
'auth': {}
},
{
'method': 'some-value',
'url': 'some-value',
'params': {},
'data': {},
'json': {},
'cookies': {},
'headers': {},
'auth': {}
}
]
res = await Api.batch(expected_requests)
expect(call_mock.call_count).to(equal(len(expected_requests)))
expect(len(res)).to(equal(len(expected_requests)))
@pytest.mark.asyncio
@asynctest.patch.object(Async, 'all')
@asynctest.patch.object(Api, 'call')
async def test_batch_async(self, *args):
# with as call_mock:
expected_requests = [
{
'method': 'some-value',
'url': 'some-value',
'params': {},
'data': {},
'json': {},
'cookies': {},
'headers': {},
'auth': {}
},
{
'method': 'some-value',
'url': 'some-value',
'params': {},
'data': {},
'json': {},
'cookies': {},
'headers': {},
'auth': {}
},
{
'method': 'some-value',
'url': 'some-value',
'params': {},
'data': {},
'json': {},
'cookies': {},
'headers': {},
'auth': {}
}
]
res = await Api.batch_async(expected_requests)
expect(args[0].call_count).to(equal(len(expected_requests)))
args[1].assert_called()
| [
"noreply@github.com"
] | noreply@github.com |
1dce40b705380b07cbe6eb7c6fbb5f9749a9ae9b | 08b4f0f914e33039d3ca408702b110fd6b24764a | /venv/Scripts/easy_install-script.py | be41f6d074beef7058f951ad44f80f761a1f7357 | [] | no_license | dakaun/wiki_crawler | 6558c0b4f4c9ec8ca94a5b7f0e24fc2bca77d1d1 | 82b526ce63e952d8364c5c553b6f00440a6be9cc | refs/heads/master | 2021-04-12T04:56:27.526150 | 2018-12-13T10:04:53 | 2018-12-13T10:04:53 | 125,738,838 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 449 | py | #!C:\Users\danielak\PycharmProjects\FIZ\venv\Scripts\python.exe
# EASY-INSTALL-ENTRY-SCRIPT: 'setuptools==28.8.0','console_scripts','easy_install'
__requires__ = 'setuptools==28.8.0'
import re
import sys
from pkg_resources import load_entry_point
if __name__ == '__main__':
sys.argv[0] = re.sub(r'(-script\.pyw?|\.exe)?$', '', sys.argv[0])
sys.exit(
load_entry_point('setuptools==28.8.0', 'console_scripts', 'easy_install')()
)
| [
"danielak@famkaun.local"
] | danielak@famkaun.local |
ccf100ecb17578bc9791263e5270183990fed468 | 0b793bce2da8c3d09b7956c0672ddbffd46feaed | /atcoder/corp/keyence2020_c.py | 9e943f94b0f860184c871b6de78e2af5092d409b | [
"MIT"
] | permissive | knuu/competitive-programming | c6c4e08fb231937d988bdc5a60a8ad6b31b97616 | 16bc68fdaedd6f96ae24310d697585ca8836ab6e | refs/heads/master | 2021-01-17T09:39:02.647688 | 2020-11-07T03:17:22 | 2020-11-07T03:17:22 | 27,886,732 | 1 | 0 | null | null | null | null | UTF-8 | Python | false | false | 201 | py | N, K, S = map(int, input().split())
if S == 1:
const = S + 1
else:
const = S - 1
ans = []
for i in range(N):
if i < K:
ans.append(S)
else:
ans.append(const)
print(*ans)
| [
"premier3next@gmail.com"
] | premier3next@gmail.com |
611af549c5585bbe5afcc4f755e6bb733fbd0ce6 | e4370c3831bf4e1324d55930870fd754df3e5e9a | /apps/survey_app/urls.py | 632d830bb37715e5ba45b8f35313f153a7322569 | [] | no_license | grommitt/Django-survey-form | d338053adcbb664ac8198182928854cab321d0ca | d7869acd0bca4b33617bde9fa51cf8e38f38d0b4 | refs/heads/master | 2020-03-17T08:33:37.926850 | 2018-05-15T01:42:09 | 2018-05-15T01:42:09 | 133,442,798 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 326 | py | from django.conf.urls import url
from . import views # This line is new!
urlpatterns = [
url(r'^$', views.index), # This line has changed! Notice that urlpatterns is a list, the comma is in
url(r'^results$', views.results), # anticipation of all the routes that will be coming soon
] | [
"gmshaughn@email.arizona.edu"
] | gmshaughn@email.arizona.edu |
57cc2f13b93c6dddb86a3a96dc0caa25754781c5 | 0c13fd0ae32c77a01a2527190d81e303d48d6fc9 | /src/snk_quiver3.0.py | ec8e47311a46d966e45d66b72ede7c3a58a9a8f3 | [] | no_license | migrau/hinosima | aa040a14381bf645158f57b5dfe738b5d47469a8 | 45a83c7c90bc8e20497c6b59179cc4fa4f21a879 | refs/heads/master | 2021-01-23T02:59:26.276261 | 2017-03-28T01:30:54 | 2017-03-28T01:30:54 | 86,038,967 | 0 | 1 | null | null | null | null | UTF-8 | Python | false | false | 5,800 | py | ##########################################################################################################
# snk.quiver3.0.py #
# #
## Script to polish a PacBio assembly using Quiver with SMRTANALYSIS V3.0 #
# 1. Convert bax.h5 files to bam #
# 2. Run pbalign with each bam file. #
# 3. Merge all the pbalign bam files output in a single bam file. #
# 4. (sort/index bam and index fasta #
# 5. run Quiver. #
# #
## Requirements: #
# - pacbio assembly (from canu, falcon, etc) #
# - pacbio reads (.bax.h5 format) #
# #
## Example run: #
# one node, cpus=24 [1 run with 24threads] #
# (dry run) $ snakemake --snakefile snk.quiver3.0.py -j 1 --config rdir=raw assembly=assembly.fasta -np #
# #
# multi-node [max 80 jobs at once, each one with threads=24] #
# (dry run) $ snakemake -j 80 --snakefile snk.quiver3.0.py --cluster-config cluster.json #
# --cluster "sbatch --partition=compute --cpus-per-task=1 --time=14-0 --job-name=snkmk --mem=10GB" #
# --config rdir=raw assembly=assembly.fasta -np #
# #
##########################################################################################################
import subprocess,glob
from os.path import join
import os,re,sys
from Bio import SeqIO
from os.path import basename
# Globals ---------------------------------------------------------------------
#PATH of PacificBiosciences pitchfork SMRTANALYSIS3.0
SMRTloc="/apps/pitchfork"
# Full path to a folder that holds PacBio reads in bax.h5 format. Use symlinks in case separate folders.
BAX_DIR = config["rdir"]
# assemnbly folder path
ASBLY = config["assembly"]
# Regular expression matching the bax.h5 files.
SAMPLES, = glob_wildcards(join(BAX_DIR, '{sample,[^/]+}.1.bax.h5'))
# Patterns using the 'sample' wildcard.
PATTERN = '{sample}.1.bax.h5'
fnames=ffnames=[]
fnames=glob.glob(BAX_DIR+"/*.1.bax.h5")
for f in fnames:
ffnames.append(os.path.basename(f).split(".1.bax.h5")[0])
# Rules -----------------------------------------------------------------------
rule all:
input:
fasta='quiverOut.fasta',
fastq='quiverOut.fastq'
rule createBAM:
input:
files=join(BAX_DIR, PATTERN)
output:
compliantFiles='bax2bam/{sample}.subreads.bam'
params:
outfiles='bax2bam/{sample}',
SMRT=SMRTloc
shell:"""
source {params.SMRT}/deployment/setup-env.sh
dname=$(dirname {input.files});
fname=$(basename {input.files} .1.bax.h5);
bax2bam $dname/$fname.1.bax.h5 $dname/$fname.2.bax.h5 $dname/$fname.3.bax.h5 -o {params.outfiles} --subread --pulsefeatures=DeletionQV,DeletionTag,InsertionQV,IPD,MergeQV,SubstitutionQV,PulseWidth,SubstitutionTag
"""
BAM_DIR = "bax2bam/"
SAMPLES, = glob_wildcards(join(BAM_DIR, '{sample,[^/]+}.subreads.bam'))
PATTERN = '{sample}.subreads.bam'
rule runPBALIGN:
input:
files=join(BAM_DIR, PATTERN)
output:
'bax2bam/{sample}_aligned.bam'
params:
assembly=ASBLY
shell:"""
pbalign --nproc 12 {input.files} {params.assembly} {output}
"""
rule mergePBALIGN:
input:
expand("bax2bam/{sample}_aligned.bam", sample=ffnames)
output:
'bax2bam/all.bam'
shell:"""
din=$(echo {input} | sed 's/ / -in /g')
bamtools merge -in $din -out {output};
"""
rule sortBAM:
input:
'bax2bam/all.bam'
output:
'bax2bam/all_sort.bam'
shell:"""
bamtools sort -in {input} -out {output}
"""
rule indexBAM:
input:
'bax2bam/all_sort.bam'
output:
'bax2bam/all_sort.bam.bai'
shell:"""
bamtools index -in {input}
"""
rule indexFasta:
input:
ay=ASBLY,
bai='bax2bam/all_sort.bam.bai'
output:
ASBLY+".fai"
shell:"""
samtools faidx {input.ay}
"""
rule pbiBAM:
input:
bai='bax2bam/all_sort.bam.bai'
output:
ASBLY+".pbi"
params:
assembly=ASBLY
shell:"""
python /apps/unit/MikheyevU/miquel/GenomicConsensus/bin/makePbi.py --referenceFasta {params.assembly} bax2bam/all_sort.bam
"""
#if pbi missing file error. http://pb-falcon.readthedocs.io/en/latest/quick_start.html
rule runQuiver:
input:
bam='bax2bam/all_sort.bam',
bai='bax2bam/all_sort.bam.bai',
fai=ASBLY+".fai"
output:
fasta='quiverOut.fasta',
fastq='quiverOut.fastq'
params:
assembly=ASBLY
shell:"""
module load gcc/4.9.2
variantCaller -j 12 --algorithm=best {input.bam} --referenceFilename {params.assembly} -o {output.fasta} -o {output.fastq}
"""
| [
"migrau@uji.es"
] | migrau@uji.es |
ab60302c0ed0fb4b5e89d82951627fa21c93947d | 902aef0f2cde6c73a70c1833bec2c6f4fa1bc0b6 | /StimControl/LightStim/Text.py | 8250df16138250fa9a1abc57f9be913ac1e111ff | [
"BSD-2-Clause",
"BSD-3-Clause",
"LicenseRef-scancode-unknown-license-reference"
] | permissive | chrox/RealTimeElectrophy | 2b0c88b28cbeeb4967db5630f3dfa35764c27f54 | e1a331b23d0a034894a0185324de235091e54bf0 | refs/heads/master | 2020-04-14T12:29:47.958693 | 2013-10-07T14:12:04 | 2013-10-07T14:12:04 | 1,662,847 | 6 | 0 | null | null | null | null | UTF-8 | Python | false | false | 1,048 | py | # Text stimulus
#
# Copyright (C) 2010-2013 Huang Xin
#
# See LICENSE.TXT that came with this file.
# Taget stimuli
#
# Copyright (C) 2010-2013 Huang Xin
#
# See LICENSE.TXT that came with this file.
from VisionEgg.Text import Text
from LightData import dictattr
from Core import Stimulus
class Hint(Stimulus):
def __init__(self, params, **kwargs):
super(Hint, self).__init__(params=params, **kwargs)
self.name = 'hint'
self.parameters = dictattr()
self.set_parameters(self.parameters, params)
self.make_stimuli()
def make_stimuli(self):
position = self.viewport.deg2pix(self.parameters.xorigDeg) + self.viewport.xorig ,\
self.viewport.deg2pix(self.parameters.yorigDeg) + self.viewport.yorig
self.text = Text(text=self.parameters.text,
position=position,
color=self.parameters.color,
font_size=self.parameters.fontsize,
anchor='center')
self.stimuli = [self.text] | [
"chrox.huang@gmail.com"
] | chrox.huang@gmail.com |
8afe9cc9f4f53d06be5e718686be5cb4cf5c0cdb | c67268ac491ecfe606308a43185f1bf8073d56a1 | /unittesting/test_employee2.py | 84682a7e52ffd035b6a9a992a079c59112128dc6 | [] | no_license | jisshub/python-django-training | 3c0fad4c80c78bcfb4b61b025da60d220b502e4b | d8c61f53e3bb500b1a58a706f20108babd6a1a54 | refs/heads/master | 2020-06-21T15:07:25.704209 | 2019-09-01T19:24:02 | 2019-09-01T19:24:02 | 197,487,745 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 1,929 | py | import unittest
# here v import Employee class employee module(employee.py)
from employee import Employee
class EmployeeTest(unittest.TestCase):
def setUp(self):
print('setup\n')
# here v create two employee obj instead of creating them for each test.
# ENSURING DRY PRINCIPLE
self.emp1 = Employee('jiss', 'jose', 3000)
self.emp2 = Employee('isco', 'alarcon', 5000)
def tearDown(self):
print('teardown\n')
def test_email(self):
print('test_email\n')
var1 = self.emp1.email
var2 = self.emp2.email
self.assertEqual(var1, 'jissjose@gmail.com')
self.assertEqual(var2, 'iscoalarcon@gmail.com')
self.emp1.first = 'john'
self.emp2.last = 'james'
self.assertEqual(self.emp1.email, 'johnjose@gmail.com')
self.assertEqual(self.emp2.email, 'iscojames@gmail.com')
def test_fullname(self):
print('test_fullname\n')
self.assertEqual(self.emp1.full_name, 'jiss jose')
self.emp1.first = 'jom'
self.emp1.last = 'thomas'
self.assertEqual(self.emp1.full_name, 'jom thomas')
self.assertEqual(self.emp2.full_name, 'isco alarcon')
self.emp2.first = 'alvaro'
self.emp2.last = 'morata'
self.assertEqual(self.emp2.full_name, 'alvaro morata')
def test_pay(self):
print('test_pay\n')
self.assertEqual(self.emp1.apply_raise, 6000)
self.emp1.pay_raise = 1.5
self.assertEqual(self.emp1.apply_raise, 9000)
self.assertEqual(self.emp2.apply_raise, 10000)
self.emp2.pay_raise = .5
self.assertEqual(self.emp2.apply_raise, 5000)
if __name__ == '__main__':
unittest.main()
# here v text whether value of apply_raise and pay are equal.
# here setUp runs before each test and tearDown method runs after each test.
# order will be like
# setUp
# testmethod
# teardown
| [
"jissmon476@gmial.com"
] | jissmon476@gmial.com |
47d9a4ccfc270889c30a9622496a38134161653c | 2e8ca9eceb525c5e8649525654a931fff637ef6c | /FSF-2020/approximations-and-optimizations/Critical Points/example.py | 3a41be7d6ecb0fbd3b74b5b15c55177857163881 | [] | no_license | abuzar0013/FSF-mathematics-python-code-archive | b4d97833cac727366f7037350300d98ba71008c1 | 573819dbfec617e253c154fd3ccc6fe0c92ab149 | refs/heads/main | 2023-03-01T16:07:55.319281 | 2021-02-05T11:39:01 | 2021-02-05T11:39:01 | null | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 1,220 | py | from manimlib.imports import*
class ExampleAnimation(ThreeDScene):
def construct(self):
axes = ThreeDAxes()
f_text = TextMobject("$f(x,y) = (y-x)(1-2x-3y)$").to_corner(UL)
d = Dot(np.array([0,0,0]), color = '#800000') #---- Critical Point
d_text = TextMobject("$(0.2,0.2)$",color = '#DC143C').scale(0.5).shift(0.2*UP) #----x = 0.2, y = 0.2
r_text=TextMobject("Critical Point",color = '#00FFFF').shift(0.3*DOWN).scale(0.6)
#----f(x,y) = (y-x)(1-2x-3y)
f = ParametricSurface(
lambda u, v: np.array([
u,
v,
(v-u)*(1-2*u-3*v)
]),v_min = -1, v_max = 1, u_min = -1, u_max = 1, checkerboard_colors = [PURPLE_D, PURPLE_E],
resolution=(20, 20)).scale(1)
self.set_camera_orientation(phi = 75 * DEGREES)
self.begin_ambient_camera_rotation(rate=0.5)
self.add_fixed_in_frame_mobjects(f_text)
self.wait(1)
self.add(axes)
self.play(Write(f),Write(d))
self.wait(1)
self.add_fixed_in_frame_mobjects(d_text)
self.wait(1)
self.add_fixed_in_frame_mobjects(r_text)
self.wait(3)
| [
"noreply@github.com"
] | noreply@github.com |
96a38c411c7fb0314bce1ce412e3ee359440beb4 | a3ab76ec61f01e6bfea0925806bba57ae37ca505 | /blog/migrations/0013_auto_20200501_2232.py | 3b69304dc1c57ddb1a59105552ede3c0d1414800 | [] | no_license | kevinzhao07/django_posts | 343c92223046ffafa7673ceee053b8623c9fbca7 | baf48403bc9f6ff36d5f70a2c7359d4658747134 | refs/heads/master | 2022-12-10T00:08:47.789394 | 2020-08-24T16:35:14 | 2020-08-24T16:35:14 | 242,276,050 | 1 | 0 | null | 2022-12-08T04:04:44 | 2020-02-22T03:57:40 | HTML | UTF-8 | Python | false | false | 983 | py | # Generated by Django 2.2.12 on 2020-05-02 02:32
from django.db import migrations, models
import django.utils.timezone
class Migration(migrations.Migration):
dependencies = [
('blog', '0012_auto_20200428_1719'),
]
operations = [
migrations.AlterField(
model_name='comment',
name='date_posted',
field=models.DateTimeField(default=django.utils.timezone.now),
),
migrations.AlterField(
model_name='like',
name='date_liked',
field=models.DateTimeField(default=django.utils.timezone.now),
),
migrations.AlterField(
model_name='message',
name='date_sent',
field=models.DateTimeField(default=django.utils.timezone.now),
),
migrations.AlterField(
model_name='post',
name='date_posted',
field=models.DateTimeField(default=django.utils.timezone.now),
),
]
| [
"kevinmz@umich.edu"
] | kevinmz@umich.edu |
fb572a3e446aa09655b1e5770b4fb5b166afd177 | 9044b1cb3c9a85094cdf02aae1c788e6d39252df | /day16/run.py | da22f278d816038bc490d2355c33cc109571b8f1 | [
"MIT"
] | permissive | 1923851861/Python_Base | 5aea29dabf9531b5f5e945989f2f567042b0bff3 | c9d8224cf41c58235fc67783d73a48f82fee61a1 | refs/heads/master | 2020-03-23T12:11:38.704227 | 2018-08-10T11:01:21 | 2018-08-10T11:01:21 | null | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 2,193 | py | # 当前的执行文件
# x=1
# y=2
# 首次导入模块发生了3件事:
#1、以模块为准创造一个模块的名称空间
#2、执行模块对应的文件,将执行过程中产生的名字都丢到模块的名称空间
#3、在当前执行文件中拿到一个模块名
# import spam
# 之后的重复导入会直接引用之前创造好的结果,不会重复执行模块的文件
# import spam #spam=spam=模块名称空间的内存地址
# spam.名字
# print(x)
# spam.x
# print(spam.money)
# print(spam.read1)
# print(spam.read2)
# print(spam.change)
# money=11111111111111
# spam.read1()
# def read1():
# print('执行文件的read1')
# spam.read2()
# spam.change()
# print(spam.money)
# print(money)
# import spam as sm
# print(sm.money)
# import time,spam,os,sys
# import spam
# import os
# import sys
# import spam
# spam.money
# from ... import ...首次导入也发生了三件事:
#1、以模块为准创造一个模块的名称空间
#2、执行模块对应的文件,将执行过程中产生的名字都丢到模块的名称空间
#3、在当前执行文件的名称空间中拿到一个名字,该名字直接指向模块中的某一个名字,意味着可以不用加任何前缀而直接使用
# x=1
# y=2
#
# from spam import money,read1
# money=10
# print(money)
# print(read1)
# from .... import ... 对比 import 。。。
# 优点:不用加前缀,代码更为精简
# 缺点:容易与当前执行文件中名称空间中的名字冲突
# 相同点:
# 1、都会执行模块对应的文件,都会产生模块的名称空间
# 2、调用功能时,需要跑到定义时寻找作用域关系,与调用位置无关
# 不同点
# 1、一种需要加前缀,一种不需要加前缀
# from spam import money,read1,read2,change
# money=111111111111111111
# read1()
# def read1():
# print('当前执行文件的read1',money)
# read1()
# def read1():
# print('当前执行文件的read1',money)
# read2()
# change=1
# change()
# print(money)
# from spam import money
# from spam import read1
# from spam import read2
# from spam import change
from spam import *
print(money)
print(read1)
print(change)
# print(read2) | [
"1923851861@qq.com"
] | 1923851861@qq.com |
cb2f886ed26850bfebfaf4e3a00a9e730652e300 | cc086a96967761f520c24ce3b22bacecb673cbf2 | /chec_operator/threads/observation.py | 877c8afa6dcb8e097cf23a53a3504277d6791849 | [] | no_license | watsonjj/chec_operator | 39524405b3c6a55fe7fa3e8353da5f456f76a27d | c537a1737a53fe996652c793c09f5a33cd03e208 | refs/heads/master | 2020-04-18T02:27:52.730614 | 2019-01-23T10:41:13 | 2019-01-23T10:41:13 | 167,163,531 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 2,722 | py | import threading
from time import sleep, ctime, time
from datetime import datetime
from chec_operator.utils.enums import CameraState
class ObservingThread(threading.Thread):
def __init__(self, parent_handler, timedelta, triggerdelta):
print("Creating observation thread")
self.parent_handler = parent_handler
self.timedelta = timedelta
self.triggerdelta = triggerdelta
self.starttime = 0
self.starttrigger = 0
self.currenttimedelta = 0
self.currenttriggerdelta = 0
self.get_trigger = self.parent_handler.get_backplane_trigger_count
super(ObservingThread, self).__init__()
self._observation_interrupt = threading.Event()
self.observation_reached_end = False
self.running = False
self.lock = threading.Lock()
def _check_time(self):
if self.timedelta:
self.currenttimedelta = datetime.now() - self.starttime
return self.currenttimedelta >= self.timedelta
else:
return False
def _check_trigger(self):
if self.triggerdelta:
self.currenttriggerdelta = self.get_trigger() - self.starttrigger
return self.currenttriggerdelta >= self.triggerdelta
else:
return False
def observation_ended(self):
return self._observation_interrupt.isSet()
def interrupt_observation(self):
if self.lock.acquire(False):
print("[WARNING] Interrupting observation thread!")
self._observation_interrupt.set()
self.join()
def run(self):
self.running = True
self.starttime = datetime.now()
self.starttrigger = self.get_trigger()
print("[INFO] Starting observation thread, "
"start time = {}, timedelta = {} s, triggerdelta = {}"
.format(ctime(time()), self.timedelta, self.triggerdelta))
while not self.observation_ended():
if self._check_time() or self._check_trigger():
self._finish_run()
break
self.running = False
print("Observation Ended")
def _finish_run(self):
if self.lock.acquire(False):
print("[INFO] Observation thread complete, "
"end time = {}, duration = {}, triggers {} (end) {} (actual)"
.format(ctime(time()), self.currenttimedelta,
self.currenttriggerdelta,
self.get_trigger() - self.starttrigger))
self.observation_reached_end = True
self.parent_handler.go_to_state(CameraState.READY)
def wait_for_end(self):
self.join()
print("Observation Ended")
| [
"jason.jw@live.co.uk"
] | jason.jw@live.co.uk |
b93b498f4935633bf8a5990243e2de8805264276 | ad48d4a4107dd02eb8b155cfcfcae86e40bf5001 | /venv/Scripts/easy_install-3.6-script.py | 6a207e087202fdcfc3044244e7cf026f5af2913f | [] | no_license | Chloe1997/VGGnet_classification | 2515f69d7c0ccf09e7161e367cf8600988c880e4 | e01a1ac42aec74cb914a17a776db9ec28f641b96 | refs/heads/master | 2020-06-30T16:03:44.485820 | 2019-08-07T14:24:52 | 2019-08-07T14:24:52 | 200,879,223 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 464 | py | #!C:\Users\student1\Desktop\VGG_classification\venv\Scripts\python.exe
# EASY-INSTALL-ENTRY-SCRIPT: 'setuptools==39.1.0','console_scripts','easy_install-3.6'
__requires__ = 'setuptools==39.1.0'
import re
import sys
from pkg_resources import load_entry_point
if __name__ == '__main__':
sys.argv[0] = re.sub(r'(-script\.pyw?|\.exe)?$', '', sys.argv[0])
sys.exit(
load_entry_point('setuptools==39.1.0', 'console_scripts', 'easy_install-3.6')()
)
| [
"dellawen1997@gmail.com"
] | dellawen1997@gmail.com |
e52bdadb66f4c2ef03e7ec524da43f8558e1ba9d | df7fbefd727c081fb5d2d0e93356042f20b908fd | /index.py | 627ee82c34fddaf2605fe13cd24a7b6a08516d29 | [] | no_license | Germey/SDUCourse | 1a1c189054dea2082bcacc58f5900434c58689a5 | bea33811eca988d1eb295ada460d677a50187230 | refs/heads/master | 2020-04-06T06:53:51.280673 | 2016-09-13T07:48:42 | 2016-09-13T07:48:42 | 62,208,347 | 4 | 1 | null | null | null | null | UTF-8 | Python | false | false | 88 | py | # -*- coding: utf-8 -*-
from lib.main import main
if __name__ == '__main__':
main() | [
"1016903103@qq.com"
] | 1016903103@qq.com |
911880abd59978863d7a68579ace6207eaea3235 | 0b6f3f46b61e23792bf3cebe068c987a517f88c6 | /chapter_09/动手试一试/9-11 导入Admin类/admin.py | d6f661f4c0b34faa87b1364d8aa7bf6c66eb2ad2 | [] | no_license | honkly/pcc | 0884522c348706a9bfde401df03084f256bf4f04 | 0789b41b0558e847e35eccbd8ecd096a8a1a34e3 | refs/heads/master | 2020-04-29T18:25:48.355921 | 2019-07-25T13:12:57 | 2019-07-25T13:12:57 | 176,323,830 | 0 | 0 | null | 2019-03-18T16:13:13 | 2019-03-18T16:13:12 | null | UTF-8 | Python | false | false | 1,006 | py | class User():
"""docstring for User"""
def __init__(self, first_name, last_name):
self.first_name = first_name
self.last_name = last_name
self.login_attempts = 0
def describe_user(self):
print(self.first_name.title() + ' ' + self.last_name.title())
def greet_user(self):
print("Hello! " + self.first_name.title() + ' ' + self.last_name.title())
def increment_login_attempts(self):
self.login_attempts += 1
def reset_login_attempts(self):
self.login_attempts = 0
def print_login_attempts(self):
print("Login attempts are " + str(self.login_attempts))
class Privileges():
def __init__(self):
self.privileges = ['can add post', 'can delete post', 'can ban user']
def show_privileges(self):
print(self.privileges)
class Admin(User):
"""9-8权限"""
def __init__(self, first_name, last_name):
super().__init__(first_name, last_name)
self.privileges = Privileges() | [
"honkly@163.com"
] | honkly@163.com |
5c03758b507d6d0764e0ee096e04ba7048e30035 | da9b9f75a693d17102be45b88efc212ca6da4085 | /sdk/cosmos/azure-cosmos/azure/cosmos/container.py | 73441d19f5abd428087ba295d4936b854400a8c0 | [
"MIT",
"LicenseRef-scancode-generic-cla"
] | permissive | elraikhm/azure-sdk-for-python | e1f57b2b4d8cc196fb04eb83d81022f50ff63db7 | dcb6fdd18b0d8e0f1d7b34fdf82b27a90ee8eafc | refs/heads/master | 2021-06-21T22:01:37.063647 | 2021-05-21T23:43:56 | 2021-05-21T23:43:56 | 216,855,069 | 0 | 0 | MIT | 2019-10-22T16:05:03 | 2019-10-22T16:05:02 | null | UTF-8 | Python | false | false | 35,017 | py | # The MIT License (MIT)
# Copyright (c) 2014 Microsoft Corporation
# Permission is hereby granted, free of charge, to any person obtaining a copy
# of this software and associated documentation files (the "Software"), to deal
# in the Software without restriction, including without limitation the rights
# to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
# copies of the Software, and to permit persons to whom the Software is
# furnished to do so, subject to the following conditions:
# The above copyright notice and this permission notice shall be included in all
# copies or substantial portions of the Software.
# THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
# IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
# FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
# AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
# LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
# OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
# SOFTWARE.
"""Create, read, update and delete items in the Azure Cosmos DB SQL API service.
"""
from typing import Any, Dict, List, Optional, Union, Iterable, cast # pylint: disable=unused-import
import six
from azure.core.tracing.decorator import distributed_trace # type: ignore
from ._cosmos_client_connection import CosmosClientConnection
from ._base import build_options
from .errors import CosmosResourceNotFoundError
from .http_constants import StatusCodes
from .offer import Offer
from .scripts import ScriptsProxy
from .partition_key import NonePartitionKeyValue
__all__ = ("ContainerProxy",)
# pylint: disable=protected-access
# pylint: disable=missing-client-constructor-parameter-credential,missing-client-constructor-parameter-kwargs
class ContainerProxy(object):
"""
An interface to interact with a specific DB Container.
This class should not be instantiated directly, use :func:`DatabaseProxy.get_container_client` method.
A container in an Azure Cosmos DB SQL API database is a collection of documents,
each of which represented as an Item.
:ivar str id: ID (name) of the container
:ivar str session_token: The session token for the container.
.. note::
To create a new container in an existing database, use :func:`Database.create_container`.
"""
def __init__(self, client_connection, database_link, id, properties=None): # pylint: disable=redefined-builtin
# type: (CosmosClientConnection, str, str, Dict[str, Any]) -> None
self.client_connection = client_connection
self.id = id
self._properties = properties
self.container_link = u"{}/colls/{}".format(database_link, self.id)
self._is_system_key = None
self._scripts = None # type: Optional[ScriptsProxy]
def _get_properties(self):
# type: () -> Dict[str, Any]
if self._properties is None:
self._properties = self.read()
return self._properties
@property
def is_system_key(self):
# type: () -> bool
if self._is_system_key is None:
properties = self._get_properties()
self._is_system_key = (
properties["partitionKey"]["systemKey"] if "systemKey" in properties["partitionKey"] else False
)
return cast('bool', self._is_system_key)
@property
def scripts(self):
# type: () -> ScriptsProxy
if self._scripts is None:
self._scripts = ScriptsProxy(self.client_connection, self.container_link, self.is_system_key)
return cast('ScriptsProxy', self._scripts)
def _get_document_link(self, item_or_link):
# type: (Union[Dict[str, Any], str]) -> str
if isinstance(item_or_link, six.string_types):
return u"{}/docs/{}".format(self.container_link, item_or_link)
return item_or_link["_self"]
def _get_conflict_link(self, conflict_or_link):
# type: (Union[Dict[str, Any], str]) -> str
if isinstance(conflict_or_link, six.string_types):
return u"{}/conflicts/{}".format(self.container_link, conflict_or_link)
return conflict_or_link["_self"]
def _set_partition_key(self, partition_key):
if partition_key == NonePartitionKeyValue:
return CosmosClientConnection._return_undefined_or_empty_partition_key(self.is_system_key)
return partition_key
@distributed_trace
def read(
self,
populate_query_metrics=None, # type: Optional[bool]
populate_partition_key_range_statistics=None, # type: Optional[bool]
populate_quota_info=None, # type: Optional[bool]
**kwargs # type: Any
):
# type: (...) -> Dict[str, Any]
"""
Read the container properties
:param session_token: Token for use with Session consistency.
:param initial_headers: Initial headers to be sent as part of the request.
:param populate_query_metrics: Enable returning query metrics in response headers.
:param populate_partition_key_range_statistics: Enable returning partition key
range statistics in response headers.
:param populate_quota_info: Enable returning collection storage quota information in response headers.
:param request_options: Dictionary of additional properties to be used for the request.
:param response_hook: a callable invoked with the response metadata
:raises ~azure.cosmos.errors.CosmosHttpResponseError: Raised if the container couldn't be retrieved.
This includes if the container does not exist.
:returns: Dict representing the retrieved container.
:rtype: dict[str, Any]
"""
request_options = build_options(kwargs)
response_hook = kwargs.pop('response_hook', None)
if populate_query_metrics is not None:
request_options["populateQueryMetrics"] = populate_query_metrics
if populate_partition_key_range_statistics is not None:
request_options["populatePartitionKeyRangeStatistics"] = populate_partition_key_range_statistics
if populate_quota_info is not None:
request_options["populateQuotaInfo"] = populate_quota_info
collection_link = self.container_link
self._properties = self.client_connection.ReadContainer(
collection_link, options=request_options, **kwargs
)
if response_hook:
response_hook(self.client_connection.last_response_headers, self._properties)
return cast('Dict[str, Any]', self._properties)
@distributed_trace
def read_item(
self,
item, # type: Union[str, Dict[str, Any]]
partition_key, # type: Any
populate_query_metrics=None, # type: Optional[bool]
post_trigger_include=None, # type: Optional[str]
**kwargs # type: Any
):
# type: (...) -> Dict[str, str]
"""
Get the item identified by `item`.
:param item: The ID (name) or dict representing item to retrieve.
:param partition_key: Partition key for the item to retrieve.
:param session_token: Token for use with Session consistency.
:param initial_headers: Initial headers to be sent as part of the request.
:param populate_query_metrics: Enable returning query metrics in response headers.
:param post_trigger_include: trigger id to be used as post operation trigger.
:param request_options: Dictionary of additional properties to be used for the request.
:param response_hook: a callable invoked with the response metadata
:returns: Dict representing the item to be retrieved.
:raises ~azure.cosmos.errors.CosmosHttpResponseError: The given item couldn't be retrieved.
:rtype: dict[str, Any]
.. admonition:: Example:
.. literalinclude:: ../samples/examples.py
:start-after: [START update_item]
:end-before: [END update_item]
:language: python
:dedent: 0
:caption: Get an item from the database and update one of its properties:
:name: update_item
"""
doc_link = self._get_document_link(item)
request_options = build_options(kwargs)
response_hook = kwargs.pop('response_hook', None)
if partition_key:
request_options["partitionKey"] = self._set_partition_key(partition_key)
if populate_query_metrics is not None:
request_options["populateQueryMetrics"] = populate_query_metrics
if post_trigger_include:
request_options["postTriggerInclude"] = post_trigger_include
result = self.client_connection.ReadItem(document_link=doc_link, options=request_options, **kwargs)
if response_hook:
response_hook(self.client_connection.last_response_headers, result)
return result
@distributed_trace
def read_all_items(
self,
max_item_count=None, # type: Optional[int]
populate_query_metrics=None, # type: Optional[bool]
**kwargs # type: Any
):
# type: (...) -> Iterable[Dict[str, Any]]
"""
List all items in the container.
:param max_item_count: Max number of items to be returned in the enumeration operation.
:param session_token: Token for use with Session consistency.
:param initial_headers: Initial headers to be sent as part of the request.
:param populate_query_metrics: Enable returning query metrics in response headers.
:param feed_options: Dictionary of additional properties to be used for the request.
:param response_hook: a callable invoked with the response metadata
:returns: An Iterable of items (dicts).
:rtype: Iterable[dict[str, Any]]
"""
feed_options = build_options(kwargs)
response_hook = kwargs.pop('response_hook', None)
if max_item_count is not None:
feed_options["maxItemCount"] = max_item_count
if populate_query_metrics is not None:
feed_options["populateQueryMetrics"] = populate_query_metrics
if hasattr(response_hook, "clear"):
response_hook.clear()
items = self.client_connection.ReadItems(
collection_link=self.container_link, feed_options=feed_options, response_hook=response_hook, **kwargs
)
if response_hook:
response_hook(self.client_connection.last_response_headers, items)
return items
@distributed_trace
def query_items_change_feed(
self,
partition_key_range_id=None, # type: Optional[str]
is_start_from_beginning=False, # type: bool
continuation=None, # type: Optional[str]
max_item_count=None, # type: Optional[int]
**kwargs # type: Any
):
# type: (...) -> Iterable[Dict[str, Any]]
"""
Get a sorted list of items that were changed, in the order in which they were modified.
:param partition_key_range_id: ChangeFeed requests can be executed against specific partition key ranges.
This is used to process the change feed in parallel across multiple consumers.
:param is_start_from_beginning: Get whether change feed should start from
beginning (true) or from current (false). By default it's start from current (false).
:param continuation: e_tag value to be used as continuation for reading change feed.
:param max_item_count: Max number of items to be returned in the enumeration operation.
:param feed_options: Dictionary of additional properties to be used for the request.
:param response_hook: a callable invoked with the response metadata
:returns: An Iterable of items (dicts).
:rtype: Iterable[dict[str, Any]]
"""
feed_options = build_options(kwargs)
response_hook = kwargs.pop('response_hook', None)
if partition_key_range_id is not None:
feed_options["partitionKeyRangeId"] = partition_key_range_id
if is_start_from_beginning is not None:
feed_options["isStartFromBeginning"] = is_start_from_beginning
if max_item_count is not None:
feed_options["maxItemCount"] = max_item_count
if continuation is not None:
feed_options["continuation"] = continuation
if hasattr(response_hook, "clear"):
response_hook.clear()
result = self.client_connection.QueryItemsChangeFeed(
self.container_link, options=feed_options, response_hook=response_hook, **kwargs
)
if response_hook:
response_hook(self.client_connection.last_response_headers, result)
return result
@distributed_trace
def query_items(
self,
query, # type: str
parameters=None, # type: Optional[List[str]]
partition_key=None, # type: Optional[Any]
enable_cross_partition_query=None, # type: Optional[bool]
max_item_count=None, # type: Optional[int]
enable_scan_in_query=None, # type: Optional[bool]
populate_query_metrics=None, # type: Optional[bool]
**kwargs # type: Any
):
# type: (...) -> Iterable[Dict[str, Any]]
"""
Return all results matching the given `query`.
You can use any value for the container name in the FROM clause, but typically the container name is used.
In the examples below, the container name is "products," and is aliased as "p" for easier referencing
in the WHERE clause.
:param query: The Azure Cosmos DB SQL query to execute.
:param parameters: Optional array of parameters to the query. Ignored if no query is provided.
:param partition_key: Specifies the partition key value for the item.
:param enable_cross_partition_query: Allows sending of more than one request to
execute the query in the Azure Cosmos DB service.
More than one request is necessary if the query is not scoped to single partition key value.
:param max_item_count: Max number of items to be returned in the enumeration operation.
:param session_token: Token for use with Session consistency.
:param initial_headers: Initial headers to be sent as part of the request.
:param enable_scan_in_query: Allow scan on the queries which couldn't be served as
indexing was opted out on the requested paths.
:param populate_query_metrics: Enable returning query metrics in response headers.
:param feed_options: Dictionary of additional properties to be used for the request.
:param response_hook: a callable invoked with the response metadata
:returns: An Iterable of items (dicts).
:rtype: Iterable[dict[str, Any]]
.. admonition:: Example:
.. literalinclude:: ../samples/examples.py
:start-after: [START query_items]
:end-before: [END query_items]
:language: python
:dedent: 0
:caption: Get all products that have not been discontinued:
:name: query_items
.. literalinclude:: ../samples/examples.py
:start-after: [START query_items_param]
:end-before: [END query_items_param]
:language: python
:dedent: 0
:caption: Parameterized query to get all products that have been discontinued:
:name: query_items_param
"""
feed_options = build_options(kwargs)
response_hook = kwargs.pop('response_hook', None)
if enable_cross_partition_query is not None:
feed_options["enableCrossPartitionQuery"] = enable_cross_partition_query
if max_item_count is not None:
feed_options["maxItemCount"] = max_item_count
if populate_query_metrics is not None:
feed_options["populateQueryMetrics"] = populate_query_metrics
if partition_key is not None:
feed_options["partitionKey"] = self._set_partition_key(partition_key)
if enable_scan_in_query is not None:
feed_options["enableScanInQuery"] = enable_scan_in_query
if hasattr(response_hook, "clear"):
response_hook.clear()
items = self.client_connection.QueryItems(
database_or_container_link=self.container_link,
query=query if parameters is None else dict(query=query, parameters=parameters),
options=feed_options,
partition_key=partition_key,
response_hook=response_hook,
**kwargs
)
if response_hook:
response_hook(self.client_connection.last_response_headers, items)
return items
@distributed_trace
def replace_item(
self,
item, # type: Union[str, Dict[str, Any]]
body, # type: Dict[str, Any]
populate_query_metrics=None, # type: Optional[bool]
pre_trigger_include=None, # type: Optional[str]
post_trigger_include=None, # type: Optional[str]
**kwargs # type: Any
):
# type: (...) -> Dict[str, str]
"""
Replaces the specified item if it exists in the container.
:param item: The ID (name) or dict representing item to be replaced.
:param body: A dict-like object representing the item to replace.
:param session_token: Token for use with Session consistency.
:param initial_headers: Initial headers to be sent as part of the request.
:param access_condition: Conditions Associated with the request.
:param populate_query_metrics: Enable returning query metrics in response headers.
:param pre_trigger_include: trigger id to be used as pre operation trigger.
:param post_trigger_include: trigger id to be used as post operation trigger.
:param request_options: Dictionary of additional properties to be used for the request.
:param response_hook: a callable invoked with the response metadata
:returns: A dict representing the item after replace went through.
:raises ~azure.cosmos.errors.CosmosHttpResponseError: The replace failed or the item with
given id does not exist.
:rtype: dict[str, Any]
"""
item_link = self._get_document_link(item)
request_options = build_options(kwargs)
response_hook = kwargs.pop('response_hook', None)
request_options["disableIdGeneration"] = True
if populate_query_metrics is not None:
request_options["populateQueryMetrics"] = populate_query_metrics
if pre_trigger_include:
request_options["preTriggerInclude"] = pre_trigger_include
if post_trigger_include:
request_options["postTriggerInclude"] = post_trigger_include
result = self.client_connection.ReplaceItem(
document_link=item_link, new_document=body, options=request_options, **kwargs
)
if response_hook:
response_hook(self.client_connection.last_response_headers, result)
return result
@distributed_trace
def upsert_item(
self,
body, # type: Dict[str, Any]
populate_query_metrics=None, # type: Optional[bool]
pre_trigger_include=None, # type: Optional[str]
post_trigger_include=None, # type: Optional[str]
**kwargs # type: Any
):
# type: (...) -> Dict[str, str]
"""
Insert or update the specified item.
If the item already exists in the container, it is replaced. If it does not, it is inserted.
:param body: A dict-like object representing the item to update or insert.
:param session_token: Token for use with Session consistency.
:param initial_headers: Initial headers to be sent as part of the request.
:param access_condition: Conditions Associated with the request.
:param populate_query_metrics: Enable returning query metrics in response headers.
:param pre_trigger_include: trigger id to be used as pre operation trigger.
:param post_trigger_include: trigger id to be used as post operation trigger.
:param request_options: Dictionary of additional properties to be used for the request.
:param response_hook: a callable invoked with the response metadata
:returns: A dict representing the upserted item.
:raises ~azure.cosmos.errors.CosmosHttpResponseError: The given item could not be upserted.
:rtype: dict[str, Any]
"""
request_options = build_options(kwargs)
response_hook = kwargs.pop('response_hook', None)
request_options["disableIdGeneration"] = True
if populate_query_metrics is not None:
request_options["populateQueryMetrics"] = populate_query_metrics
if pre_trigger_include:
request_options["preTriggerInclude"] = pre_trigger_include
if post_trigger_include:
request_options["postTriggerInclude"] = post_trigger_include
result = self.client_connection.UpsertItem(
database_or_container_link=self.container_link, document=body, **kwargs)
if response_hook:
response_hook(self.client_connection.last_response_headers, result)
return result
@distributed_trace
def create_item(
self,
body, # type: Dict[str, Any]
populate_query_metrics=None, # type: Optional[bool]
pre_trigger_include=None, # type: Optional[str]
post_trigger_include=None, # type: Optional[str]
indexing_directive=None, # type: Optional[Any]
**kwargs # type: Any
):
# type: (...) -> Dict[str, str]
"""
Create an item in the container.
To update or replace an existing item, use the :func:`ContainerProxy.upsert_item` method.
:param body: A dict-like object representing the item to create.
:param session_token: Token for use with Session consistency.
:param initial_headers: Initial headers to be sent as part of the request.
:param access_condition: Conditions Associated with the request.
:param populate_query_metrics: Enable returning query metrics in response headers.
:param pre_trigger_include: trigger id to be used as pre operation trigger.
:param post_trigger_include: trigger id to be used as post operation trigger.
:param indexing_directive: Indicate whether the document should be omitted from indexing.
:param request_options: Dictionary of additional properties to be used for the request.
:param response_hook: a callable invoked with the response metadata
:returns: A dict representing the new item.
:raises ~azure.cosmos.errors.CosmosHttpResponseError: Item with the given ID already exists.
:rtype: dict[str, Any]
"""
request_options = build_options(kwargs)
response_hook = kwargs.pop('response_hook', None)
request_options["disableAutomaticIdGeneration"] = True
if populate_query_metrics:
request_options["populateQueryMetrics"] = populate_query_metrics
if pre_trigger_include:
request_options["preTriggerInclude"] = pre_trigger_include
if post_trigger_include:
request_options["postTriggerInclude"] = post_trigger_include
if indexing_directive:
request_options["indexingDirective"] = indexing_directive
result = self.client_connection.CreateItem(
database_or_container_link=self.container_link, document=body, options=request_options, **kwargs
)
if response_hook:
response_hook(self.client_connection.last_response_headers, result)
return result
@distributed_trace
def delete_item(
self,
item, # type: Union[Dict[str, Any], str]
partition_key, # type: Any
populate_query_metrics=None, # type: Optional[bool]
pre_trigger_include=None, # type: Optional[str]
post_trigger_include=None, # type: Optional[str]
**kwargs # type: Any
):
# type: (...) -> None
"""
Delete the specified item from the container.
:param item: The ID (name) or dict representing item to be deleted.
:param partition_key: Specifies the partition key value for the item.
:param session_token: Token for use with Session consistency.
:param initial_headers: Initial headers to be sent as part of the request.
:param access_condition: Conditions Associated with the request.
:param populate_query_metrics: Enable returning query metrics in response headers.
:param pre_trigger_include: trigger id to be used as pre operation trigger.
:param post_trigger_include: trigger id to be used as post operation trigger.
:param request_options: Dictionary of additional properties to be used for the request.
:param response_hook: a callable invoked with the response metadata
:raises ~azure.cosmos.errors.CosmosHttpResponseError: The item wasn't deleted successfully.
:raises ~azure.cosmos.errors.CosmosResourceNotFoundError: The item does not exist in the container.
:rtype: None
"""
request_options = build_options(kwargs)
response_hook = kwargs.pop('response_hook', None)
if partition_key:
request_options["partitionKey"] = self._set_partition_key(partition_key)
if populate_query_metrics is not None:
request_options["populateQueryMetrics"] = populate_query_metrics
if pre_trigger_include:
request_options["preTriggerInclude"] = pre_trigger_include
if post_trigger_include:
request_options["postTriggerInclude"] = post_trigger_include
document_link = self._get_document_link(item)
result = self.client_connection.DeleteItem(document_link=document_link, options=request_options, **kwargs)
if response_hook:
response_hook(self.client_connection.last_response_headers, result)
@distributed_trace
def read_offer(self, **kwargs):
# type: (Any) -> Offer
"""
Read the Offer object for this container.
:param response_hook: a callable invoked with the response metadata
:returns: Offer for the container.
:raises ~azure.cosmos.errors.CosmosHttpResponseError: No offer exists for the container or
the offer could not be retrieved.
:rtype: ~azure.cosmos.Offer
"""
response_hook = kwargs.pop('response_hook', None)
properties = self._get_properties()
link = properties["_self"]
query_spec = {
"query": "SELECT * FROM root r WHERE r.resource=@link",
"parameters": [{"name": "@link", "value": link}],
}
offers = list(self.client_connection.QueryOffers(query_spec, **kwargs))
if not offers:
raise CosmosResourceNotFoundError(
status_code=StatusCodes.NOT_FOUND,
message="Could not find Offer for container " + self.container_link)
if response_hook:
response_hook(self.client_connection.last_response_headers, offers)
return Offer(offer_throughput=offers[0]["content"]["offerThroughput"], properties=offers[0])
@distributed_trace
def replace_throughput(self, throughput, **kwargs):
# type: (int, Any) -> Offer
"""
Replace the container's throughput
:param throughput: The throughput to be set (an integer).
:param response_hook: a callable invoked with the response metadata
:returns: Offer for the container, updated with new throughput.
:raises ~azure.cosmos.errors.CosmosHttpResponseError: No offer exists for the container
or the offer could not be updated.
:rtype: ~azure.cosmos.Offer
"""
response_hook = kwargs.pop('response_hook', None)
properties = self._get_properties()
link = properties["_self"]
query_spec = {
"query": "SELECT * FROM root r WHERE r.resource=@link",
"parameters": [{"name": "@link", "value": link}],
}
offers = list(self.client_connection.QueryOffers(query_spec, **kwargs))
if not offers:
raise CosmosResourceNotFoundError(
status_code=StatusCodes.NOT_FOUND,
message="Could not find Offer for container " + self.container_link)
new_offer = offers[0].copy()
new_offer["content"]["offerThroughput"] = throughput
data = self.client_connection.ReplaceOffer(offer_link=offers[0]["_self"], offer=offers[0], **kwargs)
if response_hook:
response_hook(self.client_connection.last_response_headers, data)
return Offer(offer_throughput=data["content"]["offerThroughput"], properties=data)
@distributed_trace
def list_conflicts(self, max_item_count=None, **kwargs):
# type: (Optional[int], Any) -> Iterable[Dict[str, Any]]
"""
List all conflicts in the container.
:param max_item_count: Max number of items to be returned in the enumeration operation.
:param feed_options: Dictionary of additional properties to be used for the request.
:param response_hook: a callable invoked with the response metadata
:returns: An Iterable of conflicts (dicts).
:rtype: Iterable[dict[str, Any]]
"""
feed_options = build_options(kwargs)
response_hook = kwargs.pop('response_hook', None)
if max_item_count is not None:
feed_options["maxItemCount"] = max_item_count
result = self.client_connection.ReadConflicts(
collection_link=self.container_link, feed_options=feed_options, **kwargs
)
if response_hook:
response_hook(self.client_connection.last_response_headers, result)
return result
@distributed_trace
def query_conflicts(
self,
query, # type: str
parameters=None, # type: Optional[List[str]]
enable_cross_partition_query=None, # type: Optional[bool]
partition_key=None, # type: Optional[Any]
max_item_count=None, # type: Optional[int]
**kwargs # type: Any
):
# type: (...) -> Iterable[Dict[str, Any]]
"""
Return all conflicts matching the given `query`.
:param query: The Azure Cosmos DB SQL query to execute.
:param parameters: Optional array of parameters to the query. Ignored if no query is provided.
:param partition_key: Specifies the partition key value for the item.
:param enable_cross_partition_query: Allows sending of more than one request to execute
the query in the Azure Cosmos DB service.
More than one request is necessary if the query is not scoped to single partition key value.
:param max_item_count: Max number of items to be returned in the enumeration operation.
:param feed_options: Dictionary of additional properties to be used for the request.
:param response_hook: a callable invoked with the response metadata
:returns: An Iterable of conflicts (dicts).
:rtype: Iterable[dict[str, Any]]
"""
feed_options = build_options(kwargs)
response_hook = kwargs.pop('response_hook', None)
if max_item_count is not None:
feed_options["maxItemCount"] = max_item_count
if enable_cross_partition_query is not None:
feed_options["enableCrossPartitionQuery"] = enable_cross_partition_query
if partition_key is not None:
feed_options["partitionKey"] = self._set_partition_key(partition_key)
result = self.client_connection.QueryConflicts(
collection_link=self.container_link,
query=query if parameters is None else dict(query=query, parameters=parameters),
options=feed_options,
**kwargs
)
if response_hook:
response_hook(self.client_connection.last_response_headers, result)
return result
@distributed_trace
def get_conflict(self, conflict, partition_key, **kwargs):
# type: (Union[str, Dict[str, Any]], Any, Any) -> Dict[str, str]
"""
Get the conflict identified by `conflict`.
:param conflict: The ID (name) or dict representing the conflict to retrieve.
:param partition_key: Partition key for the conflict to retrieve.
:param request_options: Dictionary of additional properties to be used for the request.
:param response_hook: a callable invoked with the response metadata
:returns: A dict representing the retrieved conflict.
:raises ~azure.cosmos.errors.CosmosHttpResponseError: The given conflict couldn't be retrieved.
:rtype: dict[str, Any]
"""
request_options = build_options(kwargs)
response_hook = kwargs.pop('response_hook', None)
if partition_key:
request_options["partitionKey"] = self._set_partition_key(partition_key)
result = self.client_connection.ReadConflict(
conflict_link=self._get_conflict_link(conflict), options=request_options, **kwargs
)
if response_hook:
response_hook(self.client_connection.last_response_headers, result)
return result
@distributed_trace
def delete_conflict(self, conflict, partition_key, **kwargs):
# type: (Union[str, Dict[str, Any]], Any, Any) -> None
"""
Delete the specified conflict from the container.
:param conflict: The ID (name) or dict representing the conflict to be deleted.
:param partition_key: Partition key for the conflict to delete.
:param request_options: Dictionary of additional properties to be used for the request.
:param response_hook: a callable invoked with the response metadata
:raises ~azure.cosmos.errors.CosmosHttpResponseError: The conflict wasn't deleted successfully.
:raises ~azure.cosmos.errors.CosmosResourceNotFoundError: The conflict does not exist in the container.
:rtype: None
"""
request_options = build_options(kwargs)
response_hook = kwargs.pop('response_hook', None)
if partition_key:
request_options["partitionKey"] = self._set_partition_key(partition_key)
result = self.client_connection.DeleteConflict(
conflict_link=self._get_conflict_link(conflict), options=request_options, **kwargs
)
if response_hook:
response_hook(self.client_connection.last_response_headers, result)
| [
"noreply@github.com"
] | noreply@github.com |
af2729e0f3c3c35ad20460334df67ddb78436aec | 6635686859b272d291d0ba3520ccd03cdc80a349 | /DT/threadingtext.py | 5b21c27fdb9a81a7ecb35c0e0d9c9ebe52c19d32 | [] | no_license | yangrencong/web_spiders | ac15c491f60e489000e5312c999f02e6c4fdafdf | 69fdc6eeb5ad19283690c056064f8853e0256445 | refs/heads/master | 2020-03-28T18:45:50.800667 | 2018-10-26T02:50:16 | 2018-10-26T02:50:16 | 148,908,630 | 1 | 0 | null | null | null | null | UTF-8 | Python | false | false | 995 | py | #!/usr/bin/env python
# -*- coding=utf8 -*-
"""
# Author: Mr.yang
# Created Time : 2018/10/10 星期三 12:26:07
# File Name: threadingtext.py
# Description:
# Editortool: vim8.0
"""
import threading
import time
class myThread(threading.Thread):
def __init__(self ,name ,delay):
threading.Thread.__init__(self)
self.name = name
self.delay = delay
def run(self):
print("Starting " + self.name)
print_time(self.name ,self.delay)
print("Exiting " + self.name)
def print_time(threadName ,delay):
counter = 0
while counter < 3:
time.sleep(delay)
print(threadName ,time.ctime())
counter += 1
threads = []
#创建新线程
thread1 = myThread("Thread-1" ,1)
thread2 = myThread("Thread-2" ,2)
#开启新线程
thread1.start()
thread2.start()
#添加线程到线程列表
threads.append(thread1)
threads.append(thread2)
#等待所有线程完成
for t in threads:
t.join()
print("Exiting main thread")
| [
"1452581359@qq.com"
] | 1452581359@qq.com |
8b59ab52b831746af65b2e33c02a38521a75fd82 | 40c86e0b1cda09b4beba5a74099e35578fce0dd5 | /ascii.py | 50a1e4e3711773a9a73763f4367e00da7f81ff26 | [] | no_license | nikitiwari/Learning_Python | ec9b479e6cd47423d6754aa37eb1032444cd186d | 6a76dc9fbc0934d0d655e296a7777a09bcc26afb | refs/heads/master | 2020-04-05T13:05:19.686745 | 2017-07-15T19:17:26 | 2017-07-15T19:17:26 | 95,075,399 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 105 | py | c = raw_input("Enter a character:");
try :
print ord(c)
except :
print "Only characters allowed"
| [
"tiwarinikita96@gmail.com"
] | tiwarinikita96@gmail.com |
34b5e8044cc30321ccc9752daf6dcba32df3f719 | 00a6541d639d073e819f2c080e6c65610eff1982 | /coordinates.py | beab54eac858b13725db676f5022b7b6e5a8878f | [] | no_license | Wajahat0/Annotation-Transfer-Across-the-Microscope | 16fd0eb339363505449297981c8b4817e93dc594 | 911a79b3307b890bb7ec7a2822d91f4f6fa9a6d6 | refs/heads/main | 2023-07-13T12:45:14.350687 | 2021-08-27T05:15:39 | 2021-08-27T05:15:39 | null | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 2,544 | py |
import sys
import cv2
import argparse
import numpy as np
import pandas as pd
from yattag import Doc, indent
from homography_cal import homography_cal
def tranfer_annoation(data_path,t_data_path,slide_num,r_res,t_res,surf):
load_csv_path = data_path +'/'+ slide_num+'/'+ r_res+ '/' +r_res + '_labels.csv'
save_csv_path = t_data_path +'/'+ slide_num+'/'+ t_res+ '/' +t_res + '_labels.csv'
r_image_path = data_path +'/'+ slide_num+'/'+ r_res+ '/'
t_image_path = t_data_path +'/'+ slide_num+'/'+ t_res+ '/'
print(load_csv_path)
print(save_csv_path)
print(r_image_path)
print(t_image_path)
df = pd.read_csv(data_path +'/'+ slide_num+'/'+ r_res+ '/' +r_res + '_labels.csv')
new_df=df
ref_filename = str(df.iloc[0,0])
print(ref_filename)
ref=cv2.imread(r_image_path + ref_filename)
xz=ref_filename.split('_')[-1]
t_image= ref_filename.replace(xz, (t_res +'.png'))
t_filename = t_image_path + t_image #ref_filename[:-9]+t_res+ '.png'
print(t_filename)
# # print(filename)
target_image=cv2.imread(t_filename)
h,w,c = target_image.shape
# print('ref_image:-', ref_filename)
# print('target_image:-', filename)
h2=homography_cal(ref,target_image,surf)
fname = ref_filename
for i in range(0, len(df)):
# read refrence image
ref_filename = df.iloc[i,0]
ref=cv2.imread(r_image_path + ref_filename)
# # read target image
# # need change here
# xz=ref_filename.split('_')[-1]
t_image= ref_filename.replace(xz, (t_res +'.png'))
t_filename = t_image_path + t_image #ref_filename[:-9]+t_res+ '.png'
target_image=cv2.imread(t_filename)
h,w,c = target_image.shape
if fname == ref_filename:
print('skip_homograpy.....')
else:
print('compute homograpy.....')
h2=homography_cal(ref,target_image,surf)
points = np.float32([[df.iloc[i,4],df.iloc[i,5],df.iloc[i,6],df.iloc[i,7]]]).reshape(-1, 1, 2)
t_points = cv2.perspectiveTransform(points, h2).reshape(1,4)
new_df.iloc[i,0] = t_image
new_df.iloc[i,1] = w
new_df.iloc[i,2] = h
new_df.iloc[i,3] = df.iloc[i,3]
new_df.iloc[i,4] = t_points[0,0]
new_df.iloc[i,5] = t_points[0,1]
new_df.iloc[i,6] = t_points[0,2]
new_df.iloc[i,7] = t_points[0,3]
fname = ref_filename
new_df.to_csv(save_csv_path, index=False)
| [
"noreply@github.com"
] | noreply@github.com |
0e6201ff8130bc459133a9f810db8ce61eaaf2b8 | 502a801b6ac886b25e915bc70d97f258acd9db56 | /src/data/srdata.py | e23ed16185415db9df0f0005e85ef56752bedfcf | [] | no_license | Dearbreeze/SPDNet | 31e79bff29678572372c0a99ac343932759f0291 | ff805b8a008e241ed7f8bb80242cbf2abca94f10 | refs/heads/main | 2023-08-20T10:11:14.386825 | 2021-10-13T10:45:11 | 2021-10-13T10:45:11 | null | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 7,334 | py | import os
import glob
from data import common
import pickle
import numpy as np
import imageio
import torch
import torch.utils.data as data
class SRData(data.Dataset):
def __init__(self, args, name='', train=True, benchmark=False):
self.args = args
self.name = name
self.train = train
self.split = 'train' if train else 'test'
self.do_eval = True
self.benchmark = benchmark
self.scale = args.scale
self.idx_scale = 0
data_range = [r.split('-') for r in args.data_range.split('/')]
if train:
data_range = data_range[0]
else:
if args.test_only and len(data_range) == 1:
data_range = data_range[0]
else:
data_range = data_range[1]
self.begin, self.end = list(map(lambda x: int(x), data_range))
self._set_filesystem(args.dir_data)
if args.ext.find('img') < 0:
path_bin = os.path.join(self.apath, 'bin')
os.makedirs(path_bin, exist_ok=True)
list_hr, list_lr = self._scan()
if args.ext.find('bin') >= 0:
# Binary files are stored in 'bin' folder
# If the binary file exists, load it. If not, make it.
list_hr, list_lr = self._scan()
self.images_hr = self._check_and_load(
args.ext, list_hr, self._name_hrbin()
)
self.images_lr = [
self._check_and_load(args.ext, l, self._name_lrbin(s)) \
for s, l in zip(self.scale, list_lr)
]
else:
if args.ext.find('img') >= 0 or benchmark:
self.images_hr, self.images_lr = list_hr, list_lr
elif args.ext.find('sep') >= 0:
os.makedirs(
self.dir_hr.replace(self.apath, path_bin),
exist_ok=True
)
for s in self.scale:
os.makedirs(
os.path.join(
self.dir_lr.replace(self.apath, path_bin),
'X{}'.format(s)
),
exist_ok=True
)
self.images_hr, self.images_lr = [], [[] for _ in self.scale]
for h in list_hr:
b = h.replace(self.apath, path_bin)
b = b.replace(self.ext[0], '.pt')
self.images_hr.append(b)
self._check_and_load(
args.ext, [h], b, verbose=True, load=False
)
for i, ll in enumerate(list_lr):
for l in ll:
b = l.replace(self.apath, path_bin)
b = b.replace(self.ext[1], '.pt')
self.images_lr[i].append(b)
self._check_and_load(
args.ext, [l], b, verbose=True, load=False
)
if train:
self.repeat \
= args.test_every // (len(self.images_hr) // args.batch_size)
# Below functions as used to prepare images
def _scan(self):
names_hr = sorted(
glob.glob(os.path.join(self.dir_hr, '*' + self.ext[0]))
)
names_lr = [[] for _ in self.scale]
for f in names_hr:
f = f.replace('.png','x2.png')
filename, _ = os.path.splitext(os.path.basename(f))
for si, s in enumerate(self.scale):
names_lr[si].append(os.path.join(
self.dir_lr, '{}{}'.format(
filename, self.ext[1]
)
))
return names_hr, names_lr
def _set_filesystem(self, dir_data):
self.apath = os.path.join(dir_data, self.name)
self.dir_hr = os.path.join(self.apath, 'HR')
self.dir_lr = os.path.join(self.apath, 'LR_bicubic')
self.ext = ('.png', '.png')
def _name_hrbin(self):
return os.path.join(
self.apath,
'bin',
'{}_bin_HR.pt'.format(self.split)
)
def _name_lrbin(self, scale):
return os.path.join(
self.apath,
'bin',
'{}_bin_LR.pt'.format(self.split)
)
def _check_and_load(self, ext, l, f, verbose=True, load=True):
if os.path.isfile(f) and ext.find('reset') < 0:
if load:
if verbose: print('Loading {}...'.format(f))
with open(f, 'rb') as _f: ret = pickle.load(_f)
return ret
else:
return None
else:
if verbose:
if ext.find('reset') >= 0:
print('Making a new binary: {}'.format(f))
else:
print('{} does not exist. Now making binary...'.format(f))
b = [{
'name': os.path.splitext(os.path.basename(_l))[0],
'image': imageio.imread(_l)
} for _l in l]
with open(f, 'wb') as _f: pickle.dump(b, _f)
return b
def __getitem__(self, idx):
lr, hr, filename = self._load_file(idx)
lr, hr = self.get_patch(lr, hr)
lr, hr = common.set_channel(lr, hr, n_channels=self.args.n_colors)
lr_tensor, hr_tensor = common.np2Tensor(
lr, hr, rgb_range=self.args.rgb_range
)
return lr_tensor, hr_tensor, filename
def __len__(self):
if self.train:
return len(self.images_hr) * self.repeat
else:
return len(self.images_hr)
def _get_index(self, idx):
if self.train:
return idx % len(self.images_hr)
else:
return idx
def _load_file(self, idx):
idx = self._get_index(idx)
f_hr = self.images_hr[idx]
f_lr = self.images_lr[self.idx_scale][idx]
if self.args.ext.find('bin') >= 0:
filename = f_hr['name']
hr = f_hr['image']
lr = f_lr['image']
else:
filename, _ = os.path.splitext(os.path.basename(f_hr))
if self.args.ext == 'img' or self.benchmark:
hr = imageio.imread(f_hr)
lr = imageio.imread(f_lr)
elif self.args.ext.find('sep') >= 0:
with open(f_hr, 'rb') as _f: hr = np.load(_f)[0]['image']
with open(f_lr, 'rb') as _f: lr = np.load(_f)[0]['image']
return lr, hr, filename
def get_patch(self, lr, hr):
scale = self.scale[self.idx_scale]
multi_scale = len(self.scale) > 1
if self.train:
# print('****preparte data****')
lr, hr = common.get_patch(
lr,
hr,
patch_size=self.args.patch_size,
scale=scale,
multi_scale=multi_scale
)
if not self.args.no_augment:
# print('****use augment****')
lr, hr = common.augment(lr, hr)
else:
ih, iw = lr.shape[:2]
hr = hr[0:ih, 0:iw]
#hr = hr[0:ih * scale, 0:iw * scale]
return lr, hr
def set_scale(self, idx_scale):
self.idx_scale = idx_scale
| [
"noreply@github.com"
] | noreply@github.com |
4bfbe0f48777b895dcde27aa0a5fff0fa22a8110 | 20f8e01f6c574e2cc0af8a9fffadc54e5ae938cd | /util/io.py | 8c77b6067c60d3f0872960b48d75f062ee326cb2 | [
"MIT"
] | permissive | samjbaugh/CoopNets | 3c5334b5e0b2071bf0423318306ae8d2abaa47eb | 0e3ef7060cbc37c6b4e32b93d7b4613f9ace42fe | refs/heads/master | 2020-03-10T03:41:43.358144 | 2018-03-27T21:37:41 | 2018-03-27T21:37:41 | 129,171,842 | 0 | 0 | MIT | 2018-04-12T00:51:12 | 2018-04-12T00:51:11 | null | UTF-8 | Python | false | false | 733 | py | import sys
import logging
import tensorflow as tf
def init_log(path):
log = logging.getLogger()
log.setLevel(logging.INFO)
formatter_cs = logging.Formatter('%(message)s')
cs = logging.StreamHandler(sys.stdout)
cs.setLevel(logging.INFO)
cs.setFormatter(formatter_cs)
log.addHandler(cs)
log = logging.getLogger('tensorflow')
log.setLevel(logging.INFO)
log.handlers = []
formatter_fh = logging.Formatter('%(asctime)s - %(message)s')
fh = logging.FileHandler(path)
fh.setLevel(logging.DEBUG)
fh.setFormatter(formatter_fh)
log.addHandler(fh)
def make_dir(output_dir):
if not tf.gfile.Exists(output_dir):
tf.gfile.MakeDirs(output_dir)
return output_dir
| [
"erik.nijkamp@gmail.com"
] | erik.nijkamp@gmail.com |
3245d99ff8b76871f71f66e0d14bf4ef52546a21 | 1e06402f2196da5270060d2dbe8c1aafafd3b387 | /venv/bin/chardetect | 75070d1851ae937bbe1b728865f55c1df3944359 | [] | no_license | billyggroves/myBlog | c26be905861d4a9f3b32ba8fce24103866311e7d | adc759d4a1a5389ab64cbfeb400898cad086c840 | refs/heads/master | 2022-12-15T10:40:20.626606 | 2019-02-26T01:11:20 | 2019-02-26T01:11:20 | 121,530,285 | 0 | 0 | null | 2022-12-08T00:55:40 | 2018-02-14T16:01:44 | Python | UTF-8 | Python | false | false | 257 | #!/Users/newuser/Desktop/myBlog/venv/bin/python3.6
# -*- coding: utf-8 -*-
import re
import sys
from chardet.cli.chardetect import main
if __name__ == '__main__':
sys.argv[0] = re.sub(r'(-script\.pyw?|\.exe)?$', '', sys.argv[0])
sys.exit(main())
| [
"billyggroves@gmail.com"
] | billyggroves@gmail.com | |
b512faa24699a6fe7dd43a3e59774c974454b97b | 2f79b7d886eb05ccdab5cad10fc85540e72384f2 | /HW2/Q2/Q2(b).py | 1cd2fe5200d96a8bc4179440dcb5c4a119bb6ac5 | [] | no_license | nrgdoublex/DataMiningHomework | 0276f537f2a5c6c2e9458aa56d782f358e74f84f | add7fd8d785e3f5d80724b7c7a1bb651d3705ea3 | refs/heads/master | 2021-01-01T20:50:09.368008 | 2017-08-01T01:06:11 | 2017-08-01T01:06:11 | 98,943,146 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 1,066 | py | import numpy as np
import scipy.stats as stats
import pandas as pd
from Stat_function import t_test
file_name = 'Q2.csv'
df = pd.read_csv(file_name
,skiprows = 1
,encoding = 'Latin-1'
,na_values = 'No Data')
pd.set_option('display.expand_frame_repr', False)
df['number'] = df['number'].apply(lambda x: int(x) if pd.notnull(x) else x)
df['percent'] = df['percent'].div(100)
# we only care about those county with > 100000 people
df = df.loc[df['number'] >= 100000,:]
df.index = xrange(df.shape[0])
#t-score for 95% confidence interval(Bonferroni correction)
z = stats.norm.ppf(1-0.05/(df.shape[0]-1))
max = np.max(df['percent'])
argmax = np.argmax(df['percent'])
for i in range(df.shape[0]):
if i == argmax: continue
compare = np.max(df.iloc[i,:]['percent'])
sample_size = (z**2)*max*(1-max)/((max-compare)**2)
print "{0},{1} VS {2},{3} => sample size = {4}".format(
df.iloc[argmax,:]['County'],df.iloc[argmax,:]['State'],
df.iloc[i,:]['County'],df.iloc[i,:]['State'], sample_size) | [
"nrgdoublex@gmail.com"
] | nrgdoublex@gmail.com |
f3d5dcd2e5f655280d986d7d5e685dfb3b524cc2 | 06604399c457d6ec05fa5d5ae458632e2606ec98 | /torch/utils/_sympy/functions.py | 3c78e1bebb50e8e34e979cab147e57e371f418bb | [
"BSD-3-Clause",
"BSD-2-Clause",
"LicenseRef-scancode-secret-labs-2011",
"LicenseRef-scancode-generic-cla",
"BSL-1.0",
"Apache-2.0"
] | permissive | yncxcw/pytorch | 6f262f7613caef4c2ce18c85662db9adc6a2a81a | a3b72ee354031004edd9b951d0efcdd4508fd578 | refs/heads/master | 2023-07-20T21:38:00.718093 | 2023-07-13T03:54:17 | 2023-07-13T03:54:17 | 234,432,318 | 0 | 0 | NOASSERTION | 2020-01-16T23:34:42 | 2020-01-16T23:34:41 | null | UTF-8 | Python | false | false | 5,787 | py | import sympy
from sympy.core.logic import fuzzy_and, fuzzy_or
__all__ = ["FloorDiv", "ModularIndexing", "CleanDiv", "CeilDiv", "LShift", "RShift"]
class FloorDiv(sympy.Function):
"""
We maintain this so that:
1. We can use divisibility guards to simplify FloorDiv(a, b) to a / b.
2. Printing out the expression is nicer (compared to say, representing a//b as (a - a % b) / b)
"""
nargs = (2,)
precedence = 50 # precedence of mul # noqa: F811
# Default return type for SymPy assumptions.
# https://docs.sympy.org/latest/guides/assumptions.html#implementing-assumptions-handlers
is_real = True
@property
def base(self):
return self.args[0]
@property
def divisor(self):
return self.args[1]
def _sympystr(self, printer):
base = printer.parenthesize(self.base, self.precedence)
divisor = printer.parenthesize(self.divisor, self.precedence)
return f"({base}//{divisor})"
# SymPy assumptions based on argument types.
def _eval_is_real(self):
return fuzzy_or([self.base.is_real, self.divisor.is_real])
def _eval_is_integer(self):
return fuzzy_and([self.base.is_integer, self.divisor.is_integer])
# Automatic evaluation.
# https://docs.sympy.org/latest/guides/custom-functions.html#best-practices-for-eval
@classmethod
def eval(cls, base, divisor):
def check_supported_type(x):
if (x.is_integer is False and x.is_real is False and x.is_complex) or x.is_Boolean:
raise TypeError(
f"unsupported operand type(s) for //: "
f"'{type(base).__name__}' and '{type(divisor).__name__}'"
f", expected integer or real")
check_supported_type(base)
check_supported_type(divisor)
# We don't provide the same error message as in Python because SymPy
# makes it difficult to check the types.
if divisor.is_zero:
raise ZeroDivisionError("division by zero")
if base.is_zero:
return sympy.S.Zero
if base.is_integer and divisor == 1:
return base
if base.is_real and divisor == 1:
return sympy.floor(base)
if isinstance(base, sympy.Integer) and isinstance(divisor, sympy.Integer):
return base // divisor
if isinstance(base, (sympy.Integer, sympy.Float)) and isinstance(divisor, (sympy.Integer, sympy.Float)):
return sympy.floor(base / divisor)
if isinstance(base, FloorDiv):
return FloorDiv(base.args[0], base.args[1] * divisor)
if isinstance(base, sympy.Add):
for a in base.args:
gcd = sympy.gcd(a, divisor)
if gcd == divisor:
return FloorDiv(base - a, divisor) + a / gcd
gcd = sympy.gcd(base, divisor)
if gcd != 1:
return FloorDiv(
sympy.simplify(base / gcd), sympy.simplify(divisor / gcd)
)
class ModularIndexing(sympy.Function):
"""
ModularIndexing(a, b, c) => (a // b) % c
"""
nargs = (3,)
is_integer = True
@classmethod
def eval(cls, base, divisor, modulus):
if base == 0 or modulus == 1:
return sympy.Integer(0)
if (
isinstance(base, sympy.Integer)
and isinstance(divisor, sympy.Integer)
and isinstance(modulus, sympy.Integer)
):
return (base // divisor) % modulus
if divisor != 1:
gcd = sympy.gcd(base, divisor)
if gcd != 1:
return ModularIndexing(
sympy.simplify(base / gcd), sympy.simplify(divisor / gcd), modulus
)
if isinstance(base, sympy.Add):
new_terms = []
all_positive = True
for term in base.args:
if sympy.gcd(term, modulus * divisor) != modulus * divisor:
if (isinstance(term, sympy.Integer) and term < 0) or (
isinstance(term, sympy.Mul)
and isinstance(term.args[0], sympy.Integer)
and term.args[0] < 0
):
# workaround for https://github.com/openai/triton/issues/619,
# if there are negative terms, // produces wrong result
# TODO if https://github.com/openai/triton/issues/619 is fixed
# this optimization would become valid
all_positive = False
break
else:
new_terms.append(term)
if len(new_terms) != len(base.args) and all_positive:
return ModularIndexing(sum(new_terms), divisor, modulus)
if isinstance(base, FloorDiv):
return ModularIndexing(base.args[0], base.args[1] * divisor, modulus)
class CleanDiv(FloorDiv):
"""
Div where we can assume no rounding.
This is to enable future optimizations.
"""
pass
class CeilDiv(sympy.Function):
"""
Div used in indexing that rounds up.
"""
is_integer = True
def __new__(cls, base, divisor):
if sympy.gcd(base, divisor) == divisor:
return CleanDiv(base, divisor)
else:
return FloorDiv(base + (divisor - 1), divisor)
class LShift(sympy.Function):
@classmethod
def eval(cls, base, shift):
if shift < 0:
raise ValueError('negative shift count')
return base * 2 ** shift
class RShift(sympy.Function):
@classmethod
def eval(cls, base, shift):
if shift < 0:
raise ValueError('negative shift count')
return base // 2 ** shift
| [
"pytorchmergebot@users.noreply.github.com"
] | pytorchmergebot@users.noreply.github.com |
125483cf11de2cfc713c9a895380f931f1ec938b | f9b45fd409384de3c69d68c62c0f21d9196e6927 | /python_tools/FelixUtilities.py | da0b1769ef7a4288b8afdbea8bf0248b0f87b580 | [
"MIT"
] | permissive | ultimatezen/felix | 04799eb666b54e2eeea961dc983cf3721f5182de | 5a7ad298ca4dcd5f1def05c60ae3c84519ec54c4 | refs/heads/master | 2020-04-08T14:21:02.872525 | 2016-02-16T01:18:11 | 2016-02-16T01:18:11 | 159,433,144 | 0 | 1 | null | null | null | null | UTF-8 | Python | false | false | 5,611 | py | # coding: UTF8
"""
COM sever for utilities for Felix
"""
import os
import shutil
from datetime import date
import subprocess
# COM stuff
import win32com.server.register
import loc
from loc import get_local_app_data_folder, module_path
from loc import get_prefs, set_prefs
import utils
from utils import debug, log_err
from FelixMemoryServes import update_checker_felix
__version__ = "1.0"
__author__ = "Ryan Ginstrom"
__license__ = "MIT"
__progname__ = "Felix Utilities"
CHECK_INTERVAL = update_checker_felix.CHECK_INTERVAL
class FelixUtilities(object):
"""Utilities for Felix"""
_public_methods_ = """EnsureHtml CopyHtml LoadProp SaveProp
CheckUpdates CheckUpdatesRequested
HasUpdate CurrentVersion NewVersion""".split()
_public_attrs_ = []
_reg_progid_ = "Felix.Utilities"
# import pythoncom
# print pythoncom.CreateGuid()
_reg_clsid_ = "{3D368688-B9E8-4BFE-ADF6-22F1CBA8845B}"
def __init__(self):
debug("Felix.Utilities")
def HasUpdate(self):
return False
def CurrentVersion(self):
return "1"
def NewVersion(self):
return "1"
def LoadProp(self, key):
"""
Loads a key-value pair from the properties.
:param key: The key to retrieve
:return: The value corresponding to the key
"""
val = get_prefs().get(key, None)
debug("Loaded prop: %s=%s" % (key, val))
return val
def SaveProp(self, key, val):
"""
Saves a property as a key-value pair
:param key: The key
:param val: The value
"""
prefs = get_prefs()
prefs[key] = val
set_prefs(prefs)
debug("Saved prop: %s=%s" % (key, val))
def CopyHtml(self, source, dest):
"""
Forces copy of HTML files to local app folder.
:param source: The location of the backup HTML files
:param dest: Copy the HTML files here (local app data)
"""
debug(u"Forcing copy of HTML files to local app folder")
debug(u"source: %s" % source)
debug(u"destination: %s" % dest)
shutil.copyfile(source, dest)
def EnsureHtml(self):
"""
Ensure that HTML files are in local app data. If they are not,
copies them over.
"""
debug("Checking for HTML files in local app folder")
data_folder = get_local_app_data_folder()
html_folder = os.path.join(data_folder, u"html")
if not os.path.isdir(html_folder):
debug(" ... Files not found. Copying.")
head, tail = os.path.split(module_path())
src = os.path.join(head, u"html")
dst = html_folder
shutil.copytree(src, dst)
return True
return False
def CheckUpdates(self, language):
"""
Check Felix server for updates.
:param language: The UI language for showing the update dialog.
"""
try:
debug(u"Checking for updates. Language is %s" % language)
loc.language.change_language(language)
prefs = get_prefs()
check_permission(prefs, date.today())
set_prefs(prefs)
except:
LOGGER.exception("Failed to check updates")
def CheckUpdatesRequested(self, language):
"""
If the user checks for updates, we don't have to see whether the user
wants to check automatically.
:param language: The UI language for showing dialogs.
"""
try:
debug(u"Checking for updates at user request. Language is %s" % language)
loc.language.change_language(language)
check_updates()
prefs = get_prefs()
prefs["last_update_check"] = date.today()
set_prefs(prefs)
except:
LOGGER.exception("Failed to check updates")
def get_command():
modpath = loc.module_path()
if " " in modpath:
format = '"%s"'
else:
format = '%s'
if loc.we_are_frozen():
return [format % os.path.join(modpath, "CheckUpdates.exe")]
else:
return ["pythonw.exe",
format % os.path.join(modpath, "update_checker_felix.py")]
def check_updates():
cmd = get_command()
subprocess.Popen(cmd,
shell=False)
def ask_updates():
cmd = get_command()
cmd.append("ask")
subprocess.Popen(cmd,
shell=False)
def check_permission(prefs, today):
if not prefs["ask_about_updates"] and not prefs["check_updates"]:
return
last_check = prefs["last_update_check"]
if not last_check:
if prefs["ask_about_updates"]:
ask_updates()
else:
check_updates()
prefs["last_update_check"] = today
return
else:
diff = today - last_check
if diff.days >= CHECK_INTERVAL:
if prefs["ask_about_updates"]:
ask_updates()
else:
check_updates()
prefs["last_update_check"] = today
def reg():
"""Register COM servers"""
debug("Registering COM servers")
win32com.server.register.UseCommandLine(FelixUtilities)
utils.determine_redirect("felix_utilities.log")
# Add code so that when this script is run by
# Python.exe, it self-registers.
if __name__ == '__main__':
reg()
| [
"software@ginstrom.com"
] | software@ginstrom.com |
d4d00147e745d9e951765b3fc1fd6c50c016f113 | 3c9011b549dd06b6344c6235ed22b9dd483365d1 | /Agenda/contatos/migrations/0003_contato_foto.py | 097090b85a3f8c6636e29e52edc33d1acc2ee0e7 | [] | no_license | joaoo-vittor/estudo-python | 1411f4c3620bbc5f6b7c674a096cae8f90f0db8d | 5562d823dd574d7df49fddca87a1fbd319356969 | refs/heads/master | 2023-05-31T17:59:16.752835 | 2021-06-25T04:54:56 | 2021-06-25T04:54:56 | 292,372,669 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 402 | py | # Generated by Django 3.2 on 2021-05-16 01:41
from django.db import migrations, models
class Migration(migrations.Migration):
dependencies = [
('contatos', '0002_contato_mostrar'),
]
operations = [
migrations.AddField(
model_name='contato',
name='foto',
field=models.ImageField(blank=True, upload_to='fotos/%Y/%m'),
),
]
| [
"joaoo.vittor007@gmail.com"
] | joaoo.vittor007@gmail.com |
e8012164e3c46fc9c9870c0102c227d7f0d59508 | f09b27136435cb298a55514adc5b1ccbe8badb28 | /lab 13/zad 1.py | 0ebf9023e993c963ae08d5d854e8ba1bdc8c1439 | [] | no_license | Roman43407/Sem-2- | 890db6446254266d9866992bb804af906bdd87df | de3a3eb9de1477cca29fa118264729cbe90f93ff | refs/heads/main | 2023-04-26T19:58:39.122440 | 2021-05-31T12:10:50 | 2021-05-31T12:10:50 | 372,455,729 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 1,036 | py | class Car:
def __init__(self, brand, age, color, capacity, fuelType):
self.brand = brand
self.age = age
self.color = color
self.capacity = capacity
self.fuelType = fuelType
def Refueling(self):
print("Zatankowano: {} {} litrów ".format(self.fuelType, self.capacity))
def CarInfo(self):
print("Marka:", self.brand)
print("Wiek:", self.age)
print("Kolor:", self.color)
print("Pojemność baku:", self.capacity)
print("Rodzaj paliwa:", self.fuelType)
def ChangeColor(self, newColor):
print("Zmieniono kolor z {} na {}".format(self.color, newColor))
def Velocity(self):
if (self.color == "red"):
print("Bardzo szybki")
else:
print("Nie tak szybki jak czerwony")
def CallService(self):
print("Jesteś 10 w kolejce")
opelek = Car("Volvo", 10, "Gold", 50, "diesel")
opelek.Refueling()
opelek.ChangeColor("Red")
opelek.CarInfo()
opelek.Velocity()
opelek.CallService() | [
"ya.ignis@gmail.com"
] | ya.ignis@gmail.com |
7fc024f18bdc5289a4cad605dbc8a2f6fa792e74 | c4c159a21d2f1ea0d7dfaa965aeff01c8ef70dce | /flask/flaskenv/Lib/site-packages/tensorflow/contrib/eager/python/tfe.py | c441ab87be7e0aebadefe92023f89bfd67ff471e | [] | no_license | AhsonAslam/webapi | 54cf7466aac4685da1105f9fb84c686e38f92121 | 1b2bfa4614e7afdc57c9210b0674506ea70b20b5 | refs/heads/master | 2020-07-27T06:05:36.057953 | 2019-09-17T06:35:33 | 2019-09-17T06:35:33 | 208,895,450 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 129 | py | version https://git-lfs.github.com/spec/v1
oid sha256:9b7bd976495c4645582fde2d7fcf488a311648b43813cff249462fccfa19224a
size 5928
| [
"github@cuba12345"
] | github@cuba12345 |
4353deb50a51a18cfc392b8d5fada6467c849fe1 | 91d1a6968b90d9d461e9a2ece12b465486e3ccc2 | /worklink_write_3/domain_associate.py | 622afc66bafaf4062a1575d617c77e954bc7ee2e | [] | no_license | lxtxl/aws_cli | c31fc994c9a4296d6bac851e680d5adbf7e93481 | aaf35df1b7509abf5601d3f09ff1fece482facda | refs/heads/master | 2023-02-06T09:00:33.088379 | 2020-12-27T13:38:45 | 2020-12-27T13:38:45 | 318,686,394 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 1,279 | py | #!/usr/bin/python
# -*- codding: utf-8 -*-
import os
import sys
sys.path.append(os.path.dirname(os.path.abspath(os.path.dirname(__file__))))
from common.execute_command import write_three_parameter
# url : https://awscli.amazonaws.com/v2/documentation/api/latest/reference/worklink/associate-domain.html
if __name__ == '__main__':
"""
describe-domain : https://awscli.amazonaws.com/v2/documentation/api/latest/reference/worklink/describe-domain.html
disassociate-domain : https://awscli.amazonaws.com/v2/documentation/api/latest/reference/worklink/disassociate-domain.html
list-domains : https://awscli.amazonaws.com/v2/documentation/api/latest/reference/worklink/list-domains.html
"""
parameter_display_string = """
# fleet-arn : The Amazon Resource Name (ARN) of the fleet.
# domain-name : The fully qualified domain name (FQDN).
# acm-certificate-arn : The ARN of an issued ACM certificate that is valid for the domain being associated.
"""
add_option_dict = {}
add_option_dict["parameter_display_string"] = parameter_display_string
# ex: add_option_dict["no_value_parameter_list"] = "--single-parameter"
write_three_parameter("worklink", "associate-domain", "fleet-arn", "domain-name", "acm-certificate-arn", add_option_dict)
| [
"hcseo77@gmail.com"
] | hcseo77@gmail.com |
9adc81d26ca9708c7ee07b95c8795d117a6c05e9 | 987a82368d3a15b618ff999f28dc16b89e50f675 | /plaso/parsers/winreg_plugins/shutdown.py | f14cb1abcb7795e44c6e63f3728e75b94987c6ff | [
"Apache-2.0"
] | permissive | arunthirukkonda/plaso | 185b30ab4ec90fcc2d280b3c89c521c9eef7b7ab | 846fc2fce715e1f78b11f375f6fe4e11b5c284ba | refs/heads/master | 2021-08-30T15:21:12.267584 | 2017-12-18T12:33:08 | 2017-12-18T12:33:08 | null | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 2,893 | py | # -*- coding: utf-8 -*-
"""Windows Registry plugin for parsing the last shutdown time of a system."""
from __future__ import unicode_literals
import construct
from dfdatetime import filetime as dfdatetime_filetime
from dfdatetime import semantic_time as dfdatetime_semantic_time
from plaso.containers import events
from plaso.containers import time_events
from plaso.lib import definitions
from plaso.parsers import winreg
from plaso.parsers.winreg_plugins import interface
__author__ = 'Preston Miller, dpmforensics.com, github.com/prmiller91'
class ShutdownWindowsRegistryEventData(events.EventData):
"""Shutdown Windows Registry event data.
Attributes:
key_path (str): Windows Registry key path.
value_name (str): name of the Windows Registry value.
"""
DATA_TYPE = 'windows:registry:shutdown'
def __init__(self):
"""Initializes event data."""
super(ShutdownWindowsRegistryEventData, self).__init__(
data_type=self.DATA_TYPE)
self.key_path = None
self.value_name = None
class ShutdownPlugin(interface.WindowsRegistryPlugin):
"""Windows Registry plugin for parsing the last shutdown time of a system."""
NAME = 'windows_shutdown'
DESCRIPTION = 'Parser for ShutdownTime Registry value.'
FILTERS = frozenset([
interface.WindowsRegistryKeyPathFilter(
'HKEY_LOCAL_MACHINE\\System\\CurrentControlSet\\Control\\Windows')])
_UINT64_STRUCT = construct.ULInt64('value')
def ExtractEvents(self, parser_mediator, registry_key, **kwargs):
"""Extracts events from a ShutdownTime Windows Registry value.
Args:
parser_mediator (ParserMediator): mediates interactions between parsers
and other components, such as storage and dfvfs.
registry_key (dfwinreg.WinRegistryKey): Windows Registry key.
"""
shutdown_value = registry_key.GetValueByName('ShutdownTime')
if not shutdown_value:
return
# Directly parse the Windows Registry value data in case it is defined
# as binary data.
try:
timestamp = self._UINT64_STRUCT.parse(shutdown_value.data)
except construct.FieldError as exception:
timestamp = None
parser_mediator.ProduceExtractionError(
'unable to determine shutdown timestamp with error: {0!s}'.format(
exception))
if not timestamp:
date_time = dfdatetime_semantic_time.SemanticTime('Not set')
else:
date_time = dfdatetime_filetime.Filetime(timestamp=timestamp)
event_data = ShutdownWindowsRegistryEventData()
event_data.key_path = registry_key.path
event_data.offset = shutdown_value.offset
event_data.value_name = shutdown_value.name
event = time_events.DateTimeValuesEvent(
date_time, definitions.TIME_DESCRIPTION_LAST_SHUTDOWN)
parser_mediator.ProduceEventWithEventData(event, event_data)
winreg.WinRegistryParser.RegisterPlugin(ShutdownPlugin)
| [
"joachim.metz@gmail.com"
] | joachim.metz@gmail.com |
7e323ca57a44b81a67fc770dd4181fb866231c6b | 0c7ffe86e5060cf7cfbc6dfc3ec5eb4367c20962 | /TMB_files/learn-to-program-the-fundamentals/Week 6/a3.py | f1a1dbdaa6b4b419ea18f8d629ddac3f99395824 | [] | no_license | BreslauerTodd/telework_data_and_code | 049e2d4fb1a4479078f6edf14434034824dc6f40 | 38c39fbab481b48b6e2a15f2a130329725d2e07e | refs/heads/master | 2022-11-20T12:21:44.952847 | 2020-07-20T18:20:21 | 2020-07-20T18:20:21 | 279,941,119 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 4,920 | py | def is_valid_word(wordlist, word):
""" (list of str, str) -> bool
Return True if and only if word is an element of wordlist.
>>> is_valid_word(['ANT', 'BOX', 'SOB', 'TO'], 'TO')
True
"""
if word in wordlist:
inList = True
else:
inList = False
return inList
def make_str_from_row(board, row_index):
""" (list of list of str, int) -> str
Return the characters from the row of the board with index row_index
as a single string.
>>> make_str_from_row([['A', 'N', 'T', 'T'], ['X', 'S', 'O', 'B']], 0)
'ANTT'
"""
rowsum = ''
for i in board[row_index]:
rowsum = rowsum + i
return rowsum
def make_str_from_column(board, column_index):
""" (list of list of str, int) -> str
Return the characters from the column of the board with index column_index
as a single string.
>>> make_str_from_column([['A', 'N', 'T', 'T'], ['X', 'S', 'O', 'B']], 1)
'NS'
"""
colsum = ''
for i in range(len(board)):
colsum = colsum + board[i][column_index]
return colsum
def board_contains_word_in_row(board, word):
""" (list of list of str, str) -> bool
Return True if and only if one or more of the rows of the board contains
word.
Precondition: board has at least one row and one column, and word is a
valid word.
>>> board_contains_word_in_row([['A', 'N', 'T', 'T'], ['X', 'S', 'O', 'B']], 'SOB')
True
"""
for row_index in range(len(board)):
if word in make_str_from_row(board, row_index):
return True
return False
def board_contains_word_in_column(board, word):
""" (list of list of str, str) -> bool
Return True if and only if one or more of the columns of the board
contains word.
Precondition: board has at least one row and one column, and word is a
valid word.
>>> board_contains_word_in_column([['A', 'N', 'T', 'T'], ['X', 'S', 'O', 'B']], 'NO')
False
"""
for column_index in range(len(board[0])):
if word in make_str_from_column(board, column_index):
return True
return False
def board_contains_word(board, word):
""" (list of list of str, str) -> bool
Return True if and only if word appears in board.
Precondition: board has at least one row and one column.
>>> board_contains_word([['A', 'N', 'T', 'T'], ['X', 'S', 'O', 'B']], 'ANT')
True
"""
if board_contains_word_in_row(board, word):
return True
elif board_contains_word_in_column(board, word):
return True
else:
return False
def word_score(word):
""" (str) -> int
Return the point value the word earns.
Word length: < 3: 0 points
3-6: 1 point per character for all characters in word
7-9: 2 points per character for all characters in word
10+: 3 points per character for all characters in word
>>> word_score('DRUDGERY')
16
"""
word_length = len(word)
if word_length < 3:
return 0
elif word_length < 7:
return word_length
elif word_length < 10:
return word_length * 2
else:
return word_length * 3
def update_score(player_info, word):
""" ([str, int] list, str) -> NoneType
player_info is a list with the player's name and score. Update player_info
by adding the point value word earns to the player's score.
>>> update_score(['Jonathan', 4], 'ANT')
"""
add_score = word_score(word)
new_score = player_info[1] + add_score
player_info[1] = new_score
def num_words_on_board(board, words):
""" (list of list of str, list of str) -> int
Return how many words appear on board.
>>> num_words_on_board([['A', 'N', 'T', 'T'], ['X', 'S', 'O', 'B']], ['ANT', 'BOX', 'SOB', 'TO'])
3
"""
count = 0
for i in words:
if board_contains_word(board,i):
count = count +1
return count
def read_words(words_file):
""" (file open for reading) -> list of str
Return a list of all words (with newlines removed) from open file
words_file.
Precondition: Each line of the file contains a word in uppercase characters
from the standard English alphabet.
"""
words = []
for line in words_file:
words.append(line.rstrip('\n'))
return words
def read_board(board_file):
""" (file open for reading) -> list of list of str
Return a board read from open file board_file. The board file will contain
one row of the board per line. Newlines are not included in the board.
"""
board = []
for line in board_file:
baseline = line.rstrip('\n')
lineList = []
for i in range(0,len(baseline)):
lineList.append(baseline[i])
board.append(lineList)
return board
| [
"noreply@github.com"
] | noreply@github.com |
d5aa6095ffe361c6c24f7e7ace9e878dcd34a356 | 8a452b71e3942d762fc2e86e49e72eac951b7eba | /leetcode/editor/en/[1065]Index Pairs of a String.py | 9596550ca2c48f4cb14e3df379385e19b37fe19c | [] | no_license | tainenko/Leetcode2019 | 7bea3a6545f97c678a176b93d6622f1f87e0f0df | 8595b04cf5a024c2cd8a97f750d890a818568401 | refs/heads/master | 2023-08-02T18:10:59.542292 | 2023-08-02T17:25:49 | 2023-08-02T17:25:49 | 178,761,023 | 5 | 0 | null | 2019-08-27T10:59:12 | 2019-04-01T01:04:21 | JavaScript | UTF-8 | Python | false | false | 1,478 | py | # Given a string text and an array of strings words, return an array of all
# index pairs [i, j] so that the substring text[i...j] is in words.
#
# Return the pairs [i, j] in sorted order (i.e., sort them by their first
# coordinate, and in case of ties sort them by their second coordinate).
#
#
# Example 1:
#
#
# Input: text = "thestoryofleetcodeandme", words = ["story","fleet","leetcode"]
# Output: [[3,7],[9,13],[10,17]]
#
#
# Example 2:
#
#
# Input: text = "ababa", words = ["aba","ab"]
# Output: [[0,1],[0,2],[2,3],[2,4]]
# Explanation: Notice that matches can overlap, see "aba" is found in [0,2] and
# [2,4].
#
#
#
# Constraints:
#
#
# 1 <= text.length <= 100
# 1 <= words.length <= 20
# 1 <= words[i].length <= 50
# text and words[i] consist of lowercase English letters.
# All the strings of words are unique.
#
# Related Topics Array String Trie Sorting 👍 203 👎 73
# leetcode submit region begin(Prohibit modification and deletion)
class Solution:
def indexPairs(self, text: str, words: List[str]) -> List[List[int]]:
res = []
words.sort(key=lambda x: len(x))
for i in range(len(text)):
for word in words:
if i + len(word) > len(text):
continue
if word == text[i:i + len(word)]:
res.append([i, i + len(word) - 1])
return res
# leetcode submit region end(Prohibit modification and deletion)
| [
"31752048+tainenko@users.noreply.github.com"
] | 31752048+tainenko@users.noreply.github.com |
65dd6e05846806bcfc8c999115ef687d45c97af7 | 3160c1b057319c66e3fe30af52ddabc5555b347a | /venv/Scripts/rst2pseudoxml.py | 67178034fb66411b820e6bc1291d00d95015ebc9 | [] | no_license | bopopescu/TLDR_AI | acfb35f0a0677e03f48b03805ff44fd78303f126 | d468e599fc13828f6e754e2d1d99b6a3e3c1be21 | refs/heads/master | 2022-11-25T17:40:08.513723 | 2019-04-09T06:52:06 | 2019-04-09T06:52:06 | 282,555,526 | 0 | 0 | null | 2020-07-26T01:41:57 | 2020-07-26T01:41:56 | null | UTF-8 | Python | false | false | 647 | py | #!C:\Users\aaron\PycharmProjects\TLDR_AI\venv\Scripts\python.exe
# $Id: rst2pseudoxml.py 4564 2006-05-21 20:44:42Z wiemann $
# Author: David Goodger <goodger@python.org>
# Copyright: This module has been placed in the public domain.
"""
A minimal front end to the Docutils Publisher, producing pseudo-XML.
"""
try:
import locale
locale.setlocale(locale.LC_ALL, '')
except:
pass
from docutils.core import publish_cmdline, default_description
description = ('Generates pseudo-XML from standalone reStructuredText '
'sources (for testing purposes). ' + default_description)
publish_cmdline(description=description)
| [
"aaron157@bellsouth.net"
] | aaron157@bellsouth.net |
73ddcb45b10df0f0c85f275e7a8190f7bd549bc2 | 677df89f5395f19b87b9f868da8c96f2c5ffbc5e | /code/pycs/client.py | 6afd66ec16bd58036a0f37d4f94cf7133e878ada | [] | no_license | vinnyzhaowenyu/mdnote | eefdffd8c526f2aa1893de3ec3891cc7452ea395 | d1f0835e31f46a6eeb9fe274778fdbac84f27570 | refs/heads/master | 2021-01-15T13:25:07.011643 | 2017-11-04T07:48:57 | 2017-11-04T07:48:57 | 92,654,000 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 535 | py | #!/usr/bin/env python
import socket
import os
ss=socket.socket(socket.AF_INET,socket.SOCK_STREAM)
ss.connect(('127.0.0.1',8123))
#f=open('aa','wb')
ss.sendall('hello serverdddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd')
os.system('sleep 1')
ss.send('EOF')
data=ss.recv(1024)
print "server dafu %s"%data
ss.close()
| [
"admin@zhaowenyu.com"
] | admin@zhaowenyu.com |
717fdc2d540f242841bacdd7c2381591a288cc15 | 464f40c696af1028aee439c31938137f43f482f1 | /api/serializers.py | 9fda532152aa3dd7233279f1ccf9660a67302a36 | [] | no_license | MarkWengSTR/drf_todo_practice | d0062bbbc56fd9833a24b9439cd2e9bcb31b4a80 | 431c3173178683f50e5dd9df6eab038dc35452c6 | refs/heads/master | 2023-04-07T05:14:02.815232 | 2021-04-21T03:49:56 | 2021-04-21T03:49:56 | 360,025,275 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 184 | py | from rest_framework import serializers
from .models import Task
class TaskSerializerser(serializers.ModelSerializer):
class Meta:
model = Task
fields = '__all__'
| [
"bskin0330@gmail.com"
] | bskin0330@gmail.com |
44cea545571a791e10617e28cfcb824365481f53 | c5d4fe5f7d01edf095541171fe2610ff62da5104 | /Mission_to_Mars/app.py | 5875d052c0755302a450c5da2bd7dba8ff3fd363 | [] | no_license | thierzon/web-scraping-challenge | 37c24477584415a53a59dd6d59026e432409feaa | cc961977e618b535464e394adf4edc2f569dfbc1 | refs/heads/main | 2023-04-16T03:31:57.133064 | 2021-04-16T13:11:26 | 2021-04-16T13:11:26 | 347,530,493 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 891 | py | from flask import Flask, render_template, redirect
import pymongo
import scrape_mars
# Create an instance of Flask
app = Flask(__name__)
# Use PyMongo to establish Mongo connection
mongo = pymongo.MongoClient("mongodb://localhost:27017/mars_app")
# Route to render index.html template using data from Mongo
@app.route("/")
def home():
# Find one record of data from the mongo database
mars = mongo.db.mars.find_one()
# Return template and data
return render_template("index.html", mars=mars)
# Route that will trigger the scrape function
@app.route("/scrape")
def scrape():
# Run the scrape function
mars_data = scrape_mars.scrape()
# Update the Mongo database using update and upsert=True
mongo.db.mars.update({}, mars_data, upsert=True)
# Redirect back to home page
return redirect("/")
if __name__ == "__main__":
app.run(debug=True)
| [
"thierryzonneveld@gmail.com"
] | thierryzonneveld@gmail.com |
c3a805e35fb8e1e5293462c79a7b65bffc9c28a6 | 87f7d232723117ad4c72574898bda957d7adfe4d | /Unfinished Loop - Bug Fixing.py | 0a796fecfa5fa1b4dc2bd9f430211a66291eff95 | [] | no_license | Taras198700/Home_Work_S_S | 876f53c615fef440ef58b7d39cf896e1e778d5de | 081a7b17a67f803556027162b0058eccc1a27788 | refs/heads/master | 2022-12-16T20:30:11.888892 | 2020-09-09T18:13:16 | 2020-09-09T18:13:16 | 278,019,575 | 0 | 1 | null | null | null | null | UTF-8 | Python | false | false | 113 | py | def create_array(n):
res = []
i = 1
while i <= n:
res.append(i)
i += 1
return res | [
"homitsky_taras@ukr.net"
] | homitsky_taras@ukr.net |
1cf5fc50d95b24370dad7de31e70a0868c09d967 | 0af9ff3662729cf1f949e91b8d8862ff739e7b5f | /números primos.py | 24d5729c603ea4e67a9383127801a93e2941912d | [] | no_license | lucasjukr94/JavaProjects | e979726564e6ed47974931c3e5e04bbd06901e17 | 9c7315402c74458de054be21e6986c710ddfd5ef | refs/heads/master | 2021-09-15T20:27:42.619406 | 2018-06-10T14:54:44 | 2018-06-10T14:54:44 | 125,927,047 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 3,688 | py | '''
╔══════════════════════════════════════════════════════════════════════════════╗
║ Instituição : Faculdade de Tecnologia de São Paulo ║
║ Departamento : Tecnologia da Informação ║
║ Curso : Análise e Desenvolvimento de Sistemas ║
║ Autor : Lucio Nunes de Lira ║
╠══════════════════════════════════════════════════════════════════════════════╣
║ Evento : Palestra (Números Primos: Conceitos & Algoritmos) ║
║ Programa : Implementações de algoritmos para cálculo de números primos ║
║ Linguagem : Python 3 ║
║ Compilador : CPython (3.6.4) ║
║ Versão : A (Rev. 0) ║
╚══════════════════════════════════════════════════════════════════════════════╝
'''
# Importação das funções necessárias
from math import sqrt, floor
# Conta os divisores entre 1 e n.
def primo_1(n):
divisores = 0
for i in range(1, n + 1):
if n % i == 0:
divisores += 1
if divisores == 2:
return True
else:
return False
# Conta os divisores entre 1 e n.
def primo_2(n):
divisores = 0
for i in range(1, n + 1):
if n % i == 0:
divisores += 1
return divisores == 2
# Conta os divisores entre 2 e n-1.
def primo_3(n):
divisores = 0
for i in range(2, n):
if n % i == 0:
divisores += 1
return n > 1 and divisores == 0
# Termina as tentativas de divisão na
# primeira ocorrência de uma divisão
# com resto zero.
def primo_4(n):
if n == 1: return False
divisor = 2
while n % divisor != 0:
divisor += 1
return divisor == n
# Testa apenas divisores ímpares.
def primo_5(n):
if n == 1: return False
if n % 2 == 0: return n == 2
divisor = 3
while n % divisor != 0:
divisor += 2
return divisor == n
# Testa divisores até a metade de n.
def primo_6(n):
if n % 2 == 0: return n == 2
divisor = 3
metade = n // 2
while divisor <= metade and n % divisor != 0:
divisor += 2
return n > 1 and divisor > metade
# Testa divisores até a raiz de n.
def primo_7(n):
if n % 2 == 0: return n == 2
divisor = 3
raiz = floor(sqrt(n))
while divisor <= raiz and n % divisor != 0:
divisor += 2
return n > 1 and divisor > raiz
# Marca os múltiplos de n com zero.
# Obs.: a partir do segundo múltiplo.
def marca_multiplos(n, lista, lim):
for i in range(n * 2, lim + 1, n):
lista[i] = 0
# Devolve uma nova lista sem zeros.
def filtra(lista):
return [n for n in lista if n != 0]
# Algoritmo "Crivo de Eratóstenes"
# lim = valor do limite
def crivo(lim):
lista = [n for n in range(lim + 1)]
lista[1] = 0
raiz = floor(sqrt(lim))
for i in range(2, raiz + 1):
marca_multiplos(i, lista, lim)
return filtra(lista)
| [
"noreply@github.com"
] | noreply@github.com |
24415f583157830a26bbdd97d7d14286cd332701 | 8b198614532b0fd6bc9861f4427299bcc2764c1e | /wsgi.py | 0b182b5796ca105033ef509ec1809d1c622e9bd6 | [] | no_license | summerandwinter/wxapppython | e9b0321b7b7c844eee39baf588b94c6a88f96351 | b345447260fcad01903771a38fbc66b7494c3389 | refs/heads/master | 2021-01-20T01:51:19.589571 | 2018-01-29T10:01:36 | 2018-01-29T10:01:36 | 89,334,998 | 27 | 8 | null | null | null | null | UTF-8 | Python | false | false | 1,144 | py | # coding: utf-8
from gevent import monkey
monkey.patch_all()
import os
import configparser
# 设置 Django 项目配置文件
os.environ.setdefault("DJANGO_SETTINGS_MODULE", "settings")
import leancloud
from gevent.pywsgi import WSGIServer
from cloud import engine
cf = configparser.ConfigParser()
cf.read("config.conf")
os.environ["WX_APP_ID"] = cf.get("wx", "app_id")
os.environ["WX_APP_SECRET"] = cf.get("wx", "app_secret")
os.environ["WXA_APP_ID"] = cf.get("wxa", "app_id")
os.environ["WXA_APP_SECRET"] = cf.get("wxa", "app_secret")
os.environ["QINIU_ACCESS_KEY"] = cf.get("qiniu", "access_key")
os.environ["QINIU_SECRET_KEY"] = cf.get("qiniu", "secret_key")
os.environ["QINIU_ACCESS_URL"] = cf.get("qiniu", "access_url")
#import logging
#logging.basicConfig(level=logging.DEBUG)
APP_ID = os.environ['LC_APP_ID']
MASTER_KEY = os.environ['LC_APP_MASTER_KEY']
PORT = int(os.environ['LC_APP_PORT'])
leancloud.init(APP_ID, master_key=MASTER_KEY)
application = engine
if __name__ == '__main__':
# 只在本地开发环境执行的代码
server = WSGIServer(('localhost', PORT), application)
server.serve_forever()
| [
"xiaopihailaotou@gmail.com"
] | xiaopihailaotou@gmail.com |
5a2126d0acf87e3690274f9ba267477d89a89241 | c46b327faa127e0803fbd40521a769ac54d8e2f9 | /code/word2vec.py | 03c0bffc124e52ab7b754b802ecc48b44382e853 | [] | no_license | lizy10/Modified-and-Annotated-Code-of--An--Unsupervised-Neural-Attention-Model-for-Aspect-Extraction-- | f74abd8cc6fd90322dc3876ab8eb250d55f60467 | 0b40ec8897398e214d6a6c7a8936cf34686ea233 | refs/heads/master | 2021-09-15T01:03:07.254804 | 2018-05-23T10:44:37 | 2018-05-23T10:44:37 | null | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 913 | py | import gensim
import codecs
# 利用生成器读取数据可以避免等待数据读取的过程,速度更快.
class MySentences(object):
def __init__(self, filename):
self.filename = filename
def __iter__(self):
for line in codecs.open(self.filename, 'r', 'utf-8'):
yield line.split()
# 将数据转换为向量形式,基于word2vec算法.
def main(domain):
source = '../preprocessed_data/%s/train.txt' % (domain)
model_file = '../preprocessed_data/%s/w2v_embedding' % (domain)
sentences = MySentences(source)
model = gensim.models.Word2Vec(sentences, size=200, window=5, min_count=10, workers=4)
# try to print something to show the effectiveness of word2vec.
print('model.mv:',model.wv['like'])
model.save(model_file)
# 分别转换restaurant和beer数据集.
print ('Pre-training word embeddings ...')
main('restaurant')
main('beer')
| [
"woshihaodayikeshu@163.com"
] | woshihaodayikeshu@163.com |
df90f052aa06361c8bda15da8154b401621d1aed | 50eb70ca3154f9a84f89a883718a87df63f78e80 | /mpServerProcess.py | aa49a8172d4fc14116a7062f5109b6281c5f3414 | [] | no_license | Rocia/Learning-multiprocessing | b0e5bd8b70afa7c3dce19feb9b188ede9869323f | 0b4dd7d1d7bf4831069b53adb1e8d970a24261d9 | refs/heads/master | 2020-06-28T01:22:44.012701 | 2017-07-14T07:32:01 | 2017-07-14T07:32:01 | 97,080,749 | 1 | 1 | null | null | null | null | UTF-8 | Python | false | false | 561 | py | from multiprocessing import Process, Manager
def A(d, l):
d[1] = '2'
d['2'] = 3
d[3.14] = None
d['alpha'] = None
d['beta'] = 'gamma'
l.reverse()
if __name__ == '__main__':
manager = Manager()
DICTIONARY = manager.dict()
LIST = manager.list(range(15))
process = Process(target=A, args=(DICTIONARY,LIST))
process.start()
process.join()
print (DICTIONARY)
print (LIST)
'''
OUTPUT:-
{1: '2', '2': 3, 3.14: None, 'alpha': None, 'beta': 'gamma'}
[14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0]
''' | [
"rocia.fernandes@gmail.com"
] | rocia.fernandes@gmail.com |
26e63b09b2482b2c6ad1c3618519251ceac2ed9a | 4dc5f671d6f8e1d22a90b8244c58473aa440236b | /definitions.py | dc126bb2175829caf0262328d5499902f8a6223e | [] | no_license | nithinmk26/PythonStuff | 00f6c5094ee09aedde0691c05ba6c8ca5c5640a5 | d54927147bebb94e14529c77f414fb77148ca3ba | refs/heads/master | 2023-01-09T07:30:51.416535 | 2020-11-04T10:23:53 | 2020-11-04T10:23:53 | 309,967,890 | 1 | 0 | null | null | null | null | UTF-8 | Python | false | false | 164 | py | def print_name(name):
print(name)
print_name("nanda")
#print_twice("nithin")
def print_twice(name):
print_name(name)
print_name(name)
print_twice("nithin")
| [
"nithinmk26@gmail.com"
] | nithinmk26@gmail.com |
e038243caf03d7f65252cbc3b38256816376cb27 | db6f7200e7d62f499b500d7e26678835df2e5771 | /app/__init__.py | 8723ea03b093e4553a46f087b4720fff63cc3d5e | [] | no_license | melaniesarah/flask-microblog2 | 0583d820152e5493456346554597146c8236cf9e | 98b6ebf206624a5189a16617b427c2b51779d2cb | refs/heads/master | 2023-03-10T01:42:05.322797 | 2021-02-24T21:45:58 | 2021-02-24T21:45:58 | 296,149,326 | 1 | 0 | null | null | null | null | UTF-8 | Python | false | false | 3,042 | py | import logging
from logging.handlers import SMTPHandler, RotatingFileHandler
import os
from flask import Flask, request, current_app
from flask_sqlalchemy import SQLAlchemy
from flask_migrate import Migrate
from flask_login import LoginManager
from flask_mail import Mail
from flask_bootstrap import Bootstrap
from flask_moment import Moment
from flask_babel import Babel, lazy_gettext as _l
from config import Config
from elasticsearch import Elasticsearch
db = SQLAlchemy()
migrate = Migrate()
login = LoginManager()
login.login_view = 'auth.login'
login.login_message = _l('Please log in to access this page.')
mail = Mail()
bootstrap = Bootstrap()
moment = Moment()
babel = Babel()
def create_app(config_class=Config):
app = Flask(__name__)
app.config.from_object(config_class)
db.init_app(app)
migrate.init_app(app, db)
login.init_app(app)
mail.init_app(app)
bootstrap.init_app(app)
moment.init_app(app)
babel.init_app(app)
app.elasticsearch = Elasticsearch([app.config['ELASTICSEARCH_URL']]) if app.config['ELASTICSEARCH_URL'] else None
from app.errors import bp as errors_bp
app.register_blueprint(errors_bp)
from app.auth import bp as auth_bp
app.register_blueprint(auth_bp, url_prefix='/auth')
from app.main import bp as main_bp
app.register_blueprint(main_bp)
if not app.debug and not app.testing:
if app.config['MAIL_SERVER']:
auth = None
if app.config['MAIL_USERNAME'] or app.config['MAIL_PASSWORD']:
auth = (app.config['MAIL_USERNAME'], app.config['MAIL_PASSWORD'])
secure = None
if app.config['MAIL_USE_TLS']:
secure = ()
mail_handler = SMTPHandler(
mailhost=(app.config['MAIL_SERVER'], app.config['MAIL_PORT']),
fromaddr='no-reply@' + app.config['MAIL_SERVER'],
toaddrs=app.config['ADMINS'], subject='Microblog Failure',
credentials=auth, secure=secure)
mail_handler.setLevel(logging.ERROR)
app.logger.addHandler(mail_handler)
if app.config['LOG_TO_STDOUT']:
strem_handler = logging.StreamHandler()
stream_handler.setLevel(logging.INFO)
app.logger.addHandler(stream_handler)
else:
if not os.path.exists('logs'):
os.mkdir('logs')
file_handler = RotatingFileHandler('logs/microblog.log', maxBytes=10240,
backupCount=10)
file_handler.setFormatter(logging.Formatter(
'%(asctime)s %(levelname)s: %(message)s [in %(pathname)s:%(lineno)d]'))
file_handler.setLevel(logging.INFO)
app.logger.addHandler(file_handler)
app.logger.setLevel(logging.INFO)
app.logger.info('Microblog startup')
return app
@babel.localeselector
def get_locale():
#return request.accept_languages.best_match(current_app.config['LANGUAGES'])
return 'es'
from app import models | [
"melaniesarah@gmail.com"
] | melaniesarah@gmail.com |
fab2b30ed6fb8e6b0e82f7b5754c46de062f13c1 | 6cffd0ad5a0e15dc7bc5c0ec78b6eaece0ef9371 | /bstest.py | 83525ffd175a259b0ca1875d622be75c475e938f | [] | no_license | asura123456/untitled1 | 79979edd7f51307e00b10e3a474031fe70f709d5 | 08382bef16ef603955c86cc55aafaecf1282117f | refs/heads/master | 2021-03-03T14:12:00.514373 | 2020-03-19T09:34:03 | 2020-03-19T09:34:03 | 203,749,765 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 408 | py | __author__ = 'Administrator'
from bs4 import BeautifulSoup
import urllib.request
#reponse=urllib.request.urlopen('BeautifulSoup.html')
soup=BeautifulSoup(open('BeautifulSoup.html'),'lxml')
#for i in soup.find_all('a',id="link2"):
# print(i)
print(soup.select('p .sister'))
print(soup.select('head>title'))
print(soup.select('a[id="link3"]'))
print(soup.select('p href="http://example.com/tillie"')) | [
"maya0026@tom.com"
] | maya0026@tom.com |
44a7c1943eb7b978a98e5947ac8a6925fbb719a5 | 5f27bc1a0460a078f6fe33a544f494a5dff7f452 | /script/puzzle_test_backup_D_20_1026/D_20_1020_vrepMoveit_jaco1.py | 9c282275c0404bbd785b9da6c406c8b27613031f | [] | no_license | A-Why-not-fork-repositories-Good-Luck/arm_move | 3e381f0310265f47da14beaac136c358fb318f92 | e2e6182cfd93df1935bd3b8e9158134964dc44fa | refs/heads/master | 2023-03-15T18:37:17.337770 | 2020-11-18T06:46:06 | 2020-11-18T06:46:06 | null | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 32,748 | py | #!/usr/bin/env python
GRID_SIZE = 0.01
G2P_SIZE = 100
import rospy
import numpy as np
import tf
import matplotlib.pyplot as plt
import copy
import time
import D_20_1020_custom_function as CUF
import D_20_1020_client_function as CLF
from D_20_1020_VFHplus_change_radius import influence
from D_20_1020_envClass4altest import EnvInfo as EI
from D_20_1020_envClass4altest import CanInfo as CI
from arm_move.srv._box_info_srv import *
from arm_move.srv._arm_move_srv import *
from arm_move.srv._work_start_srv import *
from arm_move.srv._att_hand_box_srv import *
from arm_move.srv._arm_goalJoint_srv import *
import timeit
def go_home():
# 2020.08.05 SH
move_group_name = 'panda_arm'
home_joint = [-0.7912285295667355, -1.7449968666946676, 1.6255344777637362, -2.9980328554805484, 1.552371742049853, 1.345932931635115, 0.8050298552807971]
CLF.move_joints_client_rad(move_group_name, home_joint)
def go_ready():
# 2020.08.05 SH
move_group_name = 'panda_arm'
home_joint = [-1.6238, -1.6078, -0.2229, -2.6057, 1.4646, 1.4325, -0.2159]
CLF.move_joints_client_rad(move_group_name, home_joint)
def hand_open():
# 2020.08.05 SH
CLF.panda_gripper_open()
def pick_and_place(env, pick_pose, pick_object_name, place_pose):
print"\tPICK AND PLACE ACTION => rearrange", pick_object_name
env.pick(env.obs_pos, pick_pose, place_pose)
CLF.att_box_client('hand', pick_object_name)
env.go_ready()
env.place(env.obs_pos, place_pose)#, vrep_env.get_current_joint(joint_names_jaco))
CLF.det_box_client(pick_object_name, [0, 0, 0], [0, 0, 0, 0], [0, 0, 0], 'red')
CLF.add_mesh_client(pick_object_name, [place_pose[0], place_pose[1], 0.605], [0.0, 0.0, 0.0, 0.0], [0.001, 0.001, 0.001])
env.go_ready()
print"\tEND PICK AND PLACE ACTION"
#
# # ret_pick_pose = env.pick(env.obs_pos, pick_pose, place_pose)
# env.move_to([[ret_pick_pose[0][0] - 0.03, ret_pick_pose[0][1], ret_pick_pose[0][2]], ret_pick_pose[1]])
#
# env.move_to([[ret_pick_pose[0][0] + 0.05, ret_pick_pose[0][1], ret_pick_pose[0][2]], ret_pick_pose[1]])
#
#
# env.pre_place(env.obs_pos, place_pose, vrep_env.get_current_joint(joint_names_jaco))
# ret_place_pose = env.place(env.obs_pos, place_pose, vrep_env.get_current_joint(joint_names_jaco))
# env.move_to([[ret_place_pose[0][0] - 0.1, ret_place_pose[0][1], ret_place_pose[0][2]], ret_place_pose[1]])
#
# CLF.det_box_client(pick_object_name, [env.object_z, -sel_can_pos[1], sel_can_pos[0]], obstacle_info[env.ore_order[0]][1], obstacle_info[env.ore_order[0]][2], 'blue')
#
# env.move_to([[ret_place_pose[0][0] + 0.1, ret_place_pose[0][1], ret_place_pose[0][2]], ret_place_pose[1]])
# # CLF.add_box_client(obstacle_name[env.ore_order[0]], [env.object_z, -sel_can_pos[1], sel_can_pos[0]], obstacle_info[env.ore_order[0]][1], obstacle_info[env.ore_order[0]][2], 'blue')
def test_algorithm(method, data_in):
# method
# "where" : icra2020 "where to relocate?"
# "far" : farthest method
go_ready()
hand_open()
print "start with method:", method
print "\n***STEP 1*** : env setting"
obj_h = -0.0
obj_z = 0.605 + obj_h#+ obj_h/2.0
target_name = ['target']
target_info = []
target_info.append([[data_in[0][1], -data_in[0][0], obj_z], [0, 0, 0, 0], [0.001, 0.001, 0.001]]) # for the add_mesh
# target_info.append([[data_in[0][0], data_in[0][1], obj_z], [0, 0, 0, 0], [0.06, 0.06, 0.12]]) # for the add_box
# target_info[i][0][2] = target_info[i][0][2] + 0.04
# target_info[i][2][2] = target_info[i][2][2] + 0.08
# obstacle_name = []
# for i in range(len(data_in[1])):
obstacle_name = [str(i).zfill(2) for i in range(len(data_in[1]))]
# obstacle_name.append('obstacle'+str(i))
# print obstacle_name
# obstacle_name = ['obstacle0', 'obstacle1', 'obstacle2', 'obstacle3', 'obstacle4', 'obstacle5', 'obstacle6', 'obstacle7', 'obstacle8']
obstacle_info = []
# [[obj_pos.x, obj_pos.y, obj_pos.z], [obj_ori_q.x, obj_ori_q.y, obj_ori_q.z, obj_ori_q.w], [obj_scale.x, obj_scale.y, obj_scale.z]]
for i in range(len(obstacle_name)):
obstacle_info.append([[data_in[1][i][1], -data_in[1][i][0], obj_z], [0, 0, 0, 0], [0.001, 0.001, 0.001]]) # for the add_mesh
# obstacle_info.append([[data_in[1][i][0], data_in[1][i][1], obj_z], [0, 0, 0, 0], [0.06, 0.06, 0.12]]) # for the add_box
# obstacle_info[i][0][2] = obstacle_info[i][0][2] + 0.04
# obstacle_info[i][2][2] = obstacle_info[i][2][2] + 0.08
print "\tNo. of obstacles:", len(obstacle_name)
env_name = ['shelf_gazebo']#2020.10.21: puzzle test, 'Jaco_base', 'table_ls', 'table_rs', 'table_us', 'table_bs']
env_info = []
base_position = [0.8637, 0, 0.0 + obj_h]
base_quaternion = [0, 0, 0, 1]
base_scale = [0.001, 0.001, 0.001]
CLF.add_mesh_client('shelf_gazebo', base_position, base_quaternion, base_scale)
ws_pos = [0.8637+0.5*0.45+0.03, 0.0, 0.0 + obj_h]
ws_rot = [0.0, 0.0, 0.0, 0.0]
ws_scale = [0.45, 0.91, 0.0]
env_info.append([ws_pos, ws_rot, ws_scale])
# for i in range(len(env_name)):
# env_info.append(vrep_env.get_object_info(env_name[i]))
# if i > 1:
# env_info[i][2][0] = env_info[i][2][0]+0.01
# env_info[i][2][1] = env_info[i][2][1]+0.01
# env_info[i][2][2] = env_info[i][2][2]+0.01
for i in range(len(obstacle_info)):
CLF.add_mesh_client(obstacle_name[i], obstacle_info[i][0], obstacle_info[i][1], obstacle_info[i][2])
# CLF.add_box_client(obstacle_name[i], obstacle_info[i][0], obstacle_info[i][1], obstacle_info[i][2], 'red')
for i in range(len(target_info)):
CLF.add_mesh_client(target_name[i], target_info[i][0], target_info[i][1], target_info[i][2])
# CLF.add_box_client(target_name[i], target_info[i][0], target_info[i][1], target_info[i][2], 'green')
# for i in range(len(env_info)):
# # CLF.add_mesh_client(env_name[i], env_info[i][0], env_info[i][1], env_info[i][2])
# CLF.add_box_client(env_name[i], env_info[i][0], env_info[i][1], env_info[i][2], 'gray')
ws = env_info[0]
# print"ws info", env_info[0]
ws_w = int(round(ws[2][0]*100)) # x-axes in Rviz
ws_d = int(round(ws[2][1]*100)) # y-axes in Rviz
print "\tRviz ws width, depth:", ws_w, ws_d
# GRID_SIZE = 0.01
ws_zero_pos = [round(ws[0][0] - ws[2][0]*0.5, 2), round(ws[0][1] - ws[2][1]*0.5, 2)]
print "\tRviz ws cen pos:", ws[0]
print "\tRviz ws, zero pos:", ws_zero_pos
# ws_w, ws_d = 100, 100 # get table size in the v-rep
ws_cen = [-ws[0][1], ws[0][0]]
rob_pos = [0.0, 0.0]
OBJ_R = 0.035
env = EI(rob_pos, ws_w, ws_d, ws_cen, ws_zero_pos, grid_size=GRID_SIZE, wall_r=OBJ_R)
env.set_env(obstacle_name, obstacle_info, target_name, target_info)
env.update_env(env.obs_pos, env.obs_grid)
print "\trearrangement order:", env.ore_order
if len(env.ore_order) == 0:
print "end rearrangement"
pick_and_place(env, env.tar_pos, 'target', env.tar_pos)
time.sleep(1)
# CUF.draw_grid_info(env.grid_ori)
# plt.show()
space_err = 0
rearr_cnt = 0
# env.get_env(obs_r, tar_r, min_ore)
algorithm_start = timeit.default_timer()
env.get_max_can(env.grid_ori, bt_num=1, trial_num=1000) # We get "grid_max_can", "can_grid"
# env.get_env_case1()
# env.get_max_can_case1()
'''
Make object info!
Type : target, obstacle, candidate
Info : pos, grid, A, BT, b, ORC, ORE
'''
can_info = []
for i in range(len(env.can_pos)):
can_info.append((CI('candidate', env.can_pos[i], env.can_grid[i])))
# check env info got right
# if 1:
# print "\n# of obstacles", len(env.obs_pos), "\n# of candidates", len(env.can_pos)
'''
GET candidates info
'''
t_ore_order = copy.deepcopy(env.ore_order)
# for i in range(len(can_info)):
# print "can", i, ":", can_info[i].pos
# CUF.draw_grid_info(env.grid_ori)
# CUF.draw_grid_info(env.grid_del)
# CUF.draw_grid_info(env.grid_max_can)
# for c_i in range(len(can_info)):
# plt.text(can_info[c_i].grid[0], can_info[c_i].grid[1], 'Can' + str(c_i), fontsize=20, ha='center', bbox=dict(facecolor='pink', alpha=0.8))
# for o_i in range(len(env.obs_grid)):
# plt.text(env.obs_grid[o_i][0], env.obs_grid[o_i][1], 'Obs' + str(o_i), fontsize=20, ha='center', bbox=dict(facecolor='red', alpha=0.8))
# plt.show()
method = 'mine'
method = 'far'
method = 'deep'
while len(env.ore_order): # this while loop is for the algorithm
print"\n***STEP 2*** REARRANGE ORDER => :", env.ore_order
print"\tCheck C.A"
# Check C.A : just next step
t_can_info = []
in_can_info = copy.deepcopy(can_info)
in_obs_pos = copy.deepcopy(env.obs_pos)
in_obs_pos.remove(env.obs_pos[env.ore_order[0]])
CLF.del_box_client(obstacle_name[env.ore_order[0]])
t_can_info.append(env.get_can_A(in_can_info, in_obs_pos, env.tar_pos))
CLF.add_mesh_client(obstacle_name[env.ore_order[0]], obstacle_info[env.ore_order[0]][0], obstacle_info[env.ore_order[0]][1], obstacle_info[env.ore_order[0]][2])
# CLF.add_box_client(obstacle_name[env.ore_order[0]], obstacle_info[env.ore_order[0]][0], obstacle_info[env.ore_order[0]][1], obstacle_info[env.ore_order[0]][2], 'red')
# Check C.BT
in_can_info = copy.deepcopy(t_can_info[0])
in_can_info = env.init_BT(in_can_info) # init the BT value of candidates to '0'
in_obs_pos = copy.deepcopy(env.obs_pos)
for ore_i in range(len(env.ore_order)): # after rearrange all ORE
in_obs_pos.remove(env.obs_pos[env.ore_order[ore_i]])
CLF.del_box_client(obstacle_name[env.ore_order[ore_i]])
t_can_info[0] = env.get_can_BT(in_can_info, in_obs_pos, env.tar_pos)
for ore_i in range(len(env.ore_order)):
CLF.add_mesh_client(obstacle_name[env.ore_order[ore_i]], obstacle_info[env.ore_order[ore_i]][0], obstacle_info[env.ore_order[ore_i]][1], obstacle_info[env.ore_order[ore_i]][2])
# CLF.add_box_client(obstacle_name[env.ore_order[ore_i]], obstacle_info[env.ore_order[ore_i]][0], obstacle_info[env.ore_order[ore_i]][1], obstacle_info[env.ore_order[ore_i]][2], 'red')
# Check C.BO : BO : other ORE, just before target
in_can_info = copy.deepcopy(t_can_info[0])
in_obs_pos = copy.deepcopy(env.obs_pos)
for ore_i in range(len(env.ore_order)): # after rearrange all ORE
in_obs_pos.remove(env.obs_pos[env.ore_order[ore_i]])
CLF.del_box_client(obstacle_name[env.ore_order[ore_i]])
for j in range(len(env.ore_order)): # check other ORE just before target
if j > i:
t_can_info[0] = env.get_can_BT(in_can_info, in_obs_pos,env.obs_pos[env.ore_order[j]])
for ore_i in range(len(env.ore_order)):
CLF.add_mesh_client(obstacle_name[env.ore_order[ore_i]], obstacle_info[env.ore_order[ore_i]][0], obstacle_info[env.ore_order[ore_i]][1], obstacle_info[env.ore_order[ore_i]][2])
# CLF.add_box_client(obstacle_name[env.ore_order[ore_i]], obstacle_info[env.ore_order[ore_i]][0], obstacle_info[env.ore_order[ore_i]][1], obstacle_info[env.ore_order[ore_i]][2], 'red')
s_v = []
s_v_index = []
for i in range(1):
in_can_info = copy.deepcopy(t_can_info[i])
ret_can, ret_index = env.get_cf(in_can_info)
s_v.append(ret_can)
s_v_index.append(ret_index)
# print "\n step", i, " has # of cf pos:", len(t_cf[i]), "index", t_cf_index[i]
print"\n***STEP 3*** : find valid candidates"
print "\ts_v:", len(s_v[0]), "\n\ts_v_index:", len(s_v_index[0])
# for i in range(len(s_v[0])):
# print "s_v index:", [i], s_v_index[0][i]
# See the feasibile candidate
# for i in range(len(t_cf[0])):
# print "\n Our Cf pos:", i, t_cf[0][i].pos
# See if this case if case0 or case1
# print "t_cf:", t_cf, "order", env.ore_order
if len(s_v[0]) >= len(env.ore_order):
print "\n\tenough candidate spots"
t_b = []
for i in range(1):
in_obs_pos = copy.deepcopy(env.obs_pos)
for ore_i in range(i + 1):
in_obs_pos.remove(env.obs_pos[env.ore_order[ore_i]])
t_b.append(env.get_cf_b(s_v[i], in_obs_pos))
# print "\n step", i, " has cf b:", t_b[i]
# draw_figs = 1
# if draw_figs == 1:
# for c_i in range(len(can_info)):
# plt.text(can_info[c_i].grid[0], can_info[c_i].grid[1], 'Can' + str(c_i), fontsize=20, ha='center', bbox=dict(facecolor='pink', alpha=0.8))
# for o_i in range(len(env.obs_grid)):
# plt.text(env.obs_grid[o_i][0], env.obs_grid[o_i][1], 'Obs' + str(o_i), fontsize=20, ha='center', bbox=dict(facecolor='red', alpha=0.8))
#
# for step_i in range(1):
# step_grid = copy.deepcopy(env.grid_act)
# step_obs_grid = copy.deepcopy(env.obs_grid)
# for ore_i in range(step_i + 1):
# step_obs_grid.remove(env.obs_grid[env.ore_order[ore_i]])
# for i in range(len(step_obs_grid)):
# step_grid = CUF.obstacle_circle(step_grid, [round(step_obs_grid[i][0], 2), round(step_obs_grid[i][1], 2), env.obs_r[i]], 2)
# for ci in range(len(can_info)):
# xi, yi = can_info[ci].grid
# step_grid = CUF.obstacle_circle(step_grid, [xi, yi, 0.04], 30)
#
# step_grid = CUF.obstacle_circle(step_grid, [env.tar_grid[0], env.tar_grid[1], tar_r], 4) # target
#
# for cf_i in range(len(t_b[step_i])):
# xi = (t_cf[step_i][cf_i].pos[0] - env.ws_zero[0]) * G2P_SIZE
# yi = (t_cf[step_i][cf_i].pos[1] - env.ws_zero[1]) * G2P_SIZE
# step_grid = CUF.obstacle_circle(step_grid, [xi, yi, 0.04], 3)
#
# CUF.draw_grid_info(step_grid)
#
# for cf_i in range(len(t_b[step_i])):
# xi = (t_cf[step_i][cf_i].pos[0] - env.ws_zero[0]) * G2P_SIZE
# yi = (t_cf[step_i][cf_i].pos[1] - env.ws_zero[1]) * G2P_SIZE
# plt.text(xi, yi, 'b=' + str(t_b[step_i][cf_i]), fontsize=20, ha='center', bbox=dict(facecolor='pink', alpha=0.8))
# for ci in range(len(t_can_info[step_i])):
# plt.text(t_can_info[step_i][ci].grid[0], t_can_info[step_i][ci].grid[1] - 2.0, '[A, BT] :' + str([t_can_info[step_i][ci].A, t_can_info[step_i][ci].BT]), fontsize=10, ha='center', bbox=dict(facecolor='pink', alpha=0.8))
# for o_i in range(len(env.obs_grid)):
# plt.text(env.obs_grid[o_i][0], env.obs_grid[o_i][1], 'Obs' + str(o_i), fontsize=20, ha='center', bbox=dict(facecolor='red', alpha=0.8))
# plt.title('step' + str(step_i) + " obs: " + str(env.ore_order[step_i]) + " rearranged")
elif len(s_v[0]) < len(env.ore_order):
print "\n\tnot enough candidate spots"
# print "Since we meet condition: N(CF) < N(ORE) by", len(t_cf[0]), "<", len(env.ore_order), ",\nwe have to remove additional obstacles."
## step1 : "get t_cp", check candidates which have A = 0 and BT = 0
## This means that a candidate is not reachable and it does not block the target object
# Check A for this environment state
in_can_info = copy.deepcopy(can_info)
in_obs_pos = copy.deepcopy(env.obs_pos)
t_can_add = copy.deepcopy(env.get_can_A(in_can_info, in_obs_pos, env.tar_pos))
s_e = [] # s_e: extra candidate spots
in_can_info = copy.deepcopy(t_can_add)
ret_can, ret_index = env.get_cp(in_can_info)
print "\t# of OR'", len(ret_can)
t_s_e = ret_can
t_s_e_index = ret_index
# print "t_cp:", len(t_cp), "index", t_cp_index
# for i in range(len(t_cp)):
# print "\n Our Cp:", i, t_cp[i].pos
if len(t_s_e) == 0:
print "\tno possible extra candidate exist"
space_err = 1
break
# step2 : check c_ore for each cp and pick min of it
t_s_r = [] # s_r: candidate spot relocate plan
in_can_info = copy.deepcopy(t_s_e)
# tmp_order_time_start = timeit.default_timer()
# tmp_order_time_start2 = time.clock()
t_s_r = env.get_c_ore(in_can_info)
# tmp_order_time_end = timeit.default_timer()
# tmp_order_time_end2 = time.clock()
# order_time = order_time + tmp_order_time_end - tmp_order_time_start
# order_time2 = order_time2 + tmp_order_time_end2 - tmp_order_time_start2
# order_cnt = order_cnt + 100 * len(t_s_e)
# print "\n"
# for i in range(len(t_cp)):
# print "cp", t_cp[i].pos, "\nc_ore", c_ore[i]
s_r = []
s_e_index = []
print "\n"
for i in range(len(t_s_e)):
print "can", t_s_e_index[i], "grid:", t_s_e[i].grid, ", s_r:", t_s_r[i]
for i in range(len(t_s_e)):
if t_s_r[i] != []:
s_e.append(t_s_e[i])
s_r.append(t_s_r[i])
s_e_index.append(t_s_e_index[i])
# tmp_se = copy.deepcopy(s_e)
# tmp_sr = copy.deepcopy(s_r)
# emp_sr = []
# for i in range(len(s_e)):
# if s_r[i] == []:
# print "remove empty s_e", i
# emp_sr.append(i)
#
# print "tmp se:", tmp_se, "\ntmp sr", tmp_sr
# for i in range(len(emp_sr)):
#
# print "tmp_se[emp_sr[i]]", tmp_se[emp_sr[i]].pos
# print "tmp_sr[emp_sr[i]]", tmp_sr[emp_sr[i]]
# s_e.remove(tmp_se[emp_sr[i]])
# s_r.remove(tmp_sr[emp_sr[i]])
while len(s_e):
print "# of s_e:", len(s_e), s_r
print "\n"
for i in range(len(s_e)):
print "can", s_e_index[i], "pos:", s_e[i].pos, ", s_r:", s_r[i]
min_s_r = CUF.min_len_list(s_r)
print "\nmin sr:", min_s_r
#
# print "picked ci index:", t_cp.index(t_cp[c_ore.index(min_c_ore)])
# print "picked ci address:", copy.deepcopy(t_cp[c_ore.index(min_c_ore)]).pos
cp = copy.deepcopy(s_e[s_r.index(min_s_r)])
# print "selected cp pos", cp.pos
## step3 : "get t_cf", check candidates which have A = 1 and BT' = 0
## Check A for this environment state T' is t_cp_i
in_can_info = copy.deepcopy(can_info)
in_obs_pos = copy.deepcopy(env.obs_pos)
in_tar_pos = copy.deepcopy(cp.pos)
t_can_add = copy.deepcopy(env.get_can_A(in_can_info, in_obs_pos, env.tar_pos))
# Check C.BT for this environment state
in_can_info = copy.deepcopy(t_can_add)
in_can_info = env.init_BT(in_can_info) # init the BT value of candidates to '0'
in_obs_pos = copy.deepcopy(env.obs_pos)
sorted_min_s_r = copy.deepcopy(min_s_r)
sorted_min_s_r.sort(reverse=True)
print "sorted min_s_r:", sorted_min_s_r
if sorted_min_s_r[0] == len(env.obs_pos): # if OR' has o_t ! remove s_e
print "o_t is in OR'"
s_e.remove(s_e[s_r.index(min_s_r)])
s_e_index.remove(s_e_index[s_r.index(min_s_r)])
s_r.remove(s_r[s_r.index(min_s_r)])
else:
for ore_i in range(len(min_s_r)): # after rearrange all OR'
in_obs_pos.remove(in_obs_pos[sorted_min_s_r[ore_i]])
CLF.del_box_client(obstacle_name[sorted_min_s_r[ore_i]])
in_tar_pos = copy.deepcopy(cp.pos)
t_can_add = env.get_can_BT(in_can_info, in_obs_pos, in_tar_pos)
for ore_i in range(len(min_s_r)): # after rearrange all OR'
CLF.add_box_client(obstacle_name[sorted_min_s_r[ore_i]], obstacle_info[sorted_min_s_r[ore_i]][0], obstacle_info[sorted_min_s_r[ore_i]][1], obstacle_info[sorted_min_s_r[ore_i]][2], 'red')
# for i in range(len(t_can_add)):
# print "can", i, "A:", t_can_add[i].A, "B:", t_can_add[i].BT
s_e_v = []
s_v_index = []
in_can_info = copy.deepcopy(t_can_add)
ret_can, ret_index = env.get_cf(in_can_info)
s_e_v.append(ret_can)
s_v_index.append(ret_index)
print "s_e_v: ", s_e_v
for i in range(len(s_e_v[0])):
print s_e_v[0][i].grid
if len(s_e_v[0]) >= len(min_s_r) - 1:
print "this se is possible"
if len(min_s_r) == 1:
print "only one move needed"
# t_can_info = []
# for i in range(len(env.ore_order)):
# in_can_info = copy.deepcopy(can_info)
# in_obs_pos = copy.deepcopy(env.obs_pos)
# for ore_i in range(i + 1):
# if min_s_r[0] != env.ore_order[ore_i]:
# in_obs_pos.remove(env.obs_pos[env.ore_order[ore_i]])
# in_obs_pos.remove(env.obs_pos[min_s_r[0]])
# t_can_info.append(env.get_can_A(in_can_info, in_obs_pos, env.tar_pos))
s_v = [[s_e[s_r.index(min_s_r)]]]
s_v_index = [[s_e_index[s_r.index(min_s_r)]]]
# print "se v:", s_v, s_v[0], s_v[0][0], s_v[0][0].pos
# for i in range(len(env.ore_order)):
# add_can_info = copy.deepcopy(t_can_info[i])
# ret_can, ret_index = env.get_cf(add_can_info)
# s_v.append(ret_can)
# s_v_index.append(ret_index)
t_b = [[0]]
# for i in range(1):
# in_obs_pos = copy.deepcopy(env.obs_pos)
# for ore_i in range(i+1):
# in_obs_pos.remove(env.obs_pos[env.ore_order[ore_i]])
# t_b.append(env.get_cf_b(s_v[i], in_obs_pos))
# # print "\n step", i, " has cf b:", t_b[i]
break # for out s_e loop
else:
t_b = []
in_obs_pos = copy.deepcopy(env.obs_pos)
for ore_i in range(1):
in_obs_pos.remove(env.obs_pos[min_s_r[ore_i]])
t_b.append(env.get_cf_b(s_e_v[0], in_obs_pos))
s_v[0] = s_e_v[0]
break # for out s_e loop
else: # s_e[s_r.index(min_s_r)]
print "\nremove",
print "s_e:", s_e
print "s_r:", s_r
print "s_e_index:", s_e_index
s_e.remove(s_e[s_r.index(min_s_r)])
s_e_index.remove(s_e_index[s_r.index(min_s_r)])
s_r.remove(s_r[s_r.index(min_s_r)])
if len(s_e) == 0:
# print "no possible extra candidate exist"
break
env.ore_order = min_s_r
# draw_figs = 1
# if draw_figs == 1:
# for c_i in range(len(can_info)):
# plt.text(can_info[c_i].grid[0], can_info[c_i].grid[1], 'Can' + str(c_i), fontsize=20, ha='center', bbox=dict(facecolor='pink', alpha=0.8))
# for o_i in range(len(env.obs_grid)):
# plt.text(env.obs_grid[o_i][0], env.obs_grid[o_i][1], 'Obs' + str(o_i), fontsize=20, ha='center', bbox=dict(facecolor='red', alpha=0.8))
#
# step_i = 0
# step_grid = copy.deepcopy(env.grid_act)
# step_obs_grid = copy.deepcopy(env.obs_grid)
# step_obs_grid.remove(env.obs_grid[env.ore_order[0]])
# for i in range(len(step_obs_grid)):
# # print "i:", i, "step_obs_grid [i]:", step_obs_grid[i]
# step_grid = CUF.obstacle_circle(step_grid, [round(step_obs_grid[i][0], 2), round(step_obs_grid[i][1], 2), env.obs_r[i]], 2)
# for ci in range(len(can_info)):
# xi, yi = can_info[ci].grid
# step_grid = CUF.obstacle_circle(step_grid, [xi, yi, 0.04], 30)
#
# step_grid = CUF.obstacle_circle(step_grid, [env.tar_grid[0], env.tar_grid[1], tar_r], 4) # target
#
# for cf_i in range(len(t_b[step_i])):
# xi = (t_cf[step_i][cf_i].pos[0] - env.ws_zero[0]) * G2P_SIZE
# yi = (t_cf[step_i][cf_i].pos[1] - env.ws_zero[1]) * G2P_SIZE
# step_grid = CUF.obstacle_circle(step_grid, [xi, yi, 0.04], 3)
#
# CUF.draw_grid_info(step_grid)
#
# for cf_i in range(len(t_b[step_i])):
# xi = (t_cf[step_i][cf_i].pos[0] - env.ws_zero[0]) * G2P_SIZE
# yi = (t_cf[step_i][cf_i].pos[1] - env.ws_zero[1]) * G2P_SIZE
# plt.text(xi, yi, 'b=' + str(t_b[step_i][cf_i]), fontsize=20, ha='center', bbox=dict(facecolor='pink', alpha=0.8))
# for ci in range(len(t_can_info[step_i])):
# plt.text(t_can_info[step_i][ci].grid[0], t_can_info[step_i][ci].grid[1] - 2.0, '[A, BT] :' + str([t_can_info[step_i][ci].A, t_can_info[step_i][ci].BT]), fontsize=10, ha='center', bbox=dict(facecolor='pink', alpha=0.8))
# for o_i in range(len(env.obs_grid)):
# plt.text(env.obs_grid[o_i][0], env.obs_grid[o_i][1], 'Obs' + str(o_i), fontsize=20, ha='center', bbox=dict(facecolor='red', alpha=0.8))
# plt.title('step' + str(step_i) + " obs: " + str(env.ore_order[step_i]) + " rearranged")
if space_err:
print "no possible extra candidate exist"
break
# move obstacle to can(min(b))
# print "s_v", s_v
# print "s_v[0]", s_v[0]
# print "s_v[0][0]", s_v[0][0]
# print "s_v[0][0].pos", s_v[0][0].pos
print "\tt_b[0]", t_b[0]
find_b = copy.deepcopy(t_b[0])
# print "move to c_", find_b.index(min(find_b))
if method == 'far':
t_sel_can_index = [i for i in range(len(find_b))]
elif method == 'deep':
t_sel_can_index = [i for i in range(len(find_b))]
elif method == 'mine':
t_sel_can_index = [i for i in range(len(find_b)) if find_b[i] == min(find_b)]
t_sel_can_dist = []
# print "\ntar grid: ", env.tar_grid
# print "\ntar pos: ", env.tar_pos
print "\tt sel can index", t_sel_can_index
for i in range(len(t_sel_can_index)):
# print "t_cf grid x,y:", t_sel_can_index[i], t_cf[0][t_sel_can_index[i]].grid[0], t_cf[0][t_sel_can_index[i]].grid[1]
# print "t_cf pos x,y:", t_sel_can_index[i], s_v[0][t_sel_can_index[i]].pos[0], s_v[0][t_sel_can_index[i]].pos[1]
if method == 'deep':
t_sel_can_dist.append(np.sqrt((env.rob_pos[0] - s_v[0][t_sel_can_index[i]].pos[0]) ** 2 + (env.rob_pos[1] - s_v[0][t_sel_can_index[i]].pos[1]) ** 2))
else:
t_sel_can_dist.append(np.sqrt((env.tar_pos[0] - s_v[0][t_sel_can_index[i]].pos[0]) ** 2 + (env.tar_pos[1] - s_v[0][t_sel_can_index[i]].pos[1]) ** 2))
# print "t sel can dist", t_sel_can_dist
sel_can_index = t_sel_can_index[t_sel_can_dist.index(max(t_sel_can_dist))]
# print "sel can index", sel_can_index
sel_can_pos = can_info[s_v_index[0][sel_can_index]].pos
sel_can_grid = can_info[s_v_index[0][sel_can_index]].grid
sel_obs_pos = env.obs_pos[env.ore_order[0]]
sel_obs_grid = env.obs_grid[env.ore_order[0]]
env.obs_pos[env.ore_order[0]] = sel_can_pos
env.obs_grid[env.ore_order[0]] = sel_can_grid
can_info[s_v_index[0][sel_can_index]].pos = sel_obs_pos
can_info[s_v_index[0][sel_can_index]].grid = sel_obs_grid
# tmp_order_time_start = timeit.default_timer()
# tmp_order_time_start2 = time.clock()
# env.pick_n_place()
# CLF.add_box_client(obstacle_name[env.ore_order[0]], [env.object_z, -sel_can_pos[1], sel_can_pos[0]], obstacle_info[env.ore_order[0]][1], obstacle_info[env.ore_order[0]][2], 'blue')
pick_and_place(env, sel_obs_pos, obstacle_name[env.ore_order[0]], env.obs_pos[env.ore_order[0]])
# time.sleep(1)
# obstacle_info = []
# for i in range(len(obstacle_name)):
# obstacle_info.append(vrep_env.get_object_info(obstacle_name[i]))
# # obstacle_info[i][0][2] = obstacle_info[i][0][2] + 0.04
# # obstacle_info[i][2][2] = obstacle_info[i][2][2] + 0.08
# for i in range(len(obstacle_info)):
# CLF.add_box_client(obstacle_name[i], obstacle_info[i][0], obstacle_info[i][1], obstacle_info[i][2], 'red')
# env.set_env(obstacle_name, obstacle_info, target_name, target_info)
# home_joint = [3.1415927410125732, 4.537856101989746, 5.93411922454834, -0.6108652353286743, 1.7453292608261108, -0.5235987901687622]
#
# CLF.move_joints_client_rad('arm', home_joint)
env.update_env(env.obs_pos, env.obs_grid)
# tmp_order_time_end = timeit.default_timer()
# order_time = order_time + tmp_order_time_end - tmp_order_time_start
# order_time2 = order_time2 + tmp_order_time_end2 - tmp_order_time_start2
# order_cnt = order_cnt + 1
rearr_cnt = rearr_cnt + 1
if env.order_error_flag == 0:
print "\nretry for another environment"
space_err = 1
break
print "after move order is:", env.ore_order
# CUF.draw_grid_info(env.grid_ori)
# for c_i in range(len(can_info)):
# plt.text(can_info[c_i].grid[0], can_info[c_i].grid[1], 'Can' + str(c_i), fontsize=20, ha='center', bbox=dict(facecolor='pink', alpha=0.8))
# for o_i in range(len(env.obs_grid)):
# plt.text(env.obs_grid[o_i][0], env.obs_grid[o_i][1], 'Obs' + str(o_i), fontsize=20, ha='center', bbox=dict(facecolor='red', alpha=0.8))
if len(env.ore_order) == 0:
print "end rearrangement"
pick_and_place(env, env.tar_pos, 'target', env.tar_pos)
time.sleep(1)
# plt.title('rearrangement finished')
break
# else:
# plt.title('after rearrngement')
# plt.show()
pick_and_place(env, env.tar_pos, 'target', env.tar_pos)
time.sleep(1)
algorithm_end = timeit.default_timer()
tot_time = algorithm_end - algorithm_start
print "tot time:", tot_time
if __name__ == "__main__":
X = ['0.03', '-0.01', '0.36', '0.30', '-0.19', '-0.05', '-0.29', '0.22', '0.19', '0.14', '-0.12']
Y = ['1.22', '1.11', '1.04', '1.17', '1.06', '1.31', '1.17', '1.31', '1.06', '1.19', '1.13']
data_in = []
data_in.append([-0.17, 1.22])
obs_list = []
for i in range(len(X)):
obs_list.append([float(X[i]), float(Y[i])])
data_in.append(obs_list)
# print "data:", data_in
method = "where"
test_algorithm(method, data_in) | [
"welovehun91@gmail.com"
] | welovehun91@gmail.com |
3c693e431934b383adf4d7f0be6aacd476f5da73 | 9a78d88daeef20341da3b3ed326f7be633cb2b46 | /instagram/instagramScraper.py | db788f63dd38fd3becb00f541e0c67850be53b54 | [] | no_license | muhammedfatih/SocialScraper | 6f410e3f44495ec65f63d8c50734aa24f0278bc7 | a83f53b272fae7da33e400d8409d44578f22ccd9 | refs/heads/master | 2022-11-17T21:30:12.966814 | 2020-07-18T17:59:17 | 2020-07-18T17:59:17 | 280,710,515 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 1,425 | py | from urllib2 import urlopen
import re
import json
import sys
sys.path.append("models")
from instagramUser import instagramUser
class instagramScraper:
INSTAGRAM_URL="https://www.instagram.com/"
userData=instagramUser("", "", "", 0, 0)
def fetch(self, username):
link = self.INSTAGRAM_URL + username
urlStream = urlopen(link)
urlContent = urlStream.read()
encapsulatedJsonContentAsString=re.findall("window._sharedData = .*?;</script>", urlContent)
jsonContentAsString=encapsulatedJsonContentAsString[0].replace("window._sharedData = ", "").replace(";</script>", "")
jsonContent=json.loads(jsonContentAsString)
userJsonContent=jsonContent["entry_data"]["ProfilePage"][0]["graphql"]["user"]
userData=instagramUser(
username
, userJsonContent["biography"].encode(sys.stdout.encoding, errors='replace')
, userJsonContent["business_category_name"]
, userJsonContent["edge_follow"]["count"]
, userJsonContent["edge_followed_by"]["count"]
)
for node in userJsonContent["edge_owner_to_timeline_media"]["edges"]:
userData.addPhoto(
node["node"]["display_url"]
, node["node"]["edge_liked_by"]["count"]
, node["node"]["taken_at_timestamp"]
)
return userData.__dict__ | [
"mfatihinanc@gmail.com"
] | mfatihinanc@gmail.com |
a112cf2bb050b37acf7da7c8c2b2bbaca571d4d0 | 1a3333fda7f53f6feff0c00e6b623e02211d5b27 | /tests/buildPythonApplication/channels/root/pkgs/testPackage/src/hello/__main__.py | d1691dbc5968d2404c607e812794e10d293e8bac | [] | no_license | jD91mZM2/flox-nixexprs | 6a1b7ca40c6dc8e99ae01941d3cbf77c8c896fc3 | be50d9f2374719cdf0ac762e4442e4fdd77758c9 | refs/heads/master | 2023-03-01T17:43:18.550198 | 2020-11-27T14:24:22 | 2020-11-27T14:24:22 | 328,352,754 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 59 | py | from . import hello
if __name__ == '__main__':
hello()
| [
"contact@infinisil.com"
] | contact@infinisil.com |
d53a6445ee0ad5ff6856c3aa7fb0d32ba99044a2 | 301912f90f6fbc0ebdfe0726a4900a74789a0707 | /tele-bot.py | a54604ea84c8c1853977a5a3f90d0c187cc90ce8 | [] | no_license | denyskrysko/python-bots | 51037374da09d08b6027fb855d951a86ace7215b | 8eebf065f0a98667de620ba1b99eb4407091fd5e | refs/heads/master | 2020-06-04T11:06:50.156392 | 2019-07-04T13:33:54 | 2019-07-04T13:33:54 | 191,995,562 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 2,476 | py | import time
import requests
from lxml import html
from bs4 import BeautifulSoup
import schedule
import datetime
import urllib.request
def parse(url):
try:
page = requests.get(url)
tree = html.fromstring(page.content)
name = tree.xpath('//div[@class="quote-ticker tickerLarge"]/text()')
value = tree.xpath('//div[@class="quote-price priceLarge"]/span/text()')
Arr = {
'name': name,
'value': value
}
# response = requests.get(url, timeout=5)
# content = BeautifulSoup(response.content, "html.parser")
# Arr = []
# for element in content.findAll('div', attrs={"class": "quote-details"}):
# Arr = {
# 'name': element.find('div', attrs={"class": "quote-ticker tickerLarge"}).text.encode('utf-8'),
# 'price': element.find('div', attrs={"class": "quote-price priceLarge"}).text.encode('utf-8'),
# }
return Arr
except Exception as e:
print(e)
def telegram_bot_sendtext(bot_message):
bot_token = '712175175:AAHXStBiBHKgCvtfv87hRfID2yc207j-dd4'
bot_chatID = '-337657336'
send_text = 'https://api.telegram.org/bot' + bot_token + '/sendMessage?chat_id=' \
+ bot_chatID + '&parse_mode=Markdown&text=' + bot_message
response = requests.get(send_text)
return response.json()
def report(symbol, target):
# current_dt = datetime.datetime.now().strftime('%Y-%m-%d %H:%M:%S')
try:
Arr = parse('https://web.tmxmoney.com/quote.php?qm_symbol=' + symbol + '&locale=EN')
message = "Heads up! " + str(Arr['name']).strip("['']") + " Stock level reached: " + \
": $" + str(Arr['value']).strip("['']")
value = str(Arr['value']).strip("['']")
if float(value) >= float(target):
global counter
if counter == 0:
telegram_bot_sendtext(message)
print(" --- M3lvin stock tracking---> " + message)
counter = + 1
except Exception as e:
print(e)
# Entry point - user input
stock_symbol = input("To get a quote enter the stock symbol: ")
stock_target = input("What is the Stock target? ")
counter = 0 # creating global variable for the report function
# schedule.every().day.at("16:50").do(report)
# schedule.every().day.at("13:30").do(report)
while True:
schedule.run_pending()
report(stock_symbol, stock_target)
time.sleep(5)
| [
"denyskrysko@gmail.com"
] | denyskrysko@gmail.com |
a8b8d378ab3f6f1387d52577aa1bf5431858cd0c | c41471781f65d38d9010450b6c9e17f2346a551b | /openstack-dashboard/templates/icehouse/local_settings.py | f45295ea2a3762091554852d5a4f73c472022cc5 | [
"Apache-2.0"
] | permissive | juanarturovargas/openstack-juju | b6854e2feea615404c053e9c754e4d7997c8a6a5 | 21b1aef8aa51c3c32cb1efd1b8cad7865c4d40a0 | refs/heads/master | 2022-12-13T15:31:53.383963 | 2017-05-05T19:18:55 | 2017-05-05T19:18:55 | 90,163,436 | 0 | 1 | NOASSERTION | 2022-11-20T08:41:15 | 2017-05-03T15:17:34 | Python | UTF-8 | Python | false | false | 17,009 | py | import os
from django.utils.translation import ugettext_lazy as _
{% if use_syslog %}
from logging.handlers import SysLogHandler
{% endif %}
from openstack_dashboard import exceptions
DEBUG = {{ debug }}
TEMPLATE_DEBUG = DEBUG
# Required for Django 1.5.
# If horizon is running in production (DEBUG is False), set this
# with the list of host/domain names that the application can serve.
# For more information see:
# https://docs.djangoproject.com/en/dev/ref/settings/#allowed-hosts
#ALLOWED_HOSTS = ['horizon.example.com', ]
# Set SSL proxy settings:
# For Django 1.4+ pass this header from the proxy after terminating the SSL,
# and don't forget to strip it from the client's request.
# For more information see:
# https://docs.djangoproject.com/en/1.4/ref/settings/#secure-proxy-ssl-header
# SECURE_PROXY_SSL_HEADER = ('HTTP_X_FORWARDED_PROTOCOL', 'https')
# If Horizon is being served through SSL, then uncomment the following two
# settings to better secure the cookies from security exploits
#CSRF_COOKIE_SECURE = True
#SESSION_COOKIE_SECURE = True
# Overrides for OpenStack API versions. Use this setting to force the
# OpenStack dashboard to use a specfic API version for a given service API.
# NOTE: The version should be formatted as it appears in the URL for the
# service API. For example, The identity service APIs have inconsistent
# use of the decimal point, so valid options would be "2.0" or "3".
# OPENSTACK_API_VERSIONS = {
# "identity": 3
# }
# Set this to True if running on multi-domain model. When this is enabled, it
# will require user to enter the Domain name in addition to username for login.
# OPENSTACK_KEYSTONE_MULTIDOMAIN_SUPPORT = False
# Overrides the default domain used when running on single-domain model
# with Keystone V3. All entities will be created in the default domain.
# OPENSTACK_KEYSTONE_DEFAULT_DOMAIN = 'Default'
# Set Console type:
# valid options would be "AUTO", "VNC" or "SPICE"
# CONSOLE_TYPE = "AUTO"
# Default OpenStack Dashboard configuration.
HORIZON_CONFIG = {
'dashboards': ('project', 'admin', 'settings',),
'default_dashboard': 'project',
'user_home': 'openstack_dashboard.views.get_user_home',
'ajax_queue_limit': 10,
'auto_fade_alerts': {
'delay': 3000,
'fade_duration': 1500,
'types': ['alert-success', 'alert-info']
},
'help_url': "http://docs.openstack.org",
'exceptions': {'recoverable': exceptions.RECOVERABLE,
'not_found': exceptions.NOT_FOUND,
'unauthorized': exceptions.UNAUTHORIZED},
}
# Specify a regular expression to validate user passwords.
# HORIZON_CONFIG["password_validator"] = {
# "regex": '.*',
# "help_text": _("Your password does not meet the requirements.")
# }
# Disable simplified floating IP address management for deployments with
# multiple floating IP pools or complex network requirements.
# HORIZON_CONFIG["simple_ip_management"] = False
# Turn off browser autocompletion for the login form if so desired.
# HORIZON_CONFIG["password_autocomplete"] = "off"
LOCAL_PATH = os.path.dirname(os.path.abspath(__file__))
# Set custom secret key:
# You can either set it to a specific value or you can let horizion generate a
# default secret key that is unique on this machine, e.i. regardless of the
# amount of Python WSGI workers (if used behind Apache+mod_wsgi): However, there
# may be situations where you would want to set this explicitly, e.g. when
# multiple dashboard instances are distributed on different machines (usually
# behind a load-balancer). Either you have to make sure that a session gets all
# requests routed to the same dashboard instance or you set the same SECRET_KEY
# for all of them.
SECRET_KEY = "{{ secret }}"
# We recommend you use memcached for development; otherwise after every reload
# of the django development server, you will have to login again. To use
# memcached set CACHES to something like
CACHES = {
'default': {
'BACKEND' : 'django.core.cache.backends.memcached.MemcachedCache',
'LOCATION' : '127.0.0.1:11211',
}
}
# Send email to the console by default
EMAIL_BACKEND = 'django.core.mail.backends.console.EmailBackend'
# Or send them to /dev/null
#EMAIL_BACKEND = 'django.core.mail.backends.dummy.EmailBackend'
# Configure these for your outgoing email host
# EMAIL_HOST = 'smtp.my-company.com'
# EMAIL_PORT = 25
# EMAIL_HOST_USER = 'djangomail'
# EMAIL_HOST_PASSWORD = 'top-secret!'
# For multiple regions uncomment this configuration, and add (endpoint, title).
# AVAILABLE_REGIONS = [
# ('http://cluster1.example.com:5000/v2.0', 'cluster1'),
# ('http://cluster2.example.com:5000/v2.0', 'cluster2'),
# ]
{% if regions|length > 1 -%}
AVAILABLE_REGIONS = [
{% for region in regions -%}
('{{ region.endpoint }}', '{{ region.title }}'),
{% endfor -%}
]
{% endif -%}
OPENSTACK_HOST = "{{ service_host }}"
OPENSTACK_KEYSTONE_URL = "{{ service_protocol }}://%s:{{ service_port }}/v2.0" % OPENSTACK_HOST
OPENSTACK_KEYSTONE_DEFAULT_ROLE = "{{ default_role }}"
# Disable SSL certificate checks (useful for self-signed certificates):
# OPENSTACK_SSL_NO_VERIFY = True
# The CA certificate to use to verify SSL connections
# OPENSTACK_SSL_CACERT = '/path/to/cacert.pem'
# The OPENSTACK_KEYSTONE_BACKEND settings can be used to identify the
# capabilities of the auth backend for Keystone.
# If Keystone has been configured to use LDAP as the auth backend then set
# can_edit_user to False and name to 'ldap'.
#
# TODO(tres): Remove these once Keystone has an API to identify auth backend.
OPENSTACK_KEYSTONE_BACKEND = {
'name': 'native',
'can_edit_user': True,
'can_edit_group': True,
'can_edit_project': True,
'can_edit_domain': True,
'can_edit_role': True
}
OPENSTACK_HYPERVISOR_FEATURES = {
'can_set_mount_point': False,
'can_set_password': False,
}
# The OPENSTACK_NEUTRON_NETWORK settings can be used to enable optional
# services provided by neutron. Options currenly available are load
# balancer service, security groups, quotas.
OPENSTACK_NEUTRON_NETWORK = {
'enable_lb': {{ neutron_network_lb }},
'enable_quotas': True,
'enable_security_group': True,
'enable_firewall': {{ neutron_network_firewall }},
'enable_vpn': {{ neutron_network_vpn }},
# The profile_support option is used to detect if an external router can be
# configured via the dashboard. When using specific plugins the
# profile_support can be turned on if needed.
#'profile_support': None,
#'profile_support': 'cisco', # Example of value set to support Cisco
{% if support_profile -%}
'profile_support': '{{ support_profile }}',
{% else -%}
'profile_support': None,
{% endif -%}
}
# The OPENSTACK_IMAGE_BACKEND settings can be used to customize features
# in the OpenStack Dashboard related to the Image service, such as the list
# of supported image formats.
OPENSTACK_IMAGE_BACKEND = {
'image_formats': [
('', ''),
('aki', _('AKI - Amazon Kernel Image')),
('ami', _('AMI - Amazon Machine Image')),
('ari', _('ARI - Amazon Ramdisk Image')),
('iso', _('ISO - Optical Disk Image')),
('qcow2', _('QCOW2 - QEMU Emulator')),
('raw', _('Raw')),
('vdi', _('VDI')),
('vhd', _('VHD')),
('vmdk', _('VMDK'))
]
}
# The IMAGE_CUSTOM_PROPERTY_TITLES settings is used to customize the titles for
# image custom property attributes that appear on image detail pages.
IMAGE_CUSTOM_PROPERTY_TITLES = {
"architecture": _("Architecture"),
"kernel_id": _("Kernel ID"),
"ramdisk_id": _("Ramdisk ID"),
"image_state": _("Euca2ools state"),
"project_id": _("Project ID"),
"image_type": _("Image Type")
}
# OPENSTACK_ENDPOINT_TYPE specifies the endpoint type to use for the endpoints
# in the Keystone service catalog. Use this setting when Horizon is running
# external to the OpenStack environment. The default is 'publicURL'.
#OPENSTACK_ENDPOINT_TYPE = "publicURL"
{% if primary_endpoint -%}
OPENSTACK_ENDPOINT_TYPE = "{{ primary_endpoint }}"
{% endif -%}
# SECONDARY_ENDPOINT_TYPE specifies the fallback endpoint type to use in the
# case that OPENSTACK_ENDPOINT_TYPE is not present in the endpoints
# in the Keystone service catalog. Use this setting when Horizon is running
# external to the OpenStack environment. The default is None. This
# value should differ from OPENSTACK_ENDPOINT_TYPE if used.
#SECONDARY_ENDPOINT_TYPE = "publicURL"
{% if secondary_endpoint -%}
SECONDARY_ENDPOINT_TYPE = "{{ secondary_endpoint }}"
{% endif -%}
# The number of objects (Swift containers/objects or images) to display
# on a single page before providing a paging element (a "more" link)
# to paginate results.
API_RESULT_LIMIT = 1000
API_RESULT_PAGE_SIZE = 20
# The timezone of the server. This should correspond with the timezone
# of your entire OpenStack installation, and hopefully be in UTC.
TIME_ZONE = "UTC"
# When launching an instance, the menu of available flavors is
# sorted by RAM usage, ascending. Provide a callback method here
# (and/or a flag for reverse sort) for the sorted() method if you'd
# like a different behaviour. For more info, see
# http://docs.python.org/2/library/functions.html#sorted
# CREATE_INSTANCE_FLAVOR_SORT = {
# 'key': my_awesome_callback_method,
# 'reverse': False,
# }
# The Horizon Policy Enforcement engine uses these values to load per service
# policy rule files. The content of these files should match the files the
# OpenStack services are using to determine role based access control in the
# target installation.
# Path to directory containing policy.json files
#POLICY_FILES_PATH = os.path.join(ROOT_PATH, "conf")
# Map of local copy of service policy files
#POLICY_FILES = {
# 'identity': 'keystone_policy.json',
# 'compute': 'nova_policy.json'
#}
# Trove user and database extension support. By default support for
# creating users and databases on database instances is turned on.
# To disable these extensions set the permission here to something
# unusable such as ["!"].
# TROVE_ADD_USER_PERMS = []
# TROVE_ADD_DATABASE_PERMS = []
LOGGING = {
'version': 1,
# When set to True this will disable all logging except
# for loggers specified in this configuration dictionary. Note that
# if nothing is specified here and disable_existing_loggers is True,
# django.db.backends will still log unless it is disabled explicitly.
'disable_existing_loggers': False,
'handlers': {
'null': {
'level': 'DEBUG',
'class': 'django.utils.log.NullHandler',
},
'console': {
# Set the level to "DEBUG" for verbose output logging.
'level': 'INFO',
'class': 'logging.StreamHandler',
},
{% if use_syslog %}
'syslog': {
'level': 'INFO',
'class': 'logging.handlers.SysLogHandler',
}
{% endif %}
},
'loggers': {
# Logging from django.db.backends is VERY verbose, send to null
# by default.
'django.db.backends': {
'handlers': ['null'],
'propagate': False,
},
'requests': {
'handlers': ['null'],
'propagate': False,
},
'horizon': {
{% if use_syslog %}
'handlers': ['syslog'],
{% else %}
'handlers': ['console'],
{% endif %}
'propagate': False,
},
'openstack_dashboard': {
{% if use_syslog %}
'handlers': ['syslog'],
{% else %}
'handlers': ['console'],
{% endif %}
'propagate': False,
},
'openstack_auth': {
{% if use_syslog %}
'handlers': ['syslog'],
{% else %}
'handlers': ['console'],
{% endif %}
'propagate': False,
},
'novaclient': {
{% if use_syslog %}
'handlers': ['syslog'],
{% else %}
'handlers': ['console'],
{% endif %}
'propagate': False,
},
'cinderclient': {
{% if use_syslog %}
'handlers': ['syslog'],
{% else %}
'handlers': ['console'],
{% endif %}
'propagate': False,
},
'keystoneclient': {
{% if use_syslog %}
'handlers': ['syslog'],
{% else %}
'handlers': ['console'],
{% endif %}
'propagate': False,
},
'glanceclient': {
{% if use_syslog %}
'handlers': ['syslog'],
{% else %}
'handlers': ['console'],
{% endif %}
'propagate': False,
},
'heatclient': {
{% if use_syslog %}
'handlers': ['syslog'],
{% else %}
'handlers': ['console'],
{% endif %}
'propagate': False,
},
'nose.plugins.manager': {
{% if use_syslog %}
'handlers': ['syslog'],
{% else %}
'handlers': ['console'],
{% endif %}
'propagate': False,
}
}
}
SECURITY_GROUP_RULES = {
'all_tcp': {
'name': 'ALL TCP',
'ip_protocol': 'tcp',
'from_port': '1',
'to_port': '65535',
},
'all_udp': {
'name': 'ALL UDP',
'ip_protocol': 'udp',
'from_port': '1',
'to_port': '65535',
},
'all_icmp': {
'name': 'ALL ICMP',
'ip_protocol': 'icmp',
'from_port': '-1',
'to_port': '-1',
},
'ssh': {
'name': 'SSH',
'ip_protocol': 'tcp',
'from_port': '22',
'to_port': '22',
},
'smtp': {
'name': 'SMTP',
'ip_protocol': 'tcp',
'from_port': '25',
'to_port': '25',
},
'dns': {
'name': 'DNS',
'ip_protocol': 'tcp',
'from_port': '53',
'to_port': '53',
},
'http': {
'name': 'HTTP',
'ip_protocol': 'tcp',
'from_port': '80',
'to_port': '80',
},
'pop3': {
'name': 'POP3',
'ip_protocol': 'tcp',
'from_port': '110',
'to_port': '110',
},
'imap': {
'name': 'IMAP',
'ip_protocol': 'tcp',
'from_port': '143',
'to_port': '143',
},
'ldap': {
'name': 'LDAP',
'ip_protocol': 'tcp',
'from_port': '389',
'to_port': '389',
},
'https': {
'name': 'HTTPS',
'ip_protocol': 'tcp',
'from_port': '443',
'to_port': '443',
},
'smtps': {
'name': 'SMTPS',
'ip_protocol': 'tcp',
'from_port': '465',
'to_port': '465',
},
'imaps': {
'name': 'IMAPS',
'ip_protocol': 'tcp',
'from_port': '993',
'to_port': '993',
},
'pop3s': {
'name': 'POP3S',
'ip_protocol': 'tcp',
'from_port': '995',
'to_port': '995',
},
'ms_sql': {
'name': 'MS SQL',
'ip_protocol': 'tcp',
'from_port': '1443',
'to_port': '1443',
},
'mysql': {
'name': 'MYSQL',
'ip_protocol': 'tcp',
'from_port': '3306',
'to_port': '3306',
},
'rdp': {
'name': 'RDP',
'ip_protocol': 'tcp',
'from_port': '3389',
'to_port': '3389',
},
}
FLAVOR_EXTRA_KEYS = {
'flavor_keys': [
('quota:read_bytes_sec', _('Quota: Read bytes')),
('quota:write_bytes_sec', _('Quota: Write bytes')),
('quota:cpu_quota', _('Quota: CPU')),
('quota:cpu_period', _('Quota: CPU period')),
('quota:inbound_average', _('Quota: Inbound average')),
('quota:outbound_average', _('Quota: Outbound average')),
]
}
{% if ubuntu_theme %}
# Enable the Ubuntu theme if it is present.
try:
from ubuntu_theme import *
except ImportError:
pass
{% endif %}
# Default Ubuntu apache configuration uses /horizon as the application root.
# Configure auth redirects here accordingly.
{% if webroot == "/" %}
LOGIN_URL='/auth/login/'
LOGOUT_URL='/auth/logout/'
{% else %}
LOGIN_URL='{{ webroot }}/auth/login/'
LOGOUT_URL='{{ webroot }}/auth/logout/'
{% endif %}
LOGIN_REDIRECT_URL='{{ webroot }}'
# The Ubuntu package includes pre-compressed JS and compiled CSS to allow
# offline compression by default. To enable online compression, install
# the node-less package and enable the following option.
COMPRESS_OFFLINE = {{ compress_offline }}
# By default, validation of the HTTP Host header is disabled. Production
# installations should have this set accordingly. For more information
# see https://docs.djangoproject.com/en/dev/ref/settings/.
ALLOWED_HOSTS = '*'
{% if password_retrieve %}
OPENSTACK_ENABLE_PASSWORD_RETRIEVE = True
{% endif %}
{{ settings|join('\n\n') }}
| [
"kataguruma1"
] | kataguruma1 |
cde91956a9edb08f6b4f838f37d9d754e7956ca6 | d78c504f800a7fc305db5a1e84fe303f97a5f3f4 | /14-0 Weather.py | 98d1f1ca145a59c9c8ac73455b9184447d63b248 | [] | no_license | InnaKr/Exercises | e2c44dc06e1e561cd00e31d3715e503901c58e1f | d9d4531d3c7f8aebf1e1f9c0645843569e6f07c4 | refs/heads/master | 2020-05-18T19:52:10.336057 | 2019-06-30T15:59:05 | 2019-06-30T15:59:05 | 184,618,845 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 1,049 | py | import requests
# city = "Macerata"
city = input('Введіть місто(En): ')
appid = "47212252e5b0293b1e53348aa194bdb3"
while True:
try:
res = requests.get("http://api.openweathermap.org/data/2.5/weather",
params={'q': city, 'units': 'metric', 'lang': 'ua', 'APPID': appid})
data = res.json()
print('Місто: ', data['name'],'('+str(data['sys']['country'])+')', '\n')
print("* Хмарність:", data['weather'][0]['description'])
print("* Температура повітря:", str(data['main']['temp']) + '°С')
print("* Температура повітря від ", str(data['main']['temp_min'])+'°С', 'до', str(data['main']['temp_max'])+'°С')
print('* Вологість: ', str(data['main']['humidity']) + '%', )
print('* Вітер: ', str(data['wind']['speed']) + 'km/год', )
break
except Exception:
print("Такого міста немає")
city = input('Введіть місто(En): ')
| [
"49992989+InnaKr@users.noreply.github.com"
] | 49992989+InnaKr@users.noreply.github.com |
00afe15515e8406d7267839d7d8a4be3bccea3fa | 1dbbb05b30d27c6419b9f34eea3b9a47f92582a0 | /parlai/zoo/sea/bart_base.py | 44e3581dd73c1b7ad168a64f76a5a09e3c7c18f6 | [
"MIT"
] | permissive | facebookresearch/ParlAI | 815334323d0ebef51bf9837336fe3eef6fe1655d | e1d899edfb92471552bae153f59ad30aa7fca468 | refs/heads/main | 2023-08-31T22:20:45.918129 | 2023-08-14T19:39:56 | 2023-08-14T19:39:56 | 89,266,735 | 10,943 | 2,395 | MIT | 2023-09-13T23:07:40 | 2017-04-24T17:10:44 | Python | UTF-8 | Python | false | false | 741 | py | #!/usr/bin/env python3
# Copyright (c) Facebook, Inc. and its affiliates.
# This source code is licensed under the MIT license found in the
# LICENSE file in the root directory of this source tree.
"""
Vanila BART-Large 400m parameter model with no retrieval.
"""
from parlai.core.build_data import built, download_models, get_model_dir
import os
import os.path
def download(datapath):
ddir = os.path.join(get_model_dir(datapath), 'sea')
model_type = 'bart_base'
version = 'v1.0'
if not built(os.path.join(ddir, model_type), version):
opt = {'datapath': datapath, 'model_type': model_type}
fnames = [f'model_{version}.tgz']
download_models(opt, fnames, 'sea', version=version, use_model_type=True)
| [
"noreply@github.com"
] | noreply@github.com |
dc3187b27ca89c6c79e0c886c4bcc3dea7baa847 | 6f707329e5ef1bb56e74126a1ded82252013807b | /donation_site/urls.py | 51acd5a85b2f6bfe73b9b6b75c2d48992267333d | [] | no_license | DPBefumo/donation-site | dbe612745953805c521596bbd9c7ec7310dd4f5b | 5ae49a6d1d992b1c2b9f9b0ebb149382a70d1711 | refs/heads/main | 2022-12-25T12:06:29.106759 | 2020-10-07T20:18:23 | 2020-10-07T20:18:23 | null | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 619 | py | from django.contrib import admin
from django.conf import settings
from django.urls import include, path
from core import views as core_views
from django.conf.urls.static import static
urlpatterns = [
path('admin/', admin.site.urls),
path('', core_views.index, name='index'),
path('donate/', core_views.donate, name='donate'),
path('success/<str:args>', core_views.successPage, name='success'),
]
if settings.DEBUG:
import debug_toolbar
urlpatterns = [
path('__debug__/', include(debug_toolbar.urls)),
] + urlpatterns + static(settings.MEDIA_URL, document_root=settings.MEDIA_ROOT)
| [
"DPBefumo@gmail.com"
] | DPBefumo@gmail.com |
d9d33fe8b116cb7dc30454d84a7a1097e2401020 | ad1e55b9a67c798cf4b4ce41c76b26977f8b4e8d | /vendor-local/celery/tests/test_utils/test_datastructures.py | c06e4d732e8b0cdfa46db88f55237d3206f51bd9 | [
"BSD-3-Clause"
] | permissive | kumar303/rockit | 7a6ac84bb8c37e5f3b65d7dcecf9b9c549902cf5 | fc347b5b143835ddd77fd0c1ea4e6f2007a21972 | refs/heads/master | 2021-01-10T19:51:30.638073 | 2020-07-26T19:00:37 | 2020-07-26T19:00:37 | 4,219,328 | 0 | 2 | BSD-3-Clause | 2020-07-26T19:00:38 | 2012-05-03T22:03:24 | Python | UTF-8 | Python | false | false | 8,435 | py | from __future__ import absolute_import
from __future__ import with_statement
import sys
from celery.datastructures import (ExceptionInfo, LRUCache, LimitedSet,
AttributeDict, DictAttribute,
ConfigurationView, DependencyGraph)
from celery.tests.utils import Case, WhateverIO
class Object(object):
pass
class test_DictAttribute(Case):
def test_get_set(self):
x = DictAttribute(Object())
x["foo"] = "The quick brown fox"
self.assertEqual(x["foo"], "The quick brown fox")
self.assertEqual(x["foo"], x.obj.foo)
self.assertEqual(x.get("foo"), "The quick brown fox")
self.assertIsNone(x.get("bar"))
with self.assertRaises(KeyError):
x["bar"]
def test_setdefault(self):
x = DictAttribute(Object())
self.assertEqual(x.setdefault("foo", "NEW"), "NEW")
self.assertEqual(x.setdefault("foo", "XYZ"), "NEW")
def test_contains(self):
x = DictAttribute(Object())
x["foo"] = 1
self.assertIn("foo", x)
self.assertNotIn("bar", x)
def test_items(self):
obj = Object()
obj.attr1 = 1
x = DictAttribute(obj)
x["attr2"] = 2
self.assertDictEqual(dict(x.iteritems()),
dict(attr1=1, attr2=2))
self.assertDictEqual(dict(x.items()),
dict(attr1=1, attr2=2))
class test_ConfigurationView(Case):
def setUp(self):
self.view = ConfigurationView({"changed_key": 1,
"both": 2},
[{"default_key": 1,
"both": 1}])
def test_setdefault(self):
self.assertEqual(self.view.setdefault("both", 36), 2)
self.assertEqual(self.view.setdefault("new", 36), 36)
def test_get(self):
self.assertEqual(self.view.get("both"), 2)
sp = object()
self.assertIs(self.view.get("nonexisting", sp), sp)
def test_update(self):
changes = dict(self.view.changes)
self.view.update(a=1, b=2, c=3)
self.assertDictEqual(self.view.changes,
dict(changes, a=1, b=2, c=3))
def test_contains(self):
self.assertIn("changed_key", self.view)
self.assertIn("default_key", self.view)
self.assertNotIn("new", self.view)
def test_repr(self):
self.assertIn("changed_key", repr(self.view))
self.assertIn("default_key", repr(self.view))
def test_iter(self):
expected = {"changed_key": 1,
"default_key": 1,
"both": 2}
self.assertDictEqual(dict(self.view.items()), expected)
self.assertItemsEqual(list(iter(self.view)),
expected.keys())
self.assertItemsEqual(self.view.keys(), expected.keys())
self.assertItemsEqual(self.view.values(), expected.values())
class test_ExceptionInfo(Case):
def test_exception_info(self):
try:
raise LookupError("The quick brown fox jumps...")
except LookupError:
exc_info = sys.exc_info()
einfo = ExceptionInfo(exc_info)
self.assertEqual(str(einfo), einfo.traceback)
self.assertIsInstance(einfo.exception, LookupError)
self.assertTupleEqual(einfo.exception.args,
("The quick brown fox jumps...", ))
self.assertTrue(einfo.traceback)
r = repr(einfo)
self.assertTrue(r)
class test_LimitedSet(Case):
def test_add(self):
s = LimitedSet(maxlen=2)
s.add("foo")
s.add("bar")
for n in "foo", "bar":
self.assertIn(n, s)
s.add("baz")
for n in "bar", "baz":
self.assertIn(n, s)
self.assertNotIn("foo", s)
def test_iter(self):
s = LimitedSet(maxlen=2)
items = "foo", "bar"
for item in items:
s.add(item)
l = list(iter(s))
for item in items:
self.assertIn(item, l)
def test_repr(self):
s = LimitedSet(maxlen=2)
items = "foo", "bar"
for item in items:
s.add(item)
self.assertIn("LimitedSet(", repr(s))
def test_clear(self):
s = LimitedSet(maxlen=2)
s.add("foo")
s.add("bar")
self.assertEqual(len(s), 2)
s.clear()
self.assertFalse(s)
def test_update(self):
s1 = LimitedSet(maxlen=2)
s1.add("foo")
s1.add("bar")
s2 = LimitedSet(maxlen=2)
s2.update(s1)
self.assertItemsEqual(list(s2), ["foo", "bar"])
s2.update(["bla"])
self.assertItemsEqual(list(s2), ["bla", "bar"])
s2.update(["do", "re"])
self.assertItemsEqual(list(s2), ["do", "re"])
def test_as_dict(self):
s = LimitedSet(maxlen=2)
s.add("foo")
self.assertIsInstance(s.as_dict(), dict)
class test_LRUCache(Case):
def test_expires(self):
limit = 100
x = LRUCache(limit=limit)
slots = list(xrange(limit * 2))
for i in slots:
x[i] = i
self.assertListEqual(x.keys(), list(slots[limit:]))
def test_least_recently_used(self):
x = LRUCache(3)
x[1], x[2], x[3] = 1, 2, 3
self.assertEqual(x.keys(), [1, 2, 3])
x[4], x[5] = 4, 5
self.assertEqual(x.keys(), [3, 4, 5])
# access 3, which makes it the last used key.
x[3]
x[6] = 6
self.assertEqual(x.keys(), [5, 3, 6])
x[7] = 7
self.assertEqual(x.keys(), [3, 6, 7])
def assertSafeIter(self, method, interval=0.01, size=10000):
from threading import Thread, Event
from time import sleep
x = LRUCache(size)
x.update(zip(xrange(size), xrange(size)))
class Burglar(Thread):
def __init__(self, cache):
self.cache = cache
self._is_shutdown = Event()
self._is_stopped = Event()
Thread.__init__(self)
def run(self):
while not self._is_shutdown.isSet():
try:
self.cache.data.popitem(last=False)
except KeyError:
break
self._is_stopped.set()
def stop(self):
self._is_shutdown.set()
self._is_stopped.wait()
self.join(1e10)
burglar = Burglar(x)
burglar.start()
try:
for _ in getattr(x, method)():
sleep(0.0001)
finally:
burglar.stop()
def test_safe_to_remove_while_iteritems(self):
self.assertSafeIter("iteritems")
def test_safe_to_remove_while_keys(self):
self.assertSafeIter("keys")
def test_safe_to_remove_while_itervalues(self):
self.assertSafeIter("itervalues")
def test_items(self):
c = LRUCache()
c.update(a=1, b=2, c=3)
self.assertTrue(c.items())
class test_AttributeDict(Case):
def test_getattr__setattr(self):
x = AttributeDict({"foo": "bar"})
self.assertEqual(x["foo"], "bar")
with self.assertRaises(AttributeError):
x.bar
x.bar = "foo"
self.assertEqual(x["bar"], "foo")
class test_DependencyGraph(Case):
def graph1(self):
return DependencyGraph([
("A", []),
("B", []),
("C", ["A"]),
("D", ["C", "B"]),
])
def test_repr(self):
self.assertTrue(repr(self.graph1()))
def test_topsort(self):
order = self.graph1().topsort()
# C must start before D
self.assertLess(order.index("C"), order.index("D"))
# and B must start before D
self.assertLess(order.index("B"), order.index("D"))
# and A must start before C
self.assertLess(order.index("A"), order.index("C"))
def test_edges(self):
self.assertListEqual(list(self.graph1().edges()),
["C", "D"])
def test_items(self):
self.assertDictEqual(dict(self.graph1().items()),
{"A": [], "B": [],
"C": ["A"], "D": ["C", "B"]})
def test_to_dot(self):
s = WhateverIO()
self.graph1().to_dot(s)
self.assertTrue(s.getvalue())
| [
"kumar.mcmillan@gmail.com"
] | kumar.mcmillan@gmail.com |
1bcf9326d84b2e3d2430d6b53b5c8bd59809b755 | 6a861335fc5d659abdbc8bd9163d924106271f8c | /addNewGTPRevScore.py | 09d4b50b18163530ac6a83b9fe21174683de55af | [] | no_license | krtin/ContrastiveSummaries | 2585b8c84f4d8aa5ffdbeffac65d91b226bf8c95 | fc6fad3ca6ba6b85074ebfa4cfb605eddbd68479 | refs/heads/master | 2020-03-28T06:46:06.244548 | 2018-09-07T21:10:10 | 2018-09-07T21:10:10 | 147,858,579 | 3 | 0 | null | null | null | null | UTF-8 | Python | false | false | 3,811 | py | import config
import pandas as pd
import pickle as pkl
import numpy as np
import os
import unicodedata
def is_number(s):
try:
float(s)
return True
except ValueError:
pass
try:
import unicodedata
unicodedata.numeric(s)
return True
except (TypeError, ValueError):
pass
return False
#get to the point smc scores
print('Reading SMC data and its probabilities')
smc_data = pd.read_csv(config.smc_final)
#print(len(smc_data))
prob_smc_data = pkl.load(open(config.new_gtp_smc_reverse_probfile, 'rb'))
#print(len(prob_smc_data))
if(len(smc_data)!=len(prob_smc_data)):
raise Exception("Number of generated probabilities and number of smcs are not equal")
else:
#print(smc_data.columns)
print('Adding GTP probabilities to smc data for INCORR padded token')
prob_smc_data = pd.DataFrame(prob_smc_data, columns=['counter', 'log_prob', 'avg_log_prob', 'gtp_target', 'gtp_output'])
#print(len(prob_smc_data))
#print(prob_smc_data['gtp_target'])
smc_data['new_rev_gtp_oov_words'] = prob_smc_data['gtp_target']==prob_smc_data['gtp_output']
smc_data['new_rev_gtp_log_prob'] = prob_smc_data['log_prob']
smc_data['new_rev_gtp_avg_log_prob'] = prob_smc_data['avg_log_prob']
#print(smc_data['gtp_log_prob'])
#print(prob_smc_data['log_prob'])
#print(smc_data['system'])
#print(prob_smc_data['gtp_target'])
smc_data.to_csv(config.smc_final, index=False)
print('Reading SMIC probability data')
probfiles = [f for f in os.listdir(config.new_gtp_smic_reverse_probdir) if os.path.isfile(os.path.join(config.new_gtp_smic_reverse_probdir, f))]
prob_smic_data = {}
for filename in probfiles:
filename_parts = filename.split('_')
if(len(filename_parts)==3):
filename_parts[2] = filename_parts[2].split('.')[0]
if(is_number(filename_parts[1]) and is_number(filename_parts[2])):
start = int(filename_parts[1])
end = int(filename_parts[2])
prob_smic_data[start] = pkl.load(open(os.path.join(config.new_gtp_smic_reverse_probdir, filename), 'rb'))
tmp_data = []
for key in sorted(prob_smic_data.iterkeys()):
#if(key>810000):
#tmp = pd.DataFrame(prob_smic_data[key], columns=['counter', 'gtp_log_prob', 'gtp_avg_log_prob', 'gtp_target', 'gtp_output'])
#print(tmp)
#print('Null count', len(tmp['counter'].isnull()))
#break
#print(key)
print(key, len(prob_smic_data[key]))
tmp_data.extend(prob_smic_data[key])
prob_smic_data = tmp_data
del tmp_data
print(len(prob_smic_data))
print('Reading SMIC data')
smic_data = pd.read_csv(config.smic_final)
#print(len(smic_data))
f = open(config.smicidfile_filtered_new_gtp, 'r')
smicids = f.read().split('\n')
f.close()
orig_len = len(smicids)
#prob_smic_data = prob_smic_data[0:orig_len]
smicids = smicids[0:len(prob_smic_data)]
cut_len = len(smicids)
print('Total available data: %d out of %d' % (cut_len, orig_len))
prob_smic_data = pd.DataFrame(prob_smic_data, columns=['counter', 'new_rev_gtp_log_prob', 'new_rev_gtp_avg_log_prob', 'gtp_target', 'gtp_output'])
prob_smic_data['smic_id'] = np.array(smicids).astype(int)
prob_smic_data['new_rev_oov_words'] = prob_smic_data['gtp_target']!=prob_smic_data['gtp_output']
del prob_smic_data['gtp_target']
del prob_smic_data['gtp_output']
del prob_smic_data['counter']
orig_smic_len = len(smic_data)
smic_data = pd.merge(smic_data, prob_smic_data, left_on='smic_id', right_on='smic_id', how='left')
prob_added_len = smic_data['new_rev_gtp_log_prob'].count()
if(prob_added_len!=(cut_len)):
raise Exception("Some ids are not matching during join")
if(len(smic_data)!=orig_smic_len):
raise Exception("Some issue in joining")
#print(smic_data)
smic_data.to_csv(config.smic_final, index=False)
| [
"kkumar15@CS.McGill.CA"
] | kkumar15@CS.McGill.CA |
f3a8746b07f7a4ed5bae29cf9753647930f33964 | 2f99edadfdea111b06000e01a4d15d7a5ec3264f | /15_python包.py | 6ea8368d1eb2295cc9df8b4e779e34431a0bb50b | [] | no_license | chenweilong1022/python- | 00dd1ffe0f612d8c573e93a5fe1e43eb7715f115 | e7251ab791af65589cf42cf2570286212f4636b3 | refs/heads/master | 2020-05-07T17:33:44.648974 | 2019-04-17T10:18:50 | 2019-04-17T10:18:50 | 180,731,375 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 104 | py |
from test01.package01 import package01
from test01.package02 import package02
package01()
package02() | [
"陈伟龙"
] | 陈伟龙 |
363f924942de71b7d00a8c1074ffc45f369e5ddd | d257d98b6f0116797d882191a85fbfc407b44a07 | /new submission/Core.py | 7e6cac5e3c020057298b1cb743c21e258730278b | [
"MIT"
] | permissive | AmrAshmawy2099/DL_Framework | dbe56d67dce78b5e679a10620f74a7ccc0b9e2d8 | a46e76936646e349f6e50852e1630362ae3d6d49 | refs/heads/main | 2023-02-20T00:14:29.866234 | 2021-01-25T11:17:13 | 2021-01-25T11:17:13 | 332,546,908 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 15,546 | py |
import numpy as np
import pandas as pd
#import matplotlib.pyplot as plt
from statistics import mean
import warnings
warnings.filterwarnings("error")
# foundation classes
###################################################
class SGD:
def __init__(self, parameters, alpha=0.1):
self.parameters = parameters
self.alpha = alpha
def zero(self):
for p in self.parameters:
p.grad.data *= 0
def step(self, zero=True):
for p in self.parameters:
p.data -= p.grad.data * self.alpha
if(zero):
p.grad.data *= 0
class Tensor (object):
def __init__(self,data,autograd=False, creators=None,creation_op=None,id=None):
self.data = np.array(data)
self.creators = creators
self.creation_op = creation_op
self.grad = None
self.autograd = autograd
self.children = {}
if(id is None):
id = np.random.randint(0,100000)
self.id = id
if(creators is not None):
for c in creators:
if(self.id not in c.children):
c.children[self.id] = 1
else:
c.children[self.id] += 1
def all_children_grads_accounted_for(self):
for id,cnt in self.children.items():
if(cnt != 0): return False
return True
def backward(self,grad=None, grad_origin=None):
if(grad is None):
grad = Tensor(np.ones_like(self.data))
if(self.autograd):
if(grad_origin is not None):
if(self.children[grad_origin.id] == 0):raise Exception("cannot backprop more than once")
else: self.children[grad_origin.id] -= 1
if(self.grad is None):
self.grad = grad
else:
self.grad += grad
if(self.creators is not None and(self.all_children_grads_accounted_for() or grad_origin is None)):
if(self.creation_op == "add"):
self.creators[0].backward(self.grad, self)
self.creators[1].backward(self.grad, self)
if(self.creation_op == "sub"):
new = Tensor(self.grad.data)
self.creators[0].backward(new, self)
new = Tensor(self.grad.__neg__().data)
self.creators[1].backward(new, self)
if(self.creation_op == "mul"):
new = self.grad * self.creators[1]
self.creators[0].backward(new , self)
new = self.grad * self.creators[0]
self.creators[1].backward(new, self)
if(self.creation_op == "mm"):
act = self.creators[0]
weights = self.creators[1]
new = self.grad.mm(weights.transpose())
act.backward(new)
new = self.grad.transpose().mm(act).transpose()
weights.backward(new)
if(self.creation_op == "transpose"):
self.creators[0].backward(self.grad.transpose())
if("sum" in self.creation_op):
dim = int(self.creation_op.split("_")[1])
ds = self.creators[0].data.shape[dim]
self.creators[0].backward(self.grad.expand(dim,ds))
if("expand" in self.creation_op):
dim = int(self.creation_op.split("_")[1])
self.creators[0].backward(self.grad.sum(dim))
if(self.creation_op == "neg"):
self.creators[0].backward(self.grad.__neg__())
if(self.creation_op == "sigmoid"):
ones = Tensor(np.ones_like(self.grad.data))
self.creators[0].backward(self.grad * (self * (ones - self)))
if(self.creation_op == "relu"):
self.creators[0].backward(self.grad * (self ))
if(self.creation_op == "tanh"):
ones = Tensor(np.ones_like(self.grad.data))
self.creators[0].backward(self.grad * (ones - (self * self)))
def __neg__(self):
if(self.autograd): return Tensor(self.data * -1,autograd=True, creators=[self],creation_op="neg")
return Tensor(self.data * -1)
def __add__(self, other):
if(self.autograd and other.autograd):
return Tensor(self.data + other.data, autograd=True,creators=[self,other], creation_op="add")
return Tensor(self.data + other.data)
def getdata(self):
return (self.data)
def __repr__(self):
return str(self.data.__repr__())
def __str__(self):
return str(self.data.__str__())
def __sub__(self, other):
if(self.autograd and other.autograd):
return Tensor(self.data - other.data,autograd=True,creators=[self,other],creation_op="sub")
return Tensor(self.data - other.data)
def __mul__(self, other):
if(self.autograd and other.autograd):
try:
return Tensor(self.data * other.data,autograd=True,creators=[self,other],creation_op="mul")
return Tensor(self.data * other.data)
except(RuntimeWarning):
other.data/=9875
if(self.autograd and other.autograd):
return Tensor(self.data * other.data,autograd=True,creators=[self,other],creation_op="mul")
return Tensor(self.data * other.data)
def sum(self, dim):
if(self.autograd):
return Tensor(self.data.sum(dim),autograd=True,creators=[self],creation_op="sum_"+str(dim))
return Tensor(self.data.sum(dim))
def expand(self, dim,copies):
trans_cmd = list(range(0,len(self.data.shape)))
trans_cmd.insert(dim,len(self.data.shape))
new_shape = list(self.data.shape) + [copies]
new_data = self.data.repeat(copies).reshape(new_shape)
new_data = new_data.transpose(trans_cmd)
if(self.autograd):
return Tensor(new_data,autograd=True,creators=[self],creation_op="expand_"+str(dim))
return Tensor(new_data)
def transpose(self):
if(self.autograd):return Tensor(self.data.transpose(),autograd=True,creators=[self],creation_op="transpose")
return Tensor(self.data.transpose())
def mm(self, x):
if(self.autograd):
try:
return Tensor(self.data.dot(x.data),autograd=True,creators=[self,x],creation_op="mm")
except(TypeError):
print(x.data)
print(x.data.shape)
print(self.data.shape)
return Tensor(self.data.dot(x.data))
def shape(self):
return self.data.shape
class Layer(object):
def __init__(self):
self.parameters = list()
def get_parameters(self):
return self.parameters
def change_Weights(self,para):
print( self.parameters )
print(type(self.parameters))
class Linear(Layer):
def __init__(self, n_inputs, n_outputs):
super().__init__()
W = np.random.randn(n_inputs, n_outputs)*np.sqrt(2.0/(n_inputs))
self.weight = Tensor(W, autograd=True)
self.bias = Tensor(np.zeros(n_outputs), autograd=True)
self.parameters.append(self.weight)
self.parameters.append(self.bias)
def forward(self, input):
return input.mm(self.weight)+self.bias.expand(0,len(input.data))
def change_Weights(self,para):
self.weight = Tensor(para, autograd=True)
self.parameters=[]
self.parameters.append(self.weight)
self.parameters.append(self.bias)
class MSELoss(Layer):
def __init__(self):
super().__init__()
def forward(self, pred, target):
return ((pred - target)*(pred - target)).sum(0)
class Sequential(Layer):
def __init__(self, layers=list()):
super().__init__()
self.layers = layers
def add(self, layer):
self.layers.append(layer)
def forward(self, input):
for layer in self.layers:
input = layer.forward(input)
return input
def set_parameters(self,para):
# print(isinstance(self.layers[0],Linear))#is instance to compare class tpes
for i in range (len(self.layers)):
#if(not(isinstance(self.layers[i],Tanh)or isinstance(self.layers[i],Sigmoid))):
self.layers[i].change_Weights(para[i])
def get_parameters(self):
params = list()
for l in self.layers:
params+= l.get_parameters()
return params
def validate(self,data,target):
criterion = MSELoss()
optim = SGD(parameters=model.get_parameters(),alpha=1)
loss_list=[]
for i in data:
pred = model.forward(i)
loss_list.append(criterion.forward(pred, target).getdata()[0])
return mean(loss_list)
def train(model,target,data,batch_no,alpha,validation_counter,validation_data,validation_target):
criterion = MSELoss()
optim = SGD(parameters=model.get_parameters(),alpha=alpha)
loss_list=[]
pred = model.forward(data)
loss = criterion.forward(pred, target)
loss.backward(Tensor(np.ones_like(loss.data)))
optim.step()
loss_list.append(loss.getdata()[0] )
validation_list=[]
counter=0
count=0
while((loss_list[-1])>=( .001) and count<100 ):
for i in range(batch_no):
pred = model.forward(data)
loss = criterion.forward(pred, target)
loss.backward(Tensor(np.ones_like(loss.data)))
optim.step()
loss_list.append(loss.getdata().sum(0))
count+=1
print(count)
if(len(loss_list)>20):
if(mean(loss_list[-10:-2])<=loss_list[-1] or (loss_list[-1]==loss_list[-2] and loss_list[-2]==loss_list[-3])):
print("training reached overflow")
return [loss_list,model]
## counter+=1
## if(counter==validation_counter):
## counter=0
## l=model.validate(validation_data,validation_target)
## if(l>loss_list[-1]):
## print("overfitting occured")
## return [loss_list,model]
return [loss_list,model]
def test(self,data):
return self.forward(data)
####################################################3
#activation functions syntax pass numpy array return a numpy array
#############################################
def hard_sigmoid(x):
l=np.zeros(len(x))
for i in range (len(x)):
if(x[i]>1): l[i]=1
elif(x[i]<=0): l[i]=0
else:l[i]=(x[i]+1)/2
return l
def softmax_function(x):
z = np.exp(x)
z_ = z/z.sum()
return z_
def leaky_relu_function(x):
if x<0:
return 0.01*x
else:
return x
def parametrized_relu_function(a,x):
if x<0:
return a*x
else:
return x
def elu_function(x, a):
if x<0:
return a*(np.exp(x)-1)
else:
return x
def swish_function(x):
return x/(1-np.exp(-x))
#############################################
# activation functions that call tensor
####################################### #####
def sigmoid(self):
try:
if(self.autograd):
return Tensor(1 / (1 + np.exp(-self.data)), autograd=True,creators=[self],creation_op="sigmoid")
return Tensor(1 / (1 + np.exp(-self.data)))
except(RuntimeWarning):
if(self.autograd):
return Tensor(np.array(.5), autograd=True,creators=[self],creation_op="sigmoid")
return Tensor(np.array(.5))
def tanh(self):
if(self.autograd):
return Tensor(np.tanh(self.data),autograd=True,creators=[self],creation_op="tanh")
return Tensor(np.tanh(self.data))
def relu1(self):
if(self.autograd):
return Tensor( np.maximum(0, self.data),autograd=True,creators=[self],creation_op="relu")
return Tensor(np.maximum(0, self.data))
def leaky_relu(self):
if(self.autograd):
return Tensor(np.maximum(self.data *.001, self.data),autograd=True,creators=[self],creation_op="relu")
return Tensor(np.maximum( self.data*.01, self.data))
########################################################
#activation functions classes
########################################################
class Sigmoid(Layer):
def __init__(self):
super().__init__()
def forward(self, input):
return sigmoid(input)
class Tanh(Layer):
def __init__(self):
super().__init__()
def forward(self, input):
return tanh(input)
class Relu(Layer):
def __init__(self):
super().__init__()
def forward(self, input):
return relu1(input)
class LeakyRelu(Layer):
def __init__(self):
super().__init__()
def forward(self, input):
return leaky_relu(input)
##################################################
#main part of code
################################################
#load teat data and train data
train_file = pd.read_csv('train.csv')
train_file = train_file.sample(frac=1)
train_set = train_file[0:600]
test_set = train_file[600:700]
validat_set=train_file[500:600]
#creat Y_train which have only the label Column of the train.csv file
Y_train = train_set["label"]
Y_test = test_set["label"]
Y_valid=validat_set["label"]
# creat X_train which have all columns of the train.csv file Except 'label' column
#X_train = train_file.drop(labels = ["label"],axis = 1)
X_train = train_set.drop(["label"],axis = 1)
X_test = test_set.drop(["label"],axis = 1)
X_valid=validat_set.drop(["label"],axis = 1)
#print(X_train)
print('-----------------------------')
# Normalize the data
X_train = X_train / 255.0
X_test = X_test / 255.0
X_valid=X_valid/255.0
# Reshape image in 3 dimensions (height = 28px, width = 28px , canal = 1)
#X_train = X_train.values.reshape(-1,28,28,1)
#X_test = X_test.values.reshape(-1,28,28,1)
#X_valid = X_valid.values.reshape(-1,28,28,1)
print('--------------- --------------')
arr2=np.array(Y_train.values)
print(arr2.shape)
##arr1=np.array(X_train)
##arr5=np.array(X_valid)
##print(type(arr1))
##print(arr1[0].shape)
##print(np.array([[0,0],[0,1],[1,0],[1,1]]).shape)
##print(np.array([[0,0],[0,1],[1,0],[1,1]])[0].shape)
def transformer(x):
arr=[]
for i in x:
arr.append(np.array([i]))
return arr
arr1=np.array((X_train.values))
arr2=np.array(transformer(Y_train.values))
arr3=np.array((X_test.values))
arr4=np.array(transformer(Y_test.values))
arr5=np.array((X_valid.values))
arr6=np.array(transformer(Y_valid.values))
print((arr1[0]))
print(arr2)
##print(np.array(Y_valid.values).shape)
##print(arr6[0].shape)
##print(arr6[1].shape)
##print(np.array(Y_valid.values)[2].shape)
##
##print(np.array([[0],[1],[1],[1]]).shape)
##print(np.array([[0],[1],[1],[1]])[0].shape)
##print(np.array([[0],[1],[1],[1]])[1].shape)
##print(np.array([[0],[1],[1],[1]])[2].shape)
##
data = Tensor(arr1, autograd=True)
target = Tensor(arr2, autograd=True)
model = Sequential([Linear(784,400),Tanh(),Linear(400,90),Tanh(),Linear(90,1),Tanh()])
epoch_no=20
[list1,model]=model.train(target,data,epoch_no,1,50,Tensor(arr5),Tensor(arr6))
print(list1)
print("***********************************")
print(model.test(Tensor(arr3[0])))
print(arr4[0])
print(model.test(Tensor(arr3[1])).sum(0))
print(arr4[1])
print(model.test(Tensor(arr3[2])))
print(arr4[2])
print(model.test(Tensor(arr3[3])))
print(arr4[3])
| [
"noreply@github.com"
] | noreply@github.com |
7240d4ecab15682787443877a38f85b61f1f032b | 46a09c35dda8f75d9df6af8db52f6e826e1fc1e1 | /main.py | 773769696b9b9bc14d292f4fe478bf0fabe7fe80 | [] | no_license | stevebottos/stocktwits-scraper | c4c675da029859e578b609fc8f452f346b404285 | 04609b9a33a94a989f3476fcaa7e742fcb179efb | refs/heads/master | 2023-03-13T06:50:19.636431 | 2021-03-04T07:20:42 | 2021-03-04T07:20:42 | 291,401,654 | 1 | 0 | null | null | null | null | UTF-8 | Python | false | false | 6,788 | py | """
* Needed to enable gmails "less secure apps" to allow the email portion to work
"""
import smtplib
import ssl
from email.mime.text import MIMEText
import os
import secrets
import requests
import json
from datetime import datetime
import pickle
import re
import pandas as pd
import pathlib
PATH = pathlib.Path(__file__).parent.absolute()
USERS = [
"Newsfilter",
"fla",
"cctranscripts"
]
EMAIL_LIST = ["MARA",
"NAK",
"FNKO",
"STAF",
"EMAN",
"BNED",
"TTOO",
"IDEX",
"LAC",
"OXLC",
"RNET",
"APHA",
"IEX",
"AYRO",
"GEVO",
"LTHM",
"SOLO",
"BLIN",
"FAMI",
"GRIL",
"TGC"
]
MONTHS = {1 : ["January", "Jan"],
2 : ["February", "Feb"],
3 : ["March"],
4 : ["April"],
5 : ["May"],
6 : ["June"],
7 : ["July"],
8 : ["August", "Aug"],
9 : ["September", "Sep"],
10: ["October", "Oct"],
11: ["November", "Nov"],
12: ["December", "Dec"]}
def find_catalysts(filename, headlines, search_terms:list, after_day=None):
headlines_filtered = headlines.loc[(headlines["message"].str.contains(search_terms[0])) | (headlines["message"].str.contains(search_terms[1]))]
with open(filename, "w+") as f:
if after_day:
for _, row in headlines_filtered.iterrows():
row_split = row["message"].split(" ")
messages = []
for i, r in enumerate(row_split):
if r in search_terms:
possible_num = row_split[i+1].replace(",", "").replace(".","")
try:
num = int(possible_num)
if num > after_day and num != 2021:
rows = row["message"] = row["message"].replace("\n", " ")
if len(row["message"]) >= 200:
rows = [rows[:200] + "\n\t|", rows[200:]]
else:
rows = [rows]
if rows not in messages:
rows_as_str = " ".join(rows)
s = row["tickers"] + "\t|" + rows_as_str
f.write(s)
f.write("\n----------------------------------------------------------------------------------------------------------------\n")
messages.append(rows)
except:
pass
def scrape(users_of_interest, tickers_of_interest):
scraped_messages = []
url_str = "https://api.stocktwits.com/api/2/streams/user/{}.json"
for user in users_of_interest:
url = url_str.format(user)
response = requests.request("GET", url)
parsed = json.loads(response.text)
messages = parsed["messages"]
for message in messages:
message_body = message["body"]
for ticker in tickers_of_interest:
search_key = "".join(("$", ticker, " "))
if search_key in message_body:
scraped_messages.append(message_body)
return scraped_messages
def format_email_message(email_content):
return "\n---\n".join(email_content)
def send_email(email_content):
msg = MIMEText(email_content)
msg['Subject'] = "STOCKTWITS ALERT"
msg['From'] = msg['To'] = secrets.USER
context=ssl.create_default_context()
with smtplib.SMTP("smtp.gmail.com", port=587) as smtp:
smtp.starttls(context=context)
smtp.login(secrets.USER, secrets.PASS)
smtp.send_message(msg)
def save_list_as_pickle(lst,fname='todays_entries.pkl'):
with open(os.path.join(PATH, fname), 'wb+') as f:
pickle.dump(lst, f)
def load_list_from_pickle(fname='todays_entries.pkl'):
with open(os.path.join(PATH, fname), 'rb') as f:
return pickle.load(f)
today = datetime.now()
if today.time().hour == 6 and today.time().minute <=10:
todays_entries = load_list_from_pickle()
info_dict = {}
info_dict["tickers"] = []
info_dict["message"] = []
for entry in todays_entries:
tickers = re.findall(r'[$][A-Za-z]+', entry)
for ticker in tickers:
info_dict["tickers"].append(ticker)
info_dict["message"].append(entry)
headlines = pd.DataFrame.from_dict(info_dict)
# To filter out the bullshit
headlines = headlines.loc[~headlines["message"].str.contains("Why")]
headlines = headlines.loc[~headlines["message"].str.contains("Stocks Moving")]
headlines = headlines.loc[~headlines["message"].str.contains("gainers")]
headlines = headlines.loc[~headlines["message"].str.contains("Gainers")]
headlines = headlines.loc[~headlines["message"].str.contains("movers")]
headlines = headlines.loc[~headlines["message"].str.contains("Movers")]
headlines = headlines.loc[~headlines["message"].str.contains("Trading Higher")]
headlines = headlines.loc[~headlines["message"].str.contains("Market Update")]
headlines.to_csv(os.path.join(PATH, "records", str(today).split(" ")[0]+".csv"))
todays_entries = []
save_list_as_pickle(todays_entries)
# The catalysts document
try:
filename = os.path.join(PATH, "catalysts", str(today).split(" ")[0]+".txt")
if not os.path.exists(filename):
open(filename, 'w+').close()
with open(filename, "a+") as f:
find_catalysts(filename,
headlines,
search_terms=MONTHS[today.month],
after_day=today.day)
except:
pass
todays_entries = load_list_from_pickle()
tickers_of_interest = set(secrets.WATCHLIST + EMAIL_LIST)
email_raw = scrape(USERS, tickers_of_interest)
email_raw_filtered = [email for email in email_raw if email not in todays_entries]
if len(email_raw_filtered):
todays_entries = email_raw_filtered + todays_entries
save_list_as_pickle(todays_entries)
# Only send an email if it contains one of the special tickers we're watching
email_messages = []
for ticker in EMAIL_LIST:
for entry in email_raw_filtered:
search_key = "".join(("$", ticker, " "))
if search_key in entry:
email_messages.append(entry)
if len(email_messages) > 0:
email_formatted = format_email_message(email_messages)
send_email(email_formatted)
print("SUCCESS")
| [
"steve.bottos@hotmail.ca"
] | steve.bottos@hotmail.ca |
1e6a1bb4cbd846a1560ef7c7610b9ebcc4805c88 | 0e8b844a713bdb29e47ab445e55e0470194f3fda | /venv/bin/isort | c01f421efccde16d7e0aa853783f244265ba6e87 | [] | no_license | Linazzi83/aplicacao4 | 77666efd124192298cf3db06704a164a33dc451b | 2a1d20b3c3bb2da618b533b6bdb79f72802cdc80 | refs/heads/master | 2020-07-20T15:14:02.049933 | 2019-09-05T19:11:46 | 2019-09-05T19:11:46 | 206,665,730 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 233 | #!/home/san/Dev/app4/venv/bin/python3.7
# -*- coding: utf-8 -*-
import re
import sys
from isort.main import main
if __name__ == '__main__':
sys.argv[0] = re.sub(r'(-script\.pyw?|\.exe)?$', '', sys.argv[0])
sys.exit(main())
| [
"gedegoliu@gmail.com"
] | gedegoliu@gmail.com | |
da2e6602833e17793bb83c4b4c77acb468f4abe2 | 850435d1ba6bd7c7a22f1a866af7ab75c49471b3 | /and-the-award-goes-to-master/extract/util.py | 90359e96560404c14899ca93231e3a018f995033 | [
"MIT"
] | permissive | csaund/formal-models | fcc93c7f6df0e4bf2a1f988cd9a6785122b92ae0 | 85d5f217fbbc05f373d4822093373665c256a3ee | refs/heads/master | 2020-04-17T09:14:37.197142 | 2019-04-17T14:28:37 | 2019-04-17T14:28:37 | 166,450,395 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 1,877 | py | import re
import os
import csv
import requests
from bs4 import BeautifulSoup
def only_ascii(string):
""" Returns only ASCII encoded characters """
return "".join([c for c in string if ord(c) < 127])
def check_headers(filename, fieldnames):
""" Validates input fields to file fields. """
with open(filename, 'r') as f:
header = csv.reader(f).next()
overlap = set(header).intersection(fieldnames)
if len(header) != len(overlap):
raise Exception("Input fields must match file fields.")
def write_to_csv(filename, results, isList = True):
""" Writes list of dictionaries to CSV file. """
fieldnames = results[0].keys() if isList else results.keys()
fieldnames = sorted(fieldnames)
with open(filename, 'ab+') as f:
writer = csv.DictWriter(f, fieldnames=fieldnames)
if os.path.getsize(filename) == 0:
writer.writeheader()
else:
check_headers(filename, fieldnames)
if isList:
writer.writerows(results)
else:
writer.writerow(results)
def retrieve_clean_response(href):
raw_html = requests.get(href)
raw_html = re.sub(r'\[[0-9A-Z]+\]', '', raw_html.content) # removes footnotes
dom_tree = BeautifulSoup(raw_html, "lxml")
# why do we do this?
for tag in dom_tree.find_all('sup'):
tag.replaceWith('')
# why do we do this?
for tag in dom_tree.find_all('td'):
if 'rowspan' in tag.attrs.keys():
tag.name = 'th'
return dom_tree
def find_first_number(string):
try:
return re.findall(r'\b\d+\b', string)[0]
except:
return ""
def find_nth(haystack, needle, n):
""" TBD. """
start = haystack.find(needle)
while start >= 0 and n > 1:
start = haystack.find(needle, start+len(needle))
n -= 1
return start | [
"tmartinovska117@gmail.com"
] | tmartinovska117@gmail.com |
5a1a215fc88b1c2d5c7a9729d348862c15461931 | b64687833bbbd206d871e5b20c73e5bf363c4995 | /crocs.py | 2ea505fc464101c7b928b4bbcbb3e5e9cd5a0f07 | [
"Apache-2.0"
] | permissive | barkinet/crocs | 462225eee0975c9240ec25ca1275e0f9dc991e00 | 7ab44d1eb45aac7b24ab64601255d9fb38049040 | refs/heads/master | 2020-12-02T09:12:18.988446 | 2017-07-09T20:09:52 | 2017-07-09T20:09:52 | null | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 5,547 | py | from random import choice, randint
from string import printable
import re
class RegexStr(object):
def __init__(self, value):
self.value = value
def invalid_data(self):
pass
def valid_data(self):
return self.value
def __str__(self):
return re.escape(self.value)
class RegexOperator(object):
# It may be interesting to have a base class Pattern
# that implements common methods with Group and Include, Exclude.
# Because these accept multiple arguments.
def __init__(self):
pass
def invalid_data(self):
pass
def valid_data(self):
pass
def encargs(self, args):
return [RegexStr(ind) if isinstance(ind, str) else ind
for ind in args]
def encstr(self, regex):
regex = RegexStr(regex) if isinstance(
regex, str) else regex
return regex
def test(self):
regex = str(self)
data = self.valid_data()
# It has to be search in order to work with ConsumeNext.
strc = re.search(regex, data)
print 'Regex;', regex
print 'Input:', data
print 'Group dict:', strc.groupdict()
print 'Group 0:', strc.group(0)
print 'Groups:', strc.groups()
def join(self):
return ''.join(map(lambda ind: str(ind), self.args))
def __str__(self):
pass
class NamedGroup(RegexOperator):
"""
Named groups.
(?P<name>...)
"""
def __init__(self, name, *args):
self.args = self.encargs(args)
self.name = name
def invalid_data(self):
pass
def valid_data(self):
return ''.join(map(lambda ind: \
ind.valid_data(), self.args))
def __str__(self):
return '(?P<%s>%s)' % (self.name, self.join())
class Group(RegexOperator):
"""
A normal group.
(abc).
"""
def __init__(self, *args):
self.args = self.encargs(args)
def invalid_data(self):
pass
def valid_data(self):
return ''.join(map(lambda ind: \
ind.valid_data(), self.args))
def __str__(self):
return '(%s)' % self.join()
class Times(RegexOperator):
"""
Match n, m times.
a{1, 3}
Note: The * and + are emulated by
Times(regex, 0) or Times(regex, 1)
"""
TEST_MAX = 10
def __init__(self, regex, min=0, max=''):
self.regex = self.encstr(regex)
self.min = min
self.max = max
def invalid_data(self):
pass
def valid_data(self):
count = randint(self.min, self.max
if self.max else self.TEST_MAX)
data = ''.join((self.regex.valid_data()
for ind in xrange(count)))
return data
def __str__(self):
return '%s{%s,%s}' % (self.regex,
self.min, self.max)
class ConsumeNext(RegexOperator):
"""
Lookbehind assertion.
(?<=...)
"""
def __init__(self, regex0, regex1):
self.regex0 = self.encstr(regex0)
self.regex1 = self.encstr(regex1)
def invalid_data(self):
pass
def valid_data(self):
return '%s%s' % (self.regex0.valid_data(),
self.regex1.valid_data())
def __str__(self):
return '(?<=%s)%s' % (self.regex0, self.regex1)
class ConsumeBack(RegexOperator):
"""
Lookahead assertion.
(?=...)
"""
def __init__(self, regex0, regex1):
self.regex0 = self.encstr(regex0)
self.regex1 = self.encstr(regex1)
def invalid_data(self):
pass
def valid_data(self):
return '%s%s' % (self.regex0.valid_data(),
self.regex1.valid_data())
def __str__(self):
return '%s(?=%s)' % (self.regex0, self.regex1)
class Seq(RegexOperator):
def __init__(self, start, end):
self.start = start
self.end = end
self.seq = [chr(ind) for ind in xrange(
ord(self.start), ord(self.end))]
def valid_data(self):
return ''.join(self.seq)
def __str__(self):
return '%s-%s' % (self.start, self.end)
class Include(RegexOperator):
"""
Sets.
[abc]
"""
def __init__(self, *args):
self.args = self.encargs(args)
def invalid_data(self):
pass
def valid_data(self):
chars = ''.join(map(lambda ind: \
ind.valid_data(), self.args))
char = choice(chars)
return char
def __str__(self):
return '[%s]' % self.join()
class Exclude(RegexOperator):
"""
Excluding.
[^abc]
"""
def __init__(self, *args):
self.args = self.encargs(args)
def invalid_data(self):
pass
def valid_data(self):
chars = ''.join(map(lambda ind: \
ind.valid_data(), self.args))
data = filter(lambda ind: \
not ind in chars, printable)
return choice(data)
def __str__(self):
return '[^%s]' % self.join()
class X(RegexOperator):
"""
The dot.
.
"""
TOKEN = '.'
def __init__(self):
pass
def invalid_data(self):
return ''
def valid_data(self):
char = choice(printable)
return char
def __str__(self):
return self.TOKEN
class Pattern(RegexOperator):
"""
Setup a pattern.
"""
def __init__(self, *args):
self.args = self.encargs(args)
def invalid_data(self):
pass
def valid_data(self):
return ''.join(map(lambda ind: \
ind.valid_data(), self.args))
def __str__(self):
return self.join()
| [
"ioliveira.id.uff.br"
] | ioliveira.id.uff.br |
02922e6762aefcb1ca578359e971f39a28df7916 | d0eb582894eff3c44e3de4bd50f571f9d9ab3a02 | /venv/lib/python3.7/site-packages/tox/config.py | 4d5e02be935189dbd4ec6eaf1666cc27c16ee562 | [
"MIT"
] | permissive | tdle94/app-store-scrapper | 159187ef3825213d40425215dd9c9806b415769e | ed75880bac0c9ef685b2c1bf57a6997901abface | refs/heads/master | 2022-12-20T21:10:59.621305 | 2020-10-28T00:32:21 | 2020-10-28T00:32:21 | 247,291,364 | 1 | 2 | MIT | 2022-12-08T03:53:08 | 2020-03-14T14:25:44 | Python | UTF-8 | Python | false | false | 57,366 | py | from __future__ import print_function
import argparse
import itertools
import os
import random
import re
import shlex
import string
import sys
import warnings
from collections import OrderedDict
from fnmatch import fnmatchcase
from subprocess import list2cmdline
import pkg_resources
import pluggy
import py
import toml
import tox
from tox.constants import INFO
from tox.interpreters import Interpreters, NoInterpreterInfo
hookimpl = tox.hookimpl
"""DEPRECATED - REMOVE - this is left for compatibility with plugins importing this from here.
Instead create a hookimpl in your code with:
import pluggy
hookimpl = pluggy.HookimplMarker("tox")
"""
default_factors = tox.PYTHON.DEFAULT_FACTORS
"""DEPRECATED MOVE - please update to new location."""
def get_plugin_manager(plugins=()):
# initialize plugin manager
import tox.venv
pm = pluggy.PluginManager("tox")
pm.add_hookspecs(tox.hookspecs)
pm.register(tox.config)
pm.register(tox.interpreters)
pm.register(tox.venv)
pm.register(tox.session)
from tox import package
pm.register(package)
pm.load_setuptools_entrypoints("tox")
for plugin in plugins:
pm.register(plugin)
pm.check_pending()
return pm
class Parser:
"""Command line and ini-parser control object."""
def __init__(self):
self.argparser = argparse.ArgumentParser(description="tox options", add_help=False)
self._testenv_attr = []
def add_argument(self, *args, **kwargs):
""" add argument to command line parser. This takes the
same arguments that ``argparse.ArgumentParser.add_argument``.
"""
return self.argparser.add_argument(*args, **kwargs)
def add_testenv_attribute(self, name, type, help, default=None, postprocess=None):
""" add an ini-file variable for "testenv" section.
Types are specified as strings like "bool", "line-list", "string", "argv", "path",
"argvlist".
The ``postprocess`` function will be called for each testenv
like ``postprocess(testenv_config=testenv_config, value=value)``
where ``value`` is the value as read from the ini (or the default value)
and ``testenv_config`` is a :py:class:`tox.config.TestenvConfig` instance
which will receive all ini-variables as object attributes.
Any postprocess function must return a value which will then be set
as the final value in the testenv section.
"""
self._testenv_attr.append(VenvAttribute(name, type, default, help, postprocess))
def add_testenv_attribute_obj(self, obj):
""" add an ini-file variable as an object.
This works as the ``add_testenv_attribute`` function but expects
"name", "type", "help", and "postprocess" attributes on the object.
"""
assert hasattr(obj, "name")
assert hasattr(obj, "type")
assert hasattr(obj, "help")
assert hasattr(obj, "postprocess")
self._testenv_attr.append(obj)
def parse_cli(self, args):
return self.argparser.parse_args(args)
def _format_help(self):
return self.argparser.format_help()
class VenvAttribute:
def __init__(self, name, type, default, help, postprocess):
self.name = name
self.type = type
self.default = default
self.help = help
self.postprocess = postprocess
class DepOption:
name = "deps"
type = "line-list"
help = "each line specifies a dependency in pip/setuptools format."
default = ()
def postprocess(self, testenv_config, value):
deps = []
config = testenv_config.config
for depline in value:
m = re.match(r":(\w+):\s*(\S+)", depline)
if m:
iname, name = m.groups()
ixserver = config.indexserver[iname]
else:
name = depline.strip()
ixserver = None
# we need to process options, in case they contain a space,
# as the subprocess call to pip install will otherwise fail.
# in case of a short option, we remove the space
for option in tox.PIP.INSTALL_SHORT_OPTIONS_ARGUMENT:
if name.startswith(option):
name = "{}{}".format(option, name[len(option) :].strip())
# in case of a long option, we add an equal sign
for option in tox.PIP.INSTALL_LONG_OPTIONS_ARGUMENT:
name_start = "{} ".format(option)
if name.startswith(name_start):
name = "{}={}".format(option, name[len(option) :].strip())
name = self._replace_forced_dep(name, config)
deps.append(DepConfig(name, ixserver))
return deps
def _replace_forced_dep(self, name, config):
"""Override given dependency config name. Take ``--force-dep-version`` option into account.
:param name: dep config, for example ["pkg==1.0", "other==2.0"].
:param config: ``Config`` instance
:return: the new dependency that should be used for virtual environments
"""
if not config.option.force_dep:
return name
for forced_dep in config.option.force_dep:
if self._is_same_dep(forced_dep, name):
return forced_dep
return name
@classmethod
def _is_same_dep(cls, dep1, dep2):
"""Definitions are the same if they refer to the same package, even if versions differ."""
dep1_name = pkg_resources.Requirement.parse(dep1).project_name
try:
dep2_name = pkg_resources.Requirement.parse(dep2).project_name
except pkg_resources.RequirementParseError:
# we couldn't parse a version, probably a URL
return False
return dep1_name == dep2_name
class PosargsOption:
name = "args_are_paths"
type = "bool"
default = True
help = "treat positional args in commands as paths"
def postprocess(self, testenv_config, value):
config = testenv_config.config
args = config.option.args
if args:
if value:
args = []
for arg in config.option.args:
if arg and not os.path.isabs(arg):
origpath = os.path.join(config.invocationcwd.strpath, arg)
if os.path.exists(origpath):
arg = os.path.relpath(origpath, testenv_config.changedir.strpath)
args.append(arg)
testenv_config._reader.addsubstitutions(args)
return value
class InstallcmdOption:
name = "install_command"
type = "argv"
default = "python -m pip install {opts} {packages}"
help = "install command for dependencies and package under test."
def postprocess(self, testenv_config, value):
if "{packages}" not in value:
raise tox.exception.ConfigError(
"'install_command' must contain '{packages}' substitution"
)
return value
def parseconfig(args, plugins=()):
"""Parse the configuration file and create a Config object.
:param plugins:
:param list[str] args: list of arguments.
:rtype: :class:`Config`
:raise SystemExit: toxinit file is not found
"""
pm = get_plugin_manager(plugins)
config, option = parse_cli(args, pm)
for config_file in propose_configs(option.configfile):
config_type = config_file.basename
content = None
if config_type == "pyproject.toml":
toml_content = get_py_project_toml(config_file)
try:
content = toml_content["tool"]["tox"]["legacy_tox_ini"]
except KeyError:
continue
ParseIni(config, config_file, content)
pm.hook.tox_configure(config=config) # post process config object
break
else:
msg = "tox config file (either {}) not found"
candidates = ", ".join(INFO.CONFIG_CANDIDATES)
feedback(msg.format(candidates), sysexit=not (option.help or option.helpini))
return config
def get_py_project_toml(path):
with open(str(path)) as file_handler:
config_data = toml.load(file_handler)
return config_data
def propose_configs(cli_config_file):
from_folder = py.path.local()
if cli_config_file is not None:
if os.path.isfile(cli_config_file):
yield py.path.local(cli_config_file)
return
if os.path.isdir(cli_config_file):
from_folder = py.path.local(cli_config_file)
else:
print(
"ERROR: {} is neither file or directory".format(cli_config_file), file=sys.stderr
)
return
for basename in INFO.CONFIG_CANDIDATES:
if from_folder.join(basename).isfile():
yield from_folder.join(basename)
for path in from_folder.parts(reverse=True):
ini_path = path.join(basename)
if ini_path.check():
yield ini_path
def parse_cli(args, pm):
parser = Parser()
pm.hook.tox_addoption(parser=parser)
option = parser.parse_cli(args)
if option.version:
print(get_version_info(pm))
raise SystemExit(0)
interpreters = Interpreters(hook=pm.hook)
config = Config(pluginmanager=pm, option=option, interpreters=interpreters, parser=parser)
return config, option
def feedback(msg, sysexit=False):
print("ERROR: {}".format(msg), file=sys.stderr)
if sysexit:
raise SystemExit(1)
def get_version_info(pm):
out = ["{} imported from {}".format(tox.__version__, tox.__file__)]
plugin_dist_info = pm.list_plugin_distinfo()
if plugin_dist_info:
out.append("registered plugins:")
for mod, egg_info in plugin_dist_info:
source = getattr(mod, "__file__", repr(mod))
out.append(" {}-{} at {}".format(egg_info.project_name, egg_info.version, source))
return "\n".join(out)
class SetenvDict(object):
_DUMMY = object()
def __init__(self, definitions, reader):
self.definitions = definitions
self.reader = reader
self.resolved = {}
self._lookupstack = []
def __repr__(self):
return "{}: {}".format(self.__class__.__name__, self.definitions)
def __contains__(self, name):
return name in self.definitions
def get(self, name, default=None):
try:
return self.resolved[name]
except KeyError:
try:
if name in self._lookupstack:
raise KeyError(name)
val = self.definitions[name]
except KeyError:
return os.environ.get(name, default)
self._lookupstack.append(name)
try:
self.resolved[name] = res = self.reader._replace(val)
finally:
self._lookupstack.pop()
return res
def __getitem__(self, name):
x = self.get(name, self._DUMMY)
if x is self._DUMMY:
raise KeyError(name)
return x
def keys(self):
return self.definitions.keys()
def __setitem__(self, name, value):
self.definitions[name] = value
self.resolved[name] = value
@tox.hookimpl
def tox_addoption(parser):
parser.add_argument(
"--version",
action="store_true",
dest="version",
help="report version information to stdout.",
)
parser.add_argument(
"-h", "--help", action="store_true", dest="help", help="show help about options"
)
parser.add_argument(
"--help-ini", "--hi", action="store_true", dest="helpini", help="show help about ini-names"
)
parser.add_argument(
"-v",
action="count",
dest="verbose_level",
default=0,
help="increase verbosity of reporting output."
"-vv mode turns off output redirection for package installation, "
"above level two verbosity flags are passed through to pip (with two less level)",
)
parser.add_argument(
"-q",
action="count",
dest="quiet_level",
default=0,
help="progressively silence reporting output.",
)
parser.add_argument(
"--showconfig",
action="store_true",
help="show configuration information for all environments. ",
)
parser.add_argument(
"-l",
"--listenvs",
action="store_true",
dest="listenvs",
help="show list of test environments (with description if verbose)",
)
parser.add_argument(
"-a",
"--listenvs-all",
action="store_true",
dest="listenvs_all",
help="show list of all defined environments (with description if verbose)",
)
parser.add_argument(
"-c",
action="store",
default=None,
dest="configfile",
help="config file name or directory with 'tox.ini' file.",
)
parser.add_argument(
"-e",
action="append",
dest="env",
metavar="envlist",
help="work against specified environments (ALL selects all).",
)
parser.add_argument(
"--notest", action="store_true", dest="notest", help="skip invoking test commands."
)
parser.add_argument(
"--sdistonly",
action="store_true",
dest="sdistonly",
help="only perform the sdist packaging activity.",
)
parser.add_argument(
"--parallel--safe-build",
action="store_true",
dest="parallel_safe_build",
help="(deprecated) ensure two tox builds can run in parallel "
"(uses a lock file in the tox workdir with .lock extension)",
)
parser.add_argument(
"--installpkg",
action="store",
default=None,
metavar="PATH",
help="use specified package for installation into venv, instead of creating an sdist.",
)
parser.add_argument(
"--develop",
action="store_true",
dest="develop",
help="install package in the venv using 'setup.py develop' via 'pip -e .'",
)
parser.add_argument(
"-i",
"--index-url",
action="append",
dest="indexurl",
metavar="URL",
help="set indexserver url (if URL is of form name=url set the "
"url for the 'name' indexserver, specifically)",
)
parser.add_argument(
"--pre",
action="store_true",
dest="pre",
help="install pre-releases and development versions of dependencies. "
"This will pass the --pre option to install_command "
"(pip by default).",
)
parser.add_argument(
"-r",
"--recreate",
action="store_true",
dest="recreate",
help="force recreation of virtual environments",
)
parser.add_argument(
"--result-json",
action="store",
dest="resultjson",
metavar="PATH",
help="write a json file with detailed information "
"about all commands and results involved.",
)
# We choose 1 to 4294967295 because it is the range of PYTHONHASHSEED.
parser.add_argument(
"--hashseed",
action="store",
metavar="SEED",
default=None,
help="set PYTHONHASHSEED to SEED before running commands. "
"Defaults to a random integer in the range [1, 4294967295] "
"([1, 1024] on Windows). "
"Passing 'noset' suppresses this behavior.",
)
parser.add_argument(
"--force-dep",
action="append",
metavar="REQ",
default=None,
help="Forces a certain version of one of the dependencies "
"when configuring the virtual environment. REQ Examples "
"'pytest<2.7' or 'django>=1.6'.",
)
parser.add_argument(
"--sitepackages",
action="store_true",
help="override sitepackages setting to True in all envs",
)
parser.add_argument(
"--alwayscopy", action="store_true", help="override alwayscopy setting to True in all envs"
)
cli_skip_missing_interpreter(parser)
parser.add_argument(
"--workdir",
action="store",
dest="workdir",
metavar="PATH",
default=None,
help="tox working directory",
)
parser.add_argument(
"args", nargs="*", help="additional arguments available to command positional substitution"
)
parser.add_testenv_attribute(
name="envdir",
type="path",
default="{toxworkdir}/{envname}",
help="set venv directory -- be very careful when changing this as tox "
"will remove this directory when recreating an environment",
)
# add various core venv interpreter attributes
def setenv(testenv_config, value):
setenv = value
config = testenv_config.config
if "PYTHONHASHSEED" not in setenv and config.hashseed is not None:
setenv["PYTHONHASHSEED"] = config.hashseed
setenv["TOX_ENV_NAME"] = str(testenv_config.envname)
setenv["TOX_ENV_DIR"] = str(testenv_config.envdir)
return setenv
parser.add_testenv_attribute(
name="setenv",
type="dict_setenv",
postprocess=setenv,
help="list of X=Y lines with environment variable settings",
)
def basepython_default(testenv_config, value):
"""either user set or proposed from the factor name
in both cases we check that the factor name implied python version and the resolved
python interpreter version match up; if they don't we warn, unless ignore base
python conflict is set in which case the factor name implied version if forced
"""
for factor in testenv_config.factors:
if factor in tox.PYTHON.DEFAULT_FACTORS:
implied_python = tox.PYTHON.DEFAULT_FACTORS[factor]
break
else:
implied_python, factor = None, None
if testenv_config.config.ignore_basepython_conflict and implied_python is not None:
return implied_python
proposed_python = (implied_python or sys.executable) if value is None else str(value)
if implied_python is not None and implied_python != proposed_python:
testenv_config.basepython = proposed_python
implied_version = tox.PYTHON.PY_FACTORS_RE.match(factor).group(2)
python_info_for_proposed = testenv_config.python_info
if not isinstance(python_info_for_proposed, NoInterpreterInfo):
proposed_version = "".join(
str(i) for i in python_info_for_proposed.version_info[0:2]
)
if implied_version != proposed_version:
# TODO(stephenfin): Raise an exception here in tox 4.0
warnings.warn(
"conflicting basepython version (set {}, should be {}) for env '{}';"
"resolve conflict or set ignore_basepython_conflict".format(
proposed_version, implied_version, testenv_config.envname
)
)
return proposed_python
parser.add_testenv_attribute(
name="basepython",
type="string",
default=None,
postprocess=basepython_default,
help="executable name or path of interpreter used to create a virtual test environment.",
)
def merge_description(testenv_config, value):
"""the reader by default joins generated description with new line,
replace new line with space"""
return value.replace("\n", " ")
parser.add_testenv_attribute(
name="description",
type="string",
default="",
postprocess=merge_description,
help="short description of this environment",
)
parser.add_testenv_attribute(
name="envtmpdir", type="path", default="{envdir}/tmp", help="venv temporary directory"
)
parser.add_testenv_attribute(
name="envlogdir", type="path", default="{envdir}/log", help="venv log directory"
)
parser.add_testenv_attribute(
name="downloadcache",
type="string",
default=None,
help="(ignored) has no effect anymore, pip-8 uses local caching by default",
)
parser.add_testenv_attribute(
name="changedir",
type="path",
default="{toxinidir}",
help="directory to change to when running commands",
)
parser.add_testenv_attribute_obj(PosargsOption())
parser.add_testenv_attribute(
name="skip_install",
type="bool",
default=False,
help="Do not install the current package. This can be used when you need the virtualenv "
"management but do not want to install the current package",
)
parser.add_testenv_attribute(
name="ignore_errors",
type="bool",
default=False,
help="if set to True all commands will be executed irrespective of their result error "
"status.",
)
def recreate(testenv_config, value):
if testenv_config.config.option.recreate:
return True
return value
parser.add_testenv_attribute(
name="recreate",
type="bool",
default=False,
postprocess=recreate,
help="always recreate this test environment.",
)
def passenv(testenv_config, value):
# Flatten the list to deal with space-separated values.
value = list(itertools.chain.from_iterable([x.split(" ") for x in value]))
passenv = {"PATH", "PIP_INDEX_URL", "LANG", "LANGUAGE", "LD_LIBRARY_PATH", "TOX_WORK_DIR"}
# read in global passenv settings
p = os.environ.get("TOX_TESTENV_PASSENV", None)
if p is not None:
env_values = [x for x in p.split() if x]
value.extend(env_values)
# we ensure that tmp directory settings are passed on
# we could also set it to the per-venv "envtmpdir"
# but this leads to very long paths when run with jenkins
# so we just pass it on by default for now.
if tox.INFO.IS_WIN:
passenv.add("SYSTEMDRIVE") # needed for pip6
passenv.add("SYSTEMROOT") # needed for python's crypto module
passenv.add("PATHEXT") # needed for discovering executables
passenv.add("COMSPEC") # needed for distutils cygwincompiler
passenv.add("TEMP")
passenv.add("TMP")
# for `multiprocessing.cpu_count()` on Windows (prior to Python 3.4).
passenv.add("NUMBER_OF_PROCESSORS")
passenv.add("PROCESSOR_ARCHITECTURE") # platform.machine()
passenv.add("USERPROFILE") # needed for `os.path.expanduser()`
passenv.add("MSYSTEM") # fixes #429
else:
passenv.add("TMPDIR")
for spec in value:
for name in os.environ:
if fnmatchcase(name.upper(), spec.upper()):
passenv.add(name)
return passenv
parser.add_testenv_attribute(
name="passenv",
type="line-list",
postprocess=passenv,
help="environment variables needed during executing test commands (taken from invocation "
"environment). Note that tox always passes through some basic environment variables "
"which are needed for basic functioning of the Python system. See --showconfig for the "
"eventual passenv setting.",
)
parser.add_testenv_attribute(
name="whitelist_externals",
type="line-list",
help="each lines specifies a path or basename for which tox will not warn "
"about it coming from outside the test environment.",
)
parser.add_testenv_attribute(
name="platform",
type="string",
default=".*",
help="regular expression which must match against ``sys.platform``. "
"otherwise testenv will be skipped.",
)
def sitepackages(testenv_config, value):
return testenv_config.config.option.sitepackages or value
def alwayscopy(testenv_config, value):
return testenv_config.config.option.alwayscopy or value
parser.add_testenv_attribute(
name="sitepackages",
type="bool",
default=False,
postprocess=sitepackages,
help="Set to ``True`` if you want to create virtual environments that also "
"have access to globally installed packages.",
)
parser.add_testenv_attribute(
name="alwayscopy",
type="bool",
default=False,
postprocess=alwayscopy,
help="Set to ``True`` if you want virtualenv to always copy files rather "
"than symlinking.",
)
def pip_pre(testenv_config, value):
return testenv_config.config.option.pre or value
parser.add_testenv_attribute(
name="pip_pre",
type="bool",
default=False,
postprocess=pip_pre,
help="If ``True``, adds ``--pre`` to the ``opts`` passed to the install command. ",
)
def develop(testenv_config, value):
option = testenv_config.config.option
return not option.installpkg and (value or option.develop)
parser.add_testenv_attribute(
name="usedevelop",
type="bool",
postprocess=develop,
default=False,
help="install package in develop/editable mode",
)
parser.add_testenv_attribute_obj(InstallcmdOption())
parser.add_testenv_attribute(
name="list_dependencies_command",
type="argv",
default="python -m pip freeze",
help="list dependencies for a virtual environment",
)
parser.add_testenv_attribute_obj(DepOption())
parser.add_testenv_attribute(
name="commands",
type="argvlist",
default="",
help="each line specifies a test command and can use substitution.",
)
parser.add_testenv_attribute(
name="commands_pre",
type="argvlist",
default="",
help="each line specifies a setup command action and can use substitution.",
)
parser.add_testenv_attribute(
name="commands_post",
type="argvlist",
default="",
help="each line specifies a teardown command and can use substitution.",
)
parser.add_testenv_attribute(
"ignore_outcome",
type="bool",
default=False,
help="if set to True a failing result of this testenv will not make "
"tox fail, only a warning will be produced",
)
parser.add_testenv_attribute(
"extras",
type="line-list",
help="list of extras to install with the source distribution or develop install",
)
def cli_skip_missing_interpreter(parser):
class SkipMissingInterpreterAction(argparse.Action):
def __call__(self, parser, namespace, values, option_string=None):
value = "true" if values is None else values
if value not in ("config", "true", "false"):
raise argparse.ArgumentTypeError("value must be config, true or false")
setattr(namespace, self.dest, value)
parser.add_argument(
"--skip-missing-interpreters",
default="config",
metavar="val",
nargs="?",
action=SkipMissingInterpreterAction,
help="don't fail tests for missing interpreters: {config,true,false} choice",
)
class Config(object):
"""Global Tox config object."""
def __init__(self, pluginmanager, option, interpreters, parser):
self.envconfigs = OrderedDict()
"""Mapping envname -> envconfig"""
self.invocationcwd = py.path.local()
self.interpreters = interpreters
self.pluginmanager = pluginmanager
self.option = option
self._parser = parser
self._testenv_attr = parser._testenv_attr
"""option namespace containing all parsed command line options"""
@property
def homedir(self):
homedir = get_homedir()
if homedir is None:
homedir = self.toxinidir # FIXME XXX good idea?
return homedir
class TestenvConfig:
"""Testenv Configuration object.
In addition to some core attributes/properties this config object holds all
per-testenv ini attributes as attributes, see "tox --help-ini" for an overview.
"""
def __init__(self, envname, config, factors, reader):
#: test environment name
self.envname = envname
#: global tox config object
self.config = config
#: set of factors
self.factors = factors
self._reader = reader
self.missing_subs = []
"""Holds substitutions that could not be resolved.
Pre 2.8.1 missing substitutions crashed with a ConfigError although this would not be a
problem if the env is not part of the current testrun. So we need to remember this and
check later when the testenv is actually run and crash only then.
"""
def get_envbindir(self):
"""Path to directory where scripts/binaries reside."""
if tox.INFO.IS_WIN and "jython" not in self.basepython and "pypy" not in self.basepython:
return self.envdir.join("Scripts")
else:
return self.envdir.join("bin")
@property
def envbindir(self):
return self.get_envbindir()
@property
def envpython(self):
"""Path to python executable."""
return self.get_envpython()
def get_envpython(self):
""" path to python/jython executable. """
if "jython" in str(self.basepython):
name = "jython"
else:
name = "python"
return self.envbindir.join(name)
def get_envsitepackagesdir(self):
"""Return sitepackagesdir of the virtualenv environment.
NOTE: Only available during execution, not during parsing.
"""
x = self.config.interpreters.get_sitepackagesdir(info=self.python_info, envdir=self.envdir)
return x
@property
def python_info(self):
"""Return sitepackagesdir of the virtualenv environment."""
return self.config.interpreters.get_info(envconfig=self)
def getsupportedinterpreter(self):
if tox.INFO.IS_WIN and self.basepython and "jython" in self.basepython:
raise tox.exception.UnsupportedInterpreter(
"Jython/Windows does not support installing scripts"
)
info = self.config.interpreters.get_info(envconfig=self)
if not info.executable:
raise tox.exception.InterpreterNotFound(self.basepython)
if not info.version_info:
raise tox.exception.InvocationError(
"Failed to get version_info for {}: {}".format(info.name, info.err)
)
return info.executable
testenvprefix = "testenv:"
def get_homedir():
try:
return py.path.local._gethomedir()
except Exception:
return None
def make_hashseed():
max_seed = 4294967295
if tox.INFO.IS_WIN:
max_seed = 1024
return str(random.randint(1, max_seed))
class ParseIni(object):
def __init__(self, config, ini_path, ini_data): # noqa
config.toxinipath = ini_path
config.toxinidir = config.toxinipath.dirpath()
self._cfg = py.iniconfig.IniConfig(config.toxinipath, ini_data)
config._cfg = self._cfg
self.config = config
prefix = "tox" if ini_path.basename == "setup.cfg" else None
context_name = getcontextname()
if context_name == "jenkins":
reader = SectionReader(
"tox:jenkins", self._cfg, prefix=prefix, fallbacksections=["tox"]
)
dist_share_default = "{toxworkdir}/distshare"
elif not context_name:
reader = SectionReader("tox", self._cfg, prefix=prefix)
dist_share_default = "{homedir}/.tox/distshare"
else:
raise ValueError("invalid context")
if config.option.hashseed is None:
hash_seed = make_hashseed()
elif config.option.hashseed == "noset":
hash_seed = None
else:
hash_seed = config.option.hashseed
config.hashseed = hash_seed
reader.addsubstitutions(toxinidir=config.toxinidir, homedir=config.homedir)
# As older versions of tox may have bugs or incompatibilities that
# prevent parsing of tox.ini this must be the first thing checked.
config.minversion = reader.getstring("minversion", None)
if config.minversion:
tox_version = pkg_resources.parse_version(tox.__version__)
config_min_version = pkg_resources.parse_version(self.config.minversion)
if config_min_version > tox_version:
raise tox.exception.MinVersionError(
"tox version is {}, required is at least {}".format(
tox.__version__, self.config.minversion
)
)
self.ensure_requires_satisfied(reader.getlist("requires"))
if config.option.workdir is None:
config.toxworkdir = reader.getpath("toxworkdir", "{toxinidir}/.tox")
else:
config.toxworkdir = config.toxinidir.join(config.option.workdir, abs=True)
if config.option.skip_missing_interpreters == "config":
val = reader.getbool("skip_missing_interpreters", False)
config.option.skip_missing_interpreters = "true" if val else "false"
config.ignore_basepython_conflict = reader.getbool("ignore_basepython_conflict", False)
# determine indexserver dictionary
config.indexserver = {"default": IndexServerConfig("default")}
prefix = "indexserver"
for line in reader.getlist(prefix):
name, url = map(lambda x: x.strip(), line.split("=", 1))
config.indexserver[name] = IndexServerConfig(name, url)
override = False
if config.option.indexurl:
for url_def in config.option.indexurl:
m = re.match(r"\W*(\w+)=(\S+)", url_def)
if m is None:
url = url_def
name = "default"
else:
name, url = m.groups()
if not url:
url = None
if name != "ALL":
config.indexserver[name].url = url
else:
override = url
# let ALL override all existing entries
if override:
for name in config.indexserver:
config.indexserver[name] = IndexServerConfig(name, override)
reader.addsubstitutions(toxworkdir=config.toxworkdir)
config.distdir = reader.getpath("distdir", "{toxworkdir}/dist")
reader.addsubstitutions(distdir=config.distdir)
config.distshare = reader.getpath("distshare", dist_share_default)
config.temp_dir = reader.getpath("temp_dir", "{toxworkdir}/.tmp")
reader.addsubstitutions(distshare=config.distshare)
config.sdistsrc = reader.getpath("sdistsrc", None)
config.setupdir = reader.getpath("setupdir", "{toxinidir}")
config.logdir = config.toxworkdir.join("log")
self.parse_build_isolation(config, reader)
config.envlist, all_envs = self._getenvdata(reader, config)
# factors used in config or predefined
known_factors = self._list_section_factors("testenv")
known_factors.update({"py", "python"})
# factors stated in config envlist
stated_envlist = reader.getstring("envlist", replace=False)
if stated_envlist:
for env in _split_env(stated_envlist):
known_factors.update(env.split("-"))
# configure testenvs
for name in all_envs:
section = "{}{}".format(testenvprefix, name)
factors = set(name.split("-"))
if (
section in self._cfg
or factors <= known_factors
or all(
tox.PYTHON.PY_FACTORS_RE.match(factor) for factor in factors - known_factors
)
):
config.envconfigs[name] = self.make_envconfig(name, section, reader._subs, config)
all_develop = all(
name in config.envconfigs and config.envconfigs[name].usedevelop
for name in config.envlist
)
config.skipsdist = reader.getbool("skipsdist", all_develop)
def parse_build_isolation(self, config, reader):
config.isolated_build = reader.getbool("isolated_build", False)
config.isolated_build_env = reader.getstring("isolated_build_env", ".package")
if config.isolated_build is True:
name = config.isolated_build_env
if name not in config.envconfigs:
config.envconfigs[name] = self.make_envconfig(
name, "{}{}".format(testenvprefix, name), reader._subs, config
)
@staticmethod
def ensure_requires_satisfied(specified):
missing_requirements = []
for s in specified:
try:
pkg_resources.get_distribution(s)
except pkg_resources.RequirementParseError:
raise
except Exception:
missing_requirements.append(str(pkg_resources.Requirement(s)))
if missing_requirements:
raise tox.exception.MissingRequirement(
"Packages {} need to be installed alongside tox in {}".format(
", ".join(missing_requirements), sys.executable
)
)
def _list_section_factors(self, section):
factors = set()
if section in self._cfg:
for _, value in self._cfg[section].items():
exprs = re.findall(r"^([\w{}\.!,-]+)\:\s+", value, re.M)
factors.update(*mapcat(_split_factor_expr_all, exprs))
return factors
def make_envconfig(self, name, section, subs, config, replace=True):
factors = set(name.split("-"))
reader = SectionReader(section, self._cfg, fallbacksections=["testenv"], factors=factors)
tc = TestenvConfig(name, config, factors, reader)
reader.addsubstitutions(
envname=name,
envbindir=tc.get_envbindir,
envsitepackagesdir=tc.get_envsitepackagesdir,
envpython=tc.get_envpython,
**subs
)
for env_attr in config._testenv_attr:
atype = env_attr.type
try:
if atype in ("bool", "path", "string", "dict", "dict_setenv", "argv", "argvlist"):
meth = getattr(reader, "get{}".format(atype))
res = meth(env_attr.name, env_attr.default, replace=replace)
elif atype == "space-separated-list":
res = reader.getlist(env_attr.name, sep=" ")
elif atype == "line-list":
res = reader.getlist(env_attr.name, sep="\n")
else:
raise ValueError("unknown type {!r}".format(atype))
if env_attr.postprocess:
res = env_attr.postprocess(testenv_config=tc, value=res)
except tox.exception.MissingSubstitution as e:
tc.missing_subs.append(e.name)
res = e.FLAG
setattr(tc, env_attr.name, res)
if atype in ("path", "string"):
reader.addsubstitutions(**{env_attr.name: res})
return tc
def _getenvdata(self, reader, config):
candidates = (
self.config.option.env,
os.environ.get("TOXENV"),
reader.getstring("envlist", replace=False),
)
env_str = next((i for i in candidates if i), [])
env_list = _split_env(env_str)
# collect section envs
all_envs = OrderedDict((i, None) for i in env_list)
if "ALL" in all_envs:
all_envs.pop("ALL")
for section in self._cfg:
if section.name.startswith(testenvprefix):
all_envs[section.name[len(testenvprefix) :]] = None
if not all_envs:
all_envs["python"] = None
package_env = config.isolated_build_env
if config.isolated_build is True and package_env in all_envs:
all_envs.pop(package_env)
if not env_list or "ALL" in env_list:
env_list = list(all_envs.keys())
if config.isolated_build is True and package_env in env_list:
msg = "isolated_build_env {} cannot be part of envlist".format(package_env)
raise tox.exception.ConfigError(msg)
all_env_list = list(all_envs.keys())
return env_list, all_env_list
def _split_env(env):
"""if handed a list, action="append" was used for -e """
if not isinstance(env, list):
env = [e.split("#", 1)[0].strip() for e in env.split("\n")]
env = ",".join([e for e in env if e])
env = [env]
return mapcat(_expand_envstr, env)
def _is_negated_factor(factor):
return factor.startswith("!")
def _base_factor_name(factor):
return factor[1:] if _is_negated_factor(factor) else factor
def _split_factor_expr(expr):
def split_single(e):
raw = e.split("-")
included = {_base_factor_name(factor) for factor in raw if not _is_negated_factor(factor)}
excluded = {_base_factor_name(factor) for factor in raw if _is_negated_factor(factor)}
return included, excluded
partial_envs = _expand_envstr(expr)
return [split_single(e) for e in partial_envs]
def _split_factor_expr_all(expr):
partial_envs = _expand_envstr(expr)
return [{_base_factor_name(factor) for factor in e.split("-")} for e in partial_envs]
def _expand_envstr(envstr):
# split by commas not in groups
tokens = re.split(r"((?:\{[^}]+\})+)|,", envstr)
envlist = ["".join(g).strip() for k, g in itertools.groupby(tokens, key=bool) if k]
def expand(env):
tokens = re.split(r"\{([^}]+)\}", env)
parts = [re.sub(r"\s+", "", token).split(",") for token in tokens]
return ["".join(variant) for variant in itertools.product(*parts)]
return mapcat(expand, envlist)
def mapcat(f, seq):
return list(itertools.chain.from_iterable(map(f, seq)))
class DepConfig:
def __init__(self, name, indexserver=None):
self.name = name
self.indexserver = indexserver
def __str__(self):
if self.indexserver:
if self.indexserver.name == "default":
return self.name
return ":{}:{}".format(self.indexserver.name, self.name)
return str(self.name)
__repr__ = __str__
class IndexServerConfig:
def __init__(self, name, url=None):
self.name = name
self.url = url
is_section_substitution = re.compile(r"{\[[^{}\s]+\]\S+?}").match
"""Check value matches substitution form of referencing value from other section.
E.g. {[base]commands}
"""
class SectionReader:
def __init__(self, section_name, cfgparser, fallbacksections=None, factors=(), prefix=None):
if prefix is None:
self.section_name = section_name
else:
self.section_name = "{}:{}".format(prefix, section_name)
self._cfg = cfgparser
self.fallbacksections = fallbacksections or []
self.factors = factors
self._subs = {}
self._subststack = []
self._setenv = None
def get_environ_value(self, name):
if self._setenv is None:
return os.environ.get(name)
return self._setenv.get(name)
def addsubstitutions(self, _posargs=None, **kw):
self._subs.update(kw)
if _posargs:
self.posargs = _posargs
def getpath(self, name, defaultpath, replace=True):
path = self.getstring(name, defaultpath, replace=replace)
if path is not None:
toxinidir = self._subs["toxinidir"]
return toxinidir.join(path, abs=True)
def getlist(self, name, sep="\n"):
s = self.getstring(name, None)
if s is None:
return []
return [x.strip() for x in s.split(sep) if x.strip()]
def getdict(self, name, default=None, sep="\n", replace=True):
value = self.getstring(name, None, replace=replace)
return self._getdict(value, default=default, sep=sep, replace=replace)
def getdict_setenv(self, name, default=None, sep="\n", replace=True):
value = self.getstring(name, None, replace=replace, crossonly=True)
definitions = self._getdict(value, default=default, sep=sep, replace=replace)
self._setenv = SetenvDict(definitions, reader=self)
return self._setenv
def _getdict(self, value, default, sep, replace=True):
if value is None or not replace:
return default or {}
d = {}
for line in value.split(sep):
if line.strip():
name, rest = line.split("=", 1)
d[name.strip()] = rest.strip()
return d
def getbool(self, name, default=None, replace=True):
s = self.getstring(name, default, replace=replace)
if not s or not replace:
s = default
if s is None:
raise KeyError("no config value [{}] {} found".format(self.section_name, name))
if not isinstance(s, bool):
if s.lower() == "true":
s = True
elif s.lower() == "false":
s = False
else:
raise tox.exception.ConfigError(
"{}: boolean value {!r} needs to be 'True' or 'False'".format(name, s)
)
return s
def getargvlist(self, name, default="", replace=True):
s = self.getstring(name, default, replace=False)
return _ArgvlistReader.getargvlist(self, s, replace=replace)
def getargv(self, name, default="", replace=True):
return self.getargvlist(name, default, replace=replace)[0]
def getstring(self, name, default=None, replace=True, crossonly=False):
x = None
for s in [self.section_name] + self.fallbacksections:
try:
x = self._cfg[s][name]
break
except KeyError:
continue
if x is None:
x = default
else:
# It is needed to apply factors before unwrapping
# dependencies, otherwise it can break the substitution
# process. Once they are unwrapped, we call apply factors
# again for those new dependencies.
x = self._apply_factors(x)
x = self._replace_if_needed(x, name, replace, crossonly)
x = self._apply_factors(x)
x = self._replace_if_needed(x, name, replace, crossonly)
return x
def _replace_if_needed(self, x, name, replace, crossonly):
if replace and x and hasattr(x, "replace"):
x = self._replace(x, name=name, crossonly=crossonly)
return x
def _apply_factors(self, s):
def factor_line(line):
m = re.search(r"^([\w{}\.!,-]+)\:\s+(.+)", line)
if not m:
return line
expr, line = m.groups()
if any(
included <= self.factors and not any(x in self.factors for x in excluded)
for included, excluded in _split_factor_expr(expr)
):
return line
lines = s.strip().splitlines()
return "\n".join(filter(None, map(factor_line, lines)))
def _replace(self, value, name=None, section_name=None, crossonly=False):
if "{" not in value:
return value
section_name = section_name if section_name else self.section_name
self._subststack.append((section_name, name))
try:
replaced = Replacer(self, crossonly=crossonly).do_replace(value)
assert self._subststack.pop() == (section_name, name)
except tox.exception.MissingSubstitution:
if not section_name.startswith(testenvprefix):
raise tox.exception.ConfigError(
"substitution env:{!r}: unknown or recursive definition in"
" section {!r}.".format(value, section_name)
)
raise
return replaced
class Replacer:
RE_ITEM_REF = re.compile(
r"""
(?<!\\)[{]
(?:(?P<sub_type>[^[:{}]+):)? # optional sub_type for special rules
(?P<substitution_value>(?:\[[^,{}]*\])?[^:,{}]*) # substitution key
(?::(?P<default_value>[^{}]*))? # default value
[}]
""",
re.VERBOSE,
)
def __init__(self, reader, crossonly=False):
self.reader = reader
self.crossonly = crossonly
def do_replace(self, value):
"""
Recursively expand substitutions starting from the innermost expression
"""
def substitute_once(x):
return self.RE_ITEM_REF.sub(self._replace_match, x)
expanded = substitute_once(value)
while expanded != value: # substitution found
value = expanded
expanded = substitute_once(value)
return expanded
def _replace_match(self, match):
g = match.groupdict()
sub_value = g["substitution_value"]
if self.crossonly:
if sub_value.startswith("["):
return self._substitute_from_other_section(sub_value)
# in crossonly we return all other hits verbatim
start, end = match.span()
return match.string[start:end]
# special case: all empty values means ":" which is os.pathsep
if not any(g.values()):
return os.pathsep
# special case: opts and packages. Leave {opts} and
# {packages} intact, they are replaced manually in
# _venv.VirtualEnv.run_install_command.
if sub_value in ("opts", "packages"):
return "{{{}}}".format(sub_value)
try:
sub_type = g["sub_type"]
except KeyError:
raise tox.exception.ConfigError(
"Malformed substitution; no substitution type provided"
)
if sub_type == "env":
return self._replace_env(match)
if sub_type == "tty":
if is_interactive():
return match.group("substitution_value")
return match.group("default_value")
if sub_type is not None:
raise tox.exception.ConfigError(
"No support for the {} substitution type".format(sub_type)
)
return self._replace_substitution(match)
def _replace_env(self, match):
key = match.group("substitution_value")
if not key:
raise tox.exception.ConfigError("env: requires an environment variable name")
default = match.group("default_value")
value = self.reader.get_environ_value(key)
if value is not None:
return value
if default is not None:
return default
raise tox.exception.MissingSubstitution(key)
def _substitute_from_other_section(self, key):
if key.startswith("[") and "]" in key:
i = key.find("]")
section, item = key[1:i], key[i + 1 :]
cfg = self.reader._cfg
if section in cfg and item in cfg[section]:
if (section, item) in self.reader._subststack:
raise ValueError(
"{} already in {}".format((section, item), self.reader._subststack)
)
x = str(cfg[section][item])
return self.reader._replace(
x, name=item, section_name=section, crossonly=self.crossonly
)
raise tox.exception.ConfigError("substitution key {!r} not found".format(key))
def _replace_substitution(self, match):
sub_key = match.group("substitution_value")
val = self.reader._subs.get(sub_key, None)
if val is None:
val = self._substitute_from_other_section(sub_key)
if callable(val):
val = val()
return str(val)
def is_interactive():
return sys.stdin.isatty()
class _ArgvlistReader:
@classmethod
def getargvlist(cls, reader, value, replace=True):
"""Parse ``commands`` argvlist multiline string.
:param SectionReader reader: reader to be used.
:param str value: Content stored by key.
:rtype: list[list[str]]
:raise :class:`tox.exception.ConfigError`:
line-continuation ends nowhere while resolving for specified section
"""
commands = []
current_command = ""
for line in value.splitlines():
line = line.rstrip()
if not line:
continue
if line.endswith("\\"):
current_command += " {}".format(line[:-1])
continue
current_command += line
if is_section_substitution(current_command):
replaced = reader._replace(current_command, crossonly=True)
commands.extend(cls.getargvlist(reader, replaced))
else:
commands.append(cls.processcommand(reader, current_command, replace))
current_command = ""
else:
if current_command:
raise tox.exception.ConfigError(
"line-continuation ends nowhere while resolving for [{}] {}".format(
reader.section_name, "commands"
)
)
return commands
@classmethod
def processcommand(cls, reader, command, replace=True):
posargs = getattr(reader, "posargs", "")
posargs_string = list2cmdline([x for x in posargs if x])
# Iterate through each word of the command substituting as
# appropriate to construct the new command string. This
# string is then broken up into exec argv components using
# shlex.
if replace:
newcommand = ""
for word in CommandParser(command).words():
if word == "{posargs}" or word == "[]":
newcommand += posargs_string
continue
elif word.startswith("{posargs:") and word.endswith("}"):
if posargs:
newcommand += posargs_string
continue
else:
word = word[9:-1]
new_arg = ""
new_word = reader._replace(word)
new_word = reader._replace(new_word)
new_word = new_word.replace("\\{", "{").replace("\\}", "}")
new_arg += new_word
newcommand += new_arg
else:
newcommand = command
# Construct shlex object that will not escape any values,
# use all values as is in argv.
shlexer = shlex.shlex(newcommand, posix=True)
shlexer.whitespace_split = True
shlexer.escape = ""
return list(shlexer)
class CommandParser(object):
class State(object):
def __init__(self):
self.word = ""
self.depth = 0
self.yield_words = []
def __init__(self, command):
self.command = command
def words(self):
ps = CommandParser.State()
def word_has_ended():
return (
(
cur_char in string.whitespace
and ps.word
and ps.word[-1] not in string.whitespace
)
or (cur_char == "{" and ps.depth == 0 and not ps.word.endswith("\\"))
or (ps.depth == 0 and ps.word and ps.word[-1] == "}")
or (cur_char not in string.whitespace and ps.word and ps.word.strip() == "")
)
def yield_this_word():
yieldword = ps.word
ps.word = ""
if yieldword:
ps.yield_words.append(yieldword)
def yield_if_word_ended():
if word_has_ended():
yield_this_word()
def accumulate():
ps.word += cur_char
def push_substitution():
ps.depth += 1
def pop_substitution():
ps.depth -= 1
for cur_char in self.command:
if cur_char in string.whitespace:
if ps.depth == 0:
yield_if_word_ended()
accumulate()
elif cur_char == "{":
yield_if_word_ended()
accumulate()
push_substitution()
elif cur_char == "}":
accumulate()
pop_substitution()
else:
yield_if_word_ended()
accumulate()
if ps.word.strip():
yield_this_word()
return ps.yield_words
def getcontextname():
if any(env in os.environ for env in ["JENKINS_URL", "HUDSON_URL"]):
return "jenkins"
return None
| [
"tuyenle@Admins-MacBook-Pro.local"
] | tuyenle@Admins-MacBook-Pro.local |
d870a37c961f3bdf556b4c8622bb56f0ade0454a | 400aa1b481650fee1d9e55ea0003384f5355808f | /main.py | b5e263a8f28c3550702e810f95b271958db28806 | [] | no_license | brunomarvss/moodmemoir | 13f34ff41002f10ae186ec5ae30a5cd905c7edad | a040b797ec0d56d81039f85515c1a49a2043591e | refs/heads/master | 2021-01-12T16:56:59.424680 | 2017-06-25T07:05:48 | 2017-06-25T07:05:48 | 71,473,361 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 188,216 | py | #kivy.require("1.8.1")
#Bruno
#Iringan
from kivy.app import App #create app
from kivy.uix.label import Label
from kivy.uix.floatlayout import FloatLayout #import float layout/ frame
from kivy.uix.gridlayout import GridLayout
from kivy.uix.boxlayout import BoxLayout
from kivy.core.window import Window
from kivy.uix.button import Button
from kivy.uix.label import Label
from kivy.uix.textinput import TextInput
from kivy.uix.carousel import Carousel
from kivy.uix.image import Image
from kivy.uix.popup import Popup
from kivy.uix.screenmanager import ScreenManager, Screen,SwapTransition,FadeTransition, SlideTransition
from kivy.uix.dropdown import DropDown
from kivy.uix.scrollview import ScrollView
import datetime
from garden import Graph, MeshLinePlot, SmoothLinePlot
from kivy.core.window import Window
from kivy.uix.label import Label
from decimal import*
from kivy.utils import get_color_from_hex as rgb
from math import sin
from textwrap import fill
from kivy.core.text import Label as CoreLabel
from functools import partial
from plyer import battery
from datetime import timedelta
from kivy.uix.slider import Slider
#naive Classifier module
import naive_classifier
import sqlite3
import re
import random
con = sqlite3.connect("moodJournal.db")
cursor = con.cursor()
#print(datetime.datetime.now().strftime('%Y-%m-%d %H:%M %p'))
#cursor.execute("DROP table if exists user_account")
#cursor.execute("CREATE TABLE user_info(id INTEGER PRIMARY KEY AUTOINCREMENT,thoughts_info text)")
#cursor.execute("INSERT INTO user_account(password) VALUES('8888')")
class NumRestrict(TextInput):
res = re.compile('[^0-9]')
def insert_text(self, substring, from_undo=False):
res = self.res
s = '%'.join([re.sub(res, '', s) for s in substring.split('%', 1)])
return super(NumRestrict, self).insert_text(s, from_undo=from_undo)
class MainApp(App): #Inherited Class App
def open_settings(self):
pass
def on_pause(self):
return True
def on_resume(self):
pass
def hook_keyboard(self,window,key,*largs):
if key==27:
return True
def build(self):
#Global Variables
global color_val
color_val = 'blue'
cursor.execute('SELECT COUNT(*) FROM color_tbl')
color_count = cursor.fetchone()[0]
con.commit()
if color_count == 0:
cursor.execute("INSERT INTO color_tbl(color) VALUES('"+color_val+"')")
con.commit()
elif color_count >= 1:
cursor.execute("SELECT color from color_tbl")
color_val = cursor.fetchone()[0]
con.commit()
global mood_var
global gN
global d_now
global dYear
global dMonth
global dDay
global wplot
global wplot2
global wplot3
global mood_value
wplot = SmoothLinePlot(mode="points",color=rgb('76abe6'))
wplot2 = SmoothLinePlot(mode="points",color=rgb('ff6666'))
wplot3 = SmoothLinePlot(mode="points",color=rgb('faff66'))
dYear = ' '
dMonth = ' '
dDay = '01'
mood_value = 0
#Window.size = (400,700)
sm = ScreenManager() #ManagesForms
sc1 = Screen(name='firstscreen')
sc2 = Screen(name ='secondscreen')
sc3 = Screen(name = 'thirdscreen')
scmain= Screen(name = 'mainscreen')
sc4 =Screen(name = 'fourthscreen')
sc5 = Screen(name='fifthscreen')
sc6 = Screen(name='sixthscreen')
sc7 = Screen(name='seventhscreen')
sm.add_widget(sc1)
sm.add_widget(scmain)
sm.add_widget(sc3)
sm.add_widget(sc2)
sm.add_widget(sc4)
sm.add_widget(sc5)
sm.add_widget(sc6)
sm.add_widget(sc7)
#--SCREEN1-------------------------------------
root= FloatLayout(size_hint = (1.0, 1.0))
frontBg = Image(pos= root.pos,
source="img/front.jpg",
allow_stretch=True,
keep_ratio=False)
mmlogo = Image(pos= root.pos,
source="img/mmfront.png", #img source
allow_stretch=False,
keep_ratio=False,
pos_hint={'x':0,'center_y': .7}, #scalable position
size_hint=(1,.9)) #scalable size
label_start = Label(text="[b]JOURNAL IT \nIN A SNAP[/b]\n[size=20sp]share what you feel \non the go[/size]",
pos_hint={'x':.1,'center_y': .35},
size_hint=(.5,.1),
markup=True,
font_size=(30))
bPopName = BoxLayout(orientation='vertical',padding=5,spacing = 5)
lblname= Label(text="[color=000000]Hi! What is your name?[/color]", markup=True,
font_size='10sp')
tbName=TextInput()
buttonPopName = Button(text="Confirm",
background_normal='img/blue.jpg')
popupname = Popup(title='',
background='img/white.jpg',
content=bPopName,
auto_dismiss=False,
size_hint=(0.8, 0.3))
#gender
bPopGender = BoxLayout(orientation='vertical',padding=5,spacing = 5)
lblgender= Label(text="[color=000000]Are you a Male or Female?[/color]",
markup=True,
font_size='10sp')
buttonPopGender1 = Button(text="Male",
background_normal='img/blue.jpg')
buttonPopGender2 = Button(text="Female",
background_normal='img/blue.jpg')
popupgender = Popup(title='',
background='img/white.jpg',
content=bPopGender,
auto_dismiss=False,
size_hint=(0.8, 0.3))
#help
bPopHelp = BoxLayout(orientation='vertical')
lblHelp= Label(text="[color=000000]<<< Swipe to left[/color]",
markup=True,
font_size='10sp')
tut1 = Image(source="img/tut1.jpg")
tut2 = Image(source="img/tut2.jpg")
tut3 = Image(source="img/tut3.jpg")
tut4 = Image(source="img/tut4.jpg")
carouselHelp = Carousel(direction='right')
btnHelp1 = Button(text="[color=ffffff]Okay, I got it![/color]",
markup=True,
pos_hint={'x':0.2,'center_y': 0.5},
size_hint=(0.6, 0.4))
popuphelp = Popup(title='',
background='img/white.jpg',
content=bPopHelp,
auto_dismiss=False,
size_hint=(0.9, 0.8))
def popgender():
bPopGender.add_widget(lblgender)
bPopGender.add_widget(buttonPopGender1)
bPopGender.add_widget(buttonPopGender2)
popupgender.open()
def popname():
cursor.execute("SELECT COUNT(*) from user_account")
if cursor.fetchone()[0] == 0:
bPopName.add_widget(lblname)
bPopName.add_widget(tbName)
bPopName.add_widget(buttonPopName)
popupname.open()
else:
pass
def screen1():
sm.transition = FadeTransition()
sc1.add_widget(root)
root.add_widget(frontBg)
root.add_widget(label_start)
root.add_widget(mmlogo)
screen1() #launches first screen
#-SCREEN2--------------------------------------
root2= FloatLayout(size_hint = (1.0, 1.0))
frontBg2 = Image(pos= root.pos,
source="img/lock.jpg",
allow_stretch=True,
keep_ratio=False)
#popoutpin--
bPopPin = GridLayout(cols=2,row=2,padding=10,spacing = 5)
lblPin= Label(text="[color=000000]Enter PIN[/color]",markup=True)
lblPin2= Label(text="[color=000000]Re-enter PIN[/color]",markup=True)
lblDum= Label(text=" ")
tbPin1 = NumRestrict(password = True,
password_mask ="*")
tbPin2 = NumRestrict(password = True,
password_mask ="*")
buttonPopPin = Button(text="CONFIRM",
background_normal='img/blue.jpg')
popupPin = Popup(title=' ',
content=bPopPin,
background='img/white.jpg',
auto_dismiss=False,
size_hint=(0.8, 0.3))
passimg1 = Button(pos= root.pos,
text="[color=ffffff]1[/color]",
font_size='30sp',
markup=True,
pos_hint={'x':0.2,'center_y': .6},
font_name='helv.otf',
background_color=[0,0,0,0.3],
size_hint=(0.15,0.1))
passimg2 = Button(pos= root.pos,
text="[color=ffffff]2[/color]",
font_size='30sp',
markup=True,
font_name='helv.otf',
background_color=[0,0,0,0.3],
size_hint=(0.15,0.1),
pos_hint={'x':0.45,'center_y': .6})
passimg3 = Button(pos= root.pos,
text="[color=ffffff]3[/color]",
font_size='30sp',
markup=True,
font_name='helv.otf',
background_color=[0,0,0,0.3],
pos_hint={'x':0.7,'center_y': .6},
size_hint=(0.15,0.1))
passimg4 = Button(pos= root.pos,
text="[color=ffffff]4[/color]",
font_size='30sp',
markup=True,
font_name='helv.otf',
background_color=[0,0,0,0.3],
pos_hint={'x':0.2,'center_y': .45},
size_hint=(0.15,0.1))
passimg5 = Button(pos= root.pos,
text="[color=ffffff]5[/color]",
font_size='30sp',
markup=True,
background_color=[0,0,0,0.3],
font_name='helv.otf',
pos_hint={'x':0.45,'center_y': .45},
size_hint=(0.15,0.1))
passimg6 = Button(pos= root.pos,
text="[color=ffffff]6[/color]",
font_size='30sp',
markup=True,
font_name='helv.otf',
background_color=[0,0,0,0.3],
pos_hint={'x':0.7,'center_y': .45},
size_hint=(0.15,0.1))
passimg7 = Button(pos= root.pos,
text="[color=ffffff]7[/color]",
font_size='30sp',
markup=True,
font_name='helv.otf',
background_color=[0,0,0,0.3],
pos_hint={'x':0.2,'center_y': .3},
size_hint=(0.15,0.1))
passimg8 = Button(pos= root.pos,
text="[color=ffffff]8[/color]",
font_size='30sp',
markup=True,
background_color=[0,0,0,0.3],
font_name='helv.otf',
pos_hint={'x':0.45,'center_y': .3},
size_hint=(0.15,0.1))
passimg9 = Button(pos= root.pos,
text="[color=ffffff]9[/color]",
font_size='30sp',
markup=True,
font_name='helv.otf',
background_color=[0,0,0,0.3],
pos_hint={'x':0.7,'center_y': .3},
size_hint=(0.15,0.1))
passimgclr = Button(pos= root.pos,
text="[color=ffffff]<[/color]",
font_size='25sp',
markup=True,
font_name='helv.otf',
background_color=[0,0,0,0.3],
pos_hint={'x':0.2,'center_y': .15},
size_hint=(0.15,0.1))
passimg0 = Button(pos= root.pos,
text="[color=ffffff]0[/color]",
font_size='30sp',
markup=True,
font_name='helv.otf',
background_color=[0,0,0,0.3],
pos_hint={'x':0.45,'center_y': .15},
size_hint=(0.15,0.1))
passimgok =Button(pos= root.pos,
text="[color=ffffff]OK[/color]",
font_name='helv.otf',
font_size='25sp',
markup=True,
background_color=[0,0,0,0.3],
pos_hint={'x':0.7,'center_y': .15},
size_hint=(0.15,0.1))
titlepin = Button(text="Enter PIN Code",
pos_hint={'x':0,'center_y': 0.96},
font_name='helv.otf',
background_color=[0,0,0,0.3],
font_size='25sp',
size_hint=(1,.09))
txtInp2 = NumRestrict(pos_hint={'x':.2,'center_y': .8},
font_name='helv.otf',
size_hint=(.65,.1),
foreground_color=[1,1,1,1],
background_color=[0,0,0,0.5],
font_size=(80),
password = True,
password_mask ="*",
readonly = True)
#--Screen3------------------------------------
root3= FloatLayout(size_hint = (1.0, 1.0))
frontBg3 = Image(pos= root.pos,
source="img/white.jpg",
allow_stretch=True,
keep_ratio=False)
b_sc3menu = Button(text="T",
font_name='IconsSouthSt.ttf',
font_size='20sp',
pos_hint={'x':0.02,'center_y': 0.96},
background_normal="img/"+color_val+".jpg",
size_hint=(.115,.09))
lblmood = Label(text="[color=000000]\nHow are you today?[/color]",
markup=True,
size_hint=(1, 0.3),
font_size='20sp')
bPop33 = BoxLayout()
bPop3= BoxLayout(orientation="vertical",
spacing=30,
padding=10)
popup3 = Popup(title="",
separator_color=[1,0,0,1],
background="img/white.jpg",
content=bPop3,
auto_dismiss=False,
size_hint=(0.9, 0.4))
#popoutSave--
bPopSave = BoxLayout(orientation='vertical',padding=5,spacing = 5)
lblsave= Label(text="[color=000000]Log saved![/color]",
font_size='15sp',
markup=True)
buttonPopSave = Button(text="Ok",
background_normal="img/"+color_val+".jpg")
popupsave = Popup(title='',
separator_color=[1,0,0,0.5],
background='img/white.jpg',
content=bPopSave,
auto_dismiss=False,
size_hint=(0.8, 0.3))
#-------
bHappy=Label(text="[color=ffdf00]m[/color]",
font_size='50sp',
font_name='smile.ttf',
markup=True,
size_hint=(0.2, 1),
background_color=[0,0,0,0],)
bHappy2=Label(text="[color=ffdf00]A[/color]",
font_size='50sp',
font_name='smile.ttf',
size_hint=(0.2, 1),
markup=True,
background_color=[0,0,0,0],)
bHappy3=Label(text="[color=ffdf00]C[/color]",
font_size='50sp',
font_name='smile.ttf',
size_hint=(0.2, 1),
markup=True,
background_color=[0,0,0,0],)
bSad=Label(text="[color=ffdf00]d[/color]",
font_size='50sp',
font_name='smile.ttf',
markup=True,
size_hint=(0.2,1),
background_color=[0,0,0,0],)
bAngry=Label(text="[color=ffdf00]W[/color]",
font_size='50sp',
font_name='smile.ttf',
markup=True,
size_hint=(0.2, 1),
background_color=[0,0,0,0])
b_sc3 = Button(text="[color=ffffff][b]SAVE LOG[b][/color]",
markup=True,
font_size='15sp',
background_normal="img/"+color_val+".jpg",
pos_hint={'x':.35,'center_y': .35},
size_hint=(.3,.06))
input_sc3 = TextInput(pos_hint={'x':.11,'center_y': .60},
size_hint=(.8,.3),
font_size='20sp')
titlewrite = Button(text="Write Log",
font_size='25sp',
pos_hint={'x':0,'center_y': 0.96},
background_normal="img/"+color_val+".jpg",
background_down="img/"+color_val+".jpg",
size_hint=(1,.09))
label_write = Label(text="[color=ffffff][b]How is your day today?[/b][/color]",
markup = True,
pos_hint={'x':.45,'center_y': .83},
size_hint=(.1,.1),
font_size='20sp',
size=(11,11))
btnSlider = Button(text='This is my mood',
markup = True,
background_normal="img/"+color_val+".jpg",
#font_name='helv.otf',
font_size='15sp',
size_hint=(1,1),)
def my_callback(instance, value):
global mood_value
#print('Current volume level: %0.2f' % value)
mood_value = str('%0.2f' % value)
tryslide=Slider(value_track=True,
min=-1,
max=1,
value_track_color=[1, 0, 0, 1],
size_hint=(1, 0.3),
)
tryslide.bind(value=my_callback)
#Screen6------
root6= FloatLayout(size_hint = (1.0, 1.0))
frontBg6 = Image(pos= root.pos,
source="img/coffee.jpg",
allow_stretch=True,
keep_ratio=False)
root6Title = Button(text="Home",
#font_name='helv.otf',
pos_hint={'x':0,'center_y': 0.96},
background_normal="img/"+color_val+".jpg",
background_down="img/"+color_val+".jpg",
font_size='25sp',
size_hint=(1,.09))
root6time = Label(text="[color=ffffff]"+datetime.datetime.now().strftime('%I:%M ')+"[/color]",
font_name='helv.otf',
markup=True,
pos_hint={'x':0.05,'center_y': 0.75},
font_size='100sp',
size_hint=(1,.095))
root6batt = Label(text="[color=ffffff][b]"+str(battery.status['percentage'])+"%[/b][/color]",
font_name='helv.otf',
markup=True,
pos_hint={'x':0.08,'center_y': 0.5825},
font_size='20sp',
size_hint=(1,.09))
root6lblbatt = Label(text="[color=ffffff]x[/color]",
font_name='IconsSouthSt.ttf',
markup=True,
pos_hint={'x':-0.05,'center_y': 0.6},
font_size='30sp',
size_hint=(1,.09))
root6lbldate = Label(text="[color=ffffff]"+datetime.datetime.now().strftime('%A %d %B %Y')+"[/color]",
font_name='helv.otf',
markup=True,
pos_hint={'x':0,'center_y': 0.63},
font_size='20sp',
size_hint=(1,.09))
#SCREEN7-------------------
root7= FloatLayout(size_hint = (1.0, 1.0))
frontBg7 = Image(pos= root.pos,
source="img/white.jpg",
allow_stretch=True,
keep_ratio=False)
b_sc7menu = Button(text="[color=000000]T[/color]",
markup=True,
font_name='IconsSouthSt.ttf',
font_size='20sp',
pos_hint={'x':0.02,'center_y': 0.96},
background_normal="img/white.jpg",
size_hint=(.115,.08))
userTitle = Button(text="[color=000000]User Garden[/color]",
markup=True,
pos_hint={'x':0,'center_y': 0.96},
font_name='helv.otf',
background_normal="img/white.jpg",
background_down="img/white.jpg",
font_size='25sp',
size_hint=(1,.09))
petalWhole = Image(pos= root.pos,
source="img/flowerone.jpg", #img source
allow_stretch=False,
keep_ratio=False,
pos_hint={'x':0.05,'center_y': .45}, #scalable position
size_hint=(.9,.7)) #scalable size
#--ScreenMain------------------------------------
#--sc5
root5= FloatLayout(size_hint = (1.0, 1.0))
frontBg5 = Image(pos= root.pos,
source="img/white.jpg",
allow_stretch=True,
keep_ratio=False)
b_sc5menu = Button(text="[color=000000]T[/color]",
markup=True,
font_name='IconsSouthSt.ttf',
font_size='20sp',
pos_hint={'x':0.02,'center_y': 0.96},
background_normal="img/white.jpg",
size_hint=(.115,.08))
recentTitle = Button(text="[color=000000]Logs History[/color]",
markup=True,
pos_hint={'x':0,'center_y': 0.96},
font_name='helv.otf',
background_normal="img/white.jpg",
background_down="img/white.jpg",
font_size='25sp',
size_hint=(1,.09))
grid5_history=GridLayout(cols=1,
spacing =20,
padding=30,
size_hint_y=None,
pos_hint={'center_x':.5, 'center_y':.5})
grid5_history.bind(minimum_height=grid5_history.setter('height'))
sv5_history=ScrollView(size_hint=(1,.75),
pos_hint={'center_x':.5, 'center_y':.45})
#scmain
rootmain= FloatLayout(size_hint = (1.0, 1.0))
grid_history=GridLayout(cols=1,
padding=30,
spacing =20,
size_hint_y=None,
pos_hint={'center_x':.5, 'center_y':.48})
grid_history.bind(minimum_height=grid_history.setter('height'))
sv_history=ScrollView(size_hint=(0.9,.8),
pos_hint={'center_x':.5, 'center_y':.45})
frontBgmain = Image(pos= root.pos,
source="img/trees.jpg",
allow_stretch=True,
keep_ratio=False)
b_scmainmenu = Button(text="T",
font_name='IconsSouthSt.ttf',
font_size='20sp',
pos_hint={'x':0.02,'center_y': 0.96},
background_normal="img/"+color_val+".jpg",
size_hint=(.115,.09))
#Get name from database Limit 1 for 1 user only
cursor.execute("SELECT COUNT(*) FROM user_account")
if cursor.fetchone()[0] == 0:
_name = 'guest'
else:
cursor.execute("SELECT name from user_account LIMIT 1")
_name = cursor.fetchone()[0]
con.commit()
bPop = BoxLayout(orientation='vertical',padding=5,spacing = 5)
cursor.execute("SELECT COUNT(*) FROM tipsTbl")
tips_count = cursor.fetchone()[0]
con.commit()
tips_n = random.randrange(1,tips_count)
cursor.execute("SELECT tips_info from tipsTbl WHERE tips_id="+str(tips_n) )
tips = cursor.fetchone()[0]
con.commit()
bPop.add_widget(Label(text="[color=000000]DAILY TIP\n\n"+str(tips)+"\n[/color]",
font_size='11sp',halign='center',
markup=True,
size_hint=(1,.8)))
buttonPop = Button(text="Close",
background_normal="img/roundblue.jpg",
size_hint=(1,.2))
bPop.add_widget(buttonPop)
popup = Popup(title='',separator_color=[1,0,0,0.5],
content=bPop,
background="img/white.jpg",
auto_dismiss=False,
size_hint=(0.8, 0.3))
#popoutexit--
bPopExit = BoxLayout(orientation='vertical',padding=5)
bPopExit2 = BoxLayout(orientation='horizontal', padding=5,spacing=10)
lblexit= Label(text="[color=000000]\n \n Are you sure you want to exit?[/color]",
markup=True,
font_size='15sp',size_hint=(1,.5))
buttonPopExit1 = Button(text="Yes",
size_hint=(.5,.3),
background_normal="img/roundblue.jpg")
buttonPopExit2 = Button(text="No",
size_hint=(.5,.3),
background_normal="img/roundblue.jpg")
popupexit = Popup(background="img/white.jpg",
separator_color=[1,0,0,0.5],
title="",
content=bPopExit,
auto_dismiss=True,
size_hint=(0.8, 0.3))
#popoutAbout--
bPopAb = GridLayout(cols=1,padding=5,spacing = 5)
lblabout= Label(text="[color=000000]Mood Memoir is a\n*Virtual Temperament Journal\n*Diary Archive \n*Monitor your moods"
+"[/color]",
markup=True,halign='center',
size_hint=(1,.5),
font_size='10sp')
buttonPopAb = Button(text="Close",
size_hint=(.5,.2),
background_normal="img/roundblue.jpg")
popupabout = Popup(background="img/white.jpg",
separator_color=[1,0,0,0.5],
content=bPopAb,
title="About Mood Memoir",
auto_dismiss=False,title_color=[0,0,0,1],
size_hint=(0.8, 0.4))
#popoutDev--
bPopD = GridLayout(cols=2,padding=5,spacing = 5)
lblDev1= Label(text="[color=000000]Jomari R. Iringan[/color]",
#font_name='helv.otf',
markup=True,
font_size='12sp',
size_hint=(1,.2))
lblDev2= Label(text="[color=000000]Lamberto A. Bruno, Jr.[/color]",
# font_name='helv.otf',
markup=True,
font_size='12sp')
imgBruno = Image(source="img/bruno.png",
size_hint=(1,1),
height=150,
allow_stretch=True)
imgJoms = Image(source="img/joms.png",
size_hint=(1,1),
allow_stretch=True)
lbldevblank = Label(size_hint=(.1,.1))
buttonPopD = Button(text="Close",
#font_name='helv.otf',
background_normal="img/roundblue.jpg",
size_hint=(.1,.2))
popupDev = Popup(background="img/white.jpg",
title="",
separator_color=[1,0,0,0.5],
content=bPopD,
auto_dismiss=False,
size_hint=(0.8, 0.6))
#popoutColor----
bPopColor= GridLayout(cols=1,padding=5,spacing = 5)
buttonColorRed = Button(text="Red",
# font_name='helv.otf',
background_normal="img/red.jpg",
size_hint=(.1,.2))
buttonColorBlue = Button(text="Blue",
#font_name='helv.otf',
background_normal="img/blue.jpg",
size_hint=(.1,.2))
buttonColorGreen = Button(text="Green",
#font_name='helv.otf',
background_normal="img/green.jpg",
size_hint=(.1,.2))
buttonColorOrange = Button(text="Orange",
# font_name='helv.otf',
background_normal="img/orange.jpg",
size_hint=(.1,.2))
buttonColorPurple = Button(text="Purple",
#font_name='helv.otf',
background_normal="img/purple.jpg",
size_hint=(.1,.2))
buttonColorPRed = Button(text="Pastel Red",
# font_name='helv.otf',
markup=True,
background_normal="img/pred.jpg",
size_hint=(.1,.2))
buttonColorPBlue = Button(text="Pastel Blue",
# font_name='helv.otf',
markup=True,
background_normal="img/pblue.jpg",
size_hint=(.1,.2))
buttonColorPGreen = Button(text="Pastel Green",
# font_name='helv.otf',
markup=True,
background_normal="img/pgreen.jpg",
size_hint=(.1,.2))
buttonColorPOrange = Button(text="Pastel Orange",
#font_name='helv.otf',
markup=True,
background_normal="img/porange.jpg",
size_hint=(.1,.2))
buttonColorPPurple = Button(text="Pastel Purple",
#font_name='helv.otf',
markup=True,
background_normal="img/ppurple.jpg",
size_hint=(.1,.2))
buttonPopColor = Button(text="Close",
# font_name='helv.otf',
background_normal="img/roundblue.jpg",
size_hint=(.1,.2))
popupColor = Popup(background="img/white.jpg",
title="",
separator_color=[1,0,0,0.5],
content=bPopColor,
auto_dismiss=False,
size_hint=(0.8, 0.8))
#popoutCP--
bPopCp = GridLayout(cols=2,row=2,padding=10,spacing = 5)
lblCp1= Label(text="[color=000000]Enter Old PIN[/color]",
markup=True,
size_hint=(1,.3))
#font_name='helv.otf')
lblCp2= Label(text="[color=000000]Enter New PIN[/color]",
size_hint=(1,.3),
markup=True)
#font_name='helv.otf')
lblCp3= Label(text="[color=000000]Re-enter New PIN[/color]",
size_hint=(1,.3),
markup=True)
#font_name='helv.otf')
lblblank= Label(text="",
#font_name='helv.otf',
size_hint=(1,.3))
lblnotice= Label(text="",
# font_name='helv.otf',
font_size='13sp',
size_hint=(1,.2))
tbCp1 = NumRestrict(password = True,
size_hint=(1,.3),
# font_name='helv.otf',
password_mask ="*")
tbCp2 = NumRestrict(password = True,
size_hint=(1,.3),
#font_name='helv.otf',
password_mask ="*")
tbCp3 = NumRestrict(password = True,
size_hint=(1,.3),
#font_name='helv.otf',
password_mask ="*")
buttonPopCpExit=Button(text="CLOSE",
# font_name='helv.otf',
size_hint=(.1,.2),
background_normal="img/roundblue.jpg")
buttonPopCp = Button(text="CONFIRM",
# font_name='helv.otf',
size_hint=(.1,.2),
background_normal="img/roundblue.jpg")
popupCp = Popup(title="",
separator_color=[1,0,0,0.5],
content=bPopCp,
background="img/white.jpg",
auto_dismiss=False,
size_hint=(0.8, 0.4))
#-------
b_logs = Button(text="[color=ffffff]D[/color]",
font_size='30sp',
font_name='trees.ttf',
pos_hint={'x':0.22,'center_y': 0.065},
background_color=[0,0,0,0],
markup=True,
size_hint=(.3,.1))
b_home = Label(text="[color=5e8ac9]D[/color]",
font_size='32sp',
font_name='IconsSouthSt.ttf',
pos_hint={'x':0.007,'center_y': 0.072},
markup=True,
size_hint=(.3,.1))
logsTitle = Button(text="Recent Logs",
#font_name='helv.otf',
pos_hint={'x':0,'center_y': 0.96},
background_normal="img/"+color_val+".jpg",
background_down="img/"+color_val+".jpg",
font_size='25sp',
size_hint=(1,.09))
lblHome = Label(text="[color=5e8ac9][b]Home[/b][/color]",
font_size='15sp',
#font_name='helv.otf',
pos_hint={'x':0.007,'center_y': 0.015},
markup=True,
size_hint=(.3,.1))
lblStat= Label(text="[color=ffffff]Stats[/color]",
font_size='15sp',
#font_name='helv.otf',
pos_hint={'x':0.74,'center_y': 0.015},
markup=True,
size_hint=(.3,.1))
lblLogs = Label(text="[color=ffffff]Logs[/color]",
font_size='15sp',
#font_name='helv.otf',
pos_hint={'x':0.22,'center_y': 0.015},
markup=True,
size_hint=(.3,.1))
lblWrite = Label(text="[color=ffffff]Write[/color]",
font_size='15sp',
#font_name='helv.otf',
pos_hint={'x':0.48,'center_y': 0.015},
markup=True,
size_hint=(.3,.1))
b_write = Button(text="[color=ffffff][b]h[/b][/color]",
font_size='29sp',
font_name='IconsSouthSt.ttf',
pos_hint={'x':0.5,'center_y': 0.068},
markup=True,
background_color=[0,0,0,0],
size_hint=(.3,.1))
b_stats = Button(text="[color=ffffff]W[/color]",
font_size="33sp",
font_name='IconsSouthSt.ttf',
pos_hint={'x':0.74,'center_y': 0.07},
markup=True,
background_color=[0,0,0,0],
size_hint=(.3,.1))
b_user= Button(text="[color=ffffff]W[/color]",
font_size="33sp",
font_name='IconsSouthSt.ttf',
pos_hint={'x':0.74,'center_y': 0.4},
markup=True,
background_color=[0,0,0,0],
size_hint=(.3,.1))
lblUser = Label(text="[color=ffffff]User[/color]",
font_size='15sp',
#font_name='helv.otf',
pos_hint={'x':0.74,'center_y': 0.35},
markup=True,
size_hint=(.3,.1))
dropdown = DropDown()
btnCP = Button(text='[color=ffffff]Change PIN[/color]',
#ont_name='helv.otf',
font_size='11sp',
markup = True,
background_normal = "img/"+color_val+".jpg",
size_hint=(1,None),
height=100)
btnCP.bind(on_release=lambda btn: dropdown.select(btn.text))
btnColor = Button(text='[color=ffffff]Color Scheme[/color]',
#font_name='helv.otf',
font_size='11sp',
markup = True,
background_normal = "img/"+color_val+".jpg",
size_hint=(1,None),
height=100)
btnColor.bind(on_release=lambda btn: dropdown.select(btn.text))
btnCons = Button(text='[color=ffffff]Consultant Mode[/color]',
#font_name='helv.otf',
font_size='11sp',
markup = True,
background_normal = "img/"+color_val+".jpg",
size_hint=(1,None),
height=100)
btnCons.bind(on_release=lambda btn: dropdown.select(btn.text))
btnHelp = Button(text='[color=ffffff]How-To[/color]',
#font_name='helv.otf',
font_size='15sp',
markup = True,
background_normal = "img/"+color_val+".jpg",
size_hint=(1,None),
height=100)
btnHelp.bind(on_release=lambda btn: dropdown.select(btn.text))
btnD = Button(text='[color=ffffff]Developers[/color]',
#font_name='helv.otf',
font_size='11sp',
markup = True,
background_normal = "img/"+color_val+".jpg",
size_hint=(1,None),
height=100)
btnD.bind(on_release=lambda btn: dropdown.select(btn.text))
btnA = Button(text='[color=ffffff]About[/color]',
font_size='11sp',
# font_name='helv.otf',
size_hint=(1,None),
markup = True,
background_normal = "img/"+color_val+".jpg",
height=100)
btnA.bind(on_release=lambda btn: dropdown.select(btn.text))
btnLO = Button(text='[color=ffffff]Log Out[/color]',
#font_name='helv.otf',
font_size='15sp',
markup = True,
background_normal = "img/"+color_val+".jpg",
size_hint=(1,None),
width=50,
height=100)
btnLO.bind(on_release=lambda btn: dropdown.select(btn.text))
# create a big main button
mainbutton = Button(text='[b]::[/b]',
markup=True,
font_size='20sp',
background_normal="img/"+color_val+".jpg",
size_hint=(.25,.085),
pos_hint={'x':0,'center_y': .96})
# show the dropdown menu when the main button is released
# note: all the bind() calls pass the instance of the caller (here, the
# mainbutton instance) as the first argument of the callback (here,
# dropdown.open.).
mainbutton.bind(on_press=dropdown.open)
mainbutton.bind(on_release=dropdown.open)
# one last thing, listen for the selection in the dropdown list and
# assign the data to the button text.
#dropdown.bind(on_select=lambda instance, x: setattr(mainbutton, 'text', x))
lbltest = Label(text="[color=ffffff]How are you today?[/color]",
markup = True)
#------Screen4--------------------------
root4= FloatLayout(size_hint = (1.0, 1.0))
frontBg4 = Image(pos= root.pos,
source="img/white.jpg",
allow_stretch=True,
keep_ratio=False)
lblDescript = Label(text="[color=000000]LEGEND\n[/color]"+"[color=76abe6]Blue line[/color]"+"[color=393d42] Positive Score\n"+"[color=ff0000]Red line[/color]"+" Negative Score[/color]",
font_size='11sp',
size_hint=(.20,.05),
markup=True,
pos_hint={'center_x':.18,'center_y':0.35})
lblTemp = Label(text="[color=000000]Result:[/color]"+"[color=393d42][/color]",
font_size='11sp',
size_hint=(.25,.05),
markup=True,pos_hint={'x':.37,'center_y':0.05}
)
petalG = Image(pos= root.pos,
source="img/flowerone.jpg", #img source
allow_stretch=False,
keep_ratio=False,
pos_hint={'x':0.05,'center_y': .45}, #scalable position
size_hint=(.9,.7)) #scalable size
boxStats = BoxLayout(size_hint=(0.85,0.4),
pos_hint={'center_x':.45, 'center_y':.6})
btnCont = Button(text=">",
size_hint=(0.08,0.05),
pos_hint={'center_x':.93, 'center_y':.6})
#boxStats.bind(minimum_width=boxStats.setter('width'))
gridData = GridLayout(cols=4,
padding=10,
spacing=30,
size_hint_y = None,
size_hint_x = 1,
pos_hint={'center_x':.6, 'center_y':.5})
gridData.bind(minimum_height=gridData.setter('height'))
scrollData=ScrollView(size_hint=(.52,.3),
pos_hint={'center_x':.7, 'center_y':.24})
#graph dropdown
graphDown = DropDown()
btnDaily = Button(text="[color=ffffff]Daily Stats[/color]",
font_size='14sp',
markup=True,
#font_name='helv.otf',
background_normal="img/"+color_val+".jpg",
size_hint=(1,None),
height=50)
btnDaily.bind(on_release=lambda btn: graphDown.select(btn.text))
btnWeekly = Button(text="[color=ffffff]Weekly Stats[/color]",
font_size='14sp',
#font_name='helv.otf',
markup=True,
background_normal="img/"+color_val+".jpg",
size_hint=(1,None),
height=50)
btnWeekly.bind(on_release=lambda btn: graphDown.select(btn.text))
btnMonthly = Button(text="[color=ffffff]Monthly Stats[/color]",
font_size='13sp',
#font_name='helv.otf',
markup=True,
background_normal="img/"+color_val+".jpg",
size_hint=(1,None),
height=50)
btnMonthly.bind(on_release=lambda btn: graphDown.select(btn.text))
btnAnnual = Button(text="[color=ffffff]Annual Stats[/color]",
font_size='14sp',
#font_name='helv.otf',
markup=True,
background_normal="img/"+color_val+".jpg",
size_hint=(1,None),
height=50)
btnAnnual.bind(on_release=lambda btn: graphDown.select(btn.text))
graphButton = Button(text='View Stats',
#font_name = 'helv.otf',
font_size='15sp',
background_normal="img/"+color_val+".jpg",
size_hint=(.25,.05),
pos_hint={'x':.7,'center_y':0.85})
graphButton.bind(on_press=graphDown.open)
graphButton.bind(on_release=graphDown.open)
b_sc4menu = Button(text="T",
font_name='IconsSouthSt.ttf',
font_size='20sp',
pos_hint={'x':0.02,'center_y': 0.96},
background_normal="img/"+color_val+".jpg",
size_hint=(.115,.09))
statsTitle = Button(text="User Garden",
#font_name = 'helv.otf',
pos_hint={'x':0,'center_y': 0.96},
background_normal="img/"+color_val+".jpg",
background_down="img/"+color_val+".jpg",
font_size='25sp',
size_hint=(1,.09))
btnHelpGraph = Button(text="[color=ffffff][b]What does the image mean?[/b][/color]",
markup=True,
background_normal="img/"+color_val+".jpg",
pos_hint={'x':0.15,'center_y': 0.06},
font_size='15sp',
size_hint=(0.7,.05))
popBoxLay = BoxLayout(orientation='vertical')
popuphelpme= Popup(title='',
separator_color=[0,0,1,0.5],
background='img/white.jpg',
content=popBoxLay,
auto_dismiss=False,
size_hint=(0.8, 0.6))
lblhelpme = Label(text="[color=000000] The image represents the positivity\n progress of you.\n The images has birds, flowers, petals\n (depends on your gender)\n that is total to 10% each, which will represent\n the positivity in terms of Daily, Weekly,\n Monthly and Annual progress.\n\n Ex: 8 birds/flowers/petals will represent\n 80% positivity.[/color]",
markup=True,
font_size='13sp')
btnhelpmeok = Button(text="Close",
background_normal="img/"+color_val+".jpg",
size_hint=(0.3, 0.1),
pos_hint={'x':0.355,'center_y': 0.01},
)
def helpme(self,*args):
popBoxLay.add_widget(lblhelpme)
popBoxLay.add_widget(btnhelpmeok)
popuphelpme.open()
def closehelpme(self,*args):
popBoxLay.remove_widget(lblhelpme)
popBoxLay.remove_widget(btnhelpmeok)
popuphelpme.dismiss()
btnHelpGraph.bind(on_press=helpme)
btnhelpmeok.bind(on_press=closehelpme)
#DayWidget
dropdownDay = DropDown()
btnd1 = Button(text='[color=ffffff]1[/color]',
#font_name = 'helv.otf',
markup=True,
font_size='12sp',
background_normal="img/"+color_val+".jpg",
size_hint=(1,None),
height=50)
btnd1.bind(on_release=lambda btn: dropdownDay.select(btn.text))
btnd2 = Button(text='[color=ffffff]2[/color]',
#font_name = 'helv.otf',
markup=True,
font_size='12sp',
background_normal="img/"+color_val+".jpg",
size_hint=(1,None),
height=50)
btnd2.bind(on_release=lambda btn: dropdownDay.select(btn.text))
btnDayDay= Button(text='[color=ffffff]Day v[/color]',
markup=True,
#font_name = 'helv.otf',
font_size='15sp',
background_normal="img/"+color_val+".jpg",
size_hint=(.13,.05),
pos_hint={'x':0.37,'center_y': 0.85})
dropdownYearDay=DropDown()
#dropdownYearDay.bind(on_select=lambda instance, x: setattr(btndy1, 'text', x))
btndy1 = Button(text='[color=ffffff]2016[/color]',
#font_name = 'helv.otf',
markup=True,
font_size='12sp',
background_normal="img/"+color_val+".jpg",
size_hint=(1,None),
height=50)
#btndy1.bind(on_release=lambda btn: dropdownYearDay.select(btn.text))
btndy2 = Button(text='[color=ffffff]2017[/color]',
# font_name = 'helv.otf',
markup=True,
font_size='12sp',
background_normal="img/"+color_val+".jpg",
size_hint=(1,None),
height=50)
btndy2.bind(on_release=lambda btn: dropdownYearDay.select(btn.text))
btnYearDay= Button(text='[color=ffffff]Year v[/color]',
markup=True,
#font_name = 'helv.otf',
font_size='12sp',
background_normal="img/"+color_val+".jpg",
size_hint=(.13,.05),
pos_hint={'x':0.05,'center_y': 0.85})
dropdownYearMonth=DropDown()
btnmy11 = Button(text='[color=ffffff]Jan[/color]',
#font_name = 'helv.otf',
markup=True,
font_size='12sp',
background_normal="img/"+color_val+".jpg",
size_hint=(1,None),
height=50)
btnmy11.bind(on_release=lambda btn: dropdownYearMonth.select(btn.text))
btnmy22 = Button(text='[color=ffffff]Feb[/color]',
#font_name = 'helv.otf',
markup=True,
font_size='12sp',
background_normal="img/"+color_val+".jpg",
size_hint=(1,None),
height=50)
btnmy22.bind(on_release=lambda btn: dropdownYearMonth.select(btn.text))
btnMonthDay= Button(text='[color=ffffff]Month v[/color]',
markup=True,
# font_name = 'helv.otf',
font_size='15sp',
background_normal="img/"+color_val+".jpg",
size_hint=(.15,.05),
pos_hint={'x':0.20,'center_y': 0.85})
btnDayDay.bind(on_press=dropdownDay.open)
btnDayDay.bind(on_release=dropdownDay.open)
btnYearDay.bind(on_press=dropdownYearDay.open)
btnYearDay.bind(on_release=dropdownYearDay.open)
btnMonthDay.bind(on_press=dropdownYearMonth.open)
btnMonthDay.bind(on_release=dropdownYearMonth.open)
#WeekWidgets
dropdownWeekDay = DropDown()
btnwd1 = Button(text='[color=ffffff]1[/color]',
#font_name = 'helv.otf',
markup=True,
font_size='12sp',
background_normal="img/"+color_val+".jpg",
size_hint=(1,None),
height=50)
btnwd1.bind(on_release=lambda btn: dropdownWeekDay.select(btn.text))
btnWeekDay= Button(text='[color=ffffff]Day v[/color]',
markup=True,
#font_name = 'helv.otf',
font_size='15sp',
background_normal="img/"+color_val+".jpg",
size_hint=(.13,.05),
pos_hint={'x':0.37,'center_y': 0.85})
dropdownWeekYear=DropDown()
btnwy1 = Button(text='[color=ffffff]2017[/color]',
#font_name = 'helv.otf',
markup=True,
font_size='12sp',
background_normal="img/"+color_val+".jpg",
size_hint=(1,None),
height=50)
btnwy1.bind(on_release=lambda btn: dropdownWeekYear.select(btn.text))
btnWeekYear= Button(text='[color=ffffff]Year v[/color]',
markup=True,
#font_name = 'helv.otf',
font_size='12sp',
background_normal="img/"+color_val+".jpg",
size_hint=(.13,.05),
pos_hint={'x':0.05,'center_y': 0.85})
dropdownWeekMonth=DropDown()
btnwm1 = Button(text='[color=ffffff]Feb[/color]',
#font_name = 'helv.otf',
markup=True,
font_size='12sp',
background_normal="img/"+color_val+".jpg",
size_hint=(1,None),
height=50)
btnwm1.bind(on_release=lambda btn: dropdownWeekMonth.select(btn.text))
btnWeekMonth= Button(text='[color=ffffff]Month v[/color]',
markup=True,
#font_name = 'helv.otf',
font_size='15sp',
background_normal="img/"+color_val+".jpg",
size_hint=(.15,.05),
pos_hint={'x':0.20,'center_y': 0.85})
btnWeekDay.bind(on_press=dropdownWeekDay.open)
btnWeekDay.bind(on_release=dropdownWeekDay.open)
btnWeekYear.bind(on_press=dropdownWeekYear.open)
btnWeekYear.bind(on_release=dropdownWeekYear.open)
btnWeekMonth.bind(on_press=dropdownWeekMonth.open)
btnWeekMonth.bind(on_release=dropdownWeekMonth.open)
#MonthWidgets
dropdownMonth = DropDown()
btnmm1 = Button(text='[color=ffffff]Jul[/color]',
# font_name = 'helv.otf',
markup=True,
font_size='12sp',
background_normal="img/"+color_val+".jpg",
size_hint=(1,None),
height=50)
btnmm1.bind(on_release=lambda btn: dropdownMonth.select(btn.text))
btnmm2 = Button(text='[color=ffffff]Jun[/color]',
# font_name = 'helv.otf',
markup=True,
font_size='12sp',
background_normal="img/"+color_val+".jpg",
size_hint=(1,None),
height=50)
btnmm2.bind(on_release=lambda btn: dropdownMonth.select(btn.text))
dropdownMonthYear = DropDown()
btnmy1 = Button(text='[color=ffffff]2017[/color]',
## font_name = 'helv.otf',
markup=True,
font_size='12sp',
background_normal="img/"+color_val+".jpg",
size_hint=(1,None),
height=50)
btnmy1.bind(on_release=lambda btn: dropdownMonthYear.select(btn.text))
btnmy2 = Button(text='[color=ffffff]2016[/color]',
# font_name = 'helv.otf',
markup=True,
font_size='12sp',
background_normal="img/"+color_val+".jpg",
size_hint=(1,None),
height=50)
btnmy2.bind(on_release=lambda btn: dropdownMonthYear.select(btn.text))
btnMon1= Button(text='[color=ffffff]Year v[/color]',
markup=True,
#font_name = 'helv.otf',
font_size='12sp',
background_normal="img/"+color_val+".jpg",
size_hint=(.25,.05),
pos_hint={'x':0.05,'center_y': 0.85})
btnMon2= Button(text='[color=ffffff]Month v[/color]',
markup=True,
# font_name = 'helv.otf',
font_size='12sp',
background_normal="img/"+color_val+".jpg",
size_hint=(.25,.05),
pos_hint={'x':0.35,'center_y': 0.85})
btnMon1.bind(on_press=dropdownMonth.open)
btnMon1.bind(on_release=dropdownMonth.open)
btnMon2.bind(on_press=dropdownMonthYear.open)
btnMon2.bind(on_release=dropdownMonthYear.open)
#annual
dropdownAnnual=DropDown()
btnann1 = Button(text='[color=ffffff]1988[/color]',
# font_name = 'helv.otf',
markup=True,
font_size='12sp',
background_normal="img/"+color_val+".jpg",
size_hint=(1,None),
height=50)
btnann1.bind(on_release=lambda btn: dropdownAnnual.select(btn.text))
btnAnnualYear= Button(text='[color=ffffff]Year v[/color]',
markup=True,
# font_name = 'helv.otf',
font_size='15sp',
background_normal="img/"+color_val+".jpg",
size_hint=(.15,.05),
pos_hint={'x':0.05,'center_y': 0.85})
btnAnnualYear.bind(on_press=dropdownAnnual.open)
btnAnnualYear.bind(on_release=dropdownAnnual.open)
def addWidgetsHis():
root4.add_widget(btnAnnualYear)
dropdownAnnual.add_widget(btnann1)
root4.add_widget(btnWeekDay)
dropdownWeekDay.add_widget(btnwd1)
root4.add_widget(btnWeekYear)
dropdownWeekYear.add_widget(btnwy1)
root4.add_widget(btnWeekMonth)
dropdownWeekMonth.add_widget(btnwm1)
root4.add_widget(btnMon1)
root4.add_widget(btnMon2)
dropdownMonth.add_widget(btnmm1)
dropdownMonth.add_widget(btnmm2)
dropdownMonthYear.add_widget(btnmy1)
dropdownMonthYear.add_widget(btnmy2)
root4.add_widget(btnDayDay)
dropdownDay.add_widget(btnd1)
dropdownDay.add_widget(btnd2)
root4.add_widget(btnYearDay)
dropdownYearDay.add_widget(btndy1)
dropdownYearDay.add_widget(btndy2)
root4.add_widget(btnMonthDay)
dropdownYearMonth.add_widget(btnmy11)
dropdownYearMonth.add_widget(btnmy22)
global d_now
d_now = datetime.datetime.now().strftime('%Y-%m-%d')
d_now_s = d_now.split('-')
cursor.execute("SELECT COUNT(*) FROM user_info WHERE year=? AND month=? AND day=?",(d_now_s[0], d_now_s[1], d_now_s[2]))
global _x_max
global cb_ctr
_x_max = int(cursor.fetchone()[0])
#print _x_max
con.commit()
graph_theme = {
'label_options': {
'color': rgb('000000'), # color of tick labels and titles
'bold': True}}
global nGraph
nGraph = Graph(xlabel='Daily',
ylabel='Polarity Values',
x_ticks_minor=1,
x_ticks_major=1,
y_ticks_minor=0,
y_ticks_major=0.20,
y_grid_label=True,
tick_color=(0,0,0,1),
border_color=(0,0,0,1),
x_grid_label=True,
padding=5,
x_grid=False,
y_grid=False,
xmin=0,
xmax=_x_max,
ymin=0,font_size='10sp',
ymax=1,
**graph_theme)
#start
def startGraphPlot(self, *args):
#dito maglagay
gridData.clear_widgets()
gridData.add_widget(Button(text="[color=000000]Date[/color]",font_size='9sp',markup=True,background_color=[0,0,0,0]))
gridData.add_widget(Button(text="[color=000000]Time[/color]",font_size='9sp',markup=True,background_color=[0,0,0,0]))
gridData.add_widget(Button(text="[color=000000]P_POS[/color]",font_size='9sp',markup=True,background_color=[0,0,0,0]))
gridData.add_widget(Button(text="[color=000000]P_NEG[/color]",font_size='9sp',markup=True,background_color=[0,0,0,0]))
if len(nGraph.plots) >= 1:
for plots in nGraph.plots:
nGraph.remove_plot(plots)
nGraph._clear_buffer()
#nGraph.xlabel = 'Daily'
d_now = datetime.datetime.now().strftime('%Y-%m-%d')
d_now_s = d_now.split('-')
cursor.execute("SELECT COUNT(*) FROM user_info")
if int(cursor.fetchone()[0]) == 0:
return
con.commit()
cursor.execute("SELECT COUNT(*) FROM user_info WHERE year=? AND month=? AND day=?",(d_now_s[0], d_now_s[1], d_now_s[2]))
_x_max = int(cursor.fetchone()[0])
minus = 1
while _x_max == 0:
d_now = datetime.datetime.strftime(datetime.datetime.now() - timedelta(minus), '%Y-%m-%d')
d_now_s = d_now.split('-')
cursor.execute("SELECT COUNT(*) FROM user_info WHERE year=? AND month=? AND day=?",(d_now_s[0], d_now_s[1], d_now_s[2]))
_x_max = int(cursor.fetchone()[0])
minus+=1
con.commit()
nGraph.xmax = _x_max
if _x_max >=12 and _x_max <= 24:
nGraph.x_ticks_major=3
elif _x_max >=25 and _x_max <=36:
nGraph.x_ticks_major = 6
elif _x_max >=37:
n_Graph.x_ticks_major = int(_x_max*20)
#print _x_max
wplot = SmoothLinePlot(mode="points",color=rgb('76abe6'))
wplot2 = SmoothLinePlot(mode="points",color=rgb('ff6666'))
#wplot3 = SmoothLinePlot(mode="points",color=rgb('faff66'))
#faff66
cursor.execute("SELECT pos,neg,time FROM user_info WHERE year=? AND month=? AND day=?",(d_now_s[0], d_now_s[1], d_now_s[2]))
plot_pts = [(0,0)]
plot_pts2 = [(0,0)]
plot_pts3 = [(0,0)]
p_total = 0
n_total = 0
#nplot_pts = [(0,-1)]
pola = cursor.fetchall()
#print pola
for m in range(0,len(pola),1):
temp = float(pola[m][0])
temp2 = float(pola[m][1])
p_total +=float(pola[m][0])
n_total +=float(pola[m][1])
gridData.add_widget(Button(text="[color=000000]"+str(d_now_s[1])+' '+str(d_now_s[2])+' '+str(d_now_s[0])+"[/color]",font_size='9sp',markup=True,background_color=[0,0,0,0]))
gridData.add_widget(Button(text="[color=000000]"+str(pola[m][2])+"[/color]",font_size='9sp',markup=True,background_color=[0,0,0,0]))
gridData.add_widget(Button(text="[color=000000]"+str('%.2f' % float(pola[m][0]))+"[/color]",font_size='9sp',markup=True,background_color=[0,0,0,0]))
gridData.add_widget(Button(text="[color=000000]"+str('%.2f' % float(pola[m][1]))+"[/color]",font_size='9sp',markup=True,background_color=[0,0,0,0]))
#print 'POS:{0} NEG:{1}'.format(temp,temp2)
dd = (m+1,temp)
dd2 = (m+1,temp2)
if temp>temp2:
dd3 = (m+1,temp-temp2)
elif temp2>temp:
dd3 = (m+1,temp2-temp)
plot_pts.append(dd)
plot_pts2.append(dd2)
plot_pts3.append(dd3)
petal_val = result_precision(float(p_total/len(pola)))
petal_val = float(petal_val) * 100
con.commit()
cursor.execute("SELECT gender from user_account LIMIT 1")
gender_t = cursor.fetchone()[0]
con.commit()
if gender_t == 'Male':
gN = 'wire'
if gender_t == 'Female':
gN = 'flower'
if petal_val <=0.00:
petalG.source="img/"+gN+"zero.jpg"
if petal_val>= 1.00 and petal_val <= 10:
petalG.source="img/"+gN+"one.jpg"
if petal_val >= 11.00 and petal_val<= 20.00:
petalG.source="img/"+gN+"two.jpg"
if petal_val >= 21.00 and petal_val<= 30.00:
petalG.source="img/"+gN+"three.jpg"
if petal_val >= 31.00 and petal_val <= 40.00:
petalG.source="img/"+gN+"four.jpg"
if petal_val >= 41.00 and petal_val <= 50.00:
petalG.source="img/"+gN+"five.jpg"
if petal_val >= 51.00 and petal_val <= 60.00:
petalG.source="img/"+gN+"six.jpg"
if petal_val >= 61.00 and petal_val <= 70.00:
petalG.source="img/"+gN+"seven.jpg"
if petal_val >= 71.00 and petal_val < 80.00:
petalG.source="img/"+gN+"eight.jpg"
if petal_val >= 81.00 and petal_val <= 90.00:
petalG.source="img/"+gN+"nine.jpg"
if petal_val >= 91.00 and petal_val <= 100.00:
petalG.source="img/"+gN+"whole.jpg"
lblTemp.text = "[color=000000]Average:[/color]"+"[color=393d42]Positive Score:"+result_precision(float(p_total/len(pola)))+" Negative Score:"+str(abs(float(result_precision(float(n_total/len(pola)) ))))+"[/color]"
wplot2.points = plot_pts2
wplot.points = plot_pts
wplot3.points = plot_pts3
nGraph.add_plot(wplot)
nGraph.add_plot(wplot2)
#nGraph.add_plot(wplot3)
#daily
def dailyGraphPlot(self ,*args):
dayWidgets()
cursor.execute("SELECT gender from user_account LIMIT 1")
gender_t = cursor.fetchone()[0]
con.commit()
if gender_t == 'Male':
petalG.source="img/wirezero.jpg"
if gender_t == 'Female':
petalG.source="img/flowerzero.jpg"
if len(nGraph.plots)>=1:
for plot in nGraph.plots:
nGraph.remove_plot(plot)
nGraph._clear_buffer()
for plot in nGraph.plots:
nGraph.remove_plot(plot)
nGraph._clear_buffer()
def dGraphPlot( y, mon, d,*args):
if len(nGraph.plots) >= 1:
for plots in nGraph.plots:
nGraph.remove_plot(plots)
nGraph._clear_buffer()
nGraph.xlabel = 'n Logs per Day'
nGraph.ylabel = 'Polarity Values'
d_now = datetime.datetime.now().strftime('%Y-%m-%d')
d_now_s = d_now.split('-')
cursor.execute("SELECT COUNT(*) FROM user_info WHERE year=? AND month=? AND day=?",(y,mon,d))
_x_max = int(cursor.fetchone()[0])
con.commit()
p_total = 0
n_total = 0
#print 'qweqwewq', _x_max
if _x_max != 0:
gridData.clear_widgets()
gridData.add_widget(Button(text="[color=000000]Date[/color]",font_size='9sp',markup=True,background_color=[0,0,0,0]))
gridData.add_widget(Button(text="[color=000000]Time[/color]",font_size='9sp',markup=True,background_color=[0,0,0,0]))
gridData.add_widget(Button(text="[color=000000]P_POS[/color]",font_size='9sp',markup=True,background_color=[0,0,0,0]))
gridData.add_widget(Button(text="[color=000000]P_NEG[/color]",font_size='9sp',markup=True,background_color=[0,0,0,0]))
nGraph.xmax = _x_max
#faff66
cursor.execute("SELECT pos,neg,time FROM user_info WHERE year=? AND month=? AND day=?",(y, mon, d))
plot_pts = [(0,0)]
plot_pts2 = [(0,0)]
plot_pts3 = [(0,0)]
pola = cursor.fetchall()
for m in range(0,len(pola),1):
temp = float(pola[m][0])
temp2 = float(pola[m][1])
p_total +=float(pola[m][0])
n_total +=float(pola[m][1])
gridData.add_widget(Button(text="[color=000000]"+str(mon)+' '+str(d)+' '+str(y)+"[/color]",font_size='9sp',markup=True,background_color=[0,0,0,0]))
gridData.add_widget(Button(text="[color=000000]"+str(pola[m][2])+"[/color]",font_size='9sp',markup=True,background_color=[0,0,0,0]))
gridData.add_widget(Button(text="[color=000000]"+str('%.2f' % float(pola[m][0]) )+"[/color]",font_size='9sp',markup=True,background_color=[0,0,0,0]))
gridData.add_widget(Button(text="[color=000000]"+str('%.2f '% float(pola[m][1]) )+"[/color]",font_size='9sp',markup=True,background_color=[0,0,0,0]))
dd = (m+1,temp)
dd2 = (m+1,temp2)
if temp>temp2:
dd3 = (m+1,temp-temp2)
elif temp2>temp:
dd3 = (m+1,temp2-temp)
plot_pts.append(dd)
plot_pts2.append(dd2)
plot_pts3.append(dd3)
petal_val = result_precision(float(p_total/len(pola)))
petal_val = float(petal_val) * 100
con.commit()
cursor.execute("SELECT gender from user_account LIMIT 1")
gender_t = cursor.fetchone()[0]
con.commit()
if gender_t == 'Male':
gN = 'wire'
if gender_t == 'Female':
gN = 'flower'
if petal_val <=0.00:
petalG.source="img/"+gN+"zero.jpg"
if petal_val>= 1.00 and petal_val <= 10:
petalG.source="img/"+gN+"one.jpg"
if petal_val >= 11.00 and petal_val<= 20.00:
petalG.source="img/"+gN+"two.jpg"
if petal_val >= 21.00 and petal_val<= 30.00:
petalG.source="img/"+gN+"three.jpg"
if petal_val >= 31.00 and petal_val <= 40.00:
petalG.source="img/"+gN+"four.jpg"
if petal_val >= 41.00 and petal_val <= 50.00:
petalG.source="img/"+gN+"five.jpg"
if petal_val >= 51.00 and petal_val <= 60.00:
petalG.source="img/"+gN+"six.jpg"
if petal_val >= 61.00 and petal_val <= 70.00:
petalG.source="img/"+gN+"seven.jpg"
if petal_val >= 71.00 and petal_val < 80.00:
petalG.source="img/"+gN+"eight.jpg"
if petal_val >= 81.00 and petal_val <= 90.00:
petalG.source="img/"+gN+"nine.jpg"
if petal_val >= 91.00 and petal_val <= 100.00:
petalG.source="img/"+gN+"whole.jpg"
lblTemp.text = "[color=000000]Average:[/color]"+"[color=393d42]Positive Text:"+result_precision(float(p_total/len(pola)) )+" Negative Text:"+result_precision(float(n_total/len(pola)) )+"[/color]"
wplot.points = plot_pts
wplot2.points = plot_pts2
wplot3.points = plot_pts3
nGraph.add_plot(wplot)
nGraph.add_plot(wplot2)
#nGraph.add_plot(wplot3)
else:
pass
def result_precision(number):
return '%.2f' % number
def weeklyGraphPlot(self, *args):
weekWidgets()
cursor.execute("SELECT gender from user_account LIMIT 1")
gender_t = cursor.fetchone()[0]
con.commit()
if gender_t == 'Male':
petalG.source="img/birdzero.jpg"
if gender_t == 'Female':
petalG.source="img/tulipzero.jpg"
for i in range(2):
if len(nGraph.plots) >= 1:
for plots in nGraph.plots:
nGraph.remove_plot(plots)
nGraph._clear_buffer()
def wGraphPlot( y, mon, d,*args):
if len(nGraph.plots) >= 1:
for plots in nGraph.plots:
nGraph.remove_plot(plots)
nGraph._clear_buffer()
nGraph.xlabel = 'Weekly Entries'
nGraph.ylabel = 'Polarity Values'
nGraph.xmax = 7
p_total = 0
n_total = 0
d_now = datetime.datetime.now().strftime('%Y-%m-%d')
d_now_s = d_now.split('-')
dlist = [ [] for _ in range(7)]
dlist2 = [ [] for _ in range(7)]
#jlist = [[] for _ in range(n)]
#create multiple list
cursor.execute("SELECT COUNT(*) FROM user_info WHERE year=? AND month=? AND day=?",(y,mon,d))
v_count = int(cursor.fetchone()[0])
#print 'eto {0}'.format(v_count)
con.commit()
c_list = []
days = []
if v_count != 0 :
gridData.clear_widgets()
gridData.add_widget(Button(text="[color=000000]Date[/color]",font_size='9sp',markup=True,background_color=[0,0,0,0]))
gridData.add_widget(Button(text="[color=000000]# of Entries[/color]",font_size='9sp',markup=True,background_color=[0,0,0,0]))
gridData.add_widget(Button(text="[color=000000]P_POS[/color]",font_size='9sp',markup=True,background_color=[0,0,0,0]))
gridData.add_widget(Button(text="[color=000000]P_NEG[/color]",font_size='9sp',markup=True,background_color=[0,0,0,0]))
for i in range(0,7):
cursor.execute("SELECT pos,neg,time FROM user_info WHERE year=? AND month=? AND day=?",(y,mon,d))
pola = cursor.fetchall()
days.append(int(d))
n_count = 0
for ou in range(0,len(pola),1):
temp = float(pola[ou][0])
temp2 = float(pola[ou][1])
dlist[i].append(temp)
dlist2[i].append(temp2)
n_count += 1
c_list.append(str(n_count))
d = str(format((int(d)+1), '02d'))
con.commit()
plot_pts = [(0,0)]
plot_pts2 = [(0,0)]
plot_pts3 = [(0,0)]
#print c_list
for n in range(7):
if len(dlist[n]) != 0:
plot_pts.append( ((n+1), compute_values(sum(dlist[n]), len(dlist[n])) ) )
wplot.points = plot_pts
for n in range(7):
if len(dlist2[n]) != 0:
plot_pts2.append( ((n+1), compute_values(sum(dlist2[n]), len(dlist2[n]))) )
#d = str(int(d)-len(dlist))
p_total = []
n_total = []
for pos_range in range(0,len(dlist)):
if dlist[pos_range] != []:
gridData.add_widget(Button(text="[color=000000]"+str(mon)+' '+str(format((int(days[pos_range])), '02d'))+' '+str(y)+"[/color]",font_size='9sp',markup=True,background_color=[0,0,0,0]))
gridData.add_widget(Button(text="[color=000000]"+str(c_list[pos_range])+"[/color]",font_size='9sp',markup=True,background_color=[0,0,0,0]))
gridData.add_widget(Button(text="[color=000000]"+str('%.2f' % compute_values(sum(dlist[pos_range]), len(dlist[pos_range])) )+"[/color]",font_size='9sp',markup=True,background_color=[0,0,0,0]))
gridData.add_widget(Button(text="[color=000000]"+str('%.2f '% compute_values(sum(dlist2[pos_range]), len(dlist2[pos_range])) )+"[/color]",font_size='9sp',markup=True,background_color=[0,0,0,0]))
p_total.append( compute_values(sum(dlist[pos_range]), len(dlist[pos_range])) )
n_total.append( compute_values(sum(dlist2[pos_range]), len(dlist2[pos_range])) )
wplot2.points = plot_pts2
nGraph.add_plot(wplot)
nGraph.add_plot(wplot2)
petal_val = compute_values(sum(p_total), len(p_total))
petal_val = float(petal_val) * 100
con.commit()
cursor.execute("SELECT gender from user_account LIMIT 1")
gender_t = cursor.fetchone()[0]
con.commit()
#print petal_val
if gender_t == 'Male':
gN = 'bird'
if gender_t == 'Female':
gN = 'tulip'
if petal_val <=0.00:
petalG.source="img/"+gN+"zero.jpg"
if petal_val>= 1.00 and petal_val <= 10:
petalG.source="img/"+gN+"one.jpg"
if petal_val >= 11.00 and petal_val<= 20.00:
petalG.source="img/"+gN+"two.jpg"
if petal_val >= 21.00 and petal_val<= 30.00:
petalG.source="img/"+gN+"three.jpg"
if petal_val >= 31.00 and petal_val <= 40.00:
petalG.source="img/"+gN+"four.jpg"
if petal_val >= 41.00 and petal_val <= 50.00:
petalG.source="img/"+gN+"five.jpg"
if petal_val >= 51.00 and petal_val <= 60.00:
petalG.source="img/"+gN+"six.jpg"
if petal_val >= 61.00 and petal_val <= 70.00:
petalG.source="img/"+gN+"seven.jpg"
if petal_val >= 71.00 and petal_val < 80.00:
petalG.source="img/"+gN+"eight.jpg"
if petal_val >= 81.00 and petal_val <= 90.00:
petalG.source="img/"+gN+"nine.jpg"
if petal_val >= 91.00 and petal_val <= 100.00:
petalG.source="img/"+gN+"whole.jpg"
lblTemp.text = "[color=000000]Average:[/color]"+"[color=393d42]Positive Text:"+('%.2f '% compute_values(sum(p_total), len(p_total)))+" Negative Text:"+('%.2f '% compute_values(sum(n_total), len(n_total)) )+"[/color]"
def monthlyGraphPlot(self, *args):
monthWidgets()
cursor.execute("SELECT gender from user_account LIMIT 1")
gender_t = cursor.fetchone()[0]
con.commit()
if gender_t == 'Male':
petalG.source="img/birdzero.jpg"
if gender_t == 'Female':
petalG.source="img/tulipzero.jpg"
for i in range(2):
if len(nGraph.plots) >= 1:
for plots in nGraph.plots:
nGraph.remove_plot(plots)
nGraph._clear_buffer()
def mGraph(mon, y,*args):
if len(nGraph.plots) >= 1:
for plots in nGraph.plots:
nGraph.remove_plot(plots)
nGraph._clear_buffer()
nGraph.xlabel = 'Monthly Entries'
nGraph.ylabel = 'Polarity Values'
nGraph.xmax = 31
p_total = 0
n_total = 0
d_now = datetime.datetime.now().strftime('%Y-%m-%d')
d_now_s = d_now.split('-')
dlist = [ [] for _ in range(31)]
dlist2 = [ [] for _ in range(31)]
#jlist = [[] for _ in range(n)]
#create multiple list
cursor.execute("SELECT COUNT(*) FROM user_info WHERE year=? AND month=?",(y,mon))
v_count = int(cursor.fetchone()[0])
#print 'eto {0}'.format(v_count)
con.commit()
c_list = []
days = []
if v_count != 0 :
gridData.clear_widgets()
gridData.add_widget(Button(text="[color=000000]Date[/color]",font_size='9sp',markup=True,background_color=[0,0,0,0]))
gridData.add_widget(Button(text="[color=000000]# of Entries[/color]",font_size='9sp',markup=True,background_color=[0,0,0,0]))
gridData.add_widget(Button(text="[color=000000]P_POS[/color]",font_size='9sp',markup=True,background_color=[0,0,0,0]))
gridData.add_widget(Button(text="[color=000000]P_NEG[/color]",font_size='9sp',markup=True,background_color=[0,0,0,0]))
for i in range(0,31):
da = str(format((i+1), '02d'))
cursor.execute("SELECT pos,neg,time FROM user_info WHERE year=? AND month=? AND day=?",(y,mon,da))
pola = cursor.fetchall()
days.append(da)
n_count = 0
for ou in range(0,len(pola),1):
temp = float(pola[ou][0])
temp2 = float(pola[ou][1])
dlist[i].append(temp)
dlist2[i].append(temp2)
n_count += 1
c_list.append(str(n_count))
con.commit()
plot_pts = [(0,0)]
plot_pts2 = [(0,0)]
plot_pts3 = [(0,0)]
#print c_list
for n in range(31):
if len(dlist[n]) != 0:
plot_pts.append( ((n+1), compute_values(sum(dlist[n]), len(dlist[n])) ) )
wplot.points = plot_pts
for n in range(31):
if len(dlist2[n]) != 0:
plot_pts2.append( ((n+1), compute_values(sum(dlist2[n]), len(dlist2[n]))) )
#d = str(int(d)-len(dlist))
p_total = []
n_total = []
for pos_range in range(0,len(dlist)):
if dlist[pos_range] != []:
gridData.add_widget(Button(text="[color=000000]"+str(mon)+' '+str(format((int(days[pos_range])), '02d'))+' '+str(y)+"[/color]",font_size='9sp',markup=True,background_color=[0,0,0,0]))
gridData.add_widget(Button(text="[color=000000]"+str(c_list[pos_range])+"[/color]",font_size='9sp',markup=True,background_color=[0,0,0,0]))
gridData.add_widget(Button(text="[color=000000]"+str('%.2f' % compute_values(sum(dlist[pos_range]), len(dlist[pos_range])) )+"[/color]",font_size='9sp',markup=True,background_color=[0,0,0,0]))
gridData.add_widget(Button(text="[color=000000]"+str('%.2f '% compute_values(sum(dlist2[pos_range]), len(dlist2[pos_range])) )+"[/color]",font_size='9sp',markup=True,background_color=[0,0,0,0]))
p_total.append( compute_values(sum(dlist[pos_range]), len(dlist[pos_range])) )
n_total.append( compute_values(sum(dlist2[pos_range]), len(dlist2[pos_range])) )
wplot2.points = plot_pts2
nGraph.add_plot(wplot)
nGraph.add_plot(wplot2)
petal_val = compute_values(sum(p_total), len(p_total))
petal_val = float(petal_val) * 100
con.commit()
cursor.execute("SELECT gender from user_account LIMIT 1")
gender_t = cursor.fetchone()[0]
con.commit()
if gender_t == 'Male':
gN = 'bird'
if gender_t == 'Female':
gN = 'tulip'
if petal_val <=0.00:
petalG.source="img/"+gN+"zero.jpg"
if petal_val>= 1.00 and petal_val <= 10:
petalG.source="img/"+gN+"one.jpg"
if petal_val >= 11.00 and petal_val<= 20.00:
petalG.source="img/"+gN+"two.jpg"
if petal_val >= 21.00 and petal_val<= 30.00:
petalG.source="img/"+gN+"three.jpg"
if petal_val >= 31.00 and petal_val <= 40.00:
petalG.source="img/"+gN+"four.jpg"
if petal_val >= 41.00 and petal_val <= 50.00:
petalG.source="img/"+gN+"five.jpg"
if petal_val >= 51.00 and petal_val <= 60.00:
petalG.source="img/"+gN+"six.jpg"
if petal_val >= 61.00 and petal_val <= 70.00:
petalG.source="img/"+gN+"seven.jpg"
if petal_val >= 71.00 and petal_val < 80.00:
petalG.source="img/"+gN+"eight.jpg"
if petal_val >= 81.00 and petal_val <= 90.00:
petalG.source="img/"+gN+"nine.jpg"
if petal_val >= 91.00 and petal_val <= 100.00:
petalG.source="img/"+gN+"whole.jpg"
lblTemp.text = "[color=000000]Average:[/color]"+"[color=393d42]Positive Text:"+('%.2f '% compute_values(sum(p_total), len(p_total)))+" Negative Text:"+('%.2f '% compute_values(sum(n_total), len(n_total)) )+"[/color]"
def annualGraphPlot(self,*args):
annualWidgets()
cursor.execute("SELECT gender from user_account LIMIT 1")
gender_t = cursor.fetchone()[0]
con.commit()
if gender_t == 'Male':
petalG.source="img/birdzero.jpg"
if gender_t == 'Female':
petalG.source="img/tulipzero.jpg"
for i in range(2):
if len(nGraph.plots) >= 1:
for plots in nGraph.plots:
nGraph.remove_plot(plots)
nGraph._clear_buffer()
def aGraph(y,*args):
if len(nGraph.plots) >= 1:
for plots in nGraph.plots:
nGraph.remove_plot(plots)
nGraph._clear_buffer()
nGraph.xlabel = 'Annual Entries'
nGraph.ylabel = 'Polarity Values'
nGraph.xmax = 12
p_total = 0
n_total = 0
d_now = datetime.datetime.now().strftime('%Y-%m-%d')
d_now_s = d_now.split('-')
dlist = [ [] for _ in range(12)]
dlist2 = [ [] for _ in range(12)]
#jlist = [[] for _ in range(n)]
#create multiple list
cursor.execute("SELECT COUNT(*) FROM user_info WHERE year="+str(y))
v_count = int(cursor.fetchone()[0])
#print 'eto {0}'.format(v_count)
con.commit()
m_list = []
days = []
temp_list = []
if v_count != 0 :
gridData.clear_widgets()
gridData.add_widget(Button(text="[color=000000]Date[/color]",font_size='9sp',markup=True,background_color=[0,0,0,0]))
gridData.add_widget(Button(text="[color=000000]# of Entries[/color]",font_size='9sp',markup=True,background_color=[0,0,0,0]))
gridData.add_widget(Button(text="[color=000000]P_POS[/color]",font_size='9sp',markup=True,background_color=[0,0,0,0]))
gridData.add_widget(Button(text="[color=000000]P_NEG[/color]",font_size='9sp',markup=True,background_color=[0,0,0,0]))
for m in range(12):
ma = str(format((m+1), '02d'))
m_list.append(ma)
for i in range(31):
da = str(format((i+1), '02d'))
cursor.execute("SELECT pos,neg,time FROM user_info WHERE year=? AND month=? AND day=?",(y,ma,da))
pola = cursor.fetchall()
for ou in range(0,len(pola),1):
temp = float(pola[ou][0])
temp2 = float(pola[ou][1])
temp_list.append(temp)
temp_list2.append(temp2)
t_Values = compute_values(sum(temp_list), len(temp_list))
t_Values2 = compute_values(sum(temp_list2), len(temp_list2))
dlist[m].append(t_Values)
dlist2[m].append(t_Values2)
t_Values = 0
t_Values2 = 0
temp_list2 = []
temp_list = []
con.commit()
con.commit()
plot_pts = [(0,0)]
plot_pts2 = [(0,0)]
plot_pts3 = [(0,0)]
for n in range(12):
if len(dlist[n]) != 0:
plot_pts.append( ((n+1), compute_values(sum(dlist[n]), len(dlist[n])) ) )
wplot.points = plot_pts
for n in range(12):
if len(dlist2[n]) != 0:
plot_pts2.append( ((n+1), compute_values(sum(dlist2[n]), len(dlist2[n]))) )
#d = str(int(d)-len(dlist))
p_total = []
n_total = []
for pos_range in range(0,len(dlist)):
if dlist[pos_range] != []:
gridData.add_widget(Button(text="[color=000000]"+str(m_list[pos_range])+' '+' '+str(y)+"[/color]",font_size='9sp',markup=True,background_color=[0,0,0,0]))
gridData.add_widget(Button(text="[color=000000]"+str(m_list[pos_range])+"[/color]",font_size='9sp',markup=True,background_color=[0,0,0,0]))
gridData.add_widget(Button(text="[color=000000]"+str('%.2f' % compute_values(sum(dlist[pos_range]), len(dlist[pos_range])) )+"[/color]",font_size='9sp',markup=True,background_color=[0,0,0,0]))
gridData.add_widget(Button(text="[color=000000]"+str('%.2f '% compute_values(sum(dlist2[pos_range]), len(dlist2[pos_range])) )+"[/color]",font_size='9sp',markup=True,background_color=[0,0,0,0]))
p_total.append( compute_values(sum(dlist[pos_range]), len(dlist[pos_range])) )
n_total.append( compute_values(sum(dlist2[pos_range]), len(dlist2[pos_range])) )
wplot2.points = plot_pts2
nGraph.add_plot(wplot)
nGraph.add_plot(wplot2)
petal_val = compute_values(sum(p_total), len(p_total))
petal_val = float(petal_val) * 100
con.commit()
cursor.execute("SELECT gender from user_account LIMIT 1")
gender_t = cursor.fetchone()[0]
con.commit()
if gender_t == 'Male':
gN = 'bird'
if gender_t == 'Female':
gN = 'tulip'
if petal_val <=0.00:
petalG.source="img/"+gN+"zero.jpg"
if petal_val>= 1.00 and petal_val <= 10:
petalG.source="img/"+gN+"one.jpg"
if petal_val >= 11.00 and petal_val<= 20.00:
petalG.source="img/"+gN+"two.jpg"
if petal_val >= 21.00 and petal_val<= 30.00:
petalG.source="img/"+gN+"three.jpg"
if petal_val >= 31.00 and petal_val <= 40.00:
petalG.source="img/"+gN+"four.jpg"
if petal_val >= 41.00 and petal_val <= 50.00:
petalG.source="img/"+gN+"five.jpg"
if petal_val >= 51.00 and petal_val <= 60.00:
petalG.source="img/"+gN+"six.jpg"
if petal_val >= 61.00 and petal_val <= 70.00:
petalG.source="img/"+gN+"seven.jpg"
if petal_val >= 71.00 and petal_val < 80.00:
petalG.source="img/"+gN+"eight.jpg"
if petal_val >= 81.00 and petal_val <= 90.00:
petalG.source="img/"+gN+"nine.jpg"
if petal_val >= 91.00 and petal_val <= 100.00:
petalG.source="img/"+gN+"whole.jpg"
lblTemp.text = "[color=000000]Average:[/color]"+"[color=393d42]Positive Text:"+('%.2f '% compute_values(sum(p_total), len(p_total)))+" Negative Text:"+('%.2f '% compute_values(sum(n_total), len(n_total)) )+"[/color]"
def compute_values(a, b):
if a==0:
return 0
else:
return float(a)/b
#daily Widgets + buttons on Dropdown
def dayWidgets():
root4.add_widget(btnDayDay)
root4.add_widget(btnYearDay)
root4.add_widget(btnMonthDay)
def daily_day(self,*args):
dropdownDay.clear_widgets()
cursor.execute('SELECT DISTINCT day FROM user_info')
day_values = cursor.fetchall()
for day in day_values:
day_t_btn = Button(text='[color=000000]'+str(day[0])+'[/color]',
markup=True,
font_size='12sp',
background_normal="ffffff",
size_hint=(1,None),
height=50)
day_t_btn.bind(on_release=lambda day_t_btn: dropdownDay.select(day_t_btn.text))
day_t_btn.bind(on_release=partial(dChange_day, str(day[0]) ))
dropdownDay.add_widget(day_t_btn)
con.commit()
def dChange_day(dD,*args):
global dDay
dDay = str(dD)
btnDayDay.text = dDay
dGraphPlot(dYear,dMonth,dDay)
def daily_year(self,*args):
dropdownYearDay.clear_widgets()
cursor.execute('SELECT DISTINCT year FROM user_info')
year_values = cursor.fetchall()
for year in year_values:
year_t_btn = Button(text='[color=000000]'+str(year[0])+'[/color]',
markup=True,
font_size='12sp',
background_normal="ffffff",
size_hint=(1,None),
height=50)
year_t_btn.bind(on_release=lambda year_t_btn: dropdownYearDay.select(year_t_btn.text))
year_t_btn.bind(on_release=partial(dChange_year, str(year[0])))
dropdownYearDay.add_widget(year_t_btn)
con.commit()
def dChange_year(dY,*args):
global dYear
dYear = str(dY)
btnYearDay.text = dY
dGraphPlot(dYear,dMonth,dDay)
def daily_month(self,*args):
dropdownYearMonth.clear_widgets()
cursor.execute('SELECT DISTINCT month FROM user_info')
month_values = cursor.fetchall()
for month in month_values:
month_t_btn = Button(text='[color=000000]'+str(month[0])+'[/color]',
markup=True,
font_size='12sp',
background_normal="ffffff",
size_hint=(1,None),
height=50)
month_t_btn.bind(on_release=lambda month_t_btn: dropdownYearMonth.select(month_t_btn.text))
month_t_btn.bind(on_release=partial(dChange_month, str(month[0]) ))
dropdownYearMonth.add_widget(month_t_btn)
con.commit()
def dChange_month(dM,*args):
global dMonth
dMonth = str(dM)
btnMonthDay.text = dM
dGraphPlot(dYear,dMonth,dDay)
#weekly Widgets + buttons on Dropdown
def weekWidgets():
root4.add_widget(btnWeekDay)
root4.add_widget(btnWeekYear)
root4.add_widget(btnWeekMonth)
def weekly_day(self,*args):
dropdownWeekDay.clear_widgets()
cursor.execute('SELECT DISTINCT day FROM user_info')
day_values = cursor.fetchall()
for day in day_values:
day_t_btn = Button(text='[color=000000]'+str(day[0])+'[/color]',
markup=True,
font_size='12sp',
background_normal="ffffff",
size_hint=(1,None),
height=50)
day_t_btn.bind(on_release=lambda day_t_btn: dropdownWeekDay.select(day_t_btn.text))
day_t_btn.bind(on_release=partial(wChange_day, str(day[0]) ))
dropdownWeekDay.add_widget(day_t_btn)
con.commit()
def wChange_day(wD,*args):
global dDay
dDay = str(wD)
btnWeekDay.text = wD
wGraphPlot(dYear,dMonth,dDay)
def weekly_year(self,*args):
dropdownWeekYear.clear_widgets()
cursor.execute('SELECT DISTINCT year FROM user_info')
year_values = cursor.fetchall()
for year in year_values:
year_t_btn = Button(text='[color=000000]'+str(year[0])+'[/color]',
markup=True,
font_size='12sp',
background_normal="ffffff",
size_hint=(1,None),
height=50)
year_t_btn.bind(on_release=lambda year_t_btn: dropdownWeekYear.select(year_t_btn.text))
year_t_btn.bind(on_release=partial(wChange_year, str(year[0])))
dropdownWeekYear.add_widget(year_t_btn)
con.commit()
def wChange_year(wY,*args):
global dYear
dYear = str(wY)
btnWeekYear.text = wY
wGraphPlot(dYear,dMonth,dDay)
def weekly_month(self,*args):
dropdownWeekMonth.clear_widgets()
cursor.execute('SELECT DISTINCT month FROM user_info')
month_values = cursor.fetchall()
for month in month_values:
month_t_btn = Button(text='[color=000000]'+str(month[0])+'[/color]',
markup=True,
font_size='12sp',
background_normal="ffffff",
size_hint=(1,None),
height=50)
month_t_btn.bind(on_release=lambda month_t_btn: dropdownWeekMonth.select(month_t_btn.text))
month_t_btn.bind(on_release=partial(wChange_month, str(month[0]) ))
dropdownWeekMonth.add_widget(month_t_btn)
con.commit()
def wChange_month(wM,*args):
global dMonth
dMonth = str(wM)
btnWeekMonth.text = wM
wGraphPlot(dYear,dMonth,dDay)
#Monthly Widgets + buttons on Dropdown
def monthWidgets():
root4.add_widget(btnMon1)
root4.add_widget(btnMon2)
def monthly_month(self,*args):
dropdownMonth.clear_widgets()
cursor.execute('SELECT DISTINCT year FROM user_info')
month_values = cursor.fetchall()
for month in month_values:
month_t_btn = Button(text='[color=000000]'+str(month[0])+'[/color]',
markup=True,
font_size='12sp',
background_normal="ffffff",
size_hint=(1,None),
height=50)
month_t_btn.bind(on_release=lambda month_t_btn: dropdownMonth.select(month_t_btn.text))
month_t_btn.bind(on_release=partial(mChange_month, str(month[0]) ))
dropdownMonth.add_widget(month_t_btn)
con.commit()
def mChange_month(mM,*args):
global dMonth
dMonth = str(mM)
btnMon1.text = mM
mGraph(dYear,dMonth)
def monthly_year(self,*args):
dropdownMonthYear.clear_widgets()
cursor.execute('SELECT DISTINCT month FROM user_info')
year_values = cursor.fetchall()
for year in year_values:
year_t_btn = Button(text='[color=000000]'+str(year[0])+'[/color]',
markup=True,
font_size='12sp',
background_normal="ffffff",
size_hint=(1,None),
height=50)
year_t_btn.bind(on_release=lambda year_t_btn: dropdownMonthYear.select(year_t_btn.text))
year_t_btn.bind(on_release=partial(mChange_year, str(year[0])))
dropdownMonthYear.add_widget(year_t_btn)
con.commit()
def mChange_year(mY,*args):
global dYear
dYear = str(mY)
btnMon2.text = mY
mGraph(dYear,dMonth)
#Yearly Widgets + buttons on Dropdown
def annualWidgets():
root4.add_widget(btnAnnualYear)
def yearly_year(self,*args):
dropdownAnnual.clear_widgets()
cursor.execute('SELECT DISTINCT year FROM user_info')
year_values = cursor.fetchall()
for year in year_values:
year_t_btn = Button(text='[color=000000]'+str(year[0])+'[/color]',
markup=True,
font_size='12sp',
background_normal="ffffff",
size_hint=(1,None),
height=50)
year_t_btn.bind(on_release=lambda year_t_btn: dropdownAnnual.select(year_t_btn.text))
year_t_btn.bind(on_release=partial(yChange_year, str(year[0])))
dropdownAnnual.add_widget(year_t_btn)
con.commit()
def yChange_year(yY,*args):
global dYear
dYear = str(yY)
btnAnnualYear.text = yY
aGraph(dYear)
def rWidgets(self,*args):
root4.remove_widget(petalG)
root4.remove_widget(lblDescript)
root4.remove_widget(scrollData)
scrollData.remove_widget(gridData)
boxStats.remove_widget(nGraph)
t_consul = '0'
www = 0
cursor.execute("SELECT COUNT(*) FROM consultant_account")
www = int(cursor.fetchone()[0])
con.commit()
cursor.execute("SELECT id from consultant_account")
tmp_id = cursor.fetchone()[0]
con.commit()
cursor.execute("SELECT Control FROM consultant_account WHERE ID="+str(tmp_id))
t_consul = str(cursor.fetchone()[0])
con.commit()
if t_consul == '0':
statsTitle.text = 'User Garden'
root4.add_widget(petalG)
if t_consul == '1':
root4.add_widget(lblDescript)
root4.add_widget(lblTemp)
root4.add_widget(scrollData)
scrollData.add_widget(gridData)
#graph
boxStats.add_widget(nGraph)
def removeDWM():
root4.remove_widget(btnAnnualYear)
dropdownAnnual.remove_widget(btnann1)
btnAnnualYear.text = "Year v"
root4.remove_widget(btnWeekDay)
dropdownDay.remove_widget(btnwd1)
root4.remove_widget(btnWeekYear)
dropdownYearDay.remove_widget(btnwy1)
root4.remove_widget(btnWeekMonth)
dropdownYearMonth.remove_widget(btnwm1)
btnWeekDay.text = "Day v"
btnWeekYear.text = "Year v"
btnWeekMonth.text = "Month v"
root4.remove_widget(btnMon1)
root4.remove_widget(btnMon2)
dropdownMonth.remove_widget(btnmm1)
dropdownMonth.remove_widget(btnmm2)
dropdownMonthYear.remove_widget(btnmy1)
dropdownMonthYear.remove_widget(btnmy2)
btnMon1.text = "Year v"
btnMon2.text = "Month v"
root4.remove_widget(btnDayDay)
dropdownDay.remove_widget(btnd1)
dropdownDay.remove_widget(btnd2)
root4.remove_widget(btnYearDay)
dropdownYearDay.remove_widget(btndy1)
dropdownYearDay.remove_widget(btndy2)
root4.remove_widget(btnMonthDay)
dropdownYearMonth.remove_widget(btnmy11)
dropdownYearMonth.remove_widget(btnmy22)
btnDayDay.text = "Day v"
btnMonthDay.text = "Year v"
btnYearDay.text = "Month v"
def removeDWM1(self,*args):
removeDWM()
def screen4():
sc4.add_widget(root4)
root4.add_widget(frontBg4)
root4.add_widget(statsTitle)
root4.add_widget(b_sc4menu)
root4.add_widget(btnHelpGraph)
graphDown.add_widget(btnDaily)
graphDown.add_widget(btnWeekly)
graphDown.add_widget(btnMonthly)
graphDown.add_widget(btnAnnual)
#eto tinanggal ko tapos dito rin ako nagcode
root4.add_widget(graphButton)
t_consul = '0'
www = 0
cursor.execute("SELECT COUNT(*) FROM consultant_account")
www = int(cursor.fetchone()[0])
con.commit()
cursor.execute("SELECT id from consultant_account")
tmp_id = cursor.fetchone()[0]
con.commit()
cursor.execute("SELECT Control FROM consultant_account WHERE ID="+str(tmp_id))
t_consul = str(cursor.fetchone()[0])
con.commit()
if t_consul == '0':
statsTitle.text = 'User Garden'
root4.add_widget(petalG)
if t_consul == '1':
root4.add_widget(lblDescript)
root4.add_widget(lblTemp)
root4.add_widget(scrollData)
scrollData.add_widget(gridData)
#graph
boxStats.add_widget(nGraph)
#Screen5
def screen5(t_date,t_time,*args):
#--sc5
root5= FloatLayout(size_hint = (1.0, 1.0))
frontBg5 = Image(pos= root.pos,
source="img/white.jpg",
allow_stretch=True,
keep_ratio=False)
b_sc5menu = Button(text="[color=000000]T[/color]",
markup=True,
font_name='IconsSouthSt.ttf',
font_size='20sp',
pos_hint={'x':0.02,'center_y': 0.96},
background_normal="img/white.jpg",
size_hint=(.115,.08))
recentTitle = Button(text="[color=000000]Logs History[/color]",
markup=True,
pos_hint={'x':0,'center_y': 0.96},
font_name='helv.otf',
background_normal="img/white.jpg",
background_down="img/white.jpg",
font_size='25sp',
size_hint=(1,.09))
grid5_history=GridLayout(cols=1,
padding=10,
size_hint_y=None,
pos_hint={'center_x':.5, 'center_y':.48})
grid5_history.bind(minimum_height=grid5_history.setter('height'))
sv5_history=ScrollView(size_hint=(1,.75),
pos_hint={'center_x':.5, 'center_y':.5})
sc5.add_widget(root5)
root5.add_widget(frontBg5)
root5.add_widget(recentTitle)
root5.add_widget(b_sc5menu)
root5.add_widget(sv5_history)
sv5_history.add_widget(grid5_history)
b_sc5menu.bind(on_press=callbackscmain)
#nilagay ni joms
grid5_history.clear_widgets()
#con.commit()
dis_date = t_date.split('-')
cursor.execute("SELECT info from user_info WHERE year=? AND month=? AND day=? AND time=?",(dis_date[0],dis_date[1],dis_date[2],t_time ) )
wat = cursor.fetchall()[0][0]
def formatItem(right):
wrapped = fill(right, width=50)
return '{0}'.format(wrapped)
wat = formatItem(wat)
btnn = Button(text="[color=000000]"+t_date+' '+t_time+"\n\n{0}[/color]".format(wat),
#background_normal='img/feedbord.png',
size_hint=(1,None),
font_size='13sp',
background_color=[0,0,0,0],
#font_name='helv.otf',
height=300,
markup=True)
btnn2 = Button(text="",
size_hint=(1,None),
font_size='13sp',
background_color=[0,0,0,0],
#font_name='helv.otf',
height=200,
markup=True)
btnn3 = Button(text="",
size_hint=(1,None),
font_size='13sp',
background_color=[0,0,0,0],
#font_name='helv.otf',
height=200,
markup=True)
if len(wat)>300:
grid5_history.add_widget(btnn3)
if len(wat)<300:
btnn.height = 200
if len(wat)<150:
btnn.height = 150
grid5_history.add_widget(btnn)
grid5_history.add_widget(btnn2)
def callbacksc5(d,t,*args):
sm.transition = SwapTransition()
sm.current='fifthscreen'
## grid_history.clear_widgets()
## sv_history.clear_widgets()
## sv_history.remove_widget(grid_history)
## rootmain.remove_widget(sv_history)
screen5(d,t)
def getDB():
cursor.execute("SELECT * from user_info")
dt = cursor.fetchall()
tp = ''
#con.commit()
cursor.execute("SELECT COUNT(*) from user_info")
xx = cursor.fetchone()
#cursor.execute("SELECT COUNT(DISTINCT month) from user_info")
#print cursor.fetchone()
#con.commit()
def formatItem(right):
wrapped = fill(right, width=50)
return '{0}'.format(wrapped)
for i in range(xx[0],0,-1):
wat = str(dt[i-1][5])
wat = formatItem(wat)
if len(wat)>=40:
wat = wat[:36] + '.'*3
else:
wat = wat + '.'*4
_tmp_d = str(dt[i-1][1]) +'-'+ str(dt[i-1][2])+'-'+str(dt[i-1][3])
tp = _tmp_d + (' '*40)+str(dt[i-1][4])+'\n'+str(wat)
btn = Button(text="[color=000000]{0}[/color]".format(tp),
background_normal='img/feedbord.png',
#background_color=[0,0,0,0.5],
size_hint=(1,None),
font_size='13sp',
#font_name='helv.otf',
height=100,
markup=True)
btn.bind(on_press=partial(callbacksc5, _tmp_d, dt[i-1][4] ))
#btn.bind(on_release= get_dot)
grid_history.add_widget(btn)
def screenmain():
scmain.add_widget(rootmain)
rootmain.add_widget(frontBgmain)
rootmain.add_widget(logsTitle)
rootmain.add_widget(sv_history)
sv_history.add_widget(grid_history)
rootmain.add_widget(b_scmainmenu)
getDB()
def addbatt():
root6.add_widget(root6time)
root6.add_widget(root6batt)
root6.add_widget(root6lblbatt)
root6.add_widget(root6lbldate)
def removebatt():
root6.remove_widget(root6time)
root6.remove_widget(root6batt)
root6.remove_widget(root6lblbatt)
root6.remove_widget(root6lbldate)
def screen6():
sc6.add_widget(root6)
root6.add_widget(frontBg6)
root6.add_widget(root6Title)
root6.add_widget(mainbutton)
dropdown.add_widget(btnCP)
dropdown.add_widget(btnColor)
dropdown.add_widget(btnCons)
dropdown.add_widget(btnD)
dropdown.add_widget(btnA)
dropdown.add_widget(btnHelp)
dropdown.add_widget(btnLO)
root6.add_widget(b_logs)
root6.add_widget(b_home)
root6.add_widget(b_write)
root6.add_widget(b_stats)
root6.add_widget(lblHome)
root6.add_widget(lblStat)
root6.add_widget(lblLogs)
root6.add_widget(lblWrite)
#change stat into Garden
t_consul = 0
www = 0
cursor.execute("SELECT COUNT(*) FROM consultant_account")
www = int(cursor.fetchone()[0])
con.commit()
if www != 0:
cursor.execute("SELECT Control FROM consultant_account")
t_consul = int(cursor.fetchone()[0])
con.commit()
if t_consul == 0:
lblStat.text = "[color=ffffff]Garden[/color]"
if t_consul == 1:
lblStat.text = "[color=ffffff]Stats[/color]"
addbatt()
def screen3():
sc3.add_widget(root3)
root3.add_widget(frontBg3)
root3.add_widget(b_sc3)
root3.add_widget(input_sc3)
root3.add_widget(titlewrite)
root3.add_widget(label_write)
root3.add_widget(b_sc3menu)
def popPin():
## cursor.execute("SELECT COUNT(*) from user_account")
## if cursor.fetchone()[0] == 1:
bPopPin.add_widget(lblPin)
bPopPin.add_widget(tbPin1)
bPopPin.add_widget(lblPin2)
bPopPin.add_widget(tbPin2)
bPopPin.add_widget(lblDum)
bPopPin.add_widget(buttonPopPin)
popupPin.open()
## else:
## pass
def popGenderRemove1(self,*args):
cursor.execute("SELECT id from user_account")
tmp_id = cursor.fetchone()[0]
con.commit()
cursor.execute("UPDATE user_account SET gender=? WHERE id=?",('Male',tmp_id))
con.commit()
bPopGender.remove_widget(buttonPopGender1)
bPopGender.remove_widget(buttonPopGender2)
bPopGender.remove_widget(lblgender)
popupgender.dismiss()
popPin()
def popGenderRemove2(self,*args):
cursor.execute("SELECT id from user_account")
tmp_id = cursor.fetchone()[0]
con.commit()
cursor.execute("UPDATE user_account SET gender=? WHERE id=?",('Female',tmp_id))
con.commit()
bPopGender.remove_widget(buttonPopGender1)
bPopGender.remove_widget(buttonPopGender2)
bPopGender.remove_widget(lblgender)
popupgender.dismiss()
popPin()
def popColor(self,*args):
bPopColor.add_widget(buttonColorGreen)
bPopColor.add_widget(buttonColorRed)
bPopColor.add_widget(buttonColorBlue)
bPopColor.add_widget(buttonColorOrange)
bPopColor.add_widget(buttonColorPurple)
bPopColor.add_widget(buttonColorPGreen)
bPopColor.add_widget(buttonColorPRed)
bPopColor.add_widget(buttonColorPBlue)
bPopColor.add_widget(buttonColorPOrange)
bPopColor.add_widget(buttonColorPPurple)
bPopColor.add_widget(buttonPopColor)
popupColor.open()
def popColorRemove():
bPopColor.remove_widget(buttonColorGreen)
bPopColor.remove_widget(buttonColorRed)
bPopColor.remove_widget(buttonColorBlue)
bPopColor.remove_widget(buttonColorOrange)
bPopColor.remove_widget(buttonColorPurple)
bPopColor.remove_widget(buttonColorPGreen)
bPopColor.remove_widget(buttonColorPRed)
bPopColor.remove_widget(buttonColorPBlue)
bPopColor.remove_widget(buttonColorPOrange)
bPopColor.remove_widget(buttonColorPPurple)
bPopColor.remove_widget(buttonPopColor)
def ConsultantMode(self,*args):
if btnCons.text == '[color=ffffff]User Mode[/color]':
cursor.execute("SELECT id from consultant_account")
tmp_id = cursor.fetchone()[0]
con.commit()
cursor.execute("UPDATE consultant_account SET Control=? WHERE id=?",('0',tmp_id))
con.commit()
btnCons.text = '[color=ffffff]Consultant Mode[/color]'
bPopConsult.add_widget(lblConsult)
lblConsult.text="[color=000000]You're now in User Mode[/color]"
bPopConsult.add_widget(buttonPopConsult)
popupConsult.open()
def consuldismiss(self,*args):
bPopConsult.remove_widget(lblConsult)
bPopConsult.remove_widget(buttonPopConsult)
popupConsult.dismiss()
buttonPopConsult.bind(on_press=consuldismiss)
elif btnCons.text == '[color=ffffff]Consultant Mode[/color]':
cursor.execute("SELECT id from consultant_account")
tmp_id = cursor.fetchone()[0]
con.commit()
cursor.execute("UPDATE consultant_account SET Control=? WHERE id=?",('1',tmp_id))
con.commit()
btnCons.text = '[color=ffffff]User Mode[/color]'
bPopConsult.add_widget(lblConsult)
bPopConsult.add_widget(buttonPopConsult)
popupConsult.open()
def consuldismiss1(self,*args):
bPopConsult.remove_widget(lblConsult)
bPopConsult.remove_widget(buttonPopConsult)
popupConsult.dismiss()
buttonPopConsult.bind(on_press=consuldismiss1)
def MMHelp(self,*args):
bPopHelp.add_widget(carouselHelp)
carouselHelp.add_widget(lblHelp)
carouselHelp.add_widget(tut1)
carouselHelp.add_widget(tut2)
carouselHelp.add_widget(tut3)
carouselHelp.add_widget(tut4)
carouselHelp.add_widget(btnHelp1)
popuphelp.open()
def dismisshelp(self,*args):
bPopHelp.remove_widget(carouselHelp)
carouselHelp.remove_widget(lblHelp)
carouselHelp.remove_widget(tut1)
carouselHelp.remove_widget(tut2)
carouselHelp.remove_widget(tut3)
carouselHelp.remove_widget(tut4)
carouselHelp.remove_widget(btnHelp1)
popuphelp.dismiss()
btnHelp1.bind(on_press=dismisshelp)
def screen2():
popname()
sc2.add_widget(root2)
root2.add_widget(frontBg2)
root2.add_widget(txtInp2)
root2.add_widget(passimg1)
root2.add_widget(passimg2)
root2.add_widget(passimg3)
root2.add_widget(passimg4)
root2.add_widget(passimg5)
root2.add_widget(passimg6)
root2.add_widget(passimg7)
root2.add_widget(passimg8)
root2.add_widget(passimg9)
root2.add_widget(passimgclr)
root2.add_widget(passimg0)
root2.add_widget(passimgok)
root2.add_widget(titlepin)
def screen7():
root7.remove_widget(frontBg7)
root7.remove_widget(userTitle)
root7.remove_widget(b_sc7menu)
root7.remove_widget(petalWhole)
sc7.remove_widget(root7)
sc7.add_widget(root7)
root7.add_widget(frontBg7)
root7.add_widget(userTitle)
root7.add_widget(petalWhole)
root7.add_widget(b_sc7menu)
#--EVENTS-----------------------
def p1(self,*args):
txtInp2.text += "1"
def p2(self,*args):
txtInp2.text += "2"
def p3(self,*args):
txtInp2.text += "3"
def p4(self,*args):
txtInp2.text += "4"
def p5(self,*args):
txtInp2.text += "5"
def p6(self,*args):
txtInp2.text += "6"
def p7(self,*args):
txtInp2.text += "7"
def p8(self,*args):
txtInp2.text += "8"
def p9(self,*args):
txtInp2.text += "9"
def p0(self,*args):
txtInp2.text += "0"
def pclr(self,*args):
txtInp2.text = ""
def feedremove():
#nilagay ni joms
grid5_history.clear_widgets()
sv5_history.clear_widgets()
grid_history.clear_widgets()
sv_history.clear_widgets()
scmain.remove_widget(rootmain)
rootmain.remove_widget(frontBgmain)
rootmain.remove_widget(logsTitle)
sv_history.remove_widget(grid_history)
rootmain.remove_widget(sv_history)
rootmain.remove_widget(b_scmainmenu)
def feedadd():
feedremove()
rootmain.add_widget(sv_history)
sv_history.add_widget(grid_history)
getDB()
def popsave():
bPopSave.add_widget(lblsave)
bPopSave.add_widget(buttonPopSave)
popupsave.open()
def callback_proctext(self,*args):
thoughts_text = input_sc3.text
thoughts_text = thoughts_text.lower()
thoughts_text = re.sub( '\s+', ' ', thoughts_text ).strip()
thoughts_text = re.sub(r"[^\w.,']", ' ', thoughts_text)
#thoughts_text = re.sub(r"[^\w]", ' ', thoughts_text)
#print thoughts_text
if str(thoughts_text).isspace() == False and str(thoughts_text).strip() != '':
date_now = datetime.datetime.now().strftime('%Y-%m-%d')
time_now = datetime.datetime.now().strftime('%I:%M %p')
p_pos = naive_classifier.classify_polarity(thoughts_text)[0]
p_neg = naive_classifier.classify_polarity(thoughts_text)[1]
cursor.execute("INSERT INTO user_info(year,month,day,time,info,mood,pos,neg) VALUES(?,?,?,?,?,?,?,?)",(date_now.split('-')[0],date_now.split('-')[1],date_now.split('-')[2],time_now,thoughts_text,str(mood_value),p_pos,p_neg))
con.commit()
input_sc3.text = ""
input_sc3.clear_widgets()
popsave()
feedremove()
else:
boxErrorWrite = BoxLayout(orientation="vertical")
buttonErrorWrite = Button(text="CLOSE",
background='img/blue.jpg',
size_hint=(1,.5))
lblErrorWrite2=Label(text="[color=000000]CANNOT SAVE EMPTY LOG[/color]",
markup=True)
boxErrorWrite.add_widget(lblErrorWrite2)
boxErrorWrite.add_widget(buttonErrorWrite)
popuperrorwrite= Popup(title='',
separator_color=[0,0,1,0.5],
background='img/white.jpg',
content=boxErrorWrite,
auto_dismiss=False,
size_hint=(0.6, 0.2))
popuperrorwrite.open()
def errorclick(self,*args):
popuperrorwrite.dismiss()
buttonErrorWrite.bind(on_press=errorclick)
#print 'walang laman'
def callbackmain(self,*args):
cursor.execute("SELECT password from user_account")
pwd = cursor.fetchall()[0]
#password is 8888
con.commit()
#txtInp2.text = '8888'
if txtInp2.text == pwd[0]:
sm.current='sixthscreen'
#always update into ordinary user
cursor.execute("SELECT id from consultant_account")
tmp_id = cursor.fetchone()[0]
con.commit()
cursor.execute("UPDATE consultant_account SET Control=? WHERE id=?",('0',tmp_id))
con.commit()
screen6()
popup.open()
screen3()
screen4()
else:
pass
def popsc3():
bPop3.add_widget(lblmood)
bPop3.add_widget(bPop33)
bPop33.add_widget(bAngry)
bPop33.add_widget(bSad)
bPop33.add_widget(bHappy3)
bPop33.add_widget(bHappy2)
bPop33.add_widget(bHappy)
bPop3.add_widget(tryslide)
bPop3.add_widget(btnSlider)
popup3.open()
def popexit(self,*args):
bPopExit.add_widget(lblexit)
bPopExit.add_widget(bPopExit2)
bPopExit2.add_widget(buttonPopExit1)
bPopExit2.add_widget(buttonPopExit2)
popupexit.open()
def popcp(self,*args):
bPopCp.add_widget(lblCp1)
bPopCp.add_widget(tbCp1)
bPopCp.add_widget(lblCp2)
bPopCp.add_widget(tbCp2)
bPopCp.add_widget(lblCp3)
bPopCp.add_widget(tbCp3)
bPopCp.add_widget(lblblank)
bPopCp.add_widget(lblnotice)
bPopCp.add_widget(buttonPopCpExit)
bPopCp.add_widget(buttonPopCp)
lblnotice.text=""
lblblank.text=""
popupCp.open()
def popdev(self,*args):
bPopD.add_widget(imgBruno)
bPopD.add_widget(imgJoms)
bPopD.add_widget(lblDev2)
bPopD.add_widget(lblDev1)
bPopD.add_widget(lbldevblank)
bPopD.add_widget(buttonPopD)
popupDev.open()
def popabout(self,*args):
bPopAb.add_widget(lblabout)
bPopAb.add_widget(buttonPopAb)
popupabout.open()
bPopError = GridLayout(cols=1,padding=5,spacing = 5)
lblerror= Label(text="[color=000000]Invalid Password![/color]",
# font_name='helv.otf',
markup=True,halign='center',
size_hint=(1,.5),
font_size='10sp')
buttonPopError = Button(text="Close",
#font_name='helv.otf',
size_hint=(.5,.2),
background_normal="img/roundblue.jpg")
popuperror = Popup(background="img/white.jpg",
separator_color=[1,0,0,0.5],
content=bPopError,
title="Error",
auto_dismiss=False,title_color=[0,0,0,1],
size_hint=(0.5, 0.2))
#--devmode
bPopConsult = GridLayout(cols=1,padding=5,spacing = 5)
lblConsult= Label(text="[color=000000]You're now in Consultant Mode[/color]",
# font_name='helv.otf',
markup=True,halign='center',
size_hint=(1,.5),
font_size='10sp')
buttonPopConsult = Button(text="Close",
#font_name='helv.otf',
size_hint=(.5,.2),
background_normal="img/roundblue.jpg")
popupConsult = Popup(background="img/white.jpg",
separator_color=[1,0,0,0.5],
content=bPopConsult,
title=" ",
auto_dismiss=False,
size_hint=(0.5, 0.2))
def popError(self, *args):
bPopError.add_widget(lblerror)
bPopError.add_widget(buttonPopError)
cursor.execute("SELECT password from user_account")
pwd = cursor.fetchall()[0]
con.commit()
if txtInp2.text != pwd[0]:
popuperror.open()
def callbacksc2(self,*args):
sm.current='secondscreen'
screen2()
def callbacksc3(self,*args):
sm.transition = SwapTransition()
sm.current='thirdscreen'
input_sc3.text=""
#popsc3()
def callbacksc4(self,*args):
#back to main screen from graph
sm.transition = SwapTransition()
sm.current='fourthscreen'
root4.add_widget(boxStats)
#root4.add_widget(btnCont)
wplot = SmoothLinePlot(mode="points",color=rgb('adaa0a'))
removeDWM()
def callbacksc7(self,*args):
sm.transition = SwapTransition()
sm.current='seventhscreen'
screen7()
def mooddismiss():
popup3.remove_widget(bPop3)
bPop3.remove_widget(bPop33)
bPop3.remove_widget(lblmood)
bPop33.remove_widget(bHappy)
bPop33.remove_widget(bHappy2)
bPop33.remove_widget(bHappy3)
bPop33.remove_widget(bSad)
bPop33.remove_widget(bAngry)
bPop3.remove_widget(tryslide)
bPop3.remove_widget(btnSlider)
def popexitdismiss(self,*args):
popupexit.dismiss()
bPopExit2.remove_widget(buttonPopExit1)
bPopExit2.remove_widget(buttonPopExit2)
bPopExit.remove_widget(bPopExit2)
bPopExit.remove_widget(lblexit)
def popaboutdismiss(self,*args):
popupabout.dismiss()
bPopAb.remove_widget(buttonPopAb)
bPopAb.remove_widget(lblabout)
def poperrordismiss(self,*args):
txtInp2.text = ''
popuperror.dismiss()
bPopError.remove_widget(buttonPopError)
bPopError.remove_widget(lblerror)
def popdevdismiss(self,*args):
popupDev.dismiss()
bPopD.remove_widget(buttonPopD)
bPopD.remove_widget(lblDev1)
bPopD.remove_widget(lblDev2)
bPopD.remove_widget(imgBruno)
bPopD.remove_widget(imgJoms)
bPopD.remove_widget(lbldevblank)
def popsavenotidismiss(self,*args):
popupsave.dismiss()
bPopSave.remove_widget(buttonPopSave)
bPopSave.remove_widget(lblsave)
def popcpdismiss(self,*args):
popupCp.dismiss()
tbCp1.text=""
tbCp2.text=""
tbCp3.text=""
bPopCp.remove_widget(buttonPopCp)
bPopCp.remove_widget(buttonPopCpExit)
bPopCp.remove_widget(lblCp1)
bPopCp.remove_widget(lblCp2)
bPopCp.remove_widget(lblCp3)
bPopCp.remove_widget(tbCp1)
bPopCp.remove_widget(tbCp2)
bPopCp.remove_widget(tbCp3)
bPopCp.remove_widget(lblnotice)
bPopCp.remove_widget(lblblank)
def popnamedismiss(self,*args):
if tbName.text in ['',' ']:
pass
else:
cursor.execute("INSERT INTO user_account(name) VALUES('"+str(tbName.text) +"')" )
con.commit()
popupname.dismiss()
bPopName.remove_widget(lblname)
bPopName.remove_widget(tbName)
bPopName.remove_widget(buttonPopName)
popgender()
#popPin()
def savepoppin(self, *args):
cursor.execute("SELECT id from user_account")
tmp_pass = cursor.fetchone()[0]
con.commit()
if tbPin1.text == tbPin2.text:
if tbPin1.text in ['',' ']:
pass
else:
cursor.execute("UPDATE user_account SET password=? WHERE id=?",(str(tbPin1.text),tmp_pass))
con.commit()
bPopPin.remove_widget(tbPin1)
bPopPin.remove_widget(lblPin)
bPopPin.remove_widget(lblPin2)
bPopPin.remove_widget(tbPin2)
bPopPin.remove_widget(lblDum)
bPopPin.remove_widget(buttonPopPin)
popupPin.dismiss()
else:
print 'mali pass'
pass
#Define Change Password Function
def popcpsave(self,*args):
cursor.execute("SELECT * FROM user_account")
temp_pass = cursor.fetchall()
con.commit()
#print temp_pass[0][1]
if tbCp1.text == temp_pass[0][1]:
if tbCp2.text == tbCp3.text:
if tbCp2.text and tbCp3.text == None or len(tbCp2.text) <=3:
pass
else:
cursor.execute("UPDATE user_account SET password=? WHERE id=?",(str(tbCp2.text),temp_pass[0][0] ))
con.commit()
#print 'oki doki'
popupCp.dismiss()
tbCp1.text=""
tbCp2.text=""
tbCp3.text=""
bPopCp.remove_widget(buttonPopCp)
bPopCp.remove_widget(buttonPopCpExit)
bPopCp.remove_widget(lblCp1)
bPopCp.remove_widget(lblCp2)
bPopCp.remove_widget(lblCp3)
bPopCp.remove_widget(tbCp1)
bPopCp.remove_widget(tbCp2)
bPopCp.remove_widget(tbCp3)
bPopCp.remove_widget(lblnotice)
bPopCp.remove_widget(lblblank)
else:
lblnotice.markup = True
lblblank.markup = True
lblnotice.text="[color=ff00ff][b]PINS DOES NOT MATCH![/b][/color]"
lblblank.text="[color=ff00ff][b]ERROR:[/b][/color]"
#print 'mali'
def callbackhome(self,*args):
sm.transition = SwapTransition()
sm.current='sixthscreen'
mooddismiss()
root4.remove_widget(boxStats)
root4.remove_widget(btnCont)
root4.remove_widget(petalG)
root4.remove_widget(lblDescript)
root4.remove_widget(lblTemp)
root4.remove_widget(scrollData)
scrollData.remove_widget(gridData)
boxStats.remove_widget(nGraph)
root4.remove_widget(boxStats)
removebatt()
root6time.text="[color=ffffff]"+datetime.datetime.now().strftime('%I:%M ')+"[/color]"
root6batt.text="[color=ffffff][b]"+str(battery.status['percentage'])+"%[/b][/color]"
root6lbldate.text="[color=ffffff]"+datetime.datetime.now().strftime('%A %d %B %Y')+"[/color]"
addbatt()
popupsave.dismiss()
bPopSave.remove_widget(buttonPopSave)
bPopSave.remove_widget(lblsave)
def callbacksc6(self,*args):
sm.transition = SwapTransition()
sm.current='sixthscreen'
def callbackscmain(self,*args):
sm.transition = SwapTransition()
sm.current='mainscreen'
feedremove()
screenmain()
def exitclose(self,*args):
App.get_running_app().stop()
#Define Mood Confirm Function
def moodconfirm(self,*args):
popup3.dismiss()
#color_schemes
def colorRed(self,*args):
color_val = 'red'
cursor.execute('SELECT COUNT(*) FROM color_tbl')
color_count = cursor.fetchone()[0]
if color_count == 0:
cursor.execute("INSERT INTO color_tbl(color) VALUES('"+color_val+"')")
con.commit()
elif color_count >= 1:
cursor.execute("UPDATE color_tbl SET color='"+color_val+"'")
con.commit()
btnhelpmeok.background_normal="img/red.jpg"
btnHelpGraph.background_normal="img/red.jpg"
buttonPop.background_normal="img/red.jpg"
titlewrite.background_down="img/red.jpg"
logsTitle.background_down="img/red.jpg"
recentTitle.background_down="img/red.jpg"
statsTitle.background_down="img/red.jpg"
root6Title.background_down="img/red.jpg"
root6Title.background_normal="img/red.jpg"
buttonPopExit1.background_normal="img/red.jpg"
buttonPopExit2.background_normal="img/red.jpg"
buttonPop.background_normal="img/red.jpg"
buttonPopAb.background_normal="img/red.jpg"
buttonPopD.background_normal="img/red.jpg"
buttonPopName.background_normal="img/red.jpg"
buttonPopCpExit.background_normal="img/red.jpg"
btnCP.background_normal="img/red.jpg"
btnCons.background_normal="img/red.jpg"
btnHelp.background_normal="img/red.jpg"
btnColor.background_normal="img/red.jpg"
btnD.background_normal="img/red.jpg"
btnA.background_normal="img/red.jpg"
btnLO.background_normal="img/red.jpg"
mainbutton.background_normal="img/red.jpg"
buttonPopName.background_normal="img/red.jpg"
btnDaily.background_normal="img/red.jpg"
btnWeekly.background_normal="img/red.jpg"
btnMonthly.background_normal="img/red.jpg"
btnAnnual.background_normal="img/red.jpg"
graphButton.background_normal="img/red.jpg"
b_sc4menu.background_normal="img/red.jpg"
statsTitle.background_normal="img/red.jpg"
btnd1.background_normal="img/red.jpg"
btnd2.background_normal="img/red.jpg"
btnDayDay.background_normal="img/red.jpg"
btndy1.background_normal="img/red.jpg"
btndy2.background_normal="img/red.jpg"
btnYearDay.background_normal="img/red.jpg"
btnmy11.background_normal="img/red.jpg"
btnmy22.background_normal="img/red.jpg"
btnMonthDay.background_normal="img/red.jpg"
btnwd1.background_normal="img/red.jpg"
btnWeekDay.background_normal="img/red.jpg"
btnwy1.background_normal="img/red.jpg"
btnWeekYear.background_normal="img/red.jpg"
btnwm1.background_normal="img/red.jpg"
btnWeekMonth.background_normal="img/red.jpg"
btnmm1.background_normal="img/red.jpg"
btnmm2.background_normal="img/red.jpg"
btnmy1.background_normal="img/red.jpg"
btnmy2.background_normal="img/red.jpg"
btnMon1.background_normal="img/red.jpg"
btnMon2.background_normal="img/red.jpg"
btnann1.background_normal="img/red.jpg"
btnAnnualYear.background_normal="img/red.jpg"
b_scmainmenu.background_normal="img/red.jpg"
logsTitle.background_normal="img/red.jpg"
b_sc5menu.background_normal="img/red.jpg"
recentTitle.background_normal="img/red.jpg"
b_sc3menu.background_normal="img/red.jpg"
buttonPopSave.background_normal="img/red.jpg"
b_sc3.background_normal="img/red.jpg"
titlewrite.background_normal="img/red.jpg"
btnSlider.background_normal="img/red.jpg"
buttonPopCp.background_normal="img/red.jpg"
b_home.text="[color=ff0000]D[/color]"
lblHome.text="[color=ff0000][b]Home[/b][/color]"
def colorBlue(self,*args):
color_val = 'blue'
cursor.execute('SELECT COUNT(*) FROM color_tbl')
color_count = cursor.fetchone()[0]
if color_count == 0:
cursor.execute("INSERT INTO color_tbl(color) VALUES('"+color_val+"')")
con.commit()
elif color_count >= 1:
cursor.execute("UPDATE color_tbl SET color='"+color_val+"'")
con.commit()
btnhelpmeok.background_normal="img/blue.jpg"
btnHelpGraph.background_normal="img/blue.jpg"
buttonPop.background_down="img/blue.jpg"
titlewrite.background_down="img/blue.jpg"
logsTitle.background_down="img/blue.jpg"
recentTitle.background_down="img/blue.jpg"
statsTitle.background_down="img/blue.jpg"
root6Title.background_down="img/blue.jpg"
root6Title.background_normal="img/blue.jpg"
buttonPopExit1.background_normal="img/blue.jpg"
buttonPopExit2.background_normal="img/blue.jpg"
buttonPop.background_normal="img/blue.jpg"
buttonPopAb.background_normal="img/blue.jpg"
buttonPopD.background_normal="img/blue.jpg"
buttonPopName.background_normal="img/blue.jpg"
buttonPopCpExit.background_normal="img/blue.jpg"
btnCP.background_normal="img/blue.jpg"
btnCons.background_normal="img/blue.jpg"
btnHelp.background_normal="img/blue.jpg"
btnColor.background_normal="img/blue.jpg"
btnD.background_normal="img/blue.jpg"
btnA.background_normal="img/blue.jpg"
btnLO.background_normal="img/blue.jpg"
mainbutton.background_normal="img/blue.jpg"
buttonPopName.background_normal="img/blue.jpg"
btnDaily.background_normal="img/blue.jpg"
btnWeekly.background_normal="img/blue.jpg"
btnMonthly.background_normal="img/blue.jpg"
btnAnnual.background_normal="img/blue.jpg"
graphButton.background_normal="img/blue.jpg"
b_sc4menu.background_normal="img/blue.jpg"
statsTitle.background_normal="img/blue.jpg"
btnd1.background_normal="img/blue.jpg"
btnd2.background_normal="img/blue.jpg"
btnDayDay.background_normal="img/blue.jpg"
btndy1.background_normal="img/blue.jpg"
btndy2.background_normal="img/blue.jpg"
btnYearDay.background_normal="img/blue.jpg"
btnmy11.background_normal="img/blue.jpg"
btnmy22.background_normal="img/blue.jpg"
btnMonthDay.background_normal="img/blue.jpg"
btnwd1.background_normal="img/blue.jpg"
btnWeekDay.background_normal="img/blue.jpg"
btnwy1.background_normal="img/blue.jpg"
btnWeekYear.background_normal="img/blue.jpg"
btnwm1.background_normal="img/blue.jpg"
btnWeekMonth.background_normal="img/blue.jpg"
btnmm1.background_normal="img/blue.jpg"
btnmm2.background_normal="img/blue.jpg"
btnmy1.background_normal="img/blue.jpg"
btnmy2.background_normal="img/blue.jpg"
btnMon1.background_normal="img/blue.jpg"
btnMon2.background_normal="img/blue.jpg"
btnann1.background_normal="img/blue.jpg"
btnAnnualYear.background_normal="img/blue.jpg"
b_scmainmenu.background_normal="img/blue.jpg"
logsTitle.background_normal="img/blue.jpg"
b_sc5menu.background_normal="img/blue.jpg"
recentTitle.background_normal="img/blue.jpg"
b_sc3menu.background_normal="img/blue.jpg"
buttonPopSave.background_normal="img/blue.jpg"
b_sc3.background_normal="img/blue.jpg"
titlewrite.background_normal="img/blue.jpg"
btnSlider.background_normal="img/blue.jpg"
buttonPopCp.background_normal="img/blue.jpg"
b_home.text="[color=3b5999]D[/color]"
lblHome.text="[color=3b5999][b]Home[/b][/color]"
def colorGreen(self,*args):
color_val = 'green'
cursor.execute('SELECT COUNT(*) FROM color_tbl')
color_count = cursor.fetchone()[0]
if color_count == 0:
cursor.execute("INSERT INTO color_tbl(color) VALUES('"+color_val+"')")
con.commit()
elif color_count >= 1:
cursor.execute("UPDATE color_tbl SET color='"+color_val+"'")
con.commit()
btnhelpmeok.background_normal="img/green.jpg"
btnHelpGraph.background_normal="img/green.jpg"
buttonPop.background_down="img/green.jpg"
titlewrite.background_down="img/green.jpg"
logsTitle.background_down="img/green.jpg"
recentTitle.background_down="img/green.jpg"
statsTitle.background_down="img/green.jpg"
root6Title.background_down="img/green.jpg"
root6Title.background_normal="img/green.jpg"
buttonPopExit1.background_normal="img/green.jpg"
buttonPopExit2.background_normal="img/green.jpg"
buttonPop.background_normal="img/green.jpg"
buttonPopAb.background_normal="img/green.jpg"
buttonPopD.background_normal="img/green.jpg"
buttonPopName.background_normal="img/green.jpg"
buttonPopCpExit.background_normal="img/green.jpg"
btnCP.background_normal="img/green.jpg"
btnCons.background_normal="img/green.jpg"
btnHelp.background_normal="img/green.jpg"
btnColor.background_normal="img/green.jpg"
btnD.background_normal="img/green.jpg"
btnA.background_normal="img/green.jpg"
btnLO.background_normal="img/green.jpg"
mainbutton.background_normal="img/green.jpg"
buttonPopName.background_normal="img/green.jpg"
btnDaily.background_normal="img/green.jpg"
btnWeekly.background_normal="img/green.jpg"
btnMonthly.background_normal="img/green.jpg"
btnAnnual.background_normal="img/green.jpg"
graphButton.background_normal="img/green.jpg"
b_sc4menu.background_normal="img/green.jpg"
statsTitle.background_normal="img/green.jpg"
btnd1.background_normal="img/green.jpg"
btnd2.background_normal="img/green.jpg"
btnDayDay.background_normal="img/green.jpg"
btndy1.background_normal="img/green.jpg"
btndy2.background_normal="img/green.jpg"
btnYearDay.background_normal="img/green.jpg"
btnmy11.background_normal="img/green.jpg"
btnmy22.background_normal="img/green.jpg"
btnMonthDay.background_normal="img/green.jpg"
btnwd1.background_normal="img/green.jpg"
btnWeekDay.background_normal="img/green.jpg"
btnwy1.background_normal="img/green.jpg"
btnWeekYear.background_normal="img/green.jpg"
btnwm1.background_normal="img/green.jpg"
btnWeekMonth.background_normal="img/green.jpg"
btnmm1.background_normal="img/green.jpg"
btnmm2.background_normal="img/green.jpg"
btnmy1.background_normal="img/green.jpg"
btnmy2.background_normal="img/green.jpg"
btnMon1.background_normal="img/green.jpg"
btnMon2.background_normal="img/green.jpg"
btnann1.background_normal="img/green.jpg"
btnAnnualYear.background_normal="img/green.jpg"
b_scmainmenu.background_normal="img/green.jpg"
logsTitle.background_normal="img/green.jpg"
b_sc5menu.background_normal="img/green.jpg"
recentTitle.background_normal="img/green.jpg"
b_sc3menu.background_normal="img/green.jpg"
buttonPopSave.background_normal="img/green.jpg"
b_sc3.background_normal="img/green.jpg"
titlewrite.background_normal="img/green.jpg"
btnSlider.background_normal="img/green.jpg"
buttonPopCp.background_normal="img/green.jpg"
b_home.text="[color=00ff00]D[/color]"
lblHome.text="[color=00ff00][b]Home[/b][/color]"
def colorOrange(self,*args):
color_val = 'orange'
cursor.execute('SELECT COUNT(*) FROM color_tbl')
color_count = cursor.fetchone()[0]
if color_count == 0:
cursor.execute("INSERT INTO color_tbl(color) VALUES('"+color_val+"')")
con.commit()
elif color_count >= 1:
cursor.execute("UPDATE color_tbl SET color='"+color_val+"'")
con.commit()
btnhelpmeok.background_normal="img/orange.jpg"
btnHelpGraph.background_normal="img/orange.jpg"
buttonPop.background_down="img/orange.jpg"
titlewrite.background_down="img/orange.jpg"
logsTitle.background_down="img/orange.jpg"
recentTitle.background_down="img/orange.jpg"
statsTitle.background_down="img/orange.jpg"
root6Title.background_down="img/orange.jpg"
root6Title.background_normal="img/orange.jpg"
buttonPopExit1.background_normal="img/orange.jpg"
buttonPopExit2.background_normal="img/orange.jpg"
buttonPop.background_normal="img/orange.jpg"
buttonPopAb.background_normal="img/orange.jpg"
buttonPopD.background_normal="img/orange.jpg"
buttonPopName.background_normal="img/orange.jpg"
buttonPopCpExit.background_normal="img/orange.jpg"
btnCP.background_normal="img/orange.jpg"
btnCons.background_normal="img/orange.jpg"
btnHelp.background_normal="img/orange.jpg"
btnColor.background_normal="img/orange.jpg"
btnD.background_normal="img/orange.jpg"
btnA.background_normal="img/orange.jpg"
btnLO.background_normal="img/orange.jpg"
mainbutton.background_normal="img/orange.jpg"
buttonPopName.background_normal="img/orange.jpg"
btnDaily.background_normal="img/orange.jpg"
btnWeekly.background_normal="img/orange.jpg"
btnMonthly.background_normal="img/orange.jpg"
btnAnnual.background_normal="img/orange.jpg"
graphButton.background_normal="img/orange.jpg"
b_sc4menu.background_normal="img/orange.jpg"
statsTitle.background_normal="img/orange.jpg"
btnd1.background_normal="img/orange.jpg"
btnd2.background_normal="img/orange.jpg"
btnDayDay.background_normal="img/orange.jpg"
btndy1.background_normal="img/orange.jpg"
btndy2.background_normal="img/orange.jpg"
btnYearDay.background_normal="img/orange.jpg"
btnmy11.background_normal="img/orange.jpg"
btnmy22.background_normal="img/orange.jpg"
btnMonthDay.background_normal="img/orange.jpg"
btnwd1.background_normal="img/orange.jpg"
btnWeekDay.background_normal="img/orange.jpg"
btnwy1.background_normal="img/orange.jpg"
btnWeekYear.background_normal="img/orange.jpg"
btnwm1.background_normal="img/orange.jpg"
btnWeekMonth.background_normal="img/orange.jpg"
btnmm1.background_normal="img/orange.jpg"
btnmm2.background_normal="img/orange.jpg"
btnmy1.background_normal="img/orange.jpg"
btnmy2.background_normal="img/orange.jpg"
btnMon1.background_normal="img/orange.jpg"
btnann1.background_normal="img/orange.jpg"
btnAnnualYear.background_normal="img/orange.jpg"
b_scmainmenu.background_normal="img/orange.jpg"
logsTitle.background_normal="img/orange.jpg"
b_sc5menu.background_normal="img/orange.jpg"
recentTitle.background_normal="img/orange.jpg"
b_sc3menu.background_normal="img/orange.jpg"
buttonPopSave.background_normal="img/orange.jpg"
b_sc3.background_normal="img/orange.jpg"
titlewrite.background_normal="img/orange.jpg"
btnSlider.background_normal="img/orange.jpg"
buttonPopCp.background_normal="img/orange.jpg"
b_home.text="[color=ff4500]D[/color]"
lblHome.text="[color=ff4500][b]Home[/b][/color]"
def colorPurple(self,*args):
color_val = 'purple'
cursor.execute('SELECT COUNT(*) FROM color_tbl')
color_count = cursor.fetchone()[0]
if color_count == 0:
cursor.execute("INSERT INTO color_tbl(color) VALUES('"+color_val+"')")
con.commit()
elif color_count >= 1:
cursor.execute("UPDATE color_tbl SET color='"+color_val+"'")
con.commit()
btnhelpmeok.background_normal="img/purple.jpg"
btnHelpGraph.background_normal="img/purple.jpg"
buttonPop.background_normal="img/purple.jpg"
titlewrite.background_down="img/purple.jpg"
logsTitle.background_down="img/purple.jpg"
recentTitle.background_down="img/purple.jpg"
statsTitle.background_down="img/purple.jpg"
root6Title.background_down="img/purple.jpg"
root6Title.background_normal="img/purple.jpg"
buttonPopExit1.background_normal="img/purple.jpg"
buttonPopExit2.background_normal="img/purple.jpg"
buttonPop.background_normal="img/purple.jpg"
buttonPopAb.background_normal="img/purple.jpg"
buttonPopD.background_normal="img/purple.jpg"
buttonPopName.background_normal="img/purple.jpg"
buttonPopCpExit.background_normal="img/purple.jpg"
btnCP.background_normal="img/purple.jpg"
btnCons.background_normal="img/purple.jpg"
btnHelp.background_normal="img/purple.jpg"
btnColor.background_normal="img/purple.jpg"
btnD.background_normal="img/purple.jpg"
btnA.background_normal="img/purple.jpg"
btnLO.background_normal="img/purple.jpg"
mainbutton.background_normal="img/purple.jpg"
buttonPopName.background_normal="img/purple.jpg"
btnDaily.background_normal="img/purple.jpg"
btnWeekly.background_normal="img/purple.jpg"
btnMonthly.background_normal="img/purple.jpg"
btnAnnual.background_normal="img/purple.jpg"
graphButton.background_normal="img/purple.jpg"
b_sc4menu.background_normal="img/purple.jpg"
statsTitle.background_normal="img/purple.jpg"
btnd1.background_normal="img/purple.jpg"
btnd2.background_normal="img/purple.jpg"
btnDayDay.background_normal="img/purple.jpg"
btndy1.background_normal="img/purple.jpg"
btndy2.background_normal="img/purple.jpg"
btnYearDay.background_normal="img/purple.jpg"
btnmy11.background_normal="img/purple.jpg"
btnmy22.background_normal="img/purple.jpg"
btnMonthDay.background_normal="img/purple.jpg"
btnwd1.background_normal="img/purple.jpg"
btnWeekDay.background_normal="img/purple.jpg"
btnwy1.background_normal="img/purple.jpg"
btnWeekYear.background_normal="img/purple.jpg"
btnwm1.background_normal="img/purple.jpg"
btnWeekMonth.background_normal="img/purple.jpg"
btnmm1.background_normal="img/purple.jpg"
btnmm2.background_normal="img/purple.jpg"
btnmy1.background_normal="img/purple.jpg"
btnmy2.background_normal="img/purple.jpg"
btnMon1.background_normal="img/purple.jpg"
btnann1.background_normal="img/purple.jpg"
btnAnnualYear.background_normal="img/purple.jpg"
b_scmainmenu.background_normal="img/purple.jpg"
logsTitle.background_normal="img/purple.jpg"
b_sc5menu.background_normal="img/purple.jpg"
recentTitle.background_normal="img/purple.jpg"
b_sc3menu.background_normal="img/purple.jpg"
buttonPopSave.background_normal="img/purple.jpg"
b_sc3.background_normal="img/purple.jpg"
titlewrite.background_normal="img/purple.jpg"
btnSlider.background_normal="img/purple.jpg"
buttonPopCp.background_normal="img/purple.jpg"
b_home.text="[color=ba55d3]D[/color]"
lblHome.text="[color=ba55d3][b]Home[/b][/color]"
def colorPRed(self,*args):
color_val = 'pred'
cursor.execute('SELECT COUNT(*) FROM color_tbl')
color_count = cursor.fetchone()[0]
if color_count == 0:
cursor.execute("INSERT INTO color_tbl(color) VALUES('"+color_val+"')")
con.commit()
elif color_count >= 1:
cursor.execute("UPDATE color_tbl SET color='"+color_val+"'")
con.commit()
btnhelpmeok.background_normal="img/pred.jpg"
btnHelpGraph.background_normal="img/pred.jpg"
buttonPop.background_down="img/pred.jpg"
titlewrite.background_down="img/pred.jpg"
logsTitle.background_down="img/pred.jpg"
recentTitle.background_down="img/pred.jpg"
statsTitle.background_down="img/pred.jpg"
root6Title.background_down="img/pred.jpg"
root6Title.background_normal="img/pred.jpg"
buttonPopExit1.background_normal="img/pred.jpg"
buttonPopExit2.background_normal="img/pred.jpg"
buttonPop.background_normal="img/pred.jpg"
buttonPopAb.background_normal="img/pred.jpg"
buttonPopD.background_normal="img/pred.jpg"
buttonPopName.background_normal="img/pred.jpg"
buttonPopCpExit.background_normal="img/pred.jpg"
btnCP.background_normal="img/pred.jpg"
btnCons.background_normal="img/pred.jpg"
btnHelp.background_normal="img/pred.jpg"
btnColor.background_normal="img/pred.jpg"
btnD.background_normal="img/pred.jpg"
btnA.background_normal="img/pred.jpg"
btnLO.background_normal="img/pred.jpg"
mainbutton.background_normal="img/pred.jpg"
buttonPopName.background_normal="img/pred.jpg"
btnDaily.background_normal="img/pred.jpg"
btnWeekly.background_normal="img/pred.jpg"
btnMonthly.background_normal="img/pred.jpg"
btnAnnual.background_normal="img/pred.jpg"
graphButton.background_normal="img/pred.jpg"
b_sc4menu.background_normal="img/pred.jpg"
statsTitle.background_normal="img/pred.jpg"
btnd1.background_normal="img/pred.jpg"
btnd2.background_normal="img/pred.jpg"
btnDayDay.background_normal="img/pred.jpg"
btndy1.background_normal="img/pred.jpg"
btndy2.background_normal="img/pred.jpg"
btnYearDay.background_normal="img/pred.jpg"
btnmy11.background_normal="img/pred.jpg"
btnmy22.background_normal="img/pred.jpg"
btnMonthDay.background_normal="img/pred.jpg"
btnwd1.background_normal="img/pred.jpg"
btnWeekDay.background_normal="img/pred.jpg"
btnwy1.background_normal="img/pred.jpg"
btnWeekYear.background_normal="img/pred.jpg"
btnwm1.background_normal="img/pred.jpg"
btnWeekMonth.background_normal="img/pred.jpg"
btnmm1.background_normal="img/pred.jpg"
btnmm2.background_normal="img/pred.jpg"
btnmy1.background_normal="img/pred.jpg"
btnmy2.background_normal="img/pred.jpg"
btnMon1.background_normal="img/pred.jpg"
btnMon2.background_normal="img/pred.jpg"
btnann1.background_normal="img/pred.jpg"
btnAnnualYear.background_normal="img/pred.jpg"
b_scmainmenu.background_normal="img/pred.jpg"
logsTitle.background_normal="img/pred.jpg"
b_sc5menu.background_normal="img/pred.jpg"
recentTitle.background_normal="img/pred.jpg"
b_sc3menu.background_normal="img/pred.jpg"
buttonPopSave.background_normal="img/pred.jpg"
b_sc3.background_normal="img/pred.jpg"
titlewrite.background_normal="img/pred.jpg"
btnSlider.background_normal="img/pred.jpg"
buttonPopCp.background_normal="img/pred.jpg"
b_home.text="[color=ffb4bb]D[/color]"
lblHome.text="[color=ffb4bb][b]Home[/b][/color]"
def colorPBlue(self,*args):
color_val = 'pblue'
cursor.execute('SELECT COUNT(*) FROM color_tbl')
color_count = cursor.fetchone()[0]
if color_count == 0:
cursor.execute("INSERT INTO color_tbl(color) VALUES('"+color_val+"')")
con.commit()
elif color_count >= 1:
cursor.execute("UPDATE color_tbl SET color='"+color_val+"'")
con.commit()
btnhelpmeok.background_normal="img/pblue.jpg"
btnHelpGraph.background_normal="img/pblue.jpg"
buttonPop.background_down="img/pblue.jpg"
titlewrite.background_down="img/pblue.jpg"
logsTitle.background_down="img/pblue.jpg"
recentTitle.background_down="img/pblue.jpg"
statsTitle.background_down="img/pblue.jpg"
root6Title.background_down="img/pblue.jpg"
root6Title.background_normal="img/pblue.jpg"
buttonPopExit1.background_normal="img/pblue.jpg"
buttonPopExit2.background_normal="img/pblue.jpg"
buttonPop.background_normal="img/pblue.jpg"
buttonPopAb.background_normal="img/pblue.jpg"
buttonPopD.background_normal="img/pblue.jpg"
buttonPopName.background_normal="img/pblue.jpg"
buttonPopCpExit.background_normal="img/pblue.jpg"
btnCP.background_normal="img/pblue.jpg"
btnCons.background_normal="img/pblue.jpg"
btnHelp.background_normal="img/pblue.jpg"
btnColor.background_normal="img/pblue.jpg"
btnD.background_normal="img/pblue.jpg"
btnA.background_normal="img/pblue.jpg"
btnLO.background_normal="img/pblue.jpg"
mainbutton.background_normal="img/pblue.jpg"
buttonPopName.background_normal="img/pblue.jpg"
btnDaily.background_normal="img/pblue.jpg"
btnWeekly.background_normal="img/pblue.jpg"
btnMonthly.background_normal="img/pblue.jpg"
btnAnnual.background_normal="img/pblue.jpg"
graphButton.background_normal="img/pblue.jpg"
b_sc4menu.background_normal="img/pblue.jpg"
statsTitle.background_normal="img/pblue.jpg"
btnd1.background_normal="img/pblue.jpg"
btnd2.background_normal="img/pblue.jpg"
btnDayDay.background_normal="img/pblue.jpg"
btndy1.background_normal="img/pblue.jpg"
btndy2.background_normal="img/pblue.jpg"
btnYearDay.background_normal="img/pblue.jpg"
btnmy11.background_normal="img/pblue.jpg"
btnmy22.background_normal="img/pblue.jpg"
btnMonthDay.background_normal="img/pblue.jpg"
btnwd1.background_normal="img/pblue.jpg"
btnWeekDay.background_normal="img/pblue.jpg"
btnwy1.background_normal="img/pblue.jpg"
btnWeekYear.background_normal="img/pblue.jpg"
btnwm1.background_normal="img/pblue.jpg"
btnWeekMonth.background_normal="img/pblue.jpg"
btnmm1.background_normal="img/pblue.jpg"
btnmm2.background_normal="img/pblue.jpg"
btnmy1.background_normal="img/pblue.jpg"
btnmy2.background_normal="img/pblue.jpg"
btnMon1.background_normal="img/pblue.jpg"
btnMon2.background_normal="img/pblue.jpg"
btnann1.background_normal="img/pblue.jpg"
btnAnnualYear.background_normal="img/pblue.jpg"
b_scmainmenu.background_normal="img/pblue.jpg"
logsTitle.background_normal="img/pblue.jpg"
b_sc5menu.background_normal="img/pblue.jpg"
recentTitle.background_normal="img/pblue.jpg"
b_sc3menu.background_normal="img/pblue.jpg"
buttonPopSave.background_normal="img/pblue.jpg"
b_sc3.background_normal="img/pblue.jpg"
titlewrite.background_normal="img/pblue.jpg"
btnSlider.background_normal="img/pblue.jpg"
buttonPopCp.background_normal="img/pblue.jpg"
b_home.text="[color=88cafd]D[/color]"
lblHome.text="[color=88cafd][b]Home[/b][/color]"
def colorPGreen(self,*args):
color_val = 'pgreen'
cursor.execute('SELECT COUNT(*) FROM color_tbl')
color_count = cursor.fetchone()[0]
if color_count == 0:
cursor.execute("INSERT INTO color_tbl(color) VALUES('"+color_val+"')")
con.commit()
elif color_count >= 1:
cursor.execute("UPDATE color_tbl SET color='"+color_val+"'")
con.commit()
btnhelpmeok.background_normal="img/pgreen.jpg"
btnHelpGraph.background_normal="img/pgreen.jpg"
buttonPop.background_down="img/pgreen.jpg"
titlewrite.background_down="img/pgreen.jpg"
logsTitle.background_down="img/pgreen.jpg"
recentTitle.background_down="img/pgreen.jpg"
statsTitle.background_down="img/pgreen.jpg"
root6Title.background_down="img/pgreen.jpg"
root6Title.background_normal="img/pgreen.jpg"
buttonPopExit1.background_normal="img/pgreen.jpg"
buttonPopExit2.background_normal="img/pgreen.jpg"
buttonPop.background_normal="img/pgreen.jpg"
buttonPopAb.background_normal="img/pgreen.jpg"
buttonPopD.background_normal="img/pgreen.jpg"
buttonPopName.background_normal="img/pgreen.jpg"
buttonPopCpExit.background_normal="img/pgreen.jpg"
btnCP.background_normal="img/pgreen.jpg"
btnCons.background_normal="img/pgreen.jpg"
btnHelp.background_normal="img/pgreen.jpg"
btnColor.background_normal="img/pgreen.jpg"
btnD.background_normal="img/pgreen.jpg"
btnA.background_normal="img/pgreen.jpg"
btnLO.background_normal="img/pgreen.jpg"
mainbutton.background_normal="img/pgreen.jpg"
buttonPopName.background_normal="img/pgreen.jpg"
btnDaily.background_normal="img/pgreen.jpg"
btnWeekly.background_normal="img/pgreen.jpg"
btnMonthly.background_normal="img/pgreen.jpg"
btnAnnual.background_normal="img/pgreen.jpg"
graphButton.background_normal="img/pgreen.jpg"
b_sc4menu.background_normal="img/pgreen.jpg"
statsTitle.background_normal="img/pgreen.jpg"
btnd1.background_normal="img/pgreen.jpg"
btnd2.background_normal="img/pgreen.jpg"
btnDayDay.background_normal="img/pgreen.jpg"
btndy1.background_normal="img/pgreen.jpg"
btndy2.background_normal="img/pgreen.jpg"
btnYearDay.background_normal="img/pgreen.jpg"
btnmy11.background_normal="img/pgreen.jpg"
btnmy22.background_normal="img/pgreen.jpg"
btnMonthDay.background_normal="img/pgreen.jpg"
btnwd1.background_normal="img/pgreen.jpg"
btnWeekDay.background_normal="img/pgreen.jpg"
btnwy1.background_normal="img/pgreen.jpg"
btnWeekYear.background_normal="img/pgreen.jpg"
btnwm1.background_normal="img/pgreen.jpg"
btnWeekMonth.background_normal="img/pgreen.jpg"
btnmm1.background_normal="img/pgreen.jpg"
btnmm2.background_normal="img/pgreen.jpg"
btnmy1.background_normal="img/pgreen.jpg"
btnmy2.background_normal="img/pgreen.jpg"
btnMon1.background_normal="img/pgreen.jpg"
btnMon2.background_normal="img/pgreen.jpg"
btnann1.background_normal="img/pgreen.jpg"
btnAnnualYear.background_normal="img/pgreen.jpg"
b_scmainmenu.background_normal="img/pgreen.jpg"
logsTitle.background_normal="img/pgreen.jpg"
b_sc5menu.background_normal="img/pgreen.jpg"
recentTitle.background_normal="img/pgreen.jpg"
b_sc3menu.background_normal="img/pgreen.jpg"
buttonPopSave.background_normal="img/pgreen.jpg"
b_sc3.background_normal="img/pgreen.jpg"
titlewrite.background_normal="img/pgreen.jpg"
btnSlider.background_normal="img/pgreen.jpg"
buttonPopCp.background_normal="img/pgreen.jpg"
b_home.text="[color=7ffe9a]D[/color]"
lblHome.text="[color=7ffe9a][b]Home[/b][/color]"
def colorPOrange(self,*args):
color_val = 'porange'
cursor.execute('SELECT COUNT(*) FROM color_tbl')
color_count = cursor.fetchone()[0]
if color_count == 0:
cursor.execute("INSERT INTO color_tbl(color) VALUES('"+color_val+"')")
con.commit()
elif color_count >= 1:
cursor.execute("UPDATE color_tbl SET color='"+color_val+"'")
con.commit()
btnhelpmeok.background_normal="img/porange.jpg"
btnHelpGraph.background_normal="img/porange.jpg"
buttonPop.background_down="img/porange.jpg"
titlewrite.background_down="img/porange.jpg"
logsTitle.background_down="img/porange.jpg"
recentTitle.background_down="img/porange.jpg"
statsTitle.background_down="img/porange.jpg"
root6Title.background_down="img/porange.jpg"
root6Title.background_normal="img/porange.jpg"
buttonPopExit1.background_normal="img/porange.jpg"
buttonPopExit2.background_normal="img/porange.jpg"
buttonPop.background_normal="img/porange.jpg"
buttonPopAb.background_normal="img/porange.jpg"
buttonPopD.background_normal="img/porange.jpg"
buttonPopName.background_normal="img/porange.jpg"
buttonPopCpExit.background_normal="img/porange.jpg"
btnCP.background_normal="img/porange.jpg"
btnCons.background_normal="img/porange.jpg"
btnHelp.background_normal="img/porange.jpg"
btnColor.background_normal="img/porange.jpg"
btnD.background_normal="img/porange.jpg"
btnA.background_normal="img/porange.jpg"
btnLO.background_normal="img/porange.jpg"
mainbutton.background_normal="img/porange.jpg"
buttonPopName.background_normal="img/porange.jpg"
btnDaily.background_normal="img/porange.jpg"
btnWeekly.background_normal="img/porange.jpg"
btnMonthly.background_normal="img/porange.jpg"
btnAnnual.background_normal="img/porange.jpg"
graphButton.background_normal="img/porange.jpg"
b_sc4menu.background_normal="img/porange.jpg"
statsTitle.background_normal="img/porange.jpg"
btnd1.background_normal="img/porange.jpg"
btnd2.background_normal="img/porange.jpg"
btnDayDay.background_normal="img/porange.jpg"
btndy1.background_normal="img/porange.jpg"
btndy2.background_normal="img/porange.jpg"
btnYearDay.background_normal="img/porange.jpg"
btnmy11.background_normal="img/porange.jpg"
btnmy22.background_normal="img/porange.jpg"
btnMonthDay.background_normal="img/porange.jpg"
btnwd1.background_normal="img/porange.jpg"
btnWeekDay.background_normal="img/porange.jpg"
btnwy1.background_normal="img/porange.jpg"
btnWeekYear.background_normal="img/porange.jpg"
btnwm1.background_normal="img/porange.jpg"
btnWeekMonth.background_normal="img/porange.jpg"
btnmm1.background_normal="img/porange.jpg"
btnmm2.background_normal="img/porange.jpg"
btnmy1.background_normal="img/porange.jpg"
btnmy2.background_normal="img/porange.jpg"
btnMon1.background_normal="img/porange.jpg"
btnann1.background_normal="img/porange.jpg"
btnAnnualYear.background_normal="img/porange.jpg"
b_scmainmenu.background_normal="img/porange.jpg"
logsTitle.background_normal="img/porange.jpg"
b_sc5menu.background_normal="img/porange.jpg"
recentTitle.background_normal="img/porange.jpg"
b_sc3menu.background_normal="img/porange.jpg"
buttonPopSave.background_normal="img/porange.jpg"
b_sc3.background_normal="img/porange.jpg"
titlewrite.background_normal="img/porange.jpg"
btnSlider.background_normal="img/porange.jpg"
buttonPopCp.background_normal="img/porange.jpg"
b_home.text="[color=fec47e]D[/color]"
lblHome.text="[color=fec47e][b]Home[/b][/color]"
def colorPPurple(self,*args):
color_val = 'ppurple'
cursor.execute('SELECT COUNT(*) FROM color_tbl')
color_count = cursor.fetchone()[0]
if color_count == 0:
cursor.execute("INSERT INTO color_tbl(color) VALUES('"+color_val+"')")
con.commit()
elif color_count >= 1:
cursor.execute("UPDATE color_tbl SET color='"+color_val+"'")
con.commit()
btnhelpmeok.background_normal="img/ppurple.jpg"
btnHelpGraph.background_normal="img/ppurple.jpg"
buttonPop.background_down="img/ppurple.jpg"
titlewrite.background_down="img/ppurple.jpg"
logsTitle.background_down="img/ppurple.jpg"
recentTitle.background_down="img/ppurple.jpg"
statsTitle.background_down="img/ppurple.jpg"
root6Title.background_down="img/ppurple.jpg"
root6Title.background_normal="img/ppurple.jpg"
buttonPopExit1.background_normal="img/ppurple.jpg"
buttonPopExit2.background_normal="img/ppurple.jpg"
buttonPop.background_normal="img/ppurple.jpg"
buttonPopAb.background_normal="img/ppurple.jpg"
buttonPopD.background_normal="img/ppurple.jpg"
buttonPopName.background_normal="img/ppurple.jpg"
buttonPopCpExit.background_normal="img/ppurple.jpg"
btnCP.background_normal="img/ppurple.jpg"
btnCons.background_normal="img/ppurple.jpg"
btnHelp.background_normal="img/ppurple.jpg"
btnColor.background_normal="img/ppurple.jpg"
btnD.background_normal="img/ppurple.jpg"
btnA.background_normal="img/ppurple.jpg"
btnLO.background_normal="img/ppurple.jpg"
mainbutton.background_normal="img/ppurple.jpg"
buttonPopName.background_normal="img/ppurple.jpg"
btnDaily.background_normal="img/ppurple.jpg"
btnWeekly.background_normal="img/ppurple.jpg"
btnMonthly.background_normal="img/ppurple.jpg"
btnAnnual.background_normal="img/ppurple.jpg"
graphButton.background_normal="img/ppurple.jpg"
b_sc4menu.background_normal="img/ppurple.jpg"
statsTitle.background_normal="img/ppurple.jpg"
btnd1.background_normal="img/ppurple.jpg"
btnd2.background_normal="img/ppurple.jpg"
btnDayDay.background_normal="img/ppurple.jpg"
btndy1.background_normal="img/ppurple.jpg"
btndy2.background_normal="img/ppurple.jpg"
btnYearDay.background_normal="img/ppurple.jpg"
btnmy11.background_normal="img/ppurple.jpg"
btnmy22.background_normal="img/ppurple.jpg"
btnMonthDay.background_normal="img/ppurple.jpg"
btnwd1.background_normal="img/ppurple.jpg"
btnWeekDay.background_normal="img/ppurple.jpg"
btnwy1.background_normal="img/ppurple.jpg"
btnWeekYear.background_normal="img/ppurple.jpg"
btnwm1.background_normal="img/ppurple.jpg"
btnWeekMonth.background_normal="img/ppurple.jpg"
btnmm1.background_normal="img/ppurple.jpg"
btnmm2.background_normal="img/ppurple.jpg"
btnmy1.background_normal="img/ppurple.jpg"
btnmy2.background_normal="img/ppurple.jpg"
btnMon1.background_normal="img/ppurple.jpg"
btnann1.background_normal="img/ppurple.jpg"
btnAnnualYear.background_normal="img/ppurple.jpg"
b_scmainmenu.background_normal="img/ppurple.jpg"
logsTitle.background_normal="img/ppurple.jpg"
b_sc5menu.background_normal="img/ppurple.jpg"
recentTitle.background_normal="img/ppurple.jpg"
b_sc3menu.background_normal="img/ppurple.jpg"
buttonPopSave.background_normal="img/ppurple.jpg"
b_sc3.background_normal="img/ppurple.jpg"
titlewrite.background_normal="img/ppurple.jpg"
btnSlider.background_normal="img/ppurple.jpg"
buttonPopCp.background_normal="img/ppurple.jpg"
b_home.text="[color=ba55d3]D[/color]"
lblHome.text="[color=ba55d3][b]Home[/b][/color]"
def colorClose(self,*args):
popupColor.dismiss()
popColorRemove()
if color_val =='red':
b_home.text="[color=ff0000]D[/color]"
lblHome.text="[color=ff0000][b]Home[/b][/color]"
elif color_val == 'blue':
b_home.text="[color=3b5999]D[/color]"
lblHome.text="[color=3b5999][b]Home[/b][/color]"
elif color_val == 'green':
b_home.text="[color=00ff00]D[/color]"
lblHome.text="[color=00ff00][b]Home[/b][/color]"
elif color_val == 'orange':
b_home.text="[color=ff4500]D[/color]"
lblHome.text="[color=ff4500][b]Home[/b][/color]"
elif color_val == 'purple':
b_home.text="[color=ba55d3]D[/color]"
lblHome.text="[color=ba55d3][b]Home[/b][/color]"
elif color_val == 'pred':
b_home.text="[color=ffb4bb]D[/color]"
lblHome.text="[color=ffb4bb][b]Home[/b][/color]"
elif color_val == 'pblue':
b_home.text="[color=88cafd]D[/color]"
lblHome.text="[color=88cafd][b]Home[/b][/color]"
elif color_val == 'pgreen':
b_home.text="[color=7ffe9a]D[/color]"
lblHome.text="[color=7ffe9a][b]Home[/b][/color]"
elif color_val == 'porange':
b_home.text="[color=fec47e]D[/color]"
lblHome.text="[color=fec47e][b]Home[/b][/color]"
elif color_val == 'ppurple':
b_home.text="[color=ba55d3]D[/color]"
lblHome.text="[color=ba55d3][b]Home[/b][/color]"
#--BINDINGS---------------------
root.bind(on_touch_down=callbacksc2) # button tap event
passimgok.bind(on_release=callbackmain)
passimgok.bind(on_release=popError)
passimg0.bind(on_press=p0)
passimg1.bind(on_press=p1)
passimg2.bind(on_press=p2)
passimg3.bind(on_press=p3)
passimg4.bind(on_press=p4)
passimg5.bind(on_press=p5)
passimg6.bind(on_press=p6)
passimg7.bind(on_press=p7)
passimg8.bind(on_press=p8)
passimg9.bind(on_press=p9)
passimgclr.bind(on_press=pclr)
buttonPop.bind(on_press=popup.dismiss)
b_write.bind(on_press=callbacksc3)
b_sc3menu.bind(on_release=callbackhome)
b_stats.bind(on_release=callbacksc4)
b_user.bind(on_press=callbacksc7)
b_logs.bind(on_release=callbackscmain)
b_scmainmenu.bind(on_release=callbackhome)
b_sc4menu.bind(on_release=callbackhome)
b_sc7menu.bind(on_release=callbackhome)
b_sc3.bind(on_press=callback_proctext)
#slider
btnSlider.bind(on_press=moodconfirm)
#logout
btnLO.bind(on_press=popexit)
buttonPopExit1.bind(on_press=exitclose)
buttonPopExit2.bind(on_press=popexitdismiss)
#about
buttonPopAb.bind(on_press=popaboutdismiss)
btnA.bind(on_press=popabout)
#error
buttonPopError.bind(on_press=poperrordismiss)
#devs
buttonPopD.bind(on_press=popdevdismiss)
btnD.bind(on_press=popdev)
#change pass
buttonPopCpExit.bind(on_press=popcpdismiss)
buttonPopCp.bind(on_press=popcpsave)
btnCP.bind(on_press=popcp)
btnColor.bind(on_press=popColor)
#dismissnote
buttonPopSave.bind(on_press=callbackhome)
#nameinput
buttonPopName.bind(on_press=popnamedismiss)
#save Button Pop Pin
buttonPopPin.bind(on_press=savepoppin)
#remove widgets
b_stats.bind(on_release=rWidgets)
#graph plot
btnWeekly.bind(on_release=weeklyGraphPlot)
btnDaily.bind(on_release=dailyGraphPlot)
btnMonthly.bind(on_release=monthlyGraphPlot)
btnAnnual.bind(on_release=annualGraphPlot)
b_stats.bind(on_release=startGraphPlot)
graphButton.bind(on_press=removeDWM1)
#Graph Buttons List
#Daily
btnDaily.bind(on_release=daily_month)
btnDaily.bind(on_release=daily_year)
btnDaily.bind(on_release=daily_day)
#weekly
btnWeekly.bind(on_release=weekly_month)
btnWeekly.bind(on_release=weekly_year)
btnWeekly.bind(on_release=weekly_day)
#monthly
btnMonthly.bind(on_release=monthly_month)
btnMonthly.bind(on_release=monthly_year)
#monthly
btnAnnualYear.bind(on_release=yearly_year)
#color
buttonColorRed.bind(on_press=colorRed)
buttonColorBlue.bind(on_press=colorBlue)
buttonColorGreen.bind(on_press=colorGreen)
buttonColorOrange.bind(on_press=colorOrange)
buttonColorPurple.bind(on_press=colorPurple)
buttonColorPRed.bind(on_press=colorPRed)
buttonColorPBlue.bind(on_press=colorPBlue)
buttonColorPGreen.bind(on_press=colorPGreen)
buttonColorPOrange.bind(on_press=colorPOrange)
buttonColorPPurple.bind(on_press=colorPPurple)
buttonPopColor.bind(on_press=colorClose)
#gender
buttonPopGender1.bind(on_press=popGenderRemove1)
buttonPopGender2.bind(on_press=popGenderRemove2)
# disable Back button on Android
Window.bind(on_keyboard=self.hook_keyboard)
#consultant Mode
btnCons.bind(on_release=ConsultantMode)
#help
btnHelp.bind(on_release=MMHelp)
return sm
if __name__ == "__main__": #run the class
MainApp().run()
con.commit()
| [
"noreply@github.com"
] | noreply@github.com |
417a08e70bb7a9d9dd71117edfc2d81b6f20d424 | 0686243fad8eb4e17959d8819c62d183d71adf57 | /tasks/views.py | e18347d900ef9a8032594d1c89c71cbcefe9fb5c | [] | no_license | package-coder/REPO-CS50Django | c241034d1e164d8d79a2ebf5601dce0517b7db66 | 0b1acb77800cbc55f8a7ef608a498aa813253851 | refs/heads/main | 2023-07-25T08:18:35.750988 | 2021-08-29T11:10:28 | 2021-08-29T11:10:28 | 401,021,110 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 1,000 | py | from django.shortcuts import render
from django import forms
from django.http import HttpResponseRedirect
from django.urls import reverse
tasks = []
class TaskDAOForm(forms.Form):
task = forms.CharField(label="Task")
priority = forms.IntegerField(label="Priority", min_value=1, max_value=10)
# Create your views here.
def index(request):
if "tasks" not in request.session:
request.session["tasks"] = []
return render(request, "tasks/index.html", {
"tasks": request.session["tasks"]
})
def new(request):
if request.method == 'POST':
form = TaskDAOForm(request.POST)
if form.is_valid():
task = form.cleaned_data["task"]
request.session["tasks"] += [task]
return HttpResponseRedirect(reverse("tasks:index"))
else:
return render(request, "tasks/new.html", {
"form": form
})
return render(request, "tasks/new.html",{
"forms": TaskDAOForm()
}) | [
"79316343+package-coder@users.noreply.github.com"
] | 79316343+package-coder@users.noreply.github.com |
f7c258330381012d574538f806f44342ee8b2f48 | 3d6f870af1af2bb608bfd404c80c3f8ed58d31b2 | /main.py | d7d437390c61e2d1294ee1a6c701de9cf33b46cb | [] | no_license | saness/tictactoe | 1cbd0b9fa8230f259c587b9194b7eefbdc029ab7 | 4f3ac33d8f0743d0887aa4f602dea311a69cc780 | refs/heads/master | 2022-04-25T16:50:30.106869 | 2020-04-10T06:53:32 | 2020-04-10T06:53:32 | 254,531,521 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 3,178 | py | board = [' ' for x in range(10)]
def insertLetter(letter, pos):
board[pos] = letter
def spaceIsFree(pos):
return board[pos] == ' '
def printBoard(board):
print(' | | ')
print(' ' + board[1] + '| ' + board[2] + '| ' + board[3])
print(' | | ')
print('----------- ')
print(' | | ')
print(' ' + board[4] + '| ' + board[5] + '| ' + board[6])
print(' | | ')
print('----------- ')
print(' | | ')
print(' ' + board[7] + '| ' + board[8] + '| ' + board[9])
print(' | | ')
def isWinner(bo, le):
return (bo[7] == le and bo[8] == le and bo[9] == le) or (bo[4] == le and bo[5] == le and bo[6] == le) or (bo[1] == le and bo[4] == le and bo[7] == le) or (bo[2] == le and bo[5] == le and bo[8] == le) or (bo[3] == le and bo[6] == le and bo[9] == le) or (bo[1] == le and bo[5] == le and bo[9] == le) or (bo[3] == le and bo[5] == le and bo[7] == le)
def playerMove():
run = True
while run:
move = input('Please select a position to place an \'X\' (1-9):')
try:
move = int(move)
if move > 0 and move < 10:
if spaceIsFree(move):
run = False
insertLetter('X', move)
else:
print("Sorry, this space is occupied")
else:
print('Please type a number within the range')
except:
print("Please type a number")
def selectRandom(li):
import random
ln = len(li)
r = random.randrange(0, ln)
return li[r]
def compMove():
possibleMoves = [x for x, letter in enumerate(board) if letter == ' ' and x != 0]
move = 0
for let in ['O', 'X']:
for i in possibleMoves:
boardCopy = board[:]
boardCopy[i] = let
if isWinner(boardCopy, let):
move = i
return move
cornersOpen = []
for i in possibleMoves:
if i in [1, 3, 7, 9]:
cornersOpen.append(i)
if len(cornersOpen) > 0:
move = selectRandom(cornersOpen)
return move
if 5 in possibleMoves:
move = 5
return move
edgesOpen = []
for i in possibleMoves:
if i in [2, 4, 6, 8]:
edgesOpen.append(i)
if len(edgesOpen) > 0:
move = selectRandom(edgesOpen)
return move
def isBoardFull(board):
if board.count(' ') > 1:
return False
else:
return True
def main():
print("Welcome to tic tac toe")
printBoard(board)
while not(isBoardFull(board)):
if not(isWinner(board, "O")):
playerMove()
printBoard(board)
else:
print('Sorry O\'s won thi time')
break
if not(isWinner(board, 'X')):
move = compMove()
if move == 0:
print('Tie Game')
else:
insertLetter('O', move)
print('Computer placed an \'O\' in position', move, ':')
printBoard(board)
else:
print('X\'S won this time')
break
if isBoardFull(board):
print("Tie Game")
main()
| [
"sanishsuwal3@gmail.com"
] | sanishsuwal3@gmail.com |
1c0a243fae087ba9520b940a1940a5458e5d1a61 | 15f321878face2af9317363c5f6de1e5ddd9b749 | /solutions_python/Problem_95/2103.py | a414f3ccb3b1e3142cf256e7d173e37982b1e31b | [] | no_license | dr-dos-ok/Code_Jam_Webscraper | c06fd59870842664cd79c41eb460a09553e1c80a | 26a35bf114a3aa30fc4c677ef069d95f41665cc0 | refs/heads/master | 2020-04-06T08:17:40.938460 | 2018-10-14T10:12:47 | 2018-10-14T10:12:47 | null | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 1,026 | py | #! /usr/bin/env python
import operator
from sys import stdin
in1 = "ejp mysljylc kd kxveddknmc re jsicpdrysirbcpc ypc rtcsra dkh wyfrepkym veddknkmkrkcdde kr kd eoya kw aej tysr re ujdr lkgc jvzq"
out1 = "our language is impossible to understandthere are twenty six factorial possibilitiesso it is okay if you want to just give upqz"
sample = "ejp mysljylc kd kxveddknmc re jsicpdrysi"
def getInput():
raw = stdin.readlines()
for x in range(0, len(raw)):
raw[x] = raw[x].replace('\n', '')
return raw
def makeMap(input_str, output_str):
mymap = {}
for x,y in zip(input_str, output_str):
if(x != " "):
mymap[x] = y
return mymap
def googler2english(input_str):
mymap = makeMap(in1, out1)
ret_str = ""
for x in input_str:
if x != ' ':
ret_str = ret_str + mymap[x]
else:
ret_str = ret_str + " "
return ret_str
def main():
myinput = getInput()
bound = int(myinput[0])
for x in range(1, bound + 1):
print "Case #%d: %s" % (x, googler2english(myinput[x]))
if __name__ == "__main__":
main()
| [
"miliar1732@gmail.com"
] | miliar1732@gmail.com |
7ce855509a02fabb32794d93a57361f957f8535c | 67bde32c6aeb71880a10cc67d5c2266fceb9deb8 | /Python3-Patterns_Recipes_Idioms/saved_script/Ch 28 Fronting for an implementation.py | fe941f29091702deca207bc093ec11c1a0da3102 | [] | no_license | Sprinterzzj/Effective-Python | 981ea71eb028d5c782ab85de2e66c30f724a4ddc | 0af0e1d20741c6bce0951ea17c23534c4ccb2d9a | refs/heads/master | 2021-07-01T16:22:46.682452 | 2020-10-08T10:11:39 | 2020-10-08T10:11:39 | 180,560,855 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 1,435 | py | #!/usr/bin/env python
# coding: utf-8
# 代理
#
# 代理模式 用来**控制 implementation 的访问**.
# In[ ]:
# Simple demonstration of the Proxy pattern
class Implementation(object):
def f(self):
print('***f***')
return
def g(self):
print('***g***')
return
def h(self):
print('***h***')
return
class Proxy(object):
def __init__(self):
self.__implementation = Implementation()
# Pass method calls to the implementation
def f(self):
self.__implementation.f()
def g(self):
self.__implementation.g()
def h(self):
self.__implementation.h()
# In[3]:
# 当然我们可以用 Python 的特性
class Implementation(object):
def f(self):
print('***f***')
return
class Proxy(object):
def __init__(self):
self.__implementation = Implementation()
def __getattr__(self, name):
"""重定向 __getattr__ 方法
"""
return getattr(self.__implementation, name)
# State
#
# State pattern adds more implementations to Proxy. State 模式可以切换 Implementation 的种类
# In[4]:
class State_d(object):
def __init__(self, imp):
self.__implementation = imp
def changeImp(self, newImp):
self.__implementation = newImp
def __getattr__(self, name):
return getattr(self.__implementation, name)
# In[ ]:
| [
"zhangzijie.dlut@gmail.com"
] | zhangzijie.dlut@gmail.com |
626ad41b7c2f45201c36c890134ff3c5e306ba4b | 2e7656524d3bc32a5f256973afcb3dd35fc6b764 | /matrice.py | e8f55835f226cee96bdf2a3211155629cc12b883 | [] | no_license | bondac-ion/Matrice | fa047adf91e23c3691b98f05bda79e8971f7d482 | 891e0325a8c8e605ae18129272e3cc80aabee421 | refs/heads/main | 2023-08-15T04:00:37.106089 | 2021-10-06T19:03:29 | 2021-10-06T19:03:29 | 414,337,296 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 589 | py | a=[[2,11,54,3,1],
[34,6,55,9,22],
[1,2,3,4,5],
[9,7,6,45,0],
[12,23,34,45,56]]
for i in a:
print("suma randului",i+1,"este",sum(i))
for i in range(len(a)):
col=[]
for rind in a:
col.append(rind[i])
print("suma coloanei",i+1,"este",sum(col))
d_principala=[]
d_secundara=[]
for i in range(len(a)):
for j in range(len(a[0])):
if i==j:
d_principala.append(a[i][j])
if i+j==(len(a)-1):
d_secundara.append(a[j][i])
print("Diagonala principala-",d_principala)
print("Diagonala secundara-",d_secundara) | [
"noreply@github.com"
] | noreply@github.com |
1e28c69475afb7c2f2e67c3a0a62edb27df50580 | 179f33dc6856f32db8aceffdd1cf2278000fb321 | /blog/models.py | 48d317ba3d060c202d24ea634059d8fb6d1d60b7 | [] | no_license | alejandrapiedrasanta/progracomercial | 8c9826012f9a48a132276f275152a7fdbf065163 | 68c2c68379e178e54fa827d8c84d2f695b6a01aa | refs/heads/master | 2023-08-14T09:19:14.836187 | 2021-10-14T09:32:20 | 2021-10-14T09:32:20 | 411,798,479 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 677 | py | from django.db import models
# Createfrom django.db import models
from django.utils import timezone
class Publicacion(models.Model):
autor = models.ForeignKey('auth.User', on_delete=models.CASCADE)
titulo = models.CharField(max_length=200,help_text="Ingrese el titulo")
texto = models.TextField()
fecha_creacion = models.DateTimeField(
default=timezone.now)
fecha_publicacion = models.DateTimeField(
blank=True, null=True)
def publicar(self):
self.fecha_publicacion = timezone.now()
self.save()
def __str__(self):
return self.titulo
class Meta:
verbose_name_plural = 'Publicaciones' | [
"alesarti@mesoamericana.edu.gt"
] | alesarti@mesoamericana.edu.gt |
b531a49ad9aebf02cb211cb8fd9267c0546770c9 | 8cfeb8b488987da3d955cdbded7e00fb4a63345e | /ula/views.py | 0225cc55dfd7da7a66938b4ac1490eafa887816b | [] | no_license | Vedaad-Shakib/oola | 77be7798b258898dde72b8efbe4dc4d35203af3e | 003b865a39e71d1a76c592cfe47bcc4fccc3305f | refs/heads/master | 2021-03-12T19:38:12.213770 | 2014-09-17T04:16:48 | 2014-09-17T04:16:48 | null | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 48,730 | py | ###############################################################################
## Copyright (c) 2013-2014 Bogt, Inc.
## All rights reserved.
## This source code is confidential and may not be disclosed.
###############################################################################
###############################################################################
##
## "views.py": create the web pages
##
###############################################################################
from django.shortcuts import render_to_response
from django.conf import settings
from django.template import RequestContext
from django.http import HttpResponseRedirect
from django.http import HttpResponse
from django.db.models import Q
import cookies
import math
import csv
import operator
from datetime import datetime, timedelta
import json
from forms import *
from cookies import *
from email_util import *
import paginator
#==============================================================================
#
# "mainPage": display the home page
#
#==============================================================================
def mainPage(request, checkSignup=None, checkSignin=None):
# if checking for signup
if checkSignup is not None:
form = SignupForm(request.GET.copy())
if form.is_valid():
user = form.save()
email = form.data['email']
user = User.objects.get(email=email)
SetUserCookies(request, user)
url = '/myprofile/'
if user.userType:
url = '/students/'
json = util.JsonLoad( url )
else:
json = util.JsonFormError(form)
return HttpResponse(json,
mimetype='application/json')
# if checking for signin
if checkSignin is not None:
form = SigninForm(request.GET.copy())
if form.is_valid():
user = User.objects.get(email=form.data['email'])
SetUserCookies(request, user)
url = '/myprofile/'
if user.userType:
url = '/students/'
json = util.JsonLoad( url )
else:
json = util.JsonFormError(form)
return HttpResponse(json,
mimetype='application/json')
# normal call
try:
userName = request.session['fullName']
userId = request.session['userId']
user = User.objects.get(userId=userId)
userType = user.userType
return render_to_response( 'home.html',
{'userName': userName,
'userType': userType,
},
context_instance=RequestContext(request))
except:
forup = SignupForm()
forin = SigninForm()
return render_to_response( 'home.html',
{'forup': forup,
'forin': forin,
'userName': None,
'userType': None,
# 'actionUrl':"/mainPage/"
},
context_instance=RequestContext(request))
#==============================================================================
#
# "signout":
#
#==============================================================================
def signout(request, check=None):
DelUserCookies(request)
return HttpResponseRedirect('/home/')
#==============================================================================
#
# "signin":
#
#==============================================================================
def signin(request, check=None):
if check is not None:
data = request.GET.copy( )
form = SigninForm( data )
if form.is_valid():
user = User.objects.get(email=form.data['email'])
SetUserCookies(request, user)
url = '/myprofile/'
if user.userType:
url = '/students/'
json = util.JsonLoad( url )
else:
json = util.JsonFormError(form)
return HttpResponse(json,
mimetype='application/json')
form = SigninForm( )
return render_to_response( 'signin.html',
{'form': form},
context_instance=RequestContext(request))
#==============================================================================
#
# "classPage": The main class page in which dancers can check in
#
#==============================================================================
def classPage(request):
DelUserCookies(request)
#form = ClassName()
userList = list(User.objects.all())
userNames = []
for i in userList:
userNames.append(str(i.name))
return render_to_response('class.html',
{'userNames': userNames},
context_instance=RequestContext(request))
#==============================================================================
#
# "classCheckRecent": Check if the student has recently checked in
#
#==============================================================================
def classCheckRecent(request):
jdata = {}
jdata["recent"] = False
try:
user = User.objects.get(name = request.GET["name"] )
currTime= datetime.datetime.now()
lst10Min= currTime - datetime.timedelta(minutes = 10)
lstAttendance = Attendance.objects.get(userId = user.userId,
dateTime__gt = lst10Min
)
if lstAttendance:
jdata["recent"] = True
jdata["name"] = user.name
jdata["userId"] = user.userId
jdata["attendanceId"] = lstAttendance.attendanceId
except:
pass
jsonData = json.dumps( jdata )
return HttpResponse( jsonData, mimetype = 'application/json')
#==============================================================================
#
# "classCancel": The cancellation page
#
# Reason for this page: more secure than doing this on the main class page
#
#==============================================================================
def classCancel(request, totalGuests, userId, attendanceId):
Attendance.objects.filter(attendanceId = attendanceId).delete()
user = User.objects.get(userId=userId)
user.balance = user.balance + 1 + int(str(totalGuests))
user.lastAccess = datetime.datetime.now( )
user.save()
#----------------------------------------------------------------------
# Set the recent change
#----------------------------------------------------------------------
rcntChg = RecentChange( )
rcntChg.userId = user
rcntChg.change = "Balance"
rcntChg.value = str( user.balance )
rcntChg.dateTime = datetime.datetime.now( )
rcntChg.save( )
return HttpResponseRedirect('/class/')
#==============================================================================
#
# "classSave": The save page
#
# Reason for this page: more secure than doing this on the main class page
#
#==============================================================================
def classSave(request, guests, userId, oldCount):
guests = int(str(guests))
oldCount = int(str(oldCount))
user = User.objects.get(userId=userId)
# update the counts and balance
newCount = guests + 1
user.balance = user.balance - newCount + oldCount
user.lastAccess = datetime.datetime.now( )
user.save()
# remove old records
attendanceList = list(Attendance.objects.filter(userId = user.userId))
today = datetime.datetime.today()
for i in attendanceList:
if i.dateTime.year == today.year and i.dateTime.month == today.month \
and i.dateTime.day == today.day:
Attendance.objects.filter(attendanceId = i.attendanceId).delete()
# create an attendance record
attendance = Attendance()
attendance.userId = User.objects.get(userId = user.userId)
attendance.dateTime = datetime.datetime.today()
attendance.DancerNumber = newCount
attendance.save()
#----------------------------------------------------------------------
# Set the recent change
#----------------------------------------------------------------------
rcntChg = RecentChange( )
rcntChg.userId = user
rcntChg.change = "Balance"
rcntChg.value = str( user.balance )
rcntChg.dateTime = datetime.datetime.now( )
rcntChg.save( )
rcntChg = RecentChange( )
rcntChg.userId = user
rcntChg.change = "Attendance (Dancer Number)"
rcntChg.value = str( newCount )
rcntChg.dateTime = datetime.datetime.now( )
rcntChg.save( )
return HttpResponseRedirect('/class/')
#==============================================================================
#
# "classCheckin": redirects to if the user is not in the database
#
#==============================================================================
def classCheckin(request, userId = None):
user = None
guests = 0
oldCount = 0
newCount = 0
if userId != None:
# get the parameters; come from signup (new student)
user = User.objects.get(userId=userId)
name = user.name
guests = 0
oldCount = 0
else:
# get the parameters;
try:
name = request.POST['name']
guests = int(str(request.POST['guests']))
oldCount = int(str(request.POST['oldCount']))
except:
return render_to_response('classInvalid.html',
{"redirect": "/class/"},
context_instance=RequestContext(request))
# redirect if the user not found
try:
user = User.objects.get(name=name)
except:
request.session['name'] = name
return HttpResponseRedirect('/class/signup/')
# cancel the previous checkin
try:
lstAttendanceId = request.POST['attendanceId']
lstAttendance = Attendance.objects.get(attendanceId = lstAttendanceId)
savedCount = lstAttendance.DancerNumber
lstAttendance.delete()
user.balance = user.balance + savedCount
user.save()
except:
pass
# update the counts and balance
newCount = guests + 1
user.balance = user.balance - newCount + oldCount
user.lastAccess = datetime.datetime.now( )
user.save()
# remove old records
attendanceList = list(Attendance.objects.filter(userId = user.userId))
today = datetime.datetime.today()
for i in attendanceList:
if i.dateTime.year == today.year and i.dateTime.month == today.month \
and i.dateTime.day == today.day:
Attendance.objects.filter(attendanceId = i.attendanceId).delete()
# create an attendance record
attendance = Attendance()
attendance.userId = User.objects.get(userId = user.userId)
attendance.dateTime = datetime.datetime.today()
attendance.DancerNumber = newCount
attendance.save()
attendanceId = attendance.attendanceId
#----------------------------------------------------------------------
# Set the recent change
#----------------------------------------------------------------------
rcntChg = RecentChange( )
rcntChg.userId = user
rcntChg.change = "Balance"
rcntChg.value = str( user.balance )
rcntChg.dateTime = datetime.datetime.now( )
rcntChg.save( )
rcntChg = RecentChange( )
rcntChg.userId = user
rcntChg.change = "Attendance (Dancer Number)"
rcntChg.value = str( newCount )
rcntChg.dateTime = datetime.datetime.now( )
rcntChg.save( )
# confirm
return render_to_response("classCheckin.html",
{"name": user.name,
"balance": user.balance,
"guests": guests,
"userId": user.userId,
"oldCount": newCount,
"waiver": user.waiverSigned,
"attendanceId": attendanceId},
context_instance=RequestContext(request))
#==============================================================================
#
# "classSignup": signup page if the user is not in the database
#
#==============================================================================
def classSignup(request, check=None):
if check is not None:
form = ClassSignup(request.GET.copy())
if form.is_valid():
user = form.save()
SetUserCookies(request, user)
json = util.JsonLoad('/class/checkin/' + str( user.userId ) + "/")
else:
json = util.JsonFormError(form)
return HttpResponse(json,
mimetype='application/json')
else:
try:
name = request.session['name'].title()
except:
name = ""
form = ClassSignup( initial = {'name': name } )
return render_to_response('classSignup.html',
{'form': form,
'name': name},
context_instance=RequestContext(request))
#==============================================================================
#
# "history": The attendance and transactions page
#
#==============================================================================
def history( request, search = "" ):
user, errUrl = GetValidUser(request)
if errUrl:
return HttpResponseRedirect(errUrl)
return render_to_response( 'history.html',
{'userName': user.name,
'userType': user.userType,
'sort': "date",
'filter': "attendance",
'currPage': 1,
'search': search
},
context_instance=RequestContext(request) )
#==============================================================================
#
# "historyList": The history page
#
#==============================================================================
def historyList( request ):
# Get "get" values
sort = str(request.POST['sort'])
filter = str(request.POST['filter'])
page = int(request.POST['page'])
searchVal=str(request.POST['search']).strip()
dRange = str(request.POST['range']).split("-")
try:
date1 = datetime.datetime.strptime( dRange[0].strip(),'%b %d, %Y')
except:
date1 = datetime.datetime.strptime( dRange[0].strip(),'%B %d, %Y')
try:
date2 = datetime.datetime.strptime( dRange[1].strip()+" 23:59:59",'%b %d, %Y %H:%M:%S')
except:
date2 = datetime.datetime.strptime( dRange[1].strip()+" 23:59:59",'%B %d, %Y %H:%M:%S')
chartBYear = date1.year
chartBMonth = date1.month - 1
chartBDay = date1.day
if filter == "attendance":
dispType = 'Attendance'
chartTitle = "Student Attendance"
chartYTitle = "No. Attendees"
chartSName = "Attendees"
dispList = []
chartData = []
chkDay = date1
while chkDay <= date2:
strDay1 = chkDay.strftime("%m %d, %Y")
chkDay2 = datetime.datetime.strptime( strDay1 + " 23:59:59",'%m %d, %Y %H:%M:%S')
dayData = 0
if searchVal:
attndLst= Attendance.objects.filter( userId__name__icontains=searchVal,
dateTime__range=(chkDay, chkDay2))
else:
attndLst= Attendance.objects.filter( dateTime__range=(chkDay, chkDay2) )
for i in attndLst:
dayData += int( i.DancerNumber )
dispList.append( i )
chartData.append( dayData )
chkDay += datetime.timedelta( days=1 )
else:
dispType = 'Purchases'
chartTitle = "Student Classes Bought"
chartYTitle = "Classes Bought"
chartSName = "Purchases"
dispList = []
chartData = []
chkDay = date1
while chkDay <= date2:
strDay1 = chkDay.strftime("%m %d, %Y")
chkDay2 = datetime.datetime.strptime( strDay1 + " 23:59:59",'%m %d, %Y %H:%M:%S')
dayData = 0
if searchVal:
attndLst= Purchase.objects.filter( userId__name__icontains=searchVal,
date__range=(chkDay, chkDay2))
else:
attndLst= Purchase.objects.filter( date__range=(chkDay, chkDay2) )
for i in attndLst:
dayData += int( i.numberOfClasses )
dispList.append( i )
chartData.append( dayData )
chkDay += datetime.timedelta( days=1 )
for item in dispList:
item.name = item.userId.name
# Sort by
if sort == "name":
dispList.sort(key=operator.attrgetter("name"))
else:
if filter == "attendance":
dispList.sort(key=operator.attrgetter("dateTime"))
else:
dispList.sort(key=operator.attrgetter("date"))
dispList.reverse()
#---------------------------------------------------------------------
# Set the paginator object and get the current page data list
#---------------------------------------------------------------------
currPage = page
pgintObj = getPgintObj(request, dispList,
currPage, nCols = 1 )
dispList = pgintObj.getDataList( )
return render_to_response( 'historyInfo.html',
{'dispType': dispType,
'dispList': dispList,
'currPage': currPage,
'paginator': pgintObj,
'chartTitle': chartTitle,
'chartYTitle': chartYTitle,
'chartSName': chartSName,
'chartData': chartData,
'chartBYear': chartBYear,
'chartBMonth': chartBMonth,
'chartBDay': chartBDay,
'sort': sort,
'filter': filter,
},
context_instance=RequestContext(request) )
#==============================================================================
##
## Gets all people with birthdays within given number of days
##
#==============================================================================
def birthdaysWithin(days):
now = datetime.datetime.now()
then = now + timedelta(days)
# Build the list of month/day tuples.
monthdays = [(now.month, now.day)]
while now <= then:
monthdays.append((now.month, now.day))
now += timedelta(days=1)
# Tranform each into queryset keyword args.
monthdays = (dict(zip(("birthday__month", "birthday__day"), t))
for t in monthdays)
# Compose the djano.db.models.Q objects together for a single query.
query = reduce(operator.or_, (Q(**d) for d in monthdays))
# Run the query.
return User.objects.filter(query)
#==============================================================================
#
# "students": The student control page
#
#==============================================================================
def students( request, check=None):
user, errUrl = GetValidUser(request)
if errUrl:
return HttpResponseRedirect(errUrl)
return render_to_response( 'students.html',
{'userName': user.name,
'userType': user.userType,
'sort': "name",
'filter': None,
'currPage': 1,
},
context_instance=RequestContext(request) )
#==============================================================================
#
# "studentList": The students page
#
#==============================================================================
def studentList( request ):
# idle?
user, errUrl = GetValidUser(request)
if errUrl:
return HttpResponseRedirect(errUrl)
# Get "get" values
sort = str(request.POST['sort'])
filter = str(request.POST['filter'])
page = int(request.POST['page'])
search = str(request.POST['search']).strip()
dRange = str(request.POST['range']).split("-")
try:
date1 = datetime.datetime.strptime( dRange[0].strip(),'%b %d, %Y')
except:
date1 = datetime.datetime.strptime( dRange[0].strip(),'%B %d, %Y')
try:
date2 = datetime.datetime.strptime( dRange[1].strip()+" 23:59:59",'%b %d, %Y %H:%M:%S')
except:
date2 = datetime.datetime.strptime( dRange[1].strip()+" 23:59:59",'%B %d, %Y %H:%M:%S')
# Get user list with filters (bday, waiver, balance)
if filter == "bday":
userList = birthdaysWithin(7)
elif filter == "balance":
userList = User.objects.filter(balance__lte=3)
elif filter == "waiver":
userList = User.objects.filter(waiverSigned=False)
else:
userList = User.objects.all()
# Filter user list by search
if search:
userList = userList.filter(name__icontains=str(search))
#print userList
# Date range
try:
if date1.year != 2014 or date1.month != 1 or date1.day != 1:
attendanceLst = Attendance.objects.filter(userId__in = userList,
dateTime__range=(date1, date2))
usrAttenLst = []
for i in attendanceLst:
usrId = i.userId.userId
if usrId in usrAttenLst: continue
usrAttenLst.append( usrId )
userList = userList.filter(userId__in=usrAttenLst)
except:
pass
# Make userList a list so we can sort
userList = list(userList)
# Sort by
if str(sort) == "name":
userList.sort(key=operator.attrgetter("name"))
elif str(sort) == "activity":
userList.sort(key=operator.attrgetter("lastAccess"))
userList.reverse()
else:
userList.sort(key=operator.attrgetter("balance"))
# @WIP1@ These lines must be fixed
now = datetime.datetime.now()
for user2 in userList:
bday = user2.birthday
user2.hasBirthday = (bday.year != 1970 and bday.month != 1 and bday.day != 1)
lst7Day = now - datetime.timedelta( days = 7 )
user2.showBirthday = (bday > lst7Day) # check Bday within 7 days
user2.recent = ((now - user2.lastAccess).total_seconds() < 2*24*3600)
# @WIP1@ These lines must be fixed
nStudents = len( userList )
currPage = 4
nPages = 4
# url = request.get_full_path()[:-1]
#url = str(sortBy) + "/" + str(page) + "/" + str(range)
#---------------------------------------------------------------------
# Set the paginator object and get the current page data list
#---------------------------------------------------------------------
currPage = page
pgintObj = getPgintObj(request, userList,
currPage, nCols = 4 )
dataLst = pgintObj.getDataList( )
return render_to_response( 'studentsInfo.html',
{'userName': user.name,
'userType': user.userType,
'userList': dataLst,
'nStudents': nStudents,
'currPage': currPage,
'paginator': pgintObj,
'nPages': nPages,
#'url': url,
#'addUserForm': addUserForm,
'sort': sort,
'filter': filter,
},
context_instance=RequestContext(request) )
###############################################################################
##
## "addStudent": Add a new student; using drop down
##
###############################################################################
def addStudent( request, check = None ):
user, errUrl = GetValidUser(request)
if errUrl:
return HttpResponseRedirect(errUrl)
# Add new user validation and forms
if check is not None:
form = AddUserForm(request.GET.copy())
if form.is_valid():
user2 = form.save()
url = '/students/' # + str(sort) + "/" + str(page) + "/" + str(range) + "/"
json = util.JsonLoad( url )
else:
json = util.JsonFormError(form)
return HttpResponse(json,
mimetype='application/json')
# Make form
addUserForm = AddUserForm()
return render_to_response( 'dropDownForm.html',
{'form': addUserForm,
#'formName': 'addClient',
#'actionUrl': addDrpdnUrl,
},
context_instance=RequestContext(request))
###############################################################################
##
## "exportStudents": Create a CSV file which contains the users data
##
###############################################################################
def exportStudents( request ):
today = datetime.datetime.now( )
response = HttpResponse( mimetype = 'text/csv' )
fileName = "students_%s.csv" %today.strftime( '%y%m%d%H%M%S' )
response['Content-Disposition'] = 'attachment; filename=%s' %fileName
writer = csv.writer( response )
writer.writerow(['Name', 'E-mail', 'Address',
'Phone', 'Birthday', 'Balance',
'Waiver', 'Admin',
])
for i in User.objects.all():
address = i.address.replace( "\r\n", " " )
writer.writerow([i.name, i.email, address, i.phone,
str(i.birthday.date()), i.balance,
i.waiverSigned, i.userType
])
return response
###############################################################################
##
## "importStudents": Import a CSV file which contains the users data
##
###############################################################################
def importStudents( request ):
if request.POST:
fileName = request.FILES.copy()['fileName']
fileData = fileName.read().split( '\r' )
#----------------------------------------------------------------
# Check the file data to be correct and save it
#----------------------------------------------------------------
fileHeader = fileData[0][:55].replace( "\r", "" ) #[:55] is the header
header = "Name,E-mail,Address,Phone,Birthday,Balance,Waiver,Admin"
if fileHeader != header: return
#----------------------------------------------------------------
# Save imported file data into data base
#----------------------------------------------------------------
for row in fileData[1:]:
rowData = row.replace("\r", "" )
studentData = rowData.split( ',' )
if len( studentData ) != 8: continue
email = studentData[1]
if studentData[1] == "excelttest3@gmail.com":
file = open("/tmp/XXX.txt", "w")
file.write(", ".join(studentData))
file.close()
try:
user = User.objects.get( email__exact = email )
except User.DoesNotExist:
user = None
if user:
updateStudent( studentData, update = True)
else:
updateStudent( studentData )
return HttpResponseRedirect("/students/")
###############################################################################
##
## "updateStudent": Update student table
##
###############################################################################
def updateStudent( data, update = False ):
if update:
#----------------------------------------------------------------
# Update student data
#----------------------------------------------------------------
user = User.objects.get( email__exact = data[1] )
user.name = data[0]
user.address = data[2]
user.phone = data[3]
try:
user.birthday = datetime.datetime.strptime(data[4],'%Y-%m-%d')
except:
try:
data[4] = (data[4][:-2] + "19" + data[4][-2:]) if int(data[4][-2:]) > 20 else (data[4][:-2] + "20" + data[4][-2:]) #csv date comes in format 7/3/99
user.birthday = datetime.datetime.strptime(data[4],'%m/%d/%Y')
except:
user.birthday = datetime.datetime(1970, 1, 1)
try:
user.balance = int(data[5])
except:
user.balance = 0
user.waiverSigned = True if data[6] == "True" else False
try:
user.userType = int(data[7])
except:
user.userType = 0
user.save()
return
#--------------------------------------------------------------------
# Save New student
#--------------------------------------------------------------------
today = datetime.datetime.today()
user = User()
user.name = data[0]
user.email = data[1].lower()
user.password = util.encryptPass(str(random.random()))
user.address = data[2].lower()
user.phone = data[3]
user.lastAccess = today
try:
user.balance = int(data[5])
except:
user.balance = 0
user.waiverSigned = True if data[6] == "True" else False
user.facebook = False
user.notes = ""
user.dateCreated = today
try:
user.userType = int(data[7])
except:
user.userType = 0
user.idleTime = 10
try:
user.birthday = datetime.datetime.strptime(data[4],'%Y-%m-%d')
except:
try:
data[4] = (data[4][:-2] + "19" + data[4][-2:]) if int(data[4][-2:]) > 20 else (data[4][:-2] + "20" + data[4][-2:]) #csv date comes in format 7/3/99
user.birthday = datetime.datetime.strptime(data[4],'%m/%d/%Y')
except:
user.birthday = datetime.datetime(1970,1, 1)
birthdayAssigned = False
user.save()
#==============================================================================
#
# "clearance": A user does not have permission to be somewhere
#
#==============================================================================
def clearance(request):
try:
userName = request.session['fullName']
#userType = request.session['userType']
userId = request.session['userId']
user = User.objects.get(userId=userId)
userType = user.userType
return render_to_response('clearance.html',
{'userName': userName,
'userType': userType,
},
context_instance=RequestContext(request))
except:
if checkSignup is not None:
form = SignupForm(request.GET.copy())
if form.is_valid():
user = form.save()
email = form.data['email']
user = User.objects.get(email=email)
SetUserCookies(request, user)
json = util.JsonLoad('/students/name0/1/all')
else:
json = util.JsonFormError(form)
return HttpResponse(json,
mimetype='application/json')
elif checkSignin is not None:
form = SigninForm(request.GET.copy())
if form.is_valid():
user = User.objects.get(email=form.data['email'])
SetUserCookies(request, user)
json = util.JsonLoad('/students/name0/1/all')
else:
json = util.JsonFormError(form)
return HttpResponse(json,
mimetype='application/json')
else:
forup = SignupForm()
forin = SigninForm()
return render_to_response('clearance.html',
{'forup': forup,
'forin': forin,
'userName': None,
'userType': None,
},
context_instance=RequestContext(request))
#==============================================================================
#
# "edit": A page in which Ula can edit a student's information
#
#==============================================================================
def editStudent(request, userId, check=None):
user, errUrl = GetValidUser(request)
if errUrl:
return HttpResponseRedirect(errUrl)
## try:
## userId = request.GET['id']
## except:
## return HttpResponseRedirect('/students/name0/1/all/')
## try:
## url = request.GET['url']
## except:
## url = 'name0/1/all/'
if check is not None:
data = request.GET.copy()
form = EditUserForm(data)
if form.is_valid():
form.save()
json = util.JsonLoad( closeSecId = "editSudent" )
else:
json = util.JsonFormError(form)
return HttpResponse(json,
mimetype='application/json')
user2 = User.objects.get(userId = userId )
if str( user2.birthday.date() ) == "1970-01-01":
birthVal = ""
else:
birthVal = user2.birthday.date()
editUserForm = EditUserForm(initial = {'userId': user2.userId,
'name': user2.name,
'email': user2.email,
'address': user2.address,
'phone': user2.phone,
'birth': birthVal,
'balance': user2.balance,
'notes': user2.notes,
'waiver': user2.waiverSigned
}
)
## return render_to_response('editStudent.html',
## {'userId': user2.userId,
## 'userName': user.name,
## 'userType': user.userType,
## 'url': url,
## 'editUserForm': editUserForm,
## },
## context_instance=RequestContext(request))
##
return render_to_response( 'editStudentForm.html',
{'form': editUserForm,
'userType': int(user2.userType),
'title': 'Edit Student',
'actionUrl': "/students/edit/%d/" %int( userId ),
'submitVal': 'Save'
},
context_instance=RequestContext(request))
#==============================================================================
#
# "purchaseStudent": A purchase student page
#
#==============================================================================
def purchaseStudent(request, userId, check=None):
user, errUrl = GetValidUser(request)
if errUrl:
return HttpResponseRedirect(errUrl)
if check is not None:
data = request.GET.copy()
form = PurchaseUserForm(data)
if form.is_valid():
form.save()
json = util.JsonLoad( closeSecId = "purchaseStudent" )
else:
json = util.JsonFormError(form)
return HttpResponse(json,
mimetype='application/json')
user2 = User.objects.get(userId = userId )
purchaseUserForm = PurchaseUserForm( initial = {'userId': user2.userId })
return render_to_response( 'purchaseStudenttForm.html',
{'form': purchaseUserForm,
'title': 'Add classes to "%s"' %user2.name,
'actionUrl': "/students/purchase/%d/" %int( userId ),
'submitVal': 'Save'
},
context_instance=RequestContext(request))
#==============================================================================
#
# "rcntActivityStudent": Student recent activity page
#
#==============================================================================
def rcntActivityStudent(request, userId ):
user, errUrl = GetValidUser(request)
if errUrl:
return HttpResponseRedirect(errUrl)
user2 = User.objects.get(userId = userId )
recentLst = usrRcntChanges( user2 )
return render_to_response( 'recentActivity.html',
{'recentLst': recentLst,
'title': '"%s" recent activity' %user2.name,
},
context_instance=RequestContext(request))
#==============================================================================
#
# "forgotPassword": Forgot password page
#
#==============================================================================
def forgotPassword( request, check = None ):
if check is not None:
data = request.GET.copy()
form = ForgotPasswordForm( data )
if form.is_valid():
email = form.data['email']
user = User.objects.get( email__exact = email )
forgotPasswd= form.save( )
emailBody = "Dear %s" %user.name
emailBody +="<br/><br/>To reset your account password, please click on the following link:"
emailBody +="<br/>"
emailBody += settings.PROJECT_URL + "forgotCode/" + forgotPasswd.forgotCode
util.emailNotification( "Forgot password",
user,
emailBody )
json = util.JsonLoad( '/forgotPdSentEmail/' )
else:
json = util.JsonFormError( form )
return HttpResponse(json, mimetype = 'application/json' )
form = ForgotPasswordForm()
extraInformation = "Please enter the e-mail address registered with your account.<br/>An e-mail containing a password reset link will be sent to your email."
forup = SignupForm()
forin = SigninForm()
return render_to_response( 'forgotPassword.html',
{'form': form,
'forup': forup,
'forin': forin,
'extraInformation': extraInformation
},
context_instance=RequestContext(request))
def forgotPdSentEmail( request ):
forup = SignupForm()
forin = SigninForm()
return render_to_response( 'forgotPwdSentEmail.html',
{'forup': forup,
'forin': forin,
},
context_instance=RequestContext(request) )
##############################################################################
##
## "changePassword": Change Password page
##
###############################################################################
def changePassword( request, forgotCode = None, check = None ):
if check is not None:
data = request.GET.copy( )
form = ChangePasswordForm( data )
form.forgotCode = forgotCode
if form.is_valid():
user = form.save( )
## SetUserCookies( request, user )
emailBody = "Dear %s" %user.name
emailBody +="<br/><br/>Your passowrd has been reset on "
emailBody += settings.PROJECT_URL
emailBody +="<br/>"
util.emailNotification( "Change password",
user,
emailBody )
json = util.JsonLoad( '/passwordChanged/' )
else:
json = util.JsonFormError( form )
return HttpResponse(json, mimetype = 'application/json')
today = datetime.datetime.today( )
( validFlag, msg ) = chkValidForgotCode( forgotCode )
if validFlag:
form = ChangePasswordForm( )
return render_to_response('changePassword.html',
{
'form': form,
'actionUrl':"/forgotCode/" + forgotCode + "/"
},
context_instance=RequestContext(request) )
else:
forup = SignupForm()
forin = SigninForm()
return render_to_response('denyAccess.html',
{
'message': msg,
'forup': forup,
'forin': forin,
},
context_instance=RequestContext(request))
def passwordChanged( request ):
forup = SignupForm()
forin = SigninForm()
return render_to_response( 'passwordChanged.html',
{'forup': forup,
'forin': forin,
},
context_instance=RequestContext(request) )
###############################################################################
##
## "chkValidForgotCode": Check forgot code validation
##
###############################################################################
def chkValidForgotCode( forgotCode ):
today = datetime.datetime.today( )
validFlag = False
retMsg = ""
try:
forgotPasswd = ForgotPassword.objects.get(
forgotCode__exact = forgotCode)
if forgotPasswd.active == 'Y':
chkTime = ( today - forgotPasswd.forgotTime ).seconds
if chkTime > 21600:
retMsg = "Your time has been expired. You have only 6 hours"
retMsg +=" to change your password."
else:
validFlag= True
forgotPasswd.active = 'N'
forgotPasswd.save( )
else:
retMsg = "Sorry, the forgot code had been used before and "
retMsg +="is not valid."
except:
retMsg = "Sorry, the user info is invalid."
return ( validFlag, retMsg )
#==============================================================================
#
# "myprofile": Allows students to change their information
#
#==============================================================================
def myprofile(request, check=None):
user, errUrl = GetValidUser(request)
if errUrl:
return HttpResponseRedirect(errUrl)
if check is not None:
data = request.GET.copy( )
form = MyprofileForm( data )
if form.is_valid():
form.save( )
json = util.JsonLoad('/home/')
else:
json = util.JsonFormError(form)
return HttpResponse(json,
mimetype='application/json')
if str( user.birthday.date() ) == "1970-01-01":
birthVal = ""
else:
birthVal = user.birthday.date()
form = MyprofileForm(initial = {'userId': user.userId,
'name': user.name,
'email': user.email,
'birth': birthVal,
'phone': user.phone,
'address': user.address,
'idleTime': user.idleTime,
}
)
return render_to_response( 'myProfile.html',
{'form': form,
'userName':user.name,
'userType':user.userType,
},
context_instance=RequestContext(request))
###############################################################################
##
## "getPgintObj": Create paginator object and return it.
##
###############################################################################
def getPgintObj( request, usrLst, currPage = 1, padding = 3, nCols = 1 ):
# currUsr = getCurrUser( request )
pgintObj = paginator.PaginatorObj( usrLst,
6 * nCols,
currPage,
padding )
return pgintObj
###############################################################################
##
## "usrRcntChanges":
##
###############################################################################
def usrRcntChanges( user ):
now = datetime.datetime.now()
rcnChgs = RecentChange.objects.filter(userId = user.userId)
retVal = []
for chg in rcnChgs:
if ((now - chg.dateTime).total_seconds() < 2*24*3600):
retVal.append( [chg.change, chg.value, chg.dateTime.strftime('%Y-%m-%d') ] )
else:
chg.delete()
return retVal
| [
"vedaad799@gmail.com"
] | vedaad799@gmail.com |
6d6f0f9613d182bbd152627313f49791fd6e358f | 4fa6036d6f924105554ae02ffbd1cec23de80354 | /2020-09-09-blusa-camiseta-criterios-minimo-stock/code_low_demand_aux.py | 541fe5ff613f616a291e409bba74790936b4ff6c | [] | no_license | daryachyzhyk/Reports | a9ac7f4850a26c51dae5198a940ed786636f7acd | 93eb92dc04beb6b02ce97fb74496195d5647c77d | refs/heads/master | 2023-05-14T20:22:41.992352 | 2021-06-09T07:06:42 | 2021-06-09T07:06:42 | 277,542,689 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 20,516 | py | '''Two families: BLUSA and CAMISETA do not follow the rule of the minimum of stock that we establish on the level family-size. They have a good level of stock but the PS feedback is bad.
Investigate these families using another variables such as: color, print, basic, style, fabric, etc.
Use particular dates.
Compare % of sent units to available in stock.
'''
import os
import pandas as pd
import numpy as np
import pickle
from joblib import Parallel, delayed, parallel_backend
def opt_sum(opt, df_opt_dummy):
df_opt = pd.DataFrame([])
df_opt['opt_demanda'] = df_opt_dummy[opt] * df_opt_dummy['demanda']
df_opt['opt_stock_actual'] = df_opt_dummy[opt] * df_opt_dummy['stock_actual']
df_opt['demanda'] = df_opt_dummy['demanda']
df_opt['stock_actual'] = df_opt_dummy['stock_actual']
# df_opt['real_stock'] = df_opt_dummy['real_stock'] # TODO: delete
opt_sum = df_opt.sum()
return opt_sum
######################################################################################################################
# path
file_demanda = ('/var/lib/lookiero/stock/stock_tool/demanda_preprocessed.csv.gz')
file_product = ('/var/lib/lookiero/stock/stock_tool/productos_preprocessed.csv.gz')
file_stock = ('/var/lib/lookiero/stock/stock_tool/stock.csv.gz')
path_results = ('/home/darya/Documents/Reports/2020-09-09-blusa-camiseta-criterios-minimo-stock')
file_feedback = ('/home/darya/Documents/Reports/2020-09-09-blusa-camiseta-criterios-minimo-stock/Stock_Situacion.csv')
#######################################
# feedback
df_feedback = pd.read_csv(file_feedback, usecols=['Familia', 'Talla', 'Fecha', 'Stock NOK'])
df_feedback = df_feedback.rename(columns={'Familia': 'family_desc',
'Talla': 'size',
'Fecha': 'date',
'Stock NOK': 'stock_nok'})
df_feedback['stock_nok'] = df_feedback['stock_nok'].fillna(0)
df_feedback.loc[df_feedback['stock_nok'] != 0, 'stock_nok'] = 1
family_list = list(set(df_feedback['family_desc']))
date_list = list(set(df_feedback['date']))
#######################################
# demanda
query_demanda_text = 'date_ps_done in @date_list and family_desc in @family_list'
df_demanda_raw = pd.read_csv(file_demanda, usecols=['reference', 'date_ps_done', 'family_desc']).query(query_demanda_text)
reference_list = list(set(df_demanda_raw['reference'].to_list()))
df_demanda = df_demanda_raw.groupby(['date_ps_done', 'reference']).size().reset_index(name='demanda')
df_demanda = df_demanda.rename(columns={'date_ps_done': 'date'})
###################################
# stock
query_stock_text = 'date in @date_list and reference in @reference_list'
df_stock_raw = pd.read_csv(file_stock).query(query_stock_text)
# visible
df_stock = df_stock_raw.drop(df_stock_raw[(df_stock_raw['es'] == 0) & (df_stock_raw['fr'] == 0) &
(df_stock_raw['gb'] == 0) & (df_stock_raw['pt'] == 0) &
(df_stock_raw['be'] == 0) & (df_stock_raw['lu'] == 0) &
(df_stock_raw['it'] == 0)].index)
df_demanda_stock = pd.merge(df_demanda,
df_stock[['date', 'real_stock', 'reference']],
on=['date', 'reference'],
how='outer')
df_demanda_stock['real_stock'] = df_demanda_stock['real_stock'].fillna(0)
df_demanda_stock['demanda'] = df_demanda_stock['demanda'].fillna(0)
######################################
# add description of the product
# product
var_list_aux = ['reference', 'family_desc', 'size']
# TODO add 'composicion' , corte, grosor, ligero -> all fields are Nan
var_list_cat = ['clima',
# 'aventurera',
'basico',
'estilo_producto',
'fit',
'uso',
# 'pattern',
'has_pattern',
'composicion',
'origen',
'color_group',
'color_category',
'price_range_product',
# 'tejido',
'acabado',
# TODO: añadir premium
#'premium',
# 'corte',
# 'grosor',
# 'ligero'
]
# var_list_opt = []
# var_list = var_list_aux + var_list_cat + var_list_opt
var_list = var_list_aux + var_list_cat
query_product_text = 'reference in @reference_list'
df_product_raw = pd.read_csv(file_product, usecols=var_list).query(query_product_text)
df_product_raw = df_product_raw.drop_duplicates('reference', keep='last')
################################
df = pd.merge(df_demanda_stock,
df_product_raw,
on=['reference'],
how='outer')
df = df[df['family_desc'].isin(family_list)]
# stock actual
df['stock_actual'] = df['real_stock']
df.loc[df['stock_actual'] < df['demanda'], 'stock_actual'] = df['demanda']
date_family_size_list = list(zip(df['date'], df['family_desc'], df['size']))
# test
# date_family_size_list = list(zip(['2020-07-24', '2020-07-24'], ['VESTIDO', 'VESTIDO'], ['M', 'XXXL']))
# df_indicators = pd.DataFrame([])
mean_weight_relative_list = []
mean_weight_abs_list = []
date_family_size_var_valor_list = []
for date_family_size in date_family_size_list:
# date_family_size = date_family_size_list[0]
dt = date_family_size[0]
family = date_family_size[1]
sz = date_family_size[2]
print(dt)
print(family)
print(sz)
df_fam_sz_var = df[(df['date'] == dt) & (df['family_desc'] == family) & (df['size'] == sz)]
for var in var_list_cat:
# print(var)
if ~df_fam_sz_var[var].isnull().all():
# dummies
df_dummy = pd.get_dummies(df_fam_sz_var[var], columns=var)
# print(df_dummy.head())
var_group_aux = ['date', 'family_desc', 'size', 'demanda', 'real_stock', 'stock_actual']
df_opt_dummy = pd.concat([df_fam_sz_var[var_group_aux], df_dummy], axis=1)
var_opt_list = df_dummy.columns.to_list()
# for each option of the variable calculate distr_abs and distr_relativa
with parallel_backend('threading', n_jobs=6):
opt_sum_paral = Parallel()(
delayed(opt_sum)(opt, df_opt_dummy) for opt in var_opt_list)
df_gr = pd.DataFrame(opt_sum_paral)
# option
# df_opt = pd.DataFrame([])
# df_gr = pd.DataFrame([])
# for opt in var_opt_list:
# df_opt['opt_demanda'] = df_opt_dummy[opt] * df_opt_dummy['demanda']
# df_opt['opt_stock_actual'] = df_opt_dummy[opt] * df_opt_dummy['stock_actual']
# df_opt['demanda'] = df_opt_dummy['demanda']
# df_opt['stock_actual'] = df_opt_dummy['stock_actual']
# # df_opt['real_stock'] = df_opt_dummy['real_stock'] # TODO: delete
# df_gr = df_gr.append(df_opt.sum(), ignore_index=True)
df_gr['pct_demanda'] = df_gr['opt_demanda'] / df_gr['demanda']
# porcentaje de demanda de opcion de demanda de variable
df_gr['pct_stock'] = df_gr['opt_stock_actual'] / df_gr['stock_actual']
df_gr['pct_demanda_stock_actual'] = df_gr['opt_demanda'] / df_gr['opt_stock_actual']
df_gr['distr_relative'] = np.where((df_gr['pct_demanda'] == 0) | (df_gr['pct_demanda'] < df_gr['pct_stock']), 0, 1)
df_gr['distr_abs'] = np.where((df_gr['demanda'] == 0) | (df_gr['pct_demanda_stock_actual'] < 1), 0, 1)
# df_gr['mean_weight'] = df_gr['mean_weight']
date_family_size_var_valor_list.append((dt, family, sz, var,
(df_gr['distr_relative'] * df_gr['pct_demanda']).sum(),
(df_gr['distr_abs'] * df_gr['pct_demanda']).sum()))
# mean_weight_relative_list.append((df_gr['distr_relative'] * df_gr['pct_demanda']).sum())
# mean_weight_abs_list.append((df_gr['distr_abs'] * df_gr['pct_demanda']).sum())
df_indicators = pd.DataFrame(date_family_size_var_valor_list, columns=['date', 'family_desc', 'size', 'variable',
'mean_weight_relative', 'mean_weight_abs'])
df_indicators_label = pd.merge(df_indicators, df_feedback,
on=['date', 'family_desc', 'size'])
df_indicators_label_gr = df_indicators_label.groupby(['date', 'family_desc', 'size']).agg({'mean_weight_relative': 'mean',
'mean_weight_abs': 'mean',
'stock_nok': 'last'}).reset_index()
# save
df_indicators_label.to_csv(os.path.join(path_results, 'date_family_size_var_mean_weight_relat_abs_psfeedback.csv'), index=False)
df_indicators_label_gr.to_csv(os.path.join(path_results, 'date_family_size_mean_var_mean_weight_relat_abs_psfeedback.csv'), index=False)
# with open(os.path.join(path_results, 'var_list.txt'), "wb") as fp: # Pickling
# pickle.dump(var_group, fp)
# df_indicators = pd.DataFrame(date_family_size_list, columns=['date', 'family_desc', 'size'])
# df_indicators['mean_weight_relative'] = mean_weight_relative_list
# df_indicators['mean_weight_abs'] = mean_weight_abs_list
##########################################
# test
#
# df_test = df.copy()
#
# df = df_test.copy()
# var = 'fit'
# df_fam_sz_var = df[(df['family_desc'].isin(['VESTIDO', 'TOP', 'ABRIGO'])) &
# (df['date'] == '2020-07-24') &
# (df['size'] == 'XXXL')]
####### end test
#
#
# for col in var_group:
# print(col)
# columns = var_group_aux + [col]
# df_var = df[columns].copy()
#
#
# df_var[col + '_demanda'] = df_var[col] * df_var['demanda']
# df_var[col + '_stock_real'] = df_var[col] * df_var['real_stock']
# df_var[col + '_stock_actual'] = df_var[col] * df_var['stock_actual']
#
# df_gr = df_var.groupby(['date', 'family_desc', 'size']).sum().reset_index()
# # stock real, stock appeared in snapshots
# # stock actual, in case when number of items in snapshots is less then in stock_real,
#
# # porcentaje de demanda de opcion de demanda de variable
# df_gr[col + '_pct_demanda_demanda'] = df_gr[col + '_demanda'] / df_gr['demanda']
#
# # porcentaje de demanda de opcion de demanda de variable
# df_gr[col + '_pct_stock_stock'] = df_gr[col + '_stock_actual'] / df_gr['stock_actual']
#
# # percentage of variable option (option "holgado" of variable "fit") shipped of all real stock (snapshot),
# # could be more then 100%
# df_gr[col + '_pct_demanda_stock_real'] = df_gr[col + '_demanda'] / df_gr[col + '_stock_real']
#
# # percentage of variable option (option "holgado" of variable "fit") shipped of all actual stock (snapshot),
# # could be 100% maximum
# df_gr[col + '_pct_demanda_stock_actual'] = df_gr[col + '_demanda'] / df_gr[col + '_stock_actual']
#
# # percentage of option (option 'holgado') of variable ('fit') stock, could be 100% maximum
# df_gr[col + '_pct_varstock'] = df_gr[col + '_stock_actual'] / df_gr['stock_actual']
#
# # percentage of option (option 'holgado') shipped of variable ('fit') stock
#
# df_gr[col + '_pct_demanda_weight'] = df_gr[col + '_pct_demanda_stock_actual'] * df_gr[col + '_pct_varstock']
#
# # TODO: distr relativa, absoluta
# # distr relativa distr abs
# # =if (OR(demanda=0, demanda de stock actual > stock actual), 0, 1)
# df_gr[col + '_distr_relativa'] = np.where((df_gr[col + '_pct_demanda_demanda'] == 0) |
# (df_gr[col + '_pct_demanda_demanda'] < df_gr[col + '_pct_stock_stock']),
# 0, 1)
#
# df_gr[col + '_distr_abs'] = np.where((df_gr[col + '_demanda'] == 0) |
# (df_gr[col + '_pct_demanda_stock_actual'] < 1),
# 0, 1)
#
#
#
# df_gr = df_gr.fillna(0)
#
# df_gr = df_gr.replace(np.inf, 1.0)
#
# # TODO: añadir nuevas columnas
# df_var_pct = df_var_pct.merge(df_gr[['date', 'family_desc', 'size',
# col + '_stock_actual',
# col + '_pct_demanda_demanda',
# col + '_pct_stock_stock',
# col + '_pct_demanda_stock_real',
# col + '_pct_demanda_stock_actual',
# col + '_pct_varstock',
# col + '_pct_demanda_weight',
# col + '_distr_relativa',
# col + '_distr_abs']],
# on=['date', 'family_desc', 'size'],
# how='outer')
#
# # save as column
#
# df_temp = df_var_pct[['date', 'family_desc', 'size']]
# df_temp['varoption_pct_demanda_stock_real'] = df_var_pct[col + '_pct_demanda_stock_real']
# df_temp['varoption_pct_demanda_stock_actual'] = df_var_pct[col + '_pct_demanda_stock_actual']
#
# df_temp['varoption_pct_varstock'] = df_var_pct[col + '_pct_varstock']
# df_temp['varoption_pct_demanda_weight'] = df_var_pct[col + '_pct_demanda_weight']
#
# df_temp['varoption'] = col
# df_var_pct_col = df_var_pct_col.append(df_temp)
#
# # TODO df_var_pct_col change inf to 10, nan to 0
#
# df_var_pct = df_var_pct.fillna(0)
# df_var_pct_col = df_var_pct_col.fillna(0)
#
# df_var_pct_col = df_var_pct_col.replace(np.inf, 1.0)
#
# df_var_pct_ps = df_var_pct.merge(df_feedback,
# on=['date', 'family_desc', 'size'],
# how='outer')
#
# df_var_pct_col_ps = df_var_pct_col.merge(df_feedback,
# on=['date', 'family_desc', 'size'],
# how='outer')
#
# df_var_pct_ps = df_var_pct_ps.fillna(0)
# df_var_pct_ps = df_var_pct_ps.replace(np.inf, 1)
#
# # TODO merge df_var_pct_col with PS labels
# ##################################################################################################################
# # save
#
# # df_var_pct_ps.to_csv(os.path.join(path_results, 'date_family_size_var_pct_psfeedback.csv'), index=False)
# # df_var_pct_col_ps.to_csv(os.path.join(path_results, 'date_family_size_var_pct_col_psfeedback.csv'), index=False)
# #
# # aa = df_var_pct_ps.groupby(['family_desc']).agg({'stock_nok': 'mean'})
# #
# # with open(os.path.join(path_results, 'var_list.txt'), "wb") as fp: # Pickling
# # pickle.dump(var_group, fp)
#
#
# test = df[(df['family_desc']=='VESTIDO') & (df['date']=='2020-07-24') & (df['size']=='XXXL')]
#
# var_list_aux = ['reference', 'family_desc', 'size']
# var_group = set(df.columns.to_list()) - set(['date', 'reference', 'demanda', 'real_stock', 'family_desc', 'size', 'stock_actual'])
#
#
#
#
# # # eliminate good dates for CAMISETA and good date for BLUSA
# #
# # df = df.drop(df[(df['family_desc'] == 'CAMISETA') & (~df['date'].isin(date_list_camiseta))].index)
# #
# # df = df.drop(df[(df['family_desc'] == 'BLUSA') & (~df['date'].isin(date_list_blusa))].index)
#
#
# df_return = pd.DataFrame([])
# df_threshold = pd.DataFrame([])
#
# var_dummies = set(df.columns.to_list()) - set(['date', 'reference', 'demanda', 'real_stock', 'family_desc', 'size'])
#
# ###############################################
# ###############################################
#
# for var_name in var_list[2:]:
# print(var_name)
# # var_name = 'aventurera'
#
# df_var = df.groupby(['date', 'family_desc', var_name]).agg({'demanda': 'sum',
# 'real_stock': 'sum'}).reset_index()
#
# df_var['demanda_pct'] = df_var['demanda'] / df_var['real_stock'] * 100
#
#
# df_var['demanda_pct_w'] = df_var['demanda_pct'] / df_var['real_stock']
#
# df_var.loc[(df_var['demanda'] == 0) & (df_var['real_stock'] != 0), 'demanda_pct'] = 0
# df_var.loc[(df_var['demanda'] != 0) & (df_var['real_stock'] == 0), 'demanda_pct'] = 1
#
# threshold_min = 20
#
# threshold_max = 80
#
# threshols_days = 0.3
#
#
# df_var_thr = df_var[(df_var['demanda_pct'] < threshold_min) | (df_var['demanda_pct'] > threshold_max)]
# df_var_thr['var_name'] = var_name
# df_var_thr = df_var_thr.rename(columns={var_name: 'var_option'})
# df_threshold = df_threshold.append(df_var_thr)
#
#
# df_var_thr_min = df_var[df_var['demanda_pct'] < threshold_min]
#
# df_var_thr_max = df_var[df_var['demanda_pct'] > threshold_max]
#
#
#
#
#
#
# df_var_thr_min['n'] = 1
# df_var_thr_max['n'] = 1
# n_div = len(df['date'].unique())
#
# df_var_thr_min_fam = df_var_thr_min.groupby(['family_desc', var_name]).agg({'n': 'sum',
# 'demanda_pct': 'mean'}).reset_index()
# df_var_thr_min_fam['n'] = df_var_thr_min_fam['n'] / n_div
#
# df_var_thr_max_fam = df_var_thr_max.groupby(['family_desc', var_name]).agg({'n': 'sum',
# 'demanda_pct': 'mean'}).reset_index()
# df_var_thr_max_fam['n'] = df_var_thr_max_fam['n'] / n_div
#
# df_var_demand_low = df_var_thr_min_fam[df_var_thr_min_fam['n'] >= threshols_days]
#
#
# df_var_demand_high = df_var_thr_max_fam[df_var_thr_max_fam['n'] >= threshols_days]
#
# df_var_demand_low['var_type'] = var_name
# df_var_demand_low['problem_type'] = 'demand_low'
#
# df_var_demand_high['var_type'] = var_name
# df_var_demand_high['problem_type'] = 'demand_high'
#
# df_var_demand_low = df_var_demand_low.rename(columns={var_name: 'var_name'})
# df_var_demand_high = df_var_demand_high.rename(columns={var_name: 'var_name'})
#
# df_return = df_return.append(df_var_demand_low)
# df_return = df_return.append(df_var_demand_high)
#
# # df_return = df_return.rename(columns={var_name: 'var_name'})
#
# # save
# df_return.to_csv(os.path.join(path_results, 'blusa_camiseta_low_hight_demand_pct.csv'))
# df_threshold.to_csv(os.path.join(path_results, 'blusa_camiseta_threshold_pct_pct.csv'))
################ END ################
#####
# brand
# df_brand = df.groupby(['date', 'family_desc', 'brand']).agg({'demanda': 'sum',
# 'real_stock': 'sum'}).reset_index()
#
# df_brand['demanda_pct'] = df_brand['demanda'] / df_brand['real_stock'] * 100
#
#
# df_brand['demanda_pct_w'] = df_brand['demanda_pct'] / df_brand['real_stock']
#
# df_brand.loc[(df_brand['demanda'] == 0) & (df_brand['real_stock'] != 0), 'demanda_pct'] = 0
# df_brand.loc[(df_brand['demanda'] != 0) & (df_brand['real_stock'] == 0), 'demanda_pct'] = 1
#
# threshold_min = 20
#
# threshold_max = 80
#
#
# df_brand_thr = df_brand[(df_brand['demanda_pct'] < threshold_min) | (df_brand['demanda_pct'] > threshold_max)]
#
#
#
# df_brand_thr_min = df_brand[df_brand['demanda_pct'] < threshold_min]
#
# df_brand_thr_max = df_brand[df_brand['demanda_pct'] > threshold_max]
#
#
#
#
#
#
# df_brand_thr_min['n'] = 1
# df_brand_thr_max['n'] = 1
# n_div = len(df['date'].unique())
#
# df_brand_thr_min_fam = df_brand_thr_min.groupby(['family_desc', 'brand']).agg({'n': 'sum'}).reset_index()
# df_brand_thr_min_fam['n'] = df_brand_thr_min_fam['n'] / n_div
#
# df_brand_thr_max_fam = df_brand_thr_max.groupby(['family_desc', 'brand']).agg({'n': 'sum'}).reset_index()
# df_brand_thr_max_fam['n'] = df_brand_thr_max_fam['n'] / n_div
#
# df_brand_demand_low = df_brand_thr_min_fam[df_brand_thr_min_fam['n'] >= 0.5]
#
#
# df_brand_demand_high = df_brand_thr_max_fam[df_brand_thr_max_fam['n'] >= 0.5]
#
# df_brand_demand_low['var_type'] = 'brand'
# df_brand_demand_low['problem_type'] = 'demand_low'
#
# df_brand_demand_high['var_type'] = 'brand'
# df_brand_demand_high['problem_type'] = 'demand_high'
#
#
#
#
# df_return = df_return.append(df_brand_demand_low)
# df_return = df_return.append(df_brand_demand_high)
#
# df_return = df_return.rename(columns={'brand': 'var_name'})
| [
"d.chyzhyk@lookiero.com"
] | d.chyzhyk@lookiero.com |
ad3ee97e6a9f4657ce8b2dda99cafee6dda6323b | 2f7f6dda2a7fc1c949450e03cbd567ac7f9a46d9 | /tools/batch/submit-job.py | 9bbbc6fe81ee4b2045ef15802d86cbf469148ea1 | [
"Apache-2.0"
] | permissive | cuulee/gluon-nlp | fe31952925a483b14df4a9faaf1e5259805f9856 | 46c9d014ac0a022b208ce335cb6ee8a54771bae4 | refs/heads/master | 2022-12-19T16:00:32.269318 | 2020-10-22T03:33:47 | 2020-10-22T03:33:47 | null | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 6,565 | py | import argparse
import random
import re
import sys
import time
from datetime import datetime
import boto3
from botocore.compat import total_seconds
instance_type_info = {
'g4dn.4x': {
'job_definition': 'gluon-nlp-g4dn_4xlarge:5',
'job_queue': 'g4dn'
},
'g4dn.8x': {
'job_definition': 'gluon-nlp-g4dn_8xlarge:5',
'job_queue': 'g4dn'
},
'g4dn.12x': {
'job_definition': 'gluon-nlp-g4dn_12xlarge:5',
'job_queue': 'g4dn-multi-gpu'
},
'p3.2x': {
'job_definition': 'gluon-nlp-p3_2xlarge:5',
'job_queue': 'p3'
},
'p3.8x': {
'job_definition': 'gluon-nlp-p3_8xlarge:5',
'job_queue': 'p3-4gpu'
},
'p3.16x': {
'job_definition': 'gluon-nlp-p3_16xlarge:5',
'job_queue': 'p3-8gpu'
},
'p3dn.24x': {
'job_definition': 'gluon-nlp-p3_24xlarge:5',
'job_queue': 'p3dn-8gpu'
},
'c5n.4x': {
'job_definition': 'gluon-nlp-c5_4xlarge:3',
'job_queue': 'c5n'
},
'c5n.18x': {
'job_definition': 'gluon-nlp-c5_18xlarge:3',
'job_queue': 'c5n'
}
}
parser = argparse.ArgumentParser(formatter_class=argparse.ArgumentDefaultsHelpFormatter)
parser.add_argument('--profile', help='profile name of aws account.', type=str,
default=None)
parser.add_argument('--region', help='Default region when creating new connections', type=str,
default=None)
parser.add_argument('--name', help='name of the job', type=str, default='dummy')
parser.add_argument('--job-type', help='type of job to submit.', type=str,
choices=instance_type_info.keys(), default='g4dn.4x')
parser.add_argument('--source-ref',
help='ref in GluonNLP main github. e.g. master, refs/pull/500/head',
type=str, default='master')
parser.add_argument('--work-dir',
help='working directory inside the repo. e.g. scripts/preprocess',
type=str, default='scripts/preprocess')
parser.add_argument('--saved-output',
help='output to be saved, relative to working directory. '
'it can be either a single file or a directory',
type=str, default='.')
parser.add_argument('--save-path',
help='s3 path where files are saved.',
type=str, default='batch/temp/{}'.format(datetime.now().isoformat()))
parser.add_argument('--command', help='command to run', type=str,
default='git rev-parse HEAD | tee stdout.log')
parser.add_argument('--remote',
help='git repo address. https://github.com/dmlc/gluon-nlp',
type=str, default="https://github.com/dmlc/gluon-nlp")
parser.add_argument('--wait', help='block wait until the job completes. '
'Non-zero exit code if job fails.', action='store_true')
parser.add_argument('--timeout', help='job timeout in seconds', default=None, type=int)
args = parser.parse_args()
session = boto3.Session(profile_name=args.profile, region_name=args.region)
batch, cloudwatch = [session.client(service_name=sn) for sn in ['batch', 'logs']]
def printLogs(logGroupName, logStreamName, startTime):
kwargs = {'logGroupName': logGroupName,
'logStreamName': logStreamName,
'startTime': startTime,
'startFromHead': True}
lastTimestamp = 0
while True:
logEvents = cloudwatch.get_log_events(**kwargs)
for event in logEvents['events']:
lastTimestamp = event['timestamp']
timestamp = datetime.utcfromtimestamp(lastTimestamp / 1000.0).isoformat()
print('[{}] {}'.format((timestamp + '.000')[:23] + 'Z', event['message']))
nextToken = logEvents['nextForwardToken']
if nextToken and kwargs.get('nextToken') != nextToken:
kwargs['nextToken'] = nextToken
else:
break
return lastTimestamp
def nowInMillis():
endTime = long(total_seconds(datetime.utcnow() - datetime(1970, 1, 1))) * 1000
return endTime
def main():
spin = ['-', '/', '|', '\\', '-', '/', '|', '\\']
logGroupName = '/aws/batch/job'
jobName = re.sub('[^A-Za-z0-9_\-]', '', args.name)[:128] # Enforce AWS Batch jobName rules
jobType = args.job_type
jobQueue = instance_type_info[jobType]['job_queue']
jobDefinition = instance_type_info[jobType]['job_definition']
command = args.command.split()
wait = args.wait
parameters = {
'SOURCE_REF': args.source_ref,
'WORK_DIR': args.work_dir,
'SAVED_OUTPUT': args.saved_output,
'SAVE_PATH': args.save_path,
'COMMAND': args.command,
'REMOTE': args.remote
}
kwargs = dict(
jobName=jobName,
jobQueue=jobQueue,
jobDefinition=jobDefinition,
parameters=parameters,
)
if args.timeout is not None:
kwargs['timeout'] = {'attemptDurationSeconds': args.timeout}
submitJobResponse = batch.submit_job(**kwargs)
jobId = submitJobResponse['jobId']
print('Submitted job [{} - {}] to the job queue [{}]'.format(jobName, jobId, jobQueue))
spinner = 0
running = False
status_set = set()
startTime = 0
logStreamName = None
while wait:
time.sleep(random.randint(5, 10))
describeJobsResponse = batch.describe_jobs(jobs=[jobId])
status = describeJobsResponse['jobs'][0]['status']
if status == 'SUCCEEDED' or status == 'FAILED':
print('=' * 80)
print('Job [{} - {}] {}'.format(jobName, jobId, status))
if logStreamName:
startTime = printLogs(logGroupName, logStreamName, startTime) + 1
sys.exit(status == 'FAILED')
elif status == 'RUNNING':
logStreamName = describeJobsResponse['jobs'][0]['container']['logStreamName']
if not running:
running = True
print('\rJob [{}, {}] is RUNNING.'.format(jobName, jobId))
if logStreamName:
print('Output [{}]:\n {}'.format(logStreamName, '=' * 80))
if logStreamName:
startTime = printLogs(logGroupName, logStreamName, startTime) + 1
elif status not in status_set:
status_set.add(status)
print('\rJob [%s - %s] is %-9s... %s' % (jobName, jobId, status, spin[spinner % len(spin)]),)
sys.stdout.flush()
spinner += 1
if __name__ == '__main__':
main()
| [
"noreply@github.com"
] | noreply@github.com |
224ec89ac931e042e1944042b2a3f733ec7c52c3 | 5894844ffd7edfe04047a1310535679904f23b31 | /app/api/serializers/lesson.py | c2e9d0608a6e2745d2a10a7d73fac58208370d86 | [] | no_license | kevbrygil/dacodes-API | 778108017584be01f88a7b74f4a5fbdd258ca886 | ec51c4e8103236e9302db419ab1bce1b14526ac1 | refs/heads/master | 2022-07-25T16:30:11.831860 | 2019-10-23T14:38:33 | 2019-10-23T14:38:33 | 216,649,135 | 0 | 0 | null | 2022-07-05T21:47:09 | 2019-10-21T19:29:20 | Python | UTF-8 | Python | false | false | 507 | py | from marshmallow import Schema, fields
from .validations.validators import validate_mandatory_courses, validate_mandatory_courses_code
class LessonSchema(Schema):
id = fields.Str()
name = fields.Str(required=True)
course_id = fields.Str()
description = fields.Str()
question_details = fields.Str()
code = fields.Str(required=True)
order = fields.Integer()
hours = fields.Integer()
score = fields.Integer(required=True)
aproval_score = fields.Integer(required=True)
| [
"tamayo_144@hotmail.com"
] | tamayo_144@hotmail.com |
2fe4cec6defc2e66ddc4db17511c536f84514dd1 | ee6acbd5fcd0fcd16230e96a4a539de41a02c97e | /operators/special-resource-operator/python/pulumi_pulumi_kubernetes_crds_operators_special_resource_operator/sro/v1alpha1/__init__.py | 7e6f12156a258138619de35e038acddf3d969e0c | [
"Apache-2.0"
] | permissive | isabella232/pulumi-kubernetes-crds | 777e78137aaf6525a44b61a02dccf91bf0d87a14 | 372c4c0182f6b899af82d6edaad521aa14f22150 | refs/heads/master | 2023-03-15T04:29:16.039753 | 2020-12-30T19:35:54 | 2020-12-30T19:35:54 | null | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 275 | py | # coding=utf-8
# *** WARNING: this file was generated by crd2pulumi. ***
# *** Do not edit by hand unless you're certain you know what you are doing! ***
# Export this package's modules as members:
from .SpecialResource import *
from ._inputs import *
from . import outputs
| [
"albertzhong0@gmail.com"
] | albertzhong0@gmail.com |
1565de3abac24dee338faefa2cd4b094f35f2ecd | 7963f09b4002249e73496c6cbf271fd6921b3d22 | /emulator_6502/instructions/sbc.py | 31e549a6575d30d766fd1bf37990f233dd92938b | [] | no_license | thales-angelino/py6502emulator | 6df908fc02f29b41fad550c8b773723a7b63c414 | 1cea28489d51d77d2dec731ab98a6fe8a515a2a8 | refs/heads/master | 2023-03-19T14:46:17.393466 | 2021-03-08T04:10:45 | 2021-03-08T04:10:45 | 345,754,473 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 3,553 | py | SBC_IMMEDIATE_OPCODE = 0xe9
SBC_ZEROPAGE_OPCODE = 0xe5
SBC_ZEROPAGEX_OPCODE = 0xf5
SBC_ABSOLUTE_OPCODE = 0xed
SBC_ABSOLUTEX_OPCODE = 0xfd
SBC_ABSOLUTEY_OPCODE = 0xf9
SBC_INDIRECTX_OPCODE = 0xe1
SBC_INDIRECTY_OPCODE = 0xf1
class SBCImmediate(object):
def __init__(self):
super(SBCImmediate, self).__init__()
def run(self, cpu):
byte_r = cpu.immediate()
print("SBC memory byte read: %s" % hex(byte_r))
print("SBC register A read: %s" % hex(cpu.a))
print("SBC processor status Carry read: %s" % hex(cpu.processor_status['carry']))
cpu.sbc(byte_r)
class SBCZeroPage(object):
"""SBC Zero Page instruction"""
def __init__(self):
super(SBCZeroPage, self).__init__()
def run(self, cpu):
byte_r = cpu.zero_page()
print("SBC zero page byte read: %s" % hex(byte_r))
print("SBC register A read: %s" % hex(cpu.a))
print("SBC processor status Carry read: %s" % hex(cpu.processor_status['carry']))
cpu.sbc(byte_r)
class SBCZeroPageX(object):
"""SBC Zero Page X instruction"""
def __init__(self):
super(SBCZeroPageX, self).__init__()
def run(self, cpu):
byte_r = cpu.zero_page_x()
print("SBC zero page X byte read: %s" % hex(byte_r))
print("SBC register A read: %s" % hex(cpu.a))
print("SBC processor status carry read: %s" % hex(cpu.processor_status['carry']))
cpu.sbc(byte_r)
class SBCAbsolute(object):
"""SBC absolute instruction"""
def __init__(self):
super(SBCAbsolute, self).__init__()
def run(self, cpu):
byte_r = cpu.absolute()
print("SBC absolute byte read: %s" % hex(byte_r))
print("SBC register A read: %s" % hex(cpu.a))
print("SBC processor status carry read: %s" % hex(cpu.processor_status['carry']))
cpu.sbc(byte_r)
class SBCAbsoluteX(object):
"""SBC absolute X instruction"""
def __init__(self):
super(SBCAbsoluteX, self).__init__()
def run(self, cpu):
byte_r = cpu.absolute_x()
print("SBC absolute x byte read: %s" % hex(byte_r))
print("SBC register A read: %s" % hex(cpu.a))
print("SBC processor status carry read: %s" % hex(cpu.processor_status['carry']))
cpu.sbc(byte_r)
class SBCAbsoluteY(object):
"""SBC absolute Y instruction"""
def __init__(self):
super(SBCAbsoluteY, self).__init__()
def run(self, cpu):
byte_r = cpu.absolute_y()
print("SBC absolute Y byte read: %s" % hex(byte_r))
print("SBC register A read: %s" % hex(cpu.a))
print("SBC processor status carry read: %s" % hex(cpu.processor_status['carry']))
cpu.sbc(byte_r)
class SBCIndirectX(object):
"""SBC indirect X instruction"""
def __init__(self):
super(SBCIndirectX, self).__init__()
def run(self, cpu):
byte_r = cpu.indirect_x()
print("SBC indirect X byte read: %s" % hex(byte_r))
print("SBC register A read: %s" % hex(cpu.a))
print("SBC processor status carry read: %s" % hex(cpu.processor_status['carry']))
cpu.sbc(byte_r)
class SBCIndirectY(object):
"""SBC Indirect Y instruction"""
def __init__(self):
super(SBCIndirectY, self).__init__()
def run(self, cpu):
byte_r = cpu.indirect_y()
print("SBC indirect Y byte read: %s" % hex(byte_r))
print("SBC register A read: %s" % hex(cpu.a))
print("SBC processor status Carry read: %s" % hex(cpu.processor_status['carry']))
cpu.sbc(byte_r) | [
"thales.angelino@gmail.com"
] | thales.angelino@gmail.com |
b92efc48c054ab584bde5eced15bde64256a245a | 3951ae64b6b7ada4f035a3dd835b23cf7e624ee0 | /advert/models.py | b7ee5f7904cd6c568b07570302a340d84dec2ef5 | [] | no_license | ichinda/Tenders | bc4123be28929f827d773c267d2fac2d93662ff5 | f13111b7aeccf90bf5e4104966f76936e83ef41e | refs/heads/master | 2022-03-06T17:26:31.351951 | 2019-09-20T03:45:48 | 2019-09-20T03:45:48 | null | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 979 | py | from django.db import models
from django.utils import timezone
from django.contrib.auth.models import User
from django.urls import reverse
from phonenumber_field.modelfields import PhoneNumberField
# Create your models here.
class Advert(models.Model):
image = models.ImageField(default= 'default_business.jpg', upload_to = 'business_pics')
name = models.CharField(max_length=100)
address = models.CharField(max_length=100)
email = models.EmailField()
products = models.ManyToManyField('Item')
hours = models.CharField(max_length=100, default= '10:00AM – 08:00PM (Mon-Sat)')
details = models.TextField()
date_posted = models.DateTimeField(auto_now_add= True)
author = models.ForeignKey(User, on_delete = models.CASCADE)
def __str__(self):
return self.name
def get_absolute_url(self):
return reverse('advert-detail', kwargs= {'pk': self.pk})
class Item(models.Model):
item = models.CharField(max_length=30)
def __str__(self):
return self.item | [
"agrawal.arpit14@gmail.com"
] | agrawal.arpit14@gmail.com |
a97d0b7b5c266a837d5caf3fefb00339c7d845dc | 8fcae139173f216eba1eaa01fd055e647d13fd4e | /.history/scraper_20191220144406.py | d08dde54a93939a73cdb07eb8e08d72519375f5e | [] | no_license | EnriqueGalindo/backend-web-scraper | 68fdea5430a0ffb69cc7fb0e0d9bcce525147e53 | 895d032f4528d88d68719838a45dae4078ebcc82 | refs/heads/master | 2020-11-27T14:02:59.989697 | 2019-12-21T19:47:34 | 2019-12-21T19:47:34 | 229,475,085 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 1,693 | py | #!/usr/bin/env python
# -*- coding: utf-8 -*-
"""
Module docstring: One line description of what your program does.
There should be a blank line in between description above, and this
more detailed description. In this section you should put any caveats,
environment variable expectations, gotchas, and other notes about running
the program. Author tag (below) helps instructors keep track of who
wrote what, when grading.
"""
__author__ = "Enrique Galindo"
# Imports go at the top of your file, after the module docstring.
# One module per import line. These are for example only.
import sys
import requests
import re
def main(args):
"""Main function is declared as standalone, for testability"""
url = args[0]
response = requests.get(url)
response.raise_for_status()
url_list = re.findall(r'http[s]?://(?:[a-zA-Z]|[0-9]|[$-_@.&+]|[!*\(\),]|(?:%[0-9a-fA-F][0-9a-fA-F]))+', response.text)
01-\x08\x0regex_email = r'''(?:[a-z0-9!#$%&‘*+/=?^_`{|}~-]+(?:\.[a-z0-9!#$%&‘*+/=?^_`{|}~-]+)*|“(?:[\xb\x0c\x0e-\x1f\x21\x23-\x5b\x5d-\x7f]|\\[\x01-\x09\x0b\x0c\x0e-\x7f])*“)@(?:(?:[a-z0-9](?:[a-z0-9-]*[a-z0-9])?\.)+[a-z0-9](?:[a-z0-9-]*[a-z0-9])?|\[(?:(?:25[0-5]|2[0-4][0-9]|[01]?[0-9][0-9]?)\.){3}(?:25[0-5]|2[0-4][0-9]|[01]?[0-9][0-9]?|[a-z0-9-]*[a-z0-9]:(?:[\x01-\x08\x0b\x0c\x0e-\x1f\x21-\x5a\x53-\x7f]|\\[\x01-\x09\x0b\x0c\x0e-\x7f])+)\])'''
regex_phone = r'''(1?\W*([2-9][0-8][0-9])\W*([2-9][0-9]{2})\W*([0-9]{4})(\se?x?t?(\d*))?)'''
email_list = set(re.findall(regex_email, response.text))
phone_list = set(re.findall(re_phone, response.text))
print(email_list)
if __name__ == '__main__':
"""Docstring goes here"""
main(sys.argv[1:]) | [
"egalindo@protonmail.com"
] | egalindo@protonmail.com |
07009bf78c01bc832ea59222c306f561654e552d | ba2efb956f472e508cac8a41921d67083136e1c8 | /test.py | d54d89821dc811294a6f2b518dd498f232324426 | [] | no_license | AHBruns/Syrio | 1599c9f97b8022dd63c297e7ea6a4e88abd5cf86 | 881512ee5e2ee52ad70de5326f80c9e100736adb | refs/heads/master | 2020-05-09T15:55:12.831283 | 2019-04-14T03:45:44 | 2019-04-14T03:45:44 | 181,250,041 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 119 | py | import sqlite3
def t(s, c):
try:
c.execute(s)
except sqlite3.OperationalError as ex:
print(ex) | [
"alex.h.bruns@gmail.com"
] | alex.h.bruns@gmail.com |
b248f7b6e4a7f92757f0a8c13236f489a28b112f | ace30d0a4b1452171123c46eb0f917e106a70225 | /filesystems/vnx_rootfs_lxc_ubuntu64-16.04-v025-openstack-compute/rootfs/usr/lib/python2.7/dist-packages/oslo_privsep/tests/test_comm.py | 72f7aefceab898b5a143045d7aa771a0fc759a9f | [
"Python-2.0"
] | permissive | juancarlosdiaztorres/Ansible-OpenStack | e98aa8c1c59b0c0040c05df292964520dd796f71 | c01951b33e278de9e769c2d0609c0be61d2cb26b | refs/heads/master | 2022-11-21T18:08:21.948330 | 2018-10-15T11:39:20 | 2018-10-15T11:39:20 | 152,568,204 | 0 | 3 | null | 2022-11-19T17:38:49 | 2018-10-11T09:45:48 | Python | UTF-8 | Python | false | false | 2,698 | py | # Copyright 2015 Rackspace Inc.
#
# Licensed under the Apache License, Version 2.0 (the "License"); you may
# not use this file except in compliance with the License. You may obtain
# a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
# License for the specific language governing permissions and limitations
# under the License.
import six
from oslotest import base
from oslo_privsep import comm
class BufSock(object):
def __init__(self):
self.readpos = 0
self.buf = six.BytesIO()
def recv(self, bufsize):
if self.buf.closed:
return b''
self.buf.seek(self.readpos, 0)
data = self.buf.read(bufsize)
self.readpos += len(data)
return data
def sendall(self, data):
self.buf.seek(0, 2)
self.buf.write(data)
def shutdown(self, _flag):
self.buf.close()
class TestSerialization(base.BaseTestCase):
def setUp(self):
super(TestSerialization, self).setUp()
sock = BufSock()
self.input = comm.Serializer(sock)
self.output = iter(comm.Deserializer(sock))
def send(self, data):
self.input.send(data)
return next(self.output)
def assertSendable(self, value):
self.assertEqual(value, self.send(value))
def test_none(self):
self.assertSendable(None)
def test_bool(self):
self.assertSendable(True)
self.assertSendable(False)
def test_int(self):
self.assertSendable(42)
self.assertSendable(-84)
def test_bytes(self):
data = b'\x00\x01\x02\xfd\xfe\xff'
self.assertSendable(data)
def test_unicode(self):
data = u'\u4e09\u9df9'
self.assertSendable(data)
def test_tuple(self):
self.assertSendable((1, 'foo'))
def test_list(self):
# NB! currently lists get converted to tuples by serialization.
self.assertEqual((1, 'foo'), self.send([1, 'foo']))
def test_dict(self):
self.assertSendable(
{
'a': 'b',
1: 2,
None: None,
(1, 2): (3, 4),
}
)
def test_badobj(self):
class UnknownClass(object):
pass
obj = UnknownClass()
self.assertRaises(TypeError, self.send, obj)
def test_eof(self):
self.input.close()
self.assertRaises(StopIteration, next, self.output)
| [
"jcdiaztorres96@gmail.com"
] | jcdiaztorres96@gmail.com |
4795a56db26abe40c2a74585633599bdd4dbe46f | 87658453efcdbed898c9dd7fae77cd6c1c9dae79 | /scripts/run_multiple_tail_at_scale.py | 860b49357df4b2486ab955ce0f908ebec6f4a722 | [
"Apache-2.0"
] | permissive | matthelb/gryff | ada0fed962988bbcc32daff040de1029eb0d17f2 | faea8bf494c472d66f98c3701cd88b7c472c374f | refs/heads/master | 2021-02-05T23:52:38.728832 | 2020-04-01T20:01:30 | 2020-04-01T20:01:30 | 243,851,022 | 8 | 3 | null | null | null | null | UTF-8 | Python | false | false | 281 | py | import utils
import sys
from utils.experiment_util import *
def main():
if len(sys.argv) != 2:
sys.stderr.write('Usage: python3 %s <config_file>\n' % sys.argv[0])
sys.exit(1)
run_multiple_tail_at_scale(sys.argv[1])
if __name__ == "__main__":
main()
| [
"mlb452@cornell.edu"
] | mlb452@cornell.edu |
992131d644d1a04bd07ae41cf367774e1e0adef0 | fb7cb229a8f68f9ba3cc23ce51238008841516e8 | /Sensorslab2/Task2/first_pkg/first_pkg/publish_gps_messge.py | b99bf288a6eb3c6f23f35d8631db0fca8795ff89 | [] | no_license | RozanMagdy/ITI-Labs | 24852442c8cae3f9d0fe44e55e5995853f18a9b5 | 3e3a4b85a415492c6eb539c79be128504fefaf96 | refs/heads/master | 2023-06-04T18:07:58.256689 | 2021-06-17T11:43:30 | 2021-06-17T11:43:30 | 359,421,308 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 2,978 | py | #!/usr/bin/env python3
import rclpy
import csv
from rclpy.node import Node
from sensor_msgs.msg import NavSatFix
#TODO Import needed messages
class my_node (Node):
def __init__(self):
super().__init__("Node_name")
self.csv_file_path = "GGA_GST.csv"
self.lines = []
with open(self.csv_file_path, newline='\n') as csvfile:
self.readCSV = csv.reader(csvfile, delimiter = ',')
for row in self.readCSV:
self.lines.append(row)
self.count = 1 #Skip header
self.create_timer(5,self.timer_call)
self.obj_pub=self.create_publisher(NavSatFix,"fix",10)
def timer_call(self):
row = self.lines[self.count]
self.count +=1
if (self.count >= len(self.lines)): # repeat csv file continously
self.count = 0
#TODO get The following values from csv
latitude_value = row [2]
latitude_direction = row [3]
longitude_value = row [4]
longitude_direction = row [5]
altitude_value = row [9]
# The following functions convert the string data in degrees/minutes to float data in degrees as ROS message requires.
latitude = self.convert_latitude(latitude_value, latitude_direction)
longitude = self.convert_longitude(longitude_value, longitude_direction)
altitude = self.safe_float(altitude_value)
hdop = float(row[8])
lat_std_dev = float(row[21])
lon_std_dev = float(row[22])
alt_std_dev = float(row[23])
#TODO Fill the gps message and publish
current_fix = NavSatFix()
#current_fix.header.stamp = current_time
#current_fix.header.frame_id = frame_id
current_fix.latitude = latitude
current_fix.longitude = longitude
current_fix.altitude = altitude
current_fix.position_covariance[0] = (hdop * lon_std_dev) ** 2
current_fix.position_covariance[4] = (hdop * lat_std_dev) ** 2
current_fix.position_covariance[8] = (2 * hdop * alt_std_dev) ** 2
self._logger.info(str(current_fix))
self.obj_pub.publish(current_fix)
def convert_latitude(self, field_lat, lat_direction):
latitude = self.safe_float(field_lat[0:2]) + self.safe_float(field_lat[2:]) / 60.0
if lat_direction == 'S':
latitude = -latitude
return latitude
def convert_longitude(self, field_long, long_direction):
longitude = self.safe_float(field_long[0:2]) + self.safe_float(field_long[2:]) / 60.0
if long_direction == 'W':
longitude = -longitude
return longitude
def safe_float(self, field):
try:
return float(field)
except ValueError:
return float('NaN')
def main (args=None):
rclpy.init(args=args)
node=my_node()
rclpy.spin(node)
rclpy.shutdown()
if __name__=="__main__":
main()
| [
"rozanabdelmawla@gmail.com"
] | rozanabdelmawla@gmail.com |
5e1e1a8a01e9a4132bd94ac4745a7070a47d4718 | be0f3dfbaa2fa3d8bbe59229aef3212d032e7dd1 | /Gauss_v45r8/Gen/DecFiles/options/22114002.py | 7953c9feeb8c8d3d7f7c9d855b04a94363e3a510 | [] | no_license | Sally27/backup_cmtuser_full | 34782102ed23c6335c48650a6eaa901137355d00 | 8924bebb935b96d438ce85b384cfc132d9af90f6 | refs/heads/master | 2020-05-21T09:27:04.370765 | 2018-12-12T14:41:07 | 2018-12-12T14:41:07 | 185,989,173 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 752 | py | # file /home/hep/ss4314/cmtuser/Gauss_v45r8/Gen/DecFiles/options/22114002.py generated: Fri, 27 Mar 2015 15:48:15
#
# Event Type: 22114002
#
# ASCII decay Descriptor: [D0 -> pi+ pi- mu+ mu-]cc
#
from Configurables import Generation
Generation().EventType = 22114002
Generation().SampleGenerationTool = "SignalPlain"
from Configurables import SignalPlain
Generation().addTool( SignalPlain )
Generation().SignalPlain.ProductionTool = "PythiaProduction"
from Configurables import ToolSvc
from Configurables import EvtGenDecay
ToolSvc().addTool( EvtGenDecay )
ToolSvc().EvtGenDecay.UserDecayFile = "$DECFILESROOT/dkfiles/D0_pipimumu=DecProdCut.dec"
Generation().SignalPlain.CutTool = "DaughtersInLHCb"
Generation().SignalPlain.SignalPIDList = [ 421,-421 ]
| [
"slavomirastefkova@b2pcx39016.desy.de"
] | slavomirastefkova@b2pcx39016.desy.de |
cf3e00e18c62ceff4ee2e392754a7ae771574aa9 | 5b0a71b1b9ff85ea107406f804936035af2352a3 | /question_ask/question_temp.py | bd7a0b8148e41585de7496596c82fa2e41183678 | [] | no_license | FantasticCode2019/kg--demo-Computer-network-course | 58d7c8763d3dc530dc0d833dad734db78b6ffc9b | e8024153c4473e67cd95695bf55071faa1e6d221 | refs/heads/master | 2023-04-12T05:09:59.828476 | 2021-05-10T01:19:46 | 2021-05-10T01:19:46 | null | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 22,339 | py | # encoding=utf-8
"""
@author: xuzelin
@file: question_temp.py
@time: 2020/12/20
@desc:
设置问题模板,为每个模板设置对应的SPARQL语句。demo提供如下模板:
1. 某实体的兄弟关系有哪些
2. 某阶段之后是哪个阶段
3. 某实体包含了哪些实体
4. 与某实体内涵相同的是
5. 与某实体内涵相反的是
6. 某实体继承自哪个实体
7. 某实体参考自哪里/那本教程
8. 与某实体可以相互变换的实体有哪些
9. 与某实体有因果的实体有哪些?
10.某实体的某属性是什么
11.某实体是正确的吗?
"""
from refo import finditer, Predicate, Star, Any, Disjunction
import re
# TODO SPARQL前缀和模板
SPARQL_PREXIX = u"""
PREFIX : <http://www.semanticweb.org/yan/ontologies/2020/9/untitled-ontology-6#>
PREFIX rdf: <http://www.w3.org/1999/02/22-rdf-syntax-ns#>
PREFIX rdfs: <http://www.w3.org/2000/01/rdf-schema#>
PREFIX xsd: <http://www.w3.org/2001/XMLSchema#>
"""
SPARQL_SELECT_TEM = u"{prefix}\n" + \
u"SELECT DISTINCT {select} WHERE {{\n" + \
u"{expression}\n" + \
u"}}\n"
SPARQL_COUNT_TEM = u"{prefix}\n" + \
u"SELECT COUNT({select}) WHERE {{\n" + \
u"{expression}\n" + \
u"}}\n"
SPARQL_ASK_TEM = u"{prefix}\n" + \
u"ASK {{\n" + \
u"{expression}\n" + \
u"}}\n"
class W(Predicate):
def __init__(self, token=".*", pos=".*"):
self.token = re.compile(token + "$")
self.pos = re.compile(pos + "$")
super(W, self).__init__(self.match)
def match(self, word):
m1 = self.token.match(word.token)
m2 = self.pos.match(word.pos)
return m1 and m2
class Rule(object):
def __init__(self, condition_num, condition=None, action=None):
assert condition and action
self.condition = condition
self.action = action
self.condition_num = condition_num
def apply(self, sentence):
matches = []
for m in finditer(self.condition, sentence):
i, j = m.span()
matches.extend(sentence[i:j])
return self.action(matches), self.condition_num
class KeywordRule(object):
def __init__(self, condition=None, action=None):
assert condition and action
self.condition = condition
self.action = action
def apply(self, sentence):
matches = []
for m in finditer(self.condition, sentence):
i, j = m.span()
matches.extend(sentence[i:j])
if len(matches) == 0:
return None
else:
return self.action()
class QuestionSet:
def __init__(self):
pass
@staticmethod
def has_brother_question(word_objects):
"""
某实体的兄弟关系有哪些
:param word_objects:
:return:
"""
select = u"?x"
sparql = None
for w in word_objects:
if w.pos == pos_disanzhang:
e = u"?y :名称 '{disanzhang}'." \
u"?y :兄弟关系 ?z." \
u"?z :名称 ?x.".format(disanzhang=w.token.encode('utf-8').decode('utf-8'))
sparql = SPARQL_SELECT_TEM.format(prefix=SPARQL_PREXIX,
select=select,
expression=e)
break
return sparql
@staticmethod
def has_Successive_question(word_objects):
"""
某阶段之后是哪个阶段
:param word_objects:
:return:
"""
select = u"?x"
sparql = None
for w in word_objects:
if w.pos == pos_disanzhang:
e = u"?y :名称 '{disanzhang}'." \
u"?y :前后继关系 ?z." \
u"?z :名称 ?x.".format(disanzhang=w.token.encode('utf-8').decode('utf-8'))
sparql = SPARQL_SELECT_TEM.format(prefix=SPARQL_PREXIX,
select=select,
expression=e)
break
return sparql
@staticmethod
def has_contain_question(word_objects):
"""
某实体包含了哪些实体
:param word_objects:
:return:
"""
select = u"?x"
sparql = None
for w in word_objects:
if w.pos == pos_disanzhang:
e = u"?y :名称 '{disanzhang}'." \
u"?y :包含关系 ?z." \
u"?z :名称 ?x.".format(disanzhang=w.token.encode('utf-8').decode('utf-8'))
sparql = SPARQL_SELECT_TEM.format(prefix=SPARQL_PREXIX,
select=select,
expression=e)
break
return sparql
@staticmethod
def has_same_question(word_objects):
"""
与某实体内涵相同的是
:param word_objects:
:return:
"""
select = u"?x"
sparql = None
for w in word_objects:
if w.pos == pos_disanzhang:
e = u"?y :名称 '{disanzhang}'." \
u"?y :同一关系 ?z." \
u"?z :名称 ?x.".format(disanzhang=w.token.encode('utf-8').decode('utf-8'))
sparql = SPARQL_SELECT_TEM.format(prefix=SPARQL_PREXIX,
select=select,
expression=e)
break
return sparql
@staticmethod
def has_opposition_question(word_objects):
"""
与某实体内涵相反的是
:param word_objects:
:return:
"""
select = u"?x"
sparql = None
for w in word_objects:
if w.pos == pos_disanzhang:
e = u"?y :名称 '{disanzhang}'." \
u"?y :对立关系 ?z." \
u"?z :名称 ?x.".format(disanzhang=w.token.encode('utf-8').decode('utf-8'))
sparql = SPARQL_SELECT_TEM.format(prefix=SPARQL_PREXIX,
select=select,
expression=e)
break
return sparql
@staticmethod
def has_inherit_question(word_objects):
"""
某实体继承自哪个实体
:param word_objects:
:return:
"""
select = u"?x"
sparql = None
for w in word_objects:
if w.pos == pos_disanzhang:
e = u"?y :名称 '{disanzhang}'." \
u"?y :前后继关系 ?z." \
u"?z :名称 ?x.".format(disanzhang=w.token.encode('utf-8').decode('utf-8'))
sparql = SPARQL_SELECT_TEM.format(prefix=SPARQL_PREXIX,
select=select,
expression=e)
break
return sparql
@staticmethod
def has_reference_question(word_objects):
"""
某实体参考自哪里
:param word_objects:
:return:
"""
select = u"?x"
sparql = None
for w in word_objects:
if w.pos == pos_disanzhang:
e = u"?y :名称 '{disanzhang}'." \
u"?y :参考关系 ?z." \
u"?z :名称 ?x.".format(disanzhang=w.token.encode('utf-8').decode('utf-8'))
sparql = SPARQL_SELECT_TEM.format(prefix=SPARQL_PREXIX,
select=select,
expression=e)
break
return sparql
@staticmethod
def has_vary_question(word_objects):
"""
与某实体可以相互变换的实体有哪些
:param word_objects:
:return:
"""
select = u"?x"
sparql = None
for w in word_objects:
if w.pos == pos_disanzhang:
e = u"?y :名称 '{disanzhang}'." \
u"?y :变换关系 ?z." \
u"?z :名称 ?x.".format(disanzhang=w.token.encode('utf-8').decode('utf-8'))
sparql = SPARQL_SELECT_TEM.format(prefix=SPARQL_PREXIX,
select=select,
expression=e)
break
return sparql
@staticmethod
def has_karma_question(word_objects):
"""
与某实体有因果的实体有哪些?
:param word_objects:
:return:
"""
select = u"?x"
sparql = None
for w in word_objects:
if w.pos == pos_disanzhang:
e = u"?y :名称 '{disanzhang}'." \
u"?y :因果关系 ?z." \
u"?z :名称 ?x.".format(disanzhang=w.token.encode('utf-8').decode('utf-8'))
sparql = SPARQL_SELECT_TEM.format(prefix=SPARQL_PREXIX,
select=select,
expression=e)
break
return sparql
@staticmethod
def has_basic_disanzhang_info_question(word_objects):
"""
某实体的某属性是什么
:param word_objects:
:return:
"""
keyword = None
for r in disanzhang_basic_keyword_rules:
keyword = r.apply(word_objects)
if keyword is not None:
break
select = u"?x"
sparql = None
for w in word_objects:
if w.pos == pos_disanzhang:
e = u"?s :名称 '{disanzhang}'." \
u"?s {keyword} ?x.".format(disanzhang=w.token.encode('utf-8').decode('utf-8'), keyword=keyword)
sparql = SPARQL_SELECT_TEM.format(prefix=SPARQL_PREXIX, select=select, expression=e)
break
return sparql
@staticmethod
def is_ASKattribute_question(word_objects):
"""
某实体是正确的吗?
:param word_objects:
:return:
"""
sparql = None
for w in word_objects:
if w.pos == pos_disanzhang:
e = u"?s :名称 '{disanzhang}'." \
u"?s rdf:type :正确.".format(disanzhang=w.token.encode('utf-8').decode('utf-8'))
sparql = SPARQL_ASK_TEM.format(prefix=SPARQL_PREXIX, expression=e)
break
return sparql
class PropertyValueSet:
def __init__(self):
pass
@staticmethod
def return_dingyi_value():
return u':定义'
@staticmethod
def return_jieshao_value():
return u':介绍'
@staticmethod
def return_youdian_value():
return u':优点'
@staticmethod
def return_quedian_value():
return u':缺点'
@staticmethod
def return_zuoyong_value():
return u':作用'
@staticmethod
def return_juyou_value():
return u':具有'
@staticmethod
def return_neirong_value():
return u':内容'
@staticmethod
def return_biecheng_value():
return u':别称'
@staticmethod
def return_gongneng_value():
return u':功能'
@staticmethod
def return_baokuo_value():
return u':包括'
@staticmethod
def return_hanyi_value():
return u':含义'
@staticmethod
def return_shuyu_value():
return u':属于'
@staticmethod
def return_shuxing_value():
return u':属性'
@staticmethod
def return_xingzhi_value():
return u':性质'
@staticmethod
def return_yiyi_value():
return u':意义'
@staticmethod
def return_shijian_value():
return u':时间'
@staticmethod
def return_tezheng_value():
return u':特征'
@staticmethod
def return_tedian_value():
return u':特点'
@staticmethod
def return_zhuangtai_value():
return u':状态'
@staticmethod
def return_jiancheng_value():
return u':简称'
@staticmethod
def return_leixing_value():
return u':类型'
@staticmethod
def return_jibie_value():
return u':级别'
@staticmethod
def return_zucheng_value():
return u':组成'
@staticmethod
def return_jiegou_value():
return u':结构'
@staticmethod
def return_zhize_value():
return u':职责'
@staticmethod
def return_yingwen_value():
return u':英文'
@staticmethod
def return_biaodashi_value():
return u':表达式'
@staticmethod
def return_yaosu_value():
return u':要素'
@staticmethod
def return_guize_value():
return u':规则'
@staticmethod
def return_xiangjie_value():
return u':详解'
@staticmethod
def return_shiyi_value():
return u':释义'
@staticmethod
def return_lingyu_value():
return u':领域'
@staticmethod
def return_gainian_value():
return u':概念'
# TODO 定义关键词
pos_disanzhang = "nz"
disanzhang_entity = (W(pos=pos_disanzhang))
dingyi = W("定义")
jieshao = W("介绍")
youdian = W("优点")
quedian = W("缺点")
zuoyong = W("作用")
juyou = W("具有")
neirong = W("内容")
biecheng = W("别称")
gongneng = W("功能")
baokuo = W("包括")
hanyi = W("含义")
shuyu = W("属于")
shuxing = W("属性")
xingzhi = W("性质")
yiyi = W("意义")
shijian = W("时间")
tezheng = W("特征")
tedian = W("特点")
zhuangtai = W("状态")
jiancheng = W("简称")
leixing = W("类型")
jibie = W("级别")
zucheng = W("组成")
jiegou = W("结构")
zhize = W("职责")
yingwen = W("英文")
biaodashi = W("表达式")
yaosu = W("要素")
guize = W("规则")
xiangjie = W("详解")
shiyi = W("释义")
lingyu = W("领域")
gainian = W("概念")
attribute = (dingyi | jieshao | youdian | quedian | zuoyong | juyou
| neirong | biecheng | gongneng | baokuo | hanyi |
shuyu | shuxing | xingzhi | yiyi | shijian | tezheng |
tedian | zhuangtai | jiancheng | leixing | jibie |
zucheng | jiegou | zhize | yingwen | biaodashi |
yaosu | guize | xiangjie | shiyi | lingyu | gainian)
brother = W("兄弟")
Successive = W("阶段")
contain = W("包含")
connotation = W("内涵") | W("意思")
same = (W("相同") | W("一致") | W("一样") )
opposition = (W("相反") | W("对立") )
inherit = W("继承")
reference = W("参考")
vary = W("变换")
karma = W("因果")
zhengque = W("正确")
# TODO 问题模板/匹配规则
"""
1. 某实体的兄弟关系有哪些
2. 某阶段之后是哪个阶段
3. 某实体包含了哪些实体
4. 与某实体内涵相同的是
5. 与某实体内涵相反的是
6. 某实体继承自哪个实体
7. 某实体参考自哪里/那本教程
8. 与某实体可以相互变换的实体有哪些
9. 与某实体有因果的实体有哪些?
10.某实体的某属性是什么
11.某实体是正确的吗?
"""
rules = [
Rule(condition_num=2, condition=disanzhang_entity + Star(Any(), greedy=False) + brother + Star(Any(), greedy=False), action=QuestionSet.has_brother_question),
Rule(condition_num=2, condition=disanzhang_entity + Star(Any(), greedy=False) + Successive + Star(Any(), greedy=False), action=QuestionSet.has_Successive_question),
Rule(condition_num=2, condition=disanzhang_entity + Star(Any(), greedy=False) + contain + Star(Any(), greedy=False), action=QuestionSet.has_contain_question),
Rule(condition_num=2, condition=disanzhang_entity + Star(Any(), greedy=False) + connotation + Star(Any(), greedy=False) + same + Star(Any(), greedy=False), action=QuestionSet.has_same_question),
Rule(condition_num=2, condition=disanzhang_entity + Star(Any(), greedy=False) + connotation + Star(Any(), greedy=False) + opposition + Star(Any(), greedy=False), action=QuestionSet.has_opposition_question),
Rule(condition_num=2, condition=disanzhang_entity + Star(Any(), greedy=False) + inherit + Star(Any(), greedy=False), action=QuestionSet.has_inherit_question),
Rule(condition_num=2, condition=disanzhang_entity + Star(Any(), greedy=False) + reference + Star(Any(), greedy=False),action=QuestionSet.has_reference_question),
Rule(condition_num=2, condition=disanzhang_entity + Star(Any(), greedy=False) + vary + Star(Any(), greedy=False), action=QuestionSet.has_vary_question),
Rule(condition_num=2, condition=disanzhang_entity + Star(Any(), greedy=False) + karma + Star(Any(), greedy=False), action=QuestionSet.has_karma_question),
Rule(condition_num=2, condition=disanzhang_entity + Star(Any(), greedy=False) + attribute + Star(Any(), greedy=False),action=QuestionSet.has_basic_disanzhang_info_question),
Rule(condition_num=3, condition=disanzhang_entity + Star(Any(), greedy=False) + zhengque + Star(Any(), greedy=False),action=QuestionSet.is_ASKattribute_question)
]
# TODO 具体的属性词匹配规则
disanzhang_basic_keyword_rules = [
KeywordRule(condition=disanzhang_entity + Star(Any(), greedy=False) + dingyi + Star(Any(), greedy=False),action=PropertyValueSet.return_dingyi_value),
KeywordRule(condition=disanzhang_entity + Star(Any(), greedy=False) + jieshao + Star(Any(), greedy=False),action=PropertyValueSet.return_jieshao_value),
KeywordRule(condition=disanzhang_entity + Star(Any(), greedy=False) + youdian + Star(Any(), greedy=False),action=PropertyValueSet.return_youdian_value),
KeywordRule(condition=disanzhang_entity + Star(Any(), greedy=False) + quedian + Star(Any(), greedy=False),action=PropertyValueSet.return_quedian_value),
KeywordRule(condition=disanzhang_entity + Star(Any(), greedy=False) + zuoyong + Star(Any(), greedy=False),action=PropertyValueSet.return_zuoyong_value),
KeywordRule(condition=disanzhang_entity + Star(Any(), greedy=False) + juyou + Star(Any(), greedy=False),action=PropertyValueSet.return_juyou_value),
KeywordRule(condition=disanzhang_entity + Star(Any(), greedy=False) + neirong + Star(Any(), greedy=False),action=PropertyValueSet.return_neirong_value),
KeywordRule(condition=disanzhang_entity + Star(Any(), greedy=False) + biecheng + Star(Any(), greedy=False),action=PropertyValueSet.return_biecheng_value),
KeywordRule(condition=disanzhang_entity + Star(Any(), greedy=False) + gongneng + Star(Any(), greedy=False),action=PropertyValueSet.return_gongneng_value),
KeywordRule(condition=disanzhang_entity + Star(Any(), greedy=False) + baokuo + Star(Any(), greedy=False),action=PropertyValueSet.return_baokuo_value),
KeywordRule(condition=disanzhang_entity + Star(Any(), greedy=False) + hanyi + Star(Any(), greedy=False),action=PropertyValueSet.return_hanyi_value),
KeywordRule(condition=disanzhang_entity + Star(Any(), greedy=False) + shuyu + Star(Any(), greedy=False),action=PropertyValueSet.return_shuyu_value),
KeywordRule(condition=disanzhang_entity + Star(Any(), greedy=False) + shuxing + Star(Any(), greedy=False),action=PropertyValueSet.return_shuxing_value),
KeywordRule(condition=disanzhang_entity + Star(Any(), greedy=False) + xingzhi + Star(Any(), greedy=False),action=PropertyValueSet.return_xingzhi_value),
KeywordRule(condition=disanzhang_entity + Star(Any(), greedy=False) + yiyi + Star(Any(), greedy=False),action=PropertyValueSet.return_yiyi_value),
KeywordRule(condition=disanzhang_entity + Star(Any(), greedy=False) + shijian + Star(Any(), greedy=False),action=PropertyValueSet.return_shijian_value),
KeywordRule(condition=disanzhang_entity + Star(Any(), greedy=False) + tezheng + Star(Any(), greedy=False),action=PropertyValueSet.return_tezheng_value),
KeywordRule(condition=disanzhang_entity + Star(Any(), greedy=False) + tedian + Star(Any(), greedy=False),action=PropertyValueSet.return_tedian_value),
KeywordRule(condition=disanzhang_entity + Star(Any(), greedy=False) + zhuangtai + Star(Any(), greedy=False),action=PropertyValueSet.return_zhuangtai_value),
KeywordRule(condition=disanzhang_entity + Star(Any(), greedy=False) + jiancheng + Star(Any(), greedy=False),action=PropertyValueSet.return_jiancheng_value),
KeywordRule(condition=disanzhang_entity + Star(Any(), greedy=False) + leixing + Star(Any(), greedy=False),action=PropertyValueSet.return_leixing_value),
KeywordRule(condition=disanzhang_entity + Star(Any(), greedy=False) + jibie + Star(Any(), greedy=False),action=PropertyValueSet.return_jibie_value),
KeywordRule(condition=disanzhang_entity + Star(Any(), greedy=False) + zucheng + Star(Any(), greedy=False),action=PropertyValueSet.return_zucheng_value),
KeywordRule(condition=disanzhang_entity + Star(Any(), greedy=False) + jiegou + Star(Any(), greedy=False),action=PropertyValueSet.return_jiegou_value),
KeywordRule(condition=disanzhang_entity + Star(Any(), greedy=False) + zhize + Star(Any(), greedy=False),action=PropertyValueSet.return_zhize_value),
KeywordRule(condition=disanzhang_entity + Star(Any(), greedy=False) + yingwen + Star(Any(), greedy=False),action=PropertyValueSet.return_yingwen_value),
KeywordRule(condition=disanzhang_entity + Star(Any(), greedy=False) + biaodashi + Star(Any(), greedy=False),action=PropertyValueSet.return_biaodashi_value),
KeywordRule(condition=disanzhang_entity + Star(Any(), greedy=False) + yaosu + Star(Any(), greedy=False),action=PropertyValueSet.return_yaosu_value),
KeywordRule(condition=disanzhang_entity + Star(Any(), greedy=False) + guize + Star(Any(), greedy=False),action=PropertyValueSet.return_guize_value),
KeywordRule(condition=disanzhang_entity + Star(Any(), greedy=False) + xiangjie + Star(Any(), greedy=False),action=PropertyValueSet.return_xiangjie_value),
KeywordRule(condition=disanzhang_entity + Star(Any(), greedy=False) + shiyi + Star(Any(), greedy=False),action=PropertyValueSet.return_shiyi_value),
KeywordRule(condition=disanzhang_entity + Star(Any(), greedy=False) + lingyu + Star(Any(), greedy=False),action=PropertyValueSet.return_lingyu_value),
KeywordRule(condition=disanzhang_entity + Star(Any(), greedy=False) + gainian + Star(Any(), greedy=False),action=PropertyValueSet.return_gainian_value),
]
| [
"1127485560@qq.com"
] | 1127485560@qq.com |
24c369e604c97d4f7e16c18d93c67302fca9be3b | 47789957e8dd6ba944bd4c04341cf8add431d64f | /shoppingcart/home/migrations/0005_address.py | 0b765715896903f72bb9986d4f645cf9f59811fb | [] | no_license | binnib/EshopApp | 8c93c518d09b5dbe9e3e9f357e9a68fa2d2d4585 | 4846b7f48e4224f2a0aee8942e9c758a9f890081 | refs/heads/main | 2023-03-28T17:04:48.063283 | 2021-04-12T05:52:37 | 2021-04-12T05:52:37 | null | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 1,121 | py | # Generated by Django 3.0.6 on 2021-03-31 07:22
from django.db import migrations, models
import django.db.models.deletion
class Migration(migrations.Migration):
dependencies = [
('home', '0004_delete_address'),
]
operations = [
migrations.CreateModel(
name='Address',
fields=[
('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')),
('address_details', models.CharField(max_length=250)),
('state', models.CharField(max_length=100)),
('city', models.CharField(max_length=100)),
('country', models.CharField(max_length=100)),
('pincode', models.CharField(max_length=10)),
('landmark', models.CharField(max_length=50)),
('resident_type', models.CharField(max_length=10)),
('is_active', models.BooleanField(default=False)),
('customer', models.ForeignKey(null=True, on_delete=django.db.models.deletion.CASCADE, to='home.Customer')),
],
),
]
| [
"maheshsharma171226@gmail.com"
] | maheshsharma171226@gmail.com |
af1029b1cb2a722afefefe70be3427a8519bf118 | 10f3e91eb6b7d155f0421f836a2fd9696bd4f184 | /getWeather.py | 466b96ed80f369d7e097565df13eafab22a31e8e | [] | no_license | Claire0223/Python_demo | b61fdcd9448bc7c57a4f13241f48a732406f4ec3 | 9c266503defeca8fae717797482303e126c66ebf | refs/heads/master | 2022-11-15T04:06:19.964521 | 2020-07-10T06:09:58 | 2020-07-10T06:09:58 | 276,874,224 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 1,807 | py | #coding=utf-8
# 微信聊天+天气预报机器人
# 和风天气https://dev.heweather.com/docs/legacy/api/s6
#
import requests
import json
def weather_forecast():
city=input('请输入想要查询的城市名称,如‘江门’:')
api='https://free-api.heweather.com/s6/weather/'
weather_type='forecast'
value={
'location':city,
'key':'63d7ffe16c3743e1af28b8ad4423e5af'
}
url=api+weather_type
weather_dict=requests.get(url,params=value).json()
return weather_dict
def get_data():
weather_dict=weather_forecast()
he_weather=weather_dict['HeWeather6']#['daily_forecast']#天气预报,list
cityname=he_weather[0]['basic']['location']
daily_forecast=he_weather[0]['basic']
for i in range(len(daily_forecast)):
date=daily_forecast[i]['date']
cond_txt_d=daily_forecast[i]['cond_txt_d']
cond_txt_n=daily_forecast[i]['cond_txt_n']
tmp_max=daily_forecast[i]['tmp_max']
tmp_min=daily_forecast[i]['tmp_min']
wind_dir=daily_forecast[i]['wind_dir']
weather_data=cityname+' '+date+' 白天天气:'+cond_txt_d+' 晚上天气:'+cond_txt_n+'\n最高温:'+ tmp_max +' 最低温:'+tmp_min+' 风向:'+wind_dir
print(weather_data)
return True
if __name__=="__main__":
# date=time.strftime('%Y-%m-%d',time.localtime())
# jinshanApi='http://open.iciba.com/dsapi?date='+date
# # print(jinshanApi)
# sentence=get_sentence(jinshanApi)
# sentenceDict=json.loads(sentence)
# content=sentenceDict['content']
# note=sentenceDict['note']
weather_forecast=get_data()
# print(type(weather_forecast))
print(weather_forecast)
| [
"1327686271@qq.com"
] | 1327686271@qq.com |
2018328ca867ccc87ff04f9b378c17cac512df31 | 747f759311d404af31c0f80029e88098193f6269 | /addons/sale_payment/sale_payment.py | 807c9c55bec7e7ec51aae95ef58e50ca79e28d7c | [] | no_license | sgeerish/sirr_production | 9b0d0f7804a928c0c582ddb4ccb7fcc084469a18 | 1081f3a5ff8864a31b2dcd89406fac076a908e78 | refs/heads/master | 2020-05-19T07:21:37.047958 | 2013-09-15T13:03:36 | 2013-09-15T13:03:36 | 9,648,444 | 0 | 1 | null | null | null | null | UTF-8 | Python | false | false | 66 | py | /home/openerp/production/extra-addons/sale_payment/sale_payment.py | [
"geerish@omerp.net"
] | geerish@omerp.net |
06e27836ec9ad1be7688bc3689db92957c6d74ec | 7f2aae79774d166173b2e003bf63c88369159d6f | /game_types/Board.py | fc76f18aafb07cf33d67913ed478c918b6a7d68d | [] | no_license | enisnazif/battleships | 6605df2c8581491b302daed26f0e697fcfb47cab | d99ecb60defb69996ac32cb3fe9661e263aec56a | refs/heads/master | 2021-06-11T13:59:25.088685 | 2019-05-19T10:53:12 | 2019-05-19T10:53:12 | 180,877,603 | 0 | 0 | null | 2021-05-06T19:33:56 | 2019-04-11T21:03:35 | Python | UTF-8 | Python | false | false | 5,196 | py | from typing import Tuple, List
from config import BOARD_SIZE
from exceptions import (
InvalidShotException,
InvalidShipPlacementException,
)
from game_types.Orientation import Orientation
from game_types.Point import Point
from game_types.Ship import Ship
from game_types.ShipType import ShipType
class Board:
def __init__(self, board_size=BOARD_SIZE):
assert board_size > 0
self.board_size = board_size
self._shot_locations = set()
self._all_ship_locations = set()
self._individual_ship_locations = dict() # A dict of sets - one for each ships
@property
def board(self):
return frozenset(
[
Point(x, y)
for x in range(self.board_size)
for y in range(self.board_size)
]
)
@property
def all_ship_locations(self):
return self._all_ship_locations
@property
def individual_ship_locations(self):
return self._individual_ship_locations
@property
def shot_locations(self):
return self._shot_locations
@all_ship_locations.setter
def all_ship_locations(self, value):
self._all_ship_locations = value
@individual_ship_locations.setter
def individual_ship_locations(self, value):
self._individual_ship_locations = value
@shot_locations.setter
def shot_locations(self, value):
self._shot_locations = value
def point_is_shot(self, point: Point):
"""
Checks to see if 'point' on the board has already been shot
:param point:
:return:
"""
return point in self.shot_locations
def is_board_lost(self):
"""
Returns true if the board is currently in a losing state for the owning player (i.e, all ships have been shot)
:return:
"""
return bool(self.all_ship_locations) and bool(
not self.all_ship_locations.difference(self.shot_locations)
)
def place_ship(self, ship: Ship, location: Point, orientation: Orientation) -> None:
"""
Places a ship at the given location / orientation
:param ship:
:param location:
:param orientation:
:return:
"""
ship_point_set = ship.get_points(location, orientation)
ship_type = ship.ship_type
if self.board.issuperset(
ship.get_points(location, orientation)
) and ship_point_set.isdisjoint(self.all_ship_locations):
self.all_ship_locations.update(ship_point_set)
self.individual_ship_locations[ship_type] = set(ship_point_set)
else:
raise InvalidShipPlacementException(f'Placement of {ship} at {location} in orientation {orientation.value} is invalid')
def shoot(self, point: Point) -> Tuple[bool, bool, ShipType]:
"""
Shoot the board location given by 'point'. Will raise ShotOffBoardException if 'point' is not on the board,
and PointAlreadyShotException if 'point'
has previously been shot
:param point:
:return:
"""
# Shot off board
if not self.point_in_board(point):
raise InvalidShotException(f'{point} is not on the board')
# Point has already been shot
elif self.point_is_shot(point):
raise InvalidShotException(f'{point} has already been shot')
else:
self.shot_locations.add(point)
is_hit = True if point in self.all_ship_locations else False
is_sunk = False
ship_sunk = None
if is_hit:
# find out which one of the ships was shot
for k, v in self.individual_ship_locations.items():
# if v was the ship that was shot
if point in v:
# remove the point from v
v.remove(point)
if len(v) == 0:
is_sunk = True
ship_sunk = k
return is_hit, is_sunk, ship_sunk
@staticmethod
def is_valid_ship_placement(placements: List[Tuple[Ship, Point, Orientation]]) -> bool:
"""
A static helper function that checks to see if ship placements are valid
:param placements:
:return:
"""
all_points = set()
for ship, point, orientation in placements:
all_points.update(ship.get_points(point, orientation))
# Check there are no overlapping placements
if not len(all_points) == sum([len(s[0]) for s in placements]):
return False
# Check all points are within the board
return all_points.issubset(set([Point(x, y) for x in range(BOARD_SIZE) for y in range(BOARD_SIZE)]))
@staticmethod
def point_in_board(point: Point):
"""
Checks to see if 'point' is within the board
:param point: Tuple
:return: bool
"""
return point in frozenset(
[
Point(x, y)
for x in range(BOARD_SIZE)
for y in range(BOARD_SIZE)
]
)
| [
"e.nazif@warwick.ac.uk"
] | e.nazif@warwick.ac.uk |
e1918394a57db35a46a6856e38ebedd667af34e4 | d21112887ed1ec675b7b519cc991fc47bfa11735 | /SaleML_PreDjango/Predicting/urls.py | fc440ad929f7717a684452088ecfe3d8b3a0a1bb | [] | no_license | SydNS/DjangoML-model | 8c9ab65075b896ff129a872b087cdcd9dfc87e83 | c15474b136d592e182e707f6a73269685c3e62ad | refs/heads/master | 2023-03-02T13:27:33.809869 | 2021-02-06T09:57:34 | 2021-02-06T09:57:34 | 336,550,706 | 1 | 0 | null | null | null | null | UTF-8 | Python | false | false | 196 | py | from django.conf.urls import url
from django.contrib import admin
from django.urls import path, include
urlpatterns = [
path('products', views.product_describe_view, name='product_add'),
]
| [
"root@localhost.localdomain"
] | root@localhost.localdomain |
5fdaeecaf64cdc22e25b4913a33d91d5c7b28577 | 714306f0587cbe874a8663d80255799e91e3bff8 | /scripts/scrape.py | 4fecb8389f2a88602b52ad694bf9f8ef3d3d10ca | [] | no_license | carbonXIII/research_scraper | 65225e536512a43746f50bc38d0b007ee14ebb86 | 658b1ee616298366631a0bebe06d5b3ef3342310 | refs/heads/master | 2022-05-17T13:44:04.167376 | 2019-10-22T19:17:51 | 2019-10-22T19:17:51 | null | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 7,404 | py | # Scraper to scrape some information from a Github repository (website and cloned repo)
# From Github:
# Description
# Issues
# Pull requests
# From Cloned Repo:
# .markdown files
# in-line comments
from shutil import rmtree
import os, sys
import bs4 as soup
import requests
from myutils import onerror, choose_subset, run_proc, clone, slurp, write_as_bytes
code_extensions = ['cpp','c','cxx','py','sh','js','java','glsl','hlsl']
markdown_extensions = ['md','rst']
comment_starters = ['//','#']
to_clean = ['.','/','-']
comment_merge_distance = 5
max_issues = 20
max_inline = 20
max_markdown = 20
force_readme = True
def parse_inline(qualified):
with open(qualified, 'r') as g:
try:
lines = g.readlines()
except:
print("can't read {}, skipping".format(qualified))
lines = []
comments = []
for i,line in enumerate(lines):
stripped = line.lstrip()
for cs in comment_starters:
if stripped.startswith(cs):
if len(comments) and comments[-1][1] > i-comment_merge_distance:
stop = min(i+comment_merge_distance,len(lines))
comments[-1] = (comments[-1][0], stop)
else:
start = max(i-comment_merge_distance,0)
stop = min(i+comment_merge_distance,len(lines))
comments += [(start, stop)]
break
merged = []
for comment in comments:
merged += ['### Line ',str(comment[0]+1),'-',str(comment[1]),'\n']
merged += ['```\n']
for i in range(comment[0],comment[1]):
if len(line.strip()) == 0:
continue
line = line.replace('```',"'''")
merged += [lines[i]]
merged += ['\n```\n']
merged = ''.join(merged)
return merged
def parse_markdown(qualified):
b = None
with open(qualified, 'rb') as g:
b = g.read()
b = b.replace(b'```', b"'''")
return b'```' + b + b'```'
def extract_from_repo(repo, out):
candidates = {"markdown": [], "inline": []}
_extract_from_repo(repo, candidates, path='/.')
inline_ = choose_subset(max_inline, len(candidates["inline"]))
markdown_ = choose_subset(max_markdown, len(candidates["markdown"]))
for i in inline_:
npath = candidates["inline"][i]
print('Scraping',npath)
qualified = repo + npath
extension = npath.split('.')[-1]
cleaned_path = npath
for symbol in to_clean:
cleaned_path = '_'.join(cleaned_path.split(symbol))
while '__' in cleaned_path:
cleaned_path = cleaned_path.replace('__', '_')
merged = parse_inline(qualified)
out["inline"][cleaned_path] = merged
if force_readme:
readme_names = ['/./README.' + ext for ext in markdown_extensions]
for i,npath in enumerate(candidates["markdown"]):
if npath in readme_names:
if i not in markdown_:
markdown_ += [i]
for i in markdown_:
npath = candidates["markdown"][i]
print('Scraping',npath)
qualified = repo + npath
extension = npath.split('.')[-1]
cleaned_path = npath
for symbol in to_clean:
cleaned_path = '_'.join(cleaned_path.split(symbol))
while '__' in cleaned_path:
cleaned_path = cleaned_path.replace('__', '_')
merged = parse_markdown(qualified)
out["markdown"][cleaned_path] = merged
def _extract_from_repo(repo, out, path='/.'):
for f in os.listdir(repo + path):
npath = path + '/' + f
qualified = repo + npath
extension = f.split('.')[-1]
if os.path.isdir(qualified):
_extract_from_repo(repo, out, npath)
elif extension in markdown_extensions:
out["markdown"] += [npath]
elif extension in code_extensions:
out["inline"] += [npath]
def count_issues(url):
max_jump = 65536
guess = 0
while max_jump:
nguess = guess + max_jump
page = requests.get(url + str(nguess))
if page.status_code is 200:
guess = nguess
max_jump = max_jump // 2
return guess
def extract_from_url(url, out):
souped = slurp(url)[0]
desc = souped.find('span', {'itemprop': 'about'}).getText()
out["info"]["description"] = desc.strip() + '\n'
issue_count = count_issues(url + '/issues/')
to_scrape = choose_subset(max_issues, issue_count)
for target in to_scrape:
target_url = url + '/issues/' + str(target + 1)
page,target_url = slurp(target_url)
if type(page) is not soup.BeautifulSoup:
print(target_url,', status_code =',page,', skipping.')
continue
print('Scraping',target_url)
issue_type = target_url.split('/')[-2]
merged = []
title = page.find('span', {'class': 'js-issue-title'}).getText()
merged += ['Title:\n```\n', title, '\n```\n']
comments = page.find_all('div', {'class': 'timeline-comment-group'})
for comment in comments:
author = comment.find('a', {'class': 'author'}).getText()
text = comment.find('td', {'class': 'comment-body'}).getText()
merged += ['Author:\n```\n',
author,
'\n```\n',
'Text:\n```\n',
text,
'\n```\n']
merged = ''.join(merged)
out[issue_type][str(target+1)] = merged
def write_summary(title, extracted, out):
with open(out, 'wb') as f:
write_as_bytes(f, '# ' + title + '\n')
for heading in extracted:
heading_adj = heading.capitalize()
write_as_bytes(f, '[- {}](#{})\n\n'.format(heading_adj, heading_adj))
for path in extracted[heading]:
path_ = path.split('/')[-1]
write_as_bytes(f, '* [{}](#{})\n\n'.format(path_, path))
write_as_bytes(f, '<!-- toc -->\n\n')
for heading in extracted:
heading_adj = heading.capitalize()
write_as_bytes(f, '# ' + str(heading_adj) + '\n')
for path in extracted[heading]:
write_as_bytes(f, '## ' + str(path) + '\n')
to_write = extracted[heading][path]
if type(to_write) is str:
f.write(extracted[heading][path].encode())
else:
f.write(extracted[heading][path])
write_as_bytes(f, '\n')
def go(url, out_file):
repo = './repo'
rmtree(repo, onerror=onerror)
clone(url, repo)
extracted = {"info": {}, "markdown": {}, "inline": {}, "issues": {}, "pull": {}}
extract_from_repo(repo, extracted)
extract_from_url(url, extracted)
write_summary('/'.join(url.split('/')[-2:]), extracted, out_file)
if __name__ == '__main__':
repo_list = 'repos.txt' if len(sys.argv) < 2 else sys.argv[1]
with open(repo_list, 'r') as f:
while True:
line = f.readline()
if line.strip() == 'END':
break
url = line.split(',')[1].strip()
name = url.split('/')[-1]
print('repo:',name)
go(url, name + '.md')
| [
"r1ch4rds.j0rdan@gmail.com"
] | r1ch4rds.j0rdan@gmail.com |
07d35f66b6c9afb14eb49238a555cec69a67ef2c | 5dc77586e3e0f9de1f032fd2ca68494d8e58928f | /contrib/great_expectations_geospatial_expectations/great_expectations_geospatial_expectations/expectations/expect_column_values_geometry_to_be_within_shape.py | 4fb1ffed1fe1ffab5fafb06d1d68eb5f5526a0d0 | [
"Apache-2.0"
] | permissive | great-expectations/great_expectations | dd7c22e6277d6b08bee3ff38a015e6e8cd434df6 | b0290e2fd2aa05aec6d7d8871b91cb4478e9501d | refs/heads/develop | 2023-09-04T09:30:26.395518 | 2023-09-02T00:00:13 | 2023-09-02T00:00:13 | 103,071,520 | 8,931 | 1,535 | Apache-2.0 | 2023-09-14T19:57:16 | 2017-09-11T00:18:46 | Python | UTF-8 | Python | false | false | 12,685 | py | from typing import Optional
import pandas as pd
import pygeos as geos
from great_expectations.core.expectation_configuration import ExpectationConfiguration
from great_expectations.execution_engine import PandasExecutionEngine
from great_expectations.expectations.expectation import ColumnMapExpectation
from great_expectations.expectations.metrics import (
ColumnMapMetricProvider,
column_condition_partial,
)
# This class defines a Metric to support your Expectation.
# For most ColumnMapExpectations, the main business logic for calculation will live in this class.
class ColumnValuesGeometryWithinShape(ColumnMapMetricProvider):
# This is the id string that will be used to reference your metric.
condition_metric_name = "column_values.geometry.within_shape"
condition_value_keys = ("shape", "shape_format", "column_shape_format", "properly")
# This method implements the core logic for the PandasExecutionEngine
@column_condition_partial(engine=PandasExecutionEngine)
def _pandas(cls, column, **kwargs):
shape = kwargs.get("shape")
shape_format = kwargs.get("shape_format")
column_shape_format = kwargs.get("column_shape_format")
properly = kwargs.get("properly")
# Check that shape is given and given in the correct format
if shape is not None:
try:
if shape_format == "wkt":
shape_ref = geos.from_wkt(shape)
elif shape_format == "wkb":
shape_ref = geos.from_wkb(shape)
elif shape_format == "geojson":
shape_ref = geos.from_geojson(shape)
else:
raise NotImplementedError(
"Shape constructor method not implemented. Must be in WKT, WKB, or GeoJSON format."
)
except Exception:
raise Exception("A valid reference shape was not given.")
else:
raise Exception("A shape must be provided for this method.")
# Load the column into a pygeos Geometry vector from numpy array (Series not supported).
if column_shape_format == "wkt":
shape_test = geos.from_wkt(column.to_numpy(), on_invalid="ignore")
elif column_shape_format == "wkb":
shape_test = geos.from_wkb(column.to_numpy(), on_invalid="ignore")
else:
raise NotImplementedError("Column values shape format not implemented.")
# Allow for an array of reference shapes to be provided. Return a union of all the shapes in the array (Polygon or Multipolygon)
shape_ref = geos.union_all(shape_ref)
# Prepare the geometries
geos.prepare(shape_ref)
geos.prepare(shape_test)
if properly:
return pd.Series(geos.contains_properly(shape_ref, shape_test))
else:
return pd.Series(geos.contains(shape_ref, shape_test))
# This method defines the business logic for evaluating your metric when using a SqlAlchemyExecutionEngine
# @column_condition_partial(engine=SqlAlchemyExecutionEngine)
# def _sqlalchemy(cls, column, _dialect, **kwargs):
# raise NotImplementedError
# This method defines the business logic for evaluating your metric when using a SparkDFExecutionEngine
# @column_condition_partial(engine=SparkDFExecutionEngine)
# def _spark(cls, column, **kwargs):
# raise NotImplementedError
# This class defines the Expectation itself
class ExpectColumnValuesGeometryToBeWithinShape(ColumnMapExpectation):
"""Expect that column values as geometries are within a given reference shape.
expect_column_values_geometry_to_be_within_shape is a \
[Column Map Expectation](https://docs.greatexpectations.io/docs/guides/expectations/creating_custom_expectations/how_to_create_custom_column_map_expectations).
Args:
column (str): \
The column name. \
Column values must be provided in WKT or WKB format, which are commom formats for GIS Database formats. \
WKT can be accessed thhrough the ST_AsText() or ST_AsBinary() functions in queries for PostGIS and MSSQL.
Keyword Args:
shape (str or list of str): \
The reference geometry
shape_format (str): \
Geometry format for 'shape' string(s). Can be provided as 'Well Known Text' (WKT), 'Well Known Binary' (WKB), or as GeoJSON. \
Must be one of: [wkt, wkb, geojson]. Default: wkt
column_shape_format (str): \
Geometry format for 'column'. Column values must be provided in WKT or WKB format, which are commom formats for GIS Database formats. \
WKT can be accessed thhrough the ST_AsText() or ST_AsBinary() functions in queries for PostGIS and MSSQL.
properly (boolean): \
Whether the 'column' values should be properly within in the reference 'shape'. \
The method allows for shapes to be 'properly contained' within the reference, meaning no points of a given geometry can touch the boundary of the reference. \
See the pygeos docs for reference. Default: False
Returns:
An [ExpectationSuiteValidationResult](https://docs.greatexpectations.io/docs/terms/validation_result)
Notes:
* Convention is (X Y Z) for points, which would map to (Longitude Latitude Elevation) for geospatial cases.
* Any convention can be followed as long as the test and reference shapes are consistent.
* The reference shape allows for an array, but will union (merge) all the shapes into 1 and check the contains condition.
"""
# These examples will be shown in the public gallery.
# They will also be executed as unit tests for your Expectation.
examples = [
{
"data": {
"points_only": [
"POINT(1 1)",
"POINT(2 2)",
"POINT(6 4)",
"POINT(3 9)",
"POINT(8 9.999)",
],
"points_and_lines": [
"POINT(1 1)",
"POINT(2 2)",
"POINT(6 4)",
"POINT(3 9)",
"LINESTRING(5 5, 8 10)",
],
},
"tests": [
{
"title": "positive_test_with_points",
"exact_match_out": False,
"include_in_gallery": True,
"in": {
"column": "points_only",
"shape": "POLYGON ((0 0, 0 10, 10 10, 10 0, 0 0))",
"shape_format": "wkt",
"properly": False,
},
"out": {
"success": True,
},
},
{
"title": "positive_test_with_points_and_lines",
"exact_match_out": False,
"include_in_gallery": True,
"in": {
"column": "points_and_lines",
"shape": "POLYGON ((0 0, 0 10, 10 10, 10 0, 0 0))",
"shape_format": "wkt",
"properly": False,
},
"out": {
"success": True,
},
},
{
"title": "positive_test_with_points_wkb_reference_shape",
"exact_match_out": False,
"include_in_gallery": True,
"in": {
"column": "points_only",
"shape": "010300000001000000050000000000000000000000000000000000000000000000000000000000000000002440000000000000244000000000000024400000000000002440000000000000000000000000000000000000000000000000",
"shape_format": "wkb",
"properly": False,
},
"out": {
"success": True,
},
},
{
"title": "positive_test_with_points_geojson_reference_shape",
"exact_match_out": False,
"include_in_gallery": True,
"in": {
"column": "points_only",
"shape": '{"type":"Polygon","coordinates":[[[0.0,0.0],[0.0,10.0],[10.0,10.0],[10.0,0.0],[0.0,0.0]]]}',
"shape_format": "geojson",
"properly": False,
},
"out": {
"success": True,
},
},
{
"title": "negative_test_with_points",
"exact_match_out": False,
"include_in_gallery": True,
"in": {
"column": "points_only",
"shape": "POLYGON ((0 0, 0 7.5, 7.5 7.5, 7.5 0, 0 0))",
"shape_format": "wkt",
"properly": True,
},
"out": {
"success": False,
},
},
{
"title": "negative_test_with_points_and_lines_not_properly_contained",
"exact_match_out": False,
"include_in_gallery": True,
"in": {
"column": "points_and_lines",
"shape": "POLYGON ((0 0, 0 10, 10 10, 10 0, 0 0))",
"shape_format": "wkt",
"properly": True,
"mostly": 1,
},
"out": {
"success": False,
},
},
],
}
]
# This is the id string of the Metric used by this Expectation.
# For most Expectations, it will be the same as the `condition_metric_name` defined in your Metric class above.
map_metric = "column_values.geometry.within_shape"
# This is a list of parameter names that can affect whether the Expectation evaluates to True or False
success_keys = (
"mostly",
"shape",
"shape_format",
"column_shape_format",
"properly",
)
# This dictionary contains default values for any parameters that should have default values
default_kwarg_values = {
"mostly": 1,
"shape_format": "wkt",
"column_shape_format": "wkt",
"properly": False,
}
def validate_configuration(self, configuration: Optional[ExpectationConfiguration]):
"""
Validates that a configuration has been set, and sets a configuration if it has yet to be set. Ensures that
necessary configuration arguments have been provided for the validation of the expectation.
Args:
configuration (OPTIONAL[ExpectationConfiguration]): \
An optional Expectation Configuration entry that will be used to configure the expectation
Returns:
True if the configuration has been validated successfully. Otherwise, raises an exception
"""
super().validate_configuration(configuration)
configuration = configuration or self.configuration
# # Check other things in configuration.kwargs and raise Exceptions if needed
# try:
# assert (
# ...
# ), "message"
# assert (
# ...
# ), "message"
# except AssertionError as e:
# raise InvalidExpectationConfigurationError(str(e))
return True
# This object contains metadata for display in the public Gallery
library_metadata = {
"tags": [
"geospatial",
"hackathon-2022",
], # Tags for this Expectation in the Gallery
"contributors": [ # Github handles for all contributors to this Expectation.
"@pjdobson", # Don't forget to add your github handle here!
],
"requirements": ["pygeos"],
}
if __name__ == "__main__":
ExpectColumnValuesGeometryToBeWithinShape().print_diagnostic_checklist()
| [
"noreply@github.com"
] | noreply@github.com |
996df91cc5baeba8d648a84074940c67fb599b8c | 9a811b17415c7a19731c42cffd18177f15adf56e | /list_all_pixels.py | efb3ef048aebe8777b914caa5513c8d0e8e93616 | [] | no_license | sanjitk7/ImageSpacialIntensityTransformations | 2e8fcd16e896374ecf05688e425b772b6d14bffc | 337c3661fdb631afe096b697c5ae064f49db8d46 | refs/heads/master | 2022-12-24T19:52:03.391437 | 2020-09-29T11:56:14 | 2020-09-29T11:56:14 | null | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 209 | py | # Listing All The pixel Values of the Image
def list_all_pixels(im):
for i in range (im.size[0]):
for j in range(im.size[1]):
print("f(" + str(i) + "," + str(j)+") = " + str(pixelMap[i,j][0])) | [
"sanjitk2018@gmail.com"
] | sanjitk2018@gmail.com |
a0adbf0801f319434a3785fe01f994198732a1a1 | 163bbb4e0920dedd5941e3edfb2d8706ba75627d | /Code/CodeRecords/2864/60618/317011.py | a6f56f689fbf65afed239fae41b9105f1e30fc28 | [] | no_license | AdamZhouSE/pythonHomework | a25c120b03a158d60aaa9fdc5fb203b1bb377a19 | ffc5606817a666aa6241cfab27364326f5c066ff | refs/heads/master | 2022-11-24T08:05:22.122011 | 2020-07-28T16:21:24 | 2020-07-28T16:21:24 | 259,576,640 | 2 | 1 | null | null | null | null | UTF-8 | Python | false | false | 1,548 | py | class Solution:
def find(self, n, data):
if n == 1:
return data[0]
re = 0
d = dict()
for i in range(n):
d[data[i]] = data.count(data[i])
# 这两行可以好好学习一下!!!
sorted_key_list = sorted(d, reverse=True) # key从大到小
sorted_dict = sorted(d.items(), key=lambda x: x[0], reverse=True)
k = sorted_key_list
dd = []
for item in sorted_dict:
dd.append(item[1])
i = 0
while i < len(k):
if k[i] == k[i+1] + 1: # 下一个不能要了
if dd[i] * k[i] > dd[i + 1] * k[i + 1]:
re += dd[i] * k[i]
i += 1
if i == len(k) - 1:
break
if i == len(k) - 2:
re += dd[i + 1] * k[i + 1]
break
else:
re += dd[i + 1] * k[i + 1]
i += 1
if i == len(k) - 1:
break
if k[i] == k[i+1]+1:
i += 1
i += 1
else: # 下一个还能要
re += dd[i] * k[i]
i += 1
if i == len(k) - 1:
re += dd[i] * k[i]
break
return re
if __name__ == '__main__':
n = int(input())
data = [int(a) for a in input().split()]
s = Solution()
re = s.find(n, data)
print(re)
| [
"1069583789@qq.com"
] | 1069583789@qq.com |
e9da2dde9e797e1e181779543be225efdc001fb3 | 0fe003ace0c47ff027de3c5497ef9bd04925939a | /day23.py | 0d5fffbeb3b042325ecfb43ede3588fa80f58148 | [] | no_license | russford/advent2016 | 132dde45adadf2c78e1ab9544af66a72541700ac | 1e92cf36a5870e05297ddba3d5d254213e4f0aaf | refs/heads/master | 2020-06-15T14:49:43.566431 | 2017-01-02T12:26:27 | 2017-01-02T12:26:27 | 75,284,768 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 3,291 | py | test_code = """cpy 2 a
tgl a
tgl a
tgl a
cpy 1 a
dec a
dec a"""
# cpy x y copies x (either an integer or the value of a register) into register y.
# inc x increases the value of register x by one.
# dec x decreases the value of register x by one.
# jnz x y jumps to an instruction y away (positive means forward; negative means backward), but only if x is not zero.
# For one-argument instructions, inc becomes dec, and all other one-argument instructions become inc.
# For two-argument instructions, jnz becomes cpy, and all other two-instructions become jnz.
# The arguments of a toggled instruction are not affected.
# If an attempt is made to toggle an instruction outside the program, nothing happens.
# If toggling produces an invalid instruction (like cpy 1 2) and an attempt is later made to execute that instruction, skip it instead.
# If tgl toggles itself (for example, if a is 0, tgl a would target itself and become inc a), the resulting instruction is not executed until the next time it is reached.
def toggle (instr):
instr = instr.split()
if len(instr) == 2:
if instr[0] == "inc":
return "dec " + instr[1]
else: return "inc " + instr[1]
if len(instr) == 3:
if instr[0] == "jnz":
return "cpy {} {}".format(instr[1], instr[2])
else:
return "jnz {} {}".format(instr[1], instr[2])
def val(v, registers):
if "a" <= v[0] <= "d":
return registers[v]
else:
return int(v)
def exec (code, code_ptr, registers):
if code_ptr == 2:
registers["a"] = registers["a"] * registers["b"]
registers["b"] -= 1
registers["c"] = 2 * registers["b"]
registers["d"] = 0
return 14
if code_ptr == 20:
registers["a"] += 95*96
return 6
instr = code[code_ptr].split()
if instr[0] == "cpy":
if instr[2] in registers:
registers[instr[2]] = val(instr[1], registers)
if instr[0] == "inc":
if instr[1] in registers:
registers[instr[1]] += 1
if instr[0] == "dec":
if instr[1] in registers:
registers[instr[1]] -= 1
if instr[0] == "jnz":
cmp = val(instr[1], registers)
if cmp != 0:
return val(instr[2], registers)
if instr[0] == "tgl":
jmp = val(instr[1], registers)
if code_ptr+jmp < len(code):
new_ins = toggle(code[code_ptr+jmp])
print ("toggled {}:{} to {}".format(code_ptr+jmp, code[code_ptr+jmp], new_ins))
code[code_ptr+jmp] = new_ins
if code_ptr+jmp == 18:
print ('\n'.join(code[16:]))
return 0
def run_code(code):
registers = {"a": 12, "b": 0, "c": 0, "d": 0}
code_ptr = 0
print(code)
i=0
while code_ptr < len(code):
jmp = exec(code, code_ptr, registers)
if code_ptr < 2 or 21 > code_ptr > 15 or code_ptr == 10 or code_ptr > 23:
print("{:>3}: {:<8} | {}".format(code_ptr, code[code_ptr], ' '.join(["{}:{:>5}".format(k,v) for k,v in sorted(registers.items())])))
if jmp == 0: jmp = 1
code_ptr += jmp
i += 1
print (sorted(registers.items()))
with open("day23.txt", "r") as f:
file_code = [l.strip('\n') for l in f.readlines()]
run_code(file_code)
| [
"russford@gmail.com"
] | russford@gmail.com |
aa7749e5a5c46e9b294ba65e63edbafd2bdc540c | e63e8963f36689e525876dd877017352e96df12d | /DFCM_Electricity.py | 00f385ba96cff565eea8439932912a7aba1a0fba | [] | no_license | FieldDoctor/DFCM | b966d01b945e8ba0d233cc353d0c9815c743f5e2 | 7da7eecbdfd2c3b34c92ef79029de47364f2eded | refs/heads/master | 2023-03-28T23:39:19.861009 | 2021-03-26T10:17:39 | 2021-03-26T10:17:39 | 321,042,829 | 2 | 1 | null | null | null | null | UTF-8 | Python | false | false | 7,511 | py | configure = {}
configure['SourceFilePath'] = './1-Electricity/1-temp.csv'
configure['InputFilePath'] = './1-Electricity/2plus-supervisedDataSet_zscore.csv'
configure['OutputFilePath'] = './1-Electricity/6-DFCM.csv'
configure['PltFilePath'] = './1-Electricity/6-DFCM/'
configure['AllAttributes'] = 8
configure['TargetAttributes'] = 3
configure['InputAttributes'] = [1,2,3,4,5,6,7]
configure['OutputAttributes'] = [13,14,15]
configure['TimeAttributes'] = [0]
configure['Length'] = 21899
configure['global_epochs'] = 400
configure['f_batch_size'] = 25000
configure['f_epochs'] = 15
configure['hidden_layer'] = 10
configure['n_batch_size'] = 25000
configure['n_epochs'] = 15
configure['LSTM_hiddenDim'] = 15
import os
import time
from pandas import DataFrame
import pandas as pd
import numpy as np
import matplotlib.pyplot as plt
from sklearn import preprocessing
from sklearn.metrics import mean_squared_error
from keras.layers import Dense, LSTM, Dropout
from keras.models import Sequential
from keras import optimizers
#optimizers.Adam(lr=0.0001, beta_1=0.9, beta_2=0.999, epsilon=1e-08)
global_epochs = configure['global_epochs']
f_batch_size = configure['f_batch_size']
f_epochs = configure['f_epochs']
hidden_layer = configure['hidden_layer']
n_batch_size = configure['n_batch_size']
n_epochs = configure['n_epochs']
LSTM_hiddenDim = configure['LSTM_hiddenDim']
# 函数 - sigmoid
def sigmoid(x):
return 1 / (1 + np.exp(-x))
def mkdir(path):
path = path.strip()
path = path.rstrip("\\")
isExists = os.path.exists(path)
if not isExists:
os.makedirs(path)
print(path + ' 创建成功')
return True
else:
print(path + ' 目录已存在')
return False
mkdir(configure['PltFilePath'])
# 加载数据集
dataset = pd.read_csv(configure['InputFilePath'])
# 狗造训练集(70%)和测试集(30%)
values = dataset.values
n_train = int(0.7 * configure['Length'])
train = values[:n_train, :]
test = values[n_train:, :]
train_X, train_Y = sigmoid( train[:, configure['InputAttributes']] ) , train[:,configure['OutputAttributes']]
test_X, test_Y = sigmoid( test[:, configure['InputAttributes']] ), test[:,configure['OutputAttributes']]
train_U = train[:,configure['TimeAttributes']]
train_U = train_U.reshape((train_U.shape[0], 1, train_U.shape[1]))
test_U = test[:,configure['TimeAttributes']]
test_U = test_U.reshape((test_U.shape[0], 1, test_U.shape[1]))
print('Train dataset length : ' + str(len(train)) + '.')
print('Test dataset length : ' + str(len(test)) + '.')
print('------')
print('X dim : ' + str(train_X.shape[1]) + '.')
print('Y dim : ' + str(train_Y.shape[1]) + '.')
print('------')
print('train_X shape : ' + str(train_X.shape))
print('train_Y shape : ' + str(train_Y.shape))
print('train_U shape : ' + str(train_U.shape))
print('------')
print('test_X shape : ' + str(test_X.shape))
print('test_Y shape : ' + str(test_Y.shape))
print('test_U shape : ' + str(test_U.shape))
# 设计DFCM_3网络
model_f = [0 for i in range(len(configure['OutputAttributes']))]
for i in range(len(configure['OutputAttributes'])):
model_f[i] = Sequential()
model_f[i].add(Dense(hidden_layer, input_dim=train_X.shape[1], activation='relu', use_bias=False))
#model_f[i].add(Dense(hidden_layer, input_dim=hidden_layer, activation='relu', use_bias=False))
#model_f[i].add(Dense(hidden_layer, input_dim=hidden_layer, activation='relu', use_bias=False))
model_f[i].add(Dense(1, input_dim=hidden_layer, use_bias=False))
model_f[i].compile(loss='mean_squared_error', optimizer='adam')
model_u = [0 for i in range(len(configure['OutputAttributes']))]
for i in range(len(configure['OutputAttributes'])):
model_u[i] = Sequential()
model_u[i].add(LSTM(LSTM_hiddenDim, input_shape=(train_U.shape[1], train_U.shape[2])))
model_u[i].add(Dense(1, input_dim=LSTM_hiddenDim, use_bias=True))
model_u[i].compile(loss='mean_squared_error', optimizer='adam')
for i in range(global_epochs):
start = time.time()
if i == 0:
y_f = train_Y
else:
y_f = train_Y - y_u_predict
for j in range(len(configure['OutputAttributes'])):
model_f[j].fit(train_X, y_f[:, j], f_batch_size, f_epochs, verbose=0, shuffle=False)
y_f_predict = DataFrame()
for j in range(len(configure['OutputAttributes'])):
y_f_predict[str(j)] = model_f[j].predict(train_X).reshape(-1)
y_f_predict = y_f_predict.values
y_u = train_Y - y_f_predict
# for j in range(len(configure['OutputAttributes'])):
# print('f' + str(j + 1) + ' : ' + str(model_f[j].evaluate(train_X,y_f[:,j],verbose=2)))
# print('The ' + str(i + 1) + ' times f() training finished. loss:' + str( pow(abs(y_u), 2).mean().mean() ))
for j in range(len(configure['OutputAttributes'])):
model_u[j].fit(train_U, y_u[:, j], n_batch_size, n_epochs, verbose=0)
y_u_predict = DataFrame()
for j in range(len(configure['OutputAttributes'])):
y_u_predict[str(j)] = model_u[j].predict(train_U).reshape(-1)
y_u_predict = y_u_predict.values
# for j in range(len(configure['OutputAttributes'])):
# print('u' + str(j + 1) + ' : ' + str(model_u[j].evaluate(train_U, y_u[:,j],verbose=2)))
# print('The ' + str(i + 1) + ' times u() training finished. loss:' + str( pow(abs(train_Y - y_u_predict), 2).mean().mean() ))
# evaluate
yhat_f_predict = DataFrame()
for j in range(len(configure['OutputAttributes'])):
yhat_f_predict[str(j)] = model_f[j].predict(test_X).reshape(-1)
yhat_f_predict = yhat_f_predict.values
yhat_u_predict = DataFrame()
for j in range(len(configure['OutputAttributes'])):
yhat_u_predict[str(j)] = model_u[j].predict(test_U).reshape(-1)
yhat_u_predict = yhat_u_predict.values
predict_train = y_u_predict + y_f_predict
predict_test = yhat_u_predict + yhat_f_predict
real_train = train_Y
real_test = test_Y
error_train = pow(abs(real_train - predict_train), 2)
error_test = pow(abs(real_test - predict_test), 2)
# print('The ' + str(i + 1) + ' times train error: ' + str(error_train.mean().mean()))
# print('The ' + str(i + 1) + ' times test error: ' + str(error_test.mean().mean()))
print(i + 1, error_train.mean().mean(), error_test.mean().mean())
if (error_test.mean().mean() < 0.125):
break
# print('This epoch TimeCost:' + str(time.time()-start) + 's.')
# 预测 & 输出
yhat_f_predict = DataFrame()
for j in range(len(configure['OutputAttributes'])):
yhat_f_predict[str(j)] = model_f[j].predict(test_X).reshape(-1)
yhat_f_predict = yhat_f_predict.values
yhat_u_predict = DataFrame()
for j in range(len(configure['OutputAttributes'])):
yhat_u_predict[str(j)] = model_u[j].predict(test_U).reshape(-1)
yhat_u_predict = yhat_u_predict.values
yhat = yhat_u_predict + yhat_f_predict
DataFrame(yhat).to_csv(configure['OutputFilePath'],index=False)
for j in range(len(configure['OutputAttributes'])):
model_f[j].save(configure['PltFilePath'] + 'model_f_' + str(j+1) + '.h5')
model_u[j].save(configure['PltFilePath'] + 'model_u_' + str(j+1) + '.h5')
# 数据概览 - 1
values = yhat
original = test_Y
# 指定要绘制的列
groups = list(range(configure['TargetAttributes']))
i = 1
# 绘制每一列
plt.figure(figsize=(15,15))
for group in groups:
plt.subplot(len(groups), 1, i)
plt.plot(original[:, group])
plt.plot(values[:, group])
i += 1
plt.savefig(configure['PltFilePath'] + 'performance.png')
plt.show() | [
"963138743@qq.com"
] | 963138743@qq.com |
fb100ad85682045d4052339d1143167421753289 | 3cfaa57f120c53694bdeb71c3785e95a6152f8e7 | /pi_libs/macros.py | 83e7f48ea48cf340b143c79a982f8e8ae9ff69f6 | [] | no_license | QUT-Motorsport/QUTMS_CANSimulation | 810593d82963717dd82e0c8ee549c06bdb4d851d | 6f3f7acaebceae90702a9cada113f1265dc51e1b | refs/heads/master | 2022-04-08T09:23:55.848072 | 2020-03-12T01:41:27 | 2020-03-12T01:41:27 | 246,234,740 | 1 | 1 | null | null | null | null | UTF-8 | Python | false | false | 1,400 | py | """
CAB202 Teensy Library (cab202_teensy)
macros.h rewritten to Python code
B.Talbot, September 2015
L.Buckingham, September 2017
M.Sadykov, November 2019
Queensland University of Technology
"""
"""
Setting data directions in a data direction register (DDR)
"""
def SET_INPUT(portddr, pin): (portddr) &= ~(1 << (pin))
def SET_OUTPUT(portddr, pin): (portddr) |= (1 << (pin))
"""
Setting, clearing, and reading bits in registers.
reg is the name of a register; pin is the index (0..7)
of the bit to set, clear or read.
(WRITE_BIT is a combination of CLEAR_BIT & SET_BIT)
"""
def SET_BIT(reg, pin): (reg) |= (1 << (pin))
def CLEAR_BIT(reg, pin): (reg) &= ~(1 << (pin))
def WRITE_BIT(reg, pin, value):
(reg) = (((reg) & ~(1 << (pin))) | ((value) << (pin)))
def BIT_VALUE(reg, pin): (((reg) >> (pin)) & 1)
def BIT_IS_SET(reg, pin): (BIT_VALUE((reg),(pin))==1)
"""
Rudimentary math macros
"""
def ABS(x):
if ((x) >= 0):
return (x)
else:
return -(x)
def SIGN(x):
if (((x) > 0) - ((x) < 0)):
return True
else:
return False
"""
Define definitions as in docs.
"""
HIGH = 1
LOW = 0
"""
Bit Write
"""
def BV(bit):
return (1 << (bit))
| [
"mattsadov@gmail.com"
] | mattsadov@gmail.com |
950ec69546022992c15201e1068609a18c452f15 | 835566e5f9ab26bf459c47678728719775d96c33 | /venv/lib/python2.7/codecs.py | a708b1569ffa72266e1aebcd14af601581eff916 | [] | no_license | BrandonWalk/bacon | a0e15898bcd8ba0ab2871a7021a0c0e92e58cf98 | 22e674da1bc94e7f608675a022b51c99e47e58e2 | refs/heads/master | 2020-04-18T20:20:59.896319 | 2019-01-26T20:32:33 | 2019-01-26T20:32:33 | 167,735,645 | 1 | 0 | null | null | null | null | UTF-8 | Python | false | false | 54 | py | /Users/brandonwalker/anaconda2/lib/python2.7/codecs.py | [
"branwalker19@gmail.com"
] | branwalker19@gmail.com |
5ee11a1a3c24fe548935f32270ff881fa53eb1a8 | 8460f7bdc6dd01c2a4f095a9af5b3766474afb4d | /util/Vis.py | 00ae504b8892e6a281301b4bdf9bb0e87591f7b0 | [] | no_license | ggqshr/ClassificationPic | 36b38890f821549cb22dffd3e99e9fa8ebe16719 | ddcd877184e368a823b08801d67c88645541ee3a | refs/heads/master | 2020-05-03T20:44:05.650053 | 2019-04-07T06:43:46 | 2019-04-07T06:43:46 | 178,809,450 | 1 | 0 | null | null | null | null | UTF-8 | Python | false | false | 2,464 | py | import visdom
import numpy as np
import time
import torchvision as tv
class Visualizer():
"""
封装了visdom的基本操作,但是你仍然可以通过`self.vis.function`
调用原生的visdom接口
"""
def __init__(self, env='default', **kwargs):
import visdom
self.vis = visdom.Visdom(env=env, use_incoming_socket=False, **kwargs)
# 画的第几个数,相当于横座标
# 保存(’loss',23) 即loss的第23个点
self.index = {}
self.log_text = ''
def reinit(self, env='default', **kwargs):
"""
修改visdom的配置
"""
self.vis = visdom.Visdom(env=env, use_incoming_socket=False, **kwargs)
return self
def plot_many(self, d):
"""
一次plot多个
@params d: dict (name,value) i.e. ('loss',0.11)
"""
for k, v in d.items():
self.plot(k, v)
def img_many(self, d):
for k, v in d.items():
self.img(k, v)
def plot(self, name, y):
"""
self.plot('loss',1.00)
"""
x = self.index.get(name, 0)
self.vis.line(Y=np.array([y]), X=np.array([x]),
win=name,
opts=dict(title=name),
update=None if x == 0 else 'append'
)
self.index[name] = x + 1
def img(self, name, img_):
"""
self.img('input_img',t.Tensor(64,64))
"""
if len(img_.size()) < 3:
img_ = img_.cpu().unsqueeze(0)
self.vis.image(img_.cpu(),
win=name,
opts=dict(title=name)
)
def img_grid_many(self, d):
for k, v in d.items():
self.img_grid(k, v)
def img_grid(self, name, input_3d):
"""
一个batch的图片转成一个网格图,i.e. input(36,64,64)
会变成 6*6 的网格图,每个格子大小64*64
"""
self.img(name, tv.utils.make_grid(
input_3d.cpu()[0].unsqueeze(1).clamp(max=1, min=0)))
def log(self, info, win='log_text'):
"""
self.log({'loss':1,'lr':0.0001})
"""
self.log_text += ('[{time}] {info} <br>'.format(
time=time.strftime('%m%d_%H%M%S'),
info=info))
self.vis.text(self.log_text, win=win)
def __getattr__(self, name):
return getattr(self.vis, name)
| [
"ggq18663278150@sian.com"
] | ggq18663278150@sian.com |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.