code
stringlengths 2
1.05M
| repo_name
stringlengths 5
104
| path
stringlengths 4
251
| language
stringclasses 1
value | license
stringclasses 15
values | size
int32 2
1.05M
|
|---|---|---|---|---|---|
import sys
import numpy as np
#######################
#### Feature Class
### Extracts features from a labeled corpus
#######################
from eval.ner.readers.brown import prepare_cluster_map
from eval.ner.sequences.id_feature import IDFeatures
class ExtendedFeatures(IDFeatures):
def __init__(self, dataset, brown_cluster_file=None):
super().__init__(dataset)
self.brown_cluster_file = brown_cluster_file
self.w_to_clusterid = None
if brown_cluster_file:
self.w_to_clusterid = prepare_cluster_map(self.brown_cluster_file)
assert self.w_to_clusterid is not None
# use emission features:
self.alphanumeric = False
self.alldigits = False
self.brown_id = False
self.brown_id_plus1 = False
self.brown_id_plus2 = False
self.brown_id_minus1 = False
self.brown_id_minus2 = False
self.brown_prefix = False # prefix length features; same for all brown_id
self.brown_prefix_lengths = []
self.capitalized = False
self.cappattern = False
self.hyphen = False
self.id = False
self.id_plus1 = False
self.id_plus2 = False
self.id_minus1 = False
self.id_minus2 = False
self.prefix = False
self.rep_id = False
self.rep_id_plus1 = False
self.rep_id_plus2 = False
self.rep_id_minus1 = False
self.rep_id_minus2 = False
self.suffix = False
self.uppercased = False
def set_baseline_features(self):
"""
Use listed emission features as baseline
"""
self.capitalized = True
self.cappattern = True
self.hyphen = True
self.id = True
self.id_plus1 = True
self.id_plus2 = True
self.id_minus1 = True
self.id_minus2 = True
self.prefix = True
self.suffix = True
self.uppercased = True
def get_emission_features(self, sequence, pos, y):
"""
Handles previous emissions by expanding the feature_cache dictionaries
"""
# w
x = sequence.x[pos]
# w-1
if pos > 0:
x_min1 = sequence.x[pos-1]
else:
x_min1 = -1
# w-2
if pos > 1:
x_min2 = sequence.x[pos-2]
else:
x_min2 = -2
# w+1
if pos < len(sequence.x)-1:
x_plus1 = sequence.x[pos+1]
else:
x_plus1 = -3
# w+2
if pos < len(sequence.x)-2:
x_plus2 = sequence.x[pos+2]
else:
x_plus2 = -4
if x not in self.node_feature_cache:
self.node_feature_cache[x] = {}
if x_min1 not in self.node_feature_cache[x]:
self.node_feature_cache[x][x_min1] = {}
if x_min2 not in self.node_feature_cache[x][x_min1]:
self.node_feature_cache[x][x_min1][x_min2] = {}
if x_plus1 not in self.node_feature_cache[x][x_min1][x_min2]:
self.node_feature_cache[x][x_min1][x_min2][x_plus1] = {}
if x_plus2 not in self.node_feature_cache[x][x_min1][x_min2][x_plus1]:
self.node_feature_cache[x][x_min1][x_min2][x_plus1][x_plus2] = {}
if y not in self.node_feature_cache[x][x_min1][x_min2][x_plus1][x_plus2]:
node_idx = {}
node_idx = self.add_emission_features(sequence, pos, y, node_idx)
self.node_feature_cache[x][x_min1][x_min2][x_plus1][x_plus2][y] = node_idx
idx = self.node_feature_cache[x][x_min1][x_min2][x_plus1][x_plus2][y]
return idx
def add_emission_features(self, sequence, pos, y, features):
# w
x = sequence.x[pos]
# w-1
if pos > 0:
x_min1 = sequence.x[pos-1]
else:
x_min1 = -1
# w-2
if pos > 1:
x_min2 = sequence.x[pos-2]
else:
x_min2 = -2
# w+1
if pos < len(sequence.x)-1:
x_plus1 = sequence.x[pos+1]
else:
x_plus1 = -3
# w+2
if pos < len(sequence.x)-2:
x_plus2 = sequence.x[pos+2]
else:
x_plus2 = -4
# Get tag name from ID.
y_name = self.dataset.y_dict.get_label_name(y)
# Get word name from ID.
x_name = self.dataset.x_dict.get_label_name(x)
word = str(x_name)
# w-1
if x_min1 == -1: # if no previous word
word_min1 = "out-of-seq-left"
else:
x_min1_name = self.dataset.x_dict.get_label_name(x_min1)
word_min1 = str(x_min1_name)
# w-2
if x_min2 == -2: # if no pre-previous word
word_min2 = "out-of-seq-left"
else:
x_min2_name = self.dataset.x_dict.get_label_name(x_min2)
word_min2 = str(x_min2_name)
# w+1
if x_plus1 == -3: # if no next word
word_plus1 = "out-of-seq-right"
else:
x_plus1_name = self.dataset.x_dict.get_label_name(x_plus1)
word_plus1 = str(x_plus1_name)
# w+3
if x_plus2 == -4: # if no post-next word
word_plus2 = "out-of-seq-right"
else:
x_plus2_name = self.dataset.x_dict.get_label_name(x_plus2)
word_plus2 = str(x_plus2_name)
if self.id:
# Generate feature name.
feat_name = "id:{}::{}".format(word, y_name)
self.features_used.add("id")
# Get feature ID from name.
feat_id = self.add_feature(feat_name)
# Append feature.
if feat_id != -1:
features[feat_id] = 1
if self.id_minus1:
feat_name = "id-1:{}::{}".format(word_min1, y_name)
self.features_used.add("id-1")
# Get feature ID from name.
feat_id = self.add_feature(feat_name)
# Append feature.
if feat_id != -1:
features[feat_id] = 1
if self.id_minus2:
feat_name = "id-2:{}::{}".format(word_min2, y_name)
self.features_used.add("id-2")
# Get feature ID from name.
feat_id = self.add_feature(feat_name)
# Append feature.
if feat_id != -1:
features[feat_id] = 1
if self.id_plus1:
feat_name = "id+1:{}::{}".format(word_plus1, y_name)
self.features_used.add("id+1")
# Get feature ID from name.
feat_id = self.add_feature(feat_name)
# Append feature.
if feat_id != -1:
features[feat_id] = 1
if self.id_plus2:
feat_name = "id+2:{}::{}".format(word_plus2, y_name)
self.features_used.add("id+2")
# Get feature ID from name.
feat_id = self.add_feature(feat_name)
# Append feature.
if feat_id != -1:
features[feat_id] = 1
if self.capitalized:
# Iscapitalized
if word.istitle():
# Generate feature name.
feat_name = "capitalized::{}".format(y_name)
self.features_used.add("capitalized")
# Get feature ID from name.
feat_id = self.add_feature(feat_name)
# Append feature.
if feat_id != -1:
features[feat_id] = 1
if self.uppercased:
# Allcapitalized
if word.isupper():
# Generate feature name.
feat_name = "uppercased::{}".format(y_name)
self.features_used.add("uppercased")
# Get feature ID from name.
feat_id = self.add_feature(feat_name)
# Append feature.
if feat_id != -1:
features[feat_id] = 1
if self.cappattern:
# Capitalization pattern in window
pattern = []
# w-2
if word_min2 == "out-of-seq-left":
pattern.append("-")
else:
if word_min2.istitle():
pattern.append("C")
else:
pattern.append("N")
# w-1
if word_min1 == "out-of-seq-left":
pattern.append("-")
else:
if word_min1.istitle():
pattern.append("C")
else:
pattern.append("N")
# w
if word.istitle():
pattern.append("C")
else:
pattern.append("N")
# w+1
if word_plus1 == "out-of-seq-right":
pattern.append("-")
else:
if word_plus1.istitle():
pattern.append("C")
else:
pattern.append("N")
# w+2
if word_plus2 == "out-of-seq-right":
pattern.append("-")
else:
if word_plus2.istitle():
pattern.append("C")
else:
pattern.append("N")
# Generate feature name.
feat_name = "cappattern:{}::{}".format("".join(pattern), y_name)
self.features_used.add("cappattern")
# Get feature ID from name.
feat_id = self.add_feature(feat_name)
# Append feature.
if feat_id != -1:
features[feat_id] = 1
if self.alldigits:
# Alldigits
if word.isdigit():
# Generate feature name.
feat_name = "number::{}".format(y_name)
self.features_used.add("number")
# Get feature ID from name.
feat_id = self.add_feature(feat_name)
# Append feature.
if feat_id != -1:
features[feat_id] = 1
if self.alphanumeric:
# Alphanumeric
if word.isalnum():
# Generate feature name.
feat_name = "alphanumber::{}".format(y_name)
self.features_used.add("alphanumber")
# Get feature ID from name.
feat_id = self.add_feature(feat_name)
# Append feature.
if feat_id != -1:
features[feat_id] = 1
if self.hyphen:
# Hyphenized
if "-" in word:
# Generate feature name.
feat_name = "hyphen::{}".format(y_name)
self.features_used.add("hyphen")
# Get feature ID from name.
feat_id = self.add_feature(feat_name)
# Append feature.
if feat_id != -1:
features[feat_id] = 1
if self.suffix:
# Suffixes
max_suffix = 3
for i in range(max_suffix):
if len(word) > i+1:
suffix = word[-(i+1):]
# Generate feature name.
feat_name = "suffix:{}::{}".format(suffix, y_name)
self.features_used.add("suffix")
# Get feature ID from name.
feat_id = self.add_feature(feat_name)
# Append feature.
if feat_id != -1:
features[feat_id] = 1
if self.prefix:
# Prefixes
max_prefix = 3
for i in range(max_prefix):
if len(word) > i+1:
prefix = word[:i+1]
# Generate feature name.
feat_name = "prefix:{}::{}".format(prefix, y_name)
self.features_used.add("prefix")
# Get feature ID from name.
feat_id = self.add_feature(feat_name)
# Append feature.
if feat_id != -1:
features[feat_id] = 1
# hmm wordrep features
positions = []
if self.rep_id:
positions.append(0)
if self.rep_id_minus1:
if word_min1 != "out-of-seq-left": # if no previous word
positions.append(-1)
if self.rep_id_minus2:
if word_min2 != "out-of-seq-left": # if no previous previous word
positions.append(-2)
if self.rep_id_plus1:
if word_plus1 != "out-of-seq-right": # if no next word
positions.append(1)
if self.rep_id_plus2:
if word_plus2 != "out-of-seq-right": # if no next next word
positions.append(2)
for position in positions:
if sequence.w is not None:
rep_id = sequence.w[pos+position]
self.add_rep_feat(rep_id, position, y_name, features, is_tree=False)
#changed sequence.t to sequence.u as seq.t is deleted
if sequence.u is not None:
try:
rep_id = sequence.u[pos+1+position] #offset of 1 because 0 is root
except KeyError:
continue
self.add_rep_feat(rep_id, position, y_name, features, is_tree=True)
if self.brown_id:
# w: Brown cluster id
clusterid = None
if word in self.w_to_clusterid:
clusterid = self.w_to_clusterid[word]
elif word.lower() in self.w_to_clusterid:
clusterid = self.w_to_clusterid[word.lower()]
if clusterid:
feat_core_name = "brown_id"
feat_name = "{}::{}::{}".format(feat_core_name, clusterid, y_name)
self.features_used.add(feat_core_name)
feat_id = self.add_feature(feat_name)
if feat_id != -1:
features[feat_id] = 1
if self.brown_prefix:
features = self.add_brown_pref_feat(feat_core_name, clusterid, y_name, features, self.brown_prefix_lengths)
# w window: Brown Cluster ids
if self.brown_id_minus1:
# wcluster-1
if word_min1 != "out-of-seq-left": # if no previous word
clusterid = None
if word_min1 in self.w_to_clusterid:
clusterid = self.w_to_clusterid[word_min1]
elif word_min1.lower() in self.w_to_clusterid:
clusterid = self.w_to_clusterid[word_min1.lower()]
if clusterid:
feat_core_name = "brown_id-1"
feat_name = "{}::{}::{}".format(feat_core_name, clusterid, y_name)
self.features_used.add(feat_core_name)
feat_id = self.add_feature(feat_name)
if feat_id != -1:
features[feat_id] = 1
if self.brown_prefix:
features = self.add_brown_pref_feat(feat_core_name, clusterid, y_name, features, self.brown_prefix_lengths)
if self.brown_id_minus2:
# wcluster-2
if word_min2 != "out-of-seq-left": # if no pre-previous word
clusterid = None
if word_min2 in self.w_to_clusterid:
clusterid = self.w_to_clusterid[word_min2]
elif word_min2.lower() in self.w_to_clusterid:
clusterid = self.w_to_clusterid[word_min2.lower()]
if clusterid:
feat_core_name = "brown_id-2"
feat_name = "{}::{}::{}".format(feat_core_name, clusterid, y_name)
self.features_used.add(feat_core_name)
feat_id = self.add_feature(feat_name)
if feat_id != -1:
features[feat_id] = 1
if self.brown_prefix:
features = self.add_brown_pref_feat(feat_core_name, clusterid, y_name, features, self.brown_prefix_lengths)
if self.brown_id_plus1:
# wcluster+1
if word_plus1 != "out-of-seq-right": # if no next word
clusterid = None
if word_plus1 in self.w_to_clusterid:
clusterid = self.w_to_clusterid[word_plus1]
elif word_plus1.lower() in self.w_to_clusterid:
clusterid = self.w_to_clusterid[word_plus1.lower()]
if clusterid:
feat_core_name = "brown_id+1"
feat_name = "{}::{}::{}".format(feat_core_name, clusterid, y_name)
self.features_used.add(feat_core_name)
feat_id = self.add_feature(feat_name)
if feat_id != -1:
features[feat_id] = 1
if self.brown_prefix:
features = self.add_brown_pref_feat(feat_core_name, clusterid, y_name, features, self.brown_prefix_lengths)
if self.brown_id_plus2:
# wcluster+2
if word_plus2 != "out-of-seq-right": # if no next word
clusterid = None
if word_plus2 in self.w_to_clusterid:
clusterid = self.w_to_clusterid[word_plus2]
elif word_plus2.lower() in self.w_to_clusterid:
clusterid = self.w_to_clusterid[word_plus2.lower()]
if clusterid:
feat_core_name = "brown_id+2"
feat_name = "{}::{}::{}".format(feat_core_name, clusterid, y_name)
self.features_used.add(feat_core_name)
feat_id = self.add_feature(feat_name)
if feat_id != -1:
features[feat_id] = 1
if self.brown_prefix:
features = self.add_brown_pref_feat(feat_core_name, clusterid, y_name, features, self.brown_prefix_lengths)
return features
def add_brown_pref_feat(self, feat_core_name, clusterid, y_name, features, pref_lengths):
"""
:param pref_lengths: list containing ints representing prefix lengths
"""
for pref in pref_lengths:
feat_name = "{}_p{}::{}::{}".format(feat_core_name, pref, clusterid[:pref], y_name)
self.features_used.add("{}_p{}".format(feat_core_name, pref))
feat_id = self.add_feature(feat_name)
if feat_id != -1:
features[feat_id] = 1
return features
# def get_transition_features(self, sequence, pos, y, y_prev, y_prev_prev):
# assert (pos >= 0 and pos < len(sequence.x))
#
# if y not in self.edge_feature_cache:
# self.edge_feature_cache[y] = {}
# if y_prev not in self.edge_feature_cache[y]:
# self.edge_feature_cache[y][y_prev] = {}
# if y_prev_prev not in self.edge_feature_cache[y][y_prev]:
# edge_idx = []
# edge_idx = self.add_transition_features(sequence, pos, y, y_prev, y_prev_prev, edge_idx)
# self.edge_feature_cache[y][y_prev][y_prev_prev] = edge_idx
# idx = self.edge_feature_cache[y][y_prev][y_prev_prev]
#
# return idx[:]
# def add_transition_features(self, sequence, pos, y, y_prev, y_prev_prev, features):
# assert pos < len(sequence.x)-1
# # Get label name from ID.
# y_name = self.dataset.y_dict.get_label_name(y)
# # Get previous label names from ID.
# y_prev_name = self.dataset.y_dict.get_label_name(y_prev)
# y_prev_prev_name = self.dataset.y_dict.get_label_name(y_prev_prev)
# # Generate feature name.
# feat_name = "prev_tag:{}::{}".format(y_prev_name, y_name)
# self.features_used.add("prev_tag")
# # Get feature ID from name.
# feat_id = self.add_feature(feat_name)
# # Append feature.
# if feat_id != -1:
# features.append(feat_id)
#
# feat_name = "prev_prev_tag:{}::{}".format(y_prev_prev_name, y_name)
# self.features_used.add("prev_prev_tag")
# # Get feature ID from name.
# feat_id = self.add_feature(feat_name)
# # Append feature.
# if feat_id != -1:
# features.append(feat_id)
#
# return features
def add_rep_feat(self, rep_id, position, y_name, features, is_tree=False):
if position == 0:
position = ""
elif position == 1 or position == 2:
position = "+{}".format(position)
tree = "tree" if is_tree else ""
# discrete
if isinstance(rep_id, (int, np.int64, np.int32)):
feat_name = "{}rep_id{}::{}::{}".format(tree, position, rep_id, y_name)
self.features_used.add("{}rep_id{}".format(tree, position))
feat_id = self.add_feature(feat_name)
if feat_id != -1:
features[feat_id] = 1
# continuous
elif isinstance(rep_id, np.ndarray):
#max_i = np.max(rep_id)
for c, i in enumerate(rep_id):
#if i == max_i or i == max_i-1:
#feat_name = "cont_{}rep_id{}::{}::{}::{}".format(tree, position, c, i, y_name)
feat_name = "cont_{}rep_id{}::{}::{}".format(tree, position, c, y_name)
self.features_used.add("cont_{}rep_id{}".format(tree, position))
feat_id = self.add_feature(feat_name)
if feat_id != -1:
features[feat_id] = i
else:
raise TypeError
#sys.exit("unexpected type: {}".format(type(rep_id)))
|
rug-compling/hmm-reps
|
eval/ner/sequences/extended_feature.py
|
Python
|
mit
| 21,598
|
from django.db import models
from django.contrib.auth.models import User
# Create your models here.
class Payment(models.Model):
created = models.DateTimeField(auto_now_add=True)
amount = models.DecimalField(max_digits=16, decimal_places=2)
payer = models.ForeignKey(User, related_name='payments', null=True)
class Meta:
ordering = ('created',)
|
linkleonard/braintree-tutorial
|
myapp/mysite/models.py
|
Python
|
mit
| 373
|
import keras
import numpy as np
from keras.datasets import cifar10
from keras.preprocessing.image import ImageDataGenerator
from keras.layers.normalization import BatchNormalization
from keras.layers import Conv2D, Dense, Input, add, Activation, GlobalAveragePooling2D
from keras.initializers import he_normal
from keras.callbacks import LearningRateScheduler, TensorBoard
from keras.models import Model
from keras import optimizers
from keras import regularizers
num_classes = 10
img_rows, img_cols = 32, 32
img_channels = 3
stack_num = 18
batch_size = 128
epochs = 200
iterations = 391
weight_decay = 0.0005
log_filepath = r'./resnet50/'
def scheduler(epoch):
if epoch <= 60:
return 0.1
if epoch <= 120:
return 0.02
if epoch <= 160:
return 0.004
return 0.0008
def color_preprocessing(x_train,x_test):
x_train = x_train.astype('float32')
x_test = x_test.astype('float32')
x_train[:,:,:,0] = (x_train[:,:,:,0] - np.mean(x_train[:,:,:,0])) / np.std(x_train[:,:,:,0])
x_train[:,:,:,1] = (x_train[:,:,:,1] - np.mean(x_train[:,:,:,1])) / np.std(x_train[:,:,:,1])
x_train[:,:,:,2] = (x_train[:,:,:,2] - np.mean(x_train[:,:,:,2])) / np.std(x_train[:,:,:,2])
x_test[:,:,:,0] = (x_test[:,:,:,0] - np.mean(x_test[:,:,:,0])) / np.std(x_test[:,:,:,0])
x_test[:,:,:,1] = (x_test[:,:,:,1] - np.mean(x_test[:,:,:,1])) / np.std(x_test[:,:,:,1])
x_test[:,:,:,2] = (x_test[:,:,:,2] - np.mean(x_test[:,:,:,2])) / np.std(x_test[:,:,:,2])
return x_train, x_test
def residual_network(img_input,classes_num=10,stack_n=18):
def residual_block(x,shape,increase_filter=False):
output_filter_num = shape[1]
if increase_filter:
first_stride = (2,2)
else:
first_stride = (1,1)
pre_bn = BatchNormalization()(x)
pre_relu = Activation('relu')(pre_bn)
conv_1 = Conv2D(output_filter_num,
kernel_size=(3,3),
strides=first_stride,
padding='same',
kernel_initializer=he_normal(),
kernel_regularizer=regularizers.l2(weight_decay)
)(pre_relu)
bn_1 = BatchNormalization()(conv_1)
relu1 = Activation('relu')(bn_1)
conv_2 = Conv2D(output_filter_num,
kernel_size=(3,3),
strides=(1,1),
padding='same',
kernel_initializer=he_normal(),
kernel_regularizer=regularizers.l2(weight_decay)
)(relu1)
if increase_filter:
projection = Conv2D(output_filter_num,
kernel_size=(1,1),
strides=(2,2),
padding='same',
kernel_initializer=he_normal(),
kernel_regularizer=regularizers.l2(weight_decay)
)(x)
block = add([conv_2, projection])
else:
block = add([conv_2,x])
return block
x = Conv2D(filters=16,
kernel_size=(3,3),
strides=(1,1),
padding='same',
kernel_initializer=he_normal(),
kernel_regularizer=regularizers.l2(weight_decay)
)(img_input)
for _ in range(0,stack_n):
x = residual_block(x,[16,16])
x = residual_block(x,[16,32],increase_filter=True)
for _ in range(1,stack_n):
x = residual_block(x,[16,32])
x = residual_block(x,[32,64],increase_filter=True)
for _ in range(1,stack_n):
x = residual_block(x,[32,64])
x = BatchNormalization()(x)
x = Activation('relu')(x)
x = GlobalAveragePooling2D()(x)
x = Dense(classes_num,activation='softmax',kernel_initializer=he_normal(),kernel_regularizer=regularizers.l2(weight_decay))(x)
return x
if __name__ == '__main__':
# load data
(x_train, y_train), (x_test, y_test) = cifar10.load_data()
y_train = keras.utils.to_categorical(y_train, num_classes)
y_test = keras.utils.to_categorical(y_test, num_classes)
# color preprocessing
x_train, x_test = color_preprocessing(x_train, x_test)
# build network
img_input = Input(shape=(img_rows,img_cols,img_channels))
output = residual_network(img_input,num_classes,18)
resnet = Model(img_input, output)
print(resnet.summary())
# set optimizer
sgd = optimizers.SGD(lr=.1, momentum=0.9, nesterov=True)
resnet.compile(loss='categorical_crossentropy', optimizer=sgd, metrics=['accuracy'])
# set callback
tb_cb = TensorBoard(log_dir=log_filepath, histogram_freq=0)
change_lr = LearningRateScheduler(scheduler)
cbks = [change_lr,tb_cb]
# set data augmentation
print('Using real-time data augmentation.')
datagen = ImageDataGenerator(horizontal_flip=True,
width_shift_range=0.125,height_shift_range=0.125,fill_mode='constant',cval=0.)
datagen.fit(x_train)
# start training
resnet.fit_generator(datagen.flow(x_train, y_train,batch_size=batch_size),
steps_per_epoch=iterations,
epochs=epochs,
callbacks=cbks,
validation_data=(x_test, y_test))
resnet.save('resnet.h5')
|
WenmuZhou/cifar-10-cnn
|
4_Residual_Network/ResNet_keras.py
|
Python
|
mit
| 5,307
|
import activity
from provider import cloudfront_provider
"""
activity_InvalidateCdn.py activity
"""
class activity_InvalidateCdn(activity.activity):
def __init__(self, settings, logger, conn=None, token=None, activity_task=None):
activity.activity.__init__(self, settings, logger, conn, token, activity_task)
self.name = "InvalidateCdn"
self.pretty_name = "CloudFront Invalidate Cdn"
self.version = "1"
self.default_task_heartbeat_timeout = 30
self.default_task_schedule_to_close_timeout = 60 * 5
self.default_task_schedule_to_start_timeout = 30
self.default_task_start_to_close_timeout = 60 * 5
self.description = "Runs CloudFront Invalidation request on Cdn bucket."
self.logger = logger
def do_activity(self, data):
try:
article_id = data['article_id']
version = data['version']
run = data['run']
except Exception as e:
self.logger.error("Error retrieving basic article data. Data: %s, Exception: %s" % (str(data), str(e)))
return activity.activity.ACTIVITY_PERMANENT_FAILURE
try:
self.emit_monitor_event(self.settings, article_id, version, run,
self.pretty_name, "start", "Starting check for generation of pdf cover.")
### If we want to run Invalidation only if CDN has been previously populated
# if "files_in_cdn" in data:
# if data["files_in_cdn"] == True:
# cloudfront_provider.create_invalidation(article_id)
# dashboard_message = "CloudFront Invalidation command sent for article %s." % str(article_id)
# else:
# dashboard_message = "CloudFront Invalidation was not necessary for article %s." % str(article_id)
cloudfront_provider.create_invalidation(article_id, self.settings)
dashboard_message = "CloudFront Invalidation command sent for article %s." % str(article_id)
self.emit_monitor_event(self.settings, article_id, version, run,
self.pretty_name, "end", dashboard_message)
return activity.activity.ACTIVITY_SUCCESS
except Exception as e:
error_message = str(e)
self.logger.error(error_message)
self.emit_monitor_event(self.settings, article_id, version, run,
self.pretty_name, "error", error_message)
return activity.activity.ACTIVITY_PERMANENT_FAILURE
|
gnott/elife-bot
|
activity/activity_InvalidateCdn.py
|
Python
|
mit
| 2,596
|
import os
import time
import datetime
import random
from django.db import models
from django.conf import settings
from django.test import TestCase, LiveServerTestCase, RequestFactory
from django.test.client import Client
from django.core.urlresolvers import reverse
from django.contrib.auth.models import User, Group
from django.http import Http404
from django.utils import timezone
from model_mommy import mommy
from account.models import AcctStmt, TransType, AcctTrans, Hotel
def _randint(a=10, b=100):
return random.randint(a,b)
def _acct_stmt(hotel, year, month):
'''
Monkey-patch save() so just generating test data and not based on actual
usage from Message records.
'''
global AcctStmt
AcctStmt.save = models.Model.save
return AcctStmt.objects.create(
hotel=hotel,
year=year,
month=month,
monthly_costs=_randint(),
total_sms=_randint(),
balance=_randint()
)
def make_acct_stmts(hotel):
return [_acct_stmt(hotel=hotel, year=2014, month=m)
for m in range(1,13)]
def _acct_trans(hotel, trans_type, insert_date, amount=None):
'''
Monkey-patch save() for same reason as AcctStmt.
'init_amt', 'recharge_amt' are 1000 Stripe Credits.
'sms_used' are b/t -100 ... -10 (for testing purposes)
Positive `amount` is a `credit`, else a `debit`
'''
global AcctTrans
AcctTrans.save = models.Model.save
# transaction
if trans_type.name in ('init_amt', 'recharge_amt'):
amount = _randint(1000, 1000)
credit, debit = True, False
else:
amount = _randint(-100, -10)
credit, debit = False, True
return AcctTrans.objects.create(
hotel=hotel,
trans_type=trans_type,
amount=amount,
sms_used=_randint(),
insert_date=insert_date,
debit=debit,
credit=credit
)
def make_acct_trans(hotel):
'''
TransType: use get() b/c tests using `fixtures`.
So, generate all transaction records until current date and test format.
This factory method can be used to manually test AcctTransDetailView template
w/ ./manage.py runserver
'''
# datetime
td = datetime.timedelta(days=-30)
next_day = datetime.timedelta(days=1)
today = datetime.date.today()
last_month = today + td
# TransType
init_amt = TransType.objects.get(name='init_amt')
recharge_amt = TransType.objects.get(name='recharge_amt')
sms_used = TransType.objects.get(name='sms_used')
# set the Hotel as Created 1 month ago
hotel.created = last_month
hotel.save()
# Create `init_amt`
_acct_trans(hotel=hotel, trans_type=init_amt, insert_date=hotel.created)
# Daily usage until `today` (start populating based on `last_month` date)
insert_date = hotel.created
balance = AcctTrans.objects.filter(hotel=hotel).balance()
# Loop thro and create. Recharge Account if Balancd < 0
while insert_date < today:
trans = _acct_trans(hotel=hotel, trans_type=sms_used, insert_date=insert_date)
balance = AcctTrans.objects.filter(hotel=hotel).balance()
if balance < 0:
trans = _acct_trans(hotel=hotel, trans_type=recharge_amt, insert_date=insert_date)
insert_date += next_day
return AcctTrans.objects.all()
|
aronysidoro/django-payasyougo
|
payg/account/tests/factory.py
|
Python
|
mit
| 3,347
|
#!/bin/env python
"""
The MIT License
Copyright (c) 2010 The Chicago Tribune & Contributors
Permission is hereby granted, free of charge, to any person obtaining a copy
of this software and associated documentation files (the "Software"), to deal
in the Software without restriction, including without limitation the rights
to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
copies of the Software, and to permit persons to whom the Software is
furnished to do so, subject to the following conditions:
The above copyright notice and this permission notice shall be included in
all copies or substantial portions of the Software.
THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN
THE SOFTWARE.
"""
import bees
from urlparse import urlparse
from optparse import OptionParser, OptionGroup
def parse_options():
"""
Handle the command line arguments for spinning up bees
"""
parser = OptionParser(usage="""
bees COMMAND [options]
Bees with Machine Guns
A utility for arming (creating) many bees (small EC2 instances) to attack
(load test) targets (web applications).
commands:
up Start a batch of load testing servers.
attack Begin the attack on a specific url.
down Shutdown and deactivate the load testing servers.
report Report the status of the load testing servers.
""")
up_group = OptionGroup(parser, "up",
"""In order to spin up new servers you will need to specify at least the -k command, which is the name of the EC2 keypair to use for creating and connecting to the new servers. The bees will expect to find a .pem file with this name in ~/.ssh/. Alternatively, bees can use SSH Agent for the key.""")
# Required
up_group.add_option('-k', '--key', metavar="KEY", nargs=1,
action='store', dest='key', type='string',
help="The ssh key pair name to use to connect to the new servers.")
up_group.add_option('-s', '--servers', metavar="SERVERS", nargs=1,
action='store', dest='servers', type='int', default=5,
help="The number of servers to start (default: 5).")
up_group.add_option('-g', '--group', metavar="GROUP", nargs=1,
action='store', dest='group', type='string', default='default',
help="The security group(s) to run the instances under (default: default).")
up_group.add_option('-z', '--zone', metavar="ZONE", nargs=1,
action='store', dest='zone', type='string', default='us-east-1d',
help="The availability zone to start the instances in (default: us-east-1d).")
up_group.add_option('-i', '--instance', metavar="INSTANCE", nargs=1,
action='store', dest='instance', type='string', default='ami-ff17fb96',
help="The instance-id to use for each server from (default: ami-ff17fb96).")
up_group.add_option('-t', '--type', metavar="TYPE", nargs=1,
action='store', dest='type', type='string', default='t1.micro',
help="The instance-type to use for each server (default: t1.micro).")
up_group.add_option('-l', '--login', metavar="LOGIN", nargs=1,
action='store', dest='login', type='string', default='newsapps',
help="The ssh username name to use to connect to the new servers (default: newsapps).")
up_group.add_option('-v', '--subnet', metavar="SUBNET", nargs=1,
action='store', dest='subnet', type='string', default=None,
help="The vpc subnet id in which the instances should be launched. (default: None).")
parser.add_option_group(up_group)
attack_group = OptionGroup(parser, "attack",
"""Beginning an attack requires only that you specify the -u option with the URL you wish to target.""")
# Required
attack_group.add_option('-u', '--url', metavar="URL", nargs=1,
action='store', dest='url', type='string',
help="URL of the target to attack.")
attack_group.add_option('-K', '--keepalive', metavar="KEEP_ALIVE", nargs=0,
action='store', dest='keep_alive', type='string', default=False,
help="Keep-Alive connection.")
attack_group.add_option('-p', '--post-file', metavar="POST_FILE", nargs=1,
action='store', dest='post_file', type='string', default=False,
help="The POST file to deliver with the bee's payload.")
attack_group.add_option('-m', '--mime-type', metavar="MIME_TYPE", nargs=1,
action='store', dest='mime_type', type='string', default='text/plain',
help="The MIME type to send with the request.")
attack_group.add_option('-n', '--number', metavar="NUMBER", nargs=1,
action='store', dest='number', type='int', default=1000,
help="The number of total connections to make to the target (default: 1000).")
attack_group.add_option('-C', '--cookies', metavar="COOKIES", nargs=1, action='store', dest='cookies',
type='string', default='',
help='Cookies to send during http requests. The cookies should be passed using standard cookie formatting, separated by semi-colons and assigned with equals signs.')
attack_group.add_option('-c', '--concurrent', metavar="CONCURRENT", nargs=1,
action='store', dest='concurrent', type='int', default=100,
help="The number of concurrent connections to make to the target (default: 100).")
attack_group.add_option('-H', '--headers', metavar="HEADERS", nargs=1,
action='store', dest='headers', type='string', default='',
help="HTTP headers to send to the target to attack. Multiple headers should be separated by semi-colons, e.g header1:value1;header2:value2")
attack_group.add_option('-e', '--csv', metavar="FILENAME", nargs=1,
action='store', dest='csv_filename', type='string', default='',
help="Store the distribution of results in a csv file for all completed bees (default: '').")
# Optional
attack_group.add_option('-T', '--tpr', metavar='TPR', nargs=1, action='store', dest='tpr', default=None, type='float',
help='The upper bounds for time per request. If this option is passed and the target is below the value a 1 will be returned with the report details (default: None).')
attack_group.add_option('-R', '--rps', metavar='RPS', nargs=1, action='store', dest='rps', default=None, type='float',
help='The lower bounds for request per second. If this option is passed and the target is above the value a 1 will be returned with the report details (default: None).')
attack_group.add_option('-A', '--basic_auth', metavar='basic_auth', nargs=1, action='store', dest='basic_auth', default='', type='string',
help='BASIC authentication credentials, format auth-username:password (default: None).')
parser.add_option_group(attack_group)
(options, args) = parser.parse_args()
if len(args) <= 0:
parser.error('Please enter a command.')
command = args[0]
if command == 'up':
if not options.key:
parser.error('To spin up new instances you need to specify a key-pair name with -k')
if options.group == 'default':
print 'New bees will use the "default" EC2 security group. Please note that port 22 (SSH) is not normally open on this group. You will need to use to the EC2 tools to open it before you will be able to attack.'
bees.up(options.servers, options.group, options.zone, options.instance, options.type, options.login, options.key, options.subnet)
elif command == 'attack':
if not options.url:
parser.error('To run an attack you need to specify a url with -u')
parsed = urlparse(options.url)
if "/" not in parsed.path:
if not parsed.scheme:
parsed = urlparse("http://" + options.url + "/")
else:
parsed = urlparse(options.url + "/")
if not parsed.scheme:
parsed = urlparse("http://" + options.url)
additional_options = dict(
cookies=options.cookies,
headers=options.headers,
post_file=options.post_file,
keep_alive=options.keep_alive,
mime_type=options.mime_type,
csv_filename=options.csv_filename,
tpr=options.tpr,
rps=options.rps,
basic_auth=options.basic_auth
)
bees.attack(options.url, options.number, options.concurrent, **additional_options)
elif command == 'down':
bees.down()
elif command == 'report':
bees.report()
def main():
parse_options()
|
tylerturk/beeswithmachineguns
|
beeswithmachineguns/main.py
|
Python
|
mit
| 9,593
|
"""
The MIT License
Copyright (c) 2007-2010 Leah Culver, Joe Stump, Mark Paschal, Vic Fryzel
Permission is hereby granted, free of charge, to any person obtaining a copy
of this software and associated documentation files (the "Software"), to deal
in the Software without restriction, including without limitation the rights
to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
copies of the Software, and to permit persons to whom the Software is
furnished to do so, subject to the following conditions:
The above copyright notice and this permission notice shall be included in
all copies or substantial portions of the Software.
THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN
THE SOFTWARE.
"""
import base64
import urllib
import time
import random
import urlparse
import hmac
import binascii
import httplib2
try:
from urlparse import parse_qs
parse_qs # placate pyflakes
except ImportError:
# fall back for Python 2.5
from cgi import parse_qs
try:
from hashlib import sha1
sha = sha1
except ImportError:
# hashlib was added in Python 2.5
import sha
import _version
__version__ = _version.__version__
OAUTH_VERSION = '1.0' # Hi Blaine!
HTTP_METHOD = 'GET'
SIGNATURE_METHOD = 'PLAINTEXT'
class Error(RuntimeError):
"""Generic exception class."""
def __init__(self, message='OAuth error occurred.'):
self._message = message
@property
def message(self):
"""A hack to get around the deprecation errors in 2.6."""
return self._message
def __str__(self):
return self._message
class MissingSignature(Error):
pass
def build_authenticate_header(realm=''):
"""Optional WWW-Authenticate header (401 error)"""
return {'WWW-Authenticate': 'OAuth realm="%s"' % realm}
def build_xoauth_string(url, consumer, token=None):
"""Build an XOAUTH string for use in SMTP/IMPA authentication."""
request = Request.from_consumer_and_token(consumer, token,
"GET", url)
signing_method = SignatureMethod_HMAC_SHA1()
request.sign_request(signing_method, consumer, token)
params = []
for k, v in sorted(request.iteritems()):
if v is not None:
params.append('%s="%s"' % (k, escape(v)))
return "%s %s %s" % ("GET", url, ','.join(params))
def to_unicode(s):
""" Convert to unicode, raise exception with instructive error
message if s is not unicode, ascii, or utf-8. """
if not isinstance(s, unicode):
if not isinstance(s, str):
raise TypeError('You are required to pass either unicode or string here, not: %r (%s)' % (type(s), s))
try:
s = s.decode('utf-8')
except UnicodeDecodeError, le:
raise TypeError('You are required to pass either a unicode object or a utf-8 string here. You passed a Python string object which contained non-utf-8: %r. The UnicodeDecodeError that resulted from attempting to interpret it as utf-8 was: %s' % (s, le,))
return s
def to_utf8(s):
return to_unicode(s).encode('utf-8')
def to_unicode_if_string(s):
if isinstance(s, basestring):
return to_unicode(s)
else:
return s
def to_utf8_if_string(s):
if isinstance(s, basestring):
return to_utf8(s)
else:
return s
def to_unicode_optional_iterator(x):
"""
Raise TypeError if x is a str containing non-utf8 bytes or if x is
an iterable which contains such a str.
"""
if isinstance(x, basestring):
return to_unicode(x)
try:
l = list(x)
except TypeError, e:
assert 'is not iterable' in str(e)
return x
else:
return [ to_unicode(e) for e in l ]
def to_utf8_optional_iterator(x):
"""
Raise TypeError if x is a str or if x is an iterable which
contains a str.
"""
if isinstance(x, basestring):
return to_utf8(x)
try:
l = list(x)
except TypeError, e:
assert 'is not iterable' in str(e)
return x
else:
return [ to_utf8_if_string(e) for e in l ]
def escape(s):
"""Escape a URL including any /."""
return urllib.quote(s.encode('utf-8'), safe='~')
def generate_timestamp():
"""Get seconds since epoch (UTC)."""
return int(time.time())
def generate_nonce(length=8):
"""Generate pseudorandom number."""
return ''.join([str(random.randint(0, 9)) for i in range(length)])
def generate_verifier(length=8):
"""Generate pseudorandom number."""
return ''.join([str(random.randint(0, 9)) for i in range(length)])
class Consumer(object):
"""A consumer of OAuth-protected services.
The OAuth consumer is a "third-party" service that wants to access
protected resources from an OAuth service provider on behalf of an end
user. It's kind of the OAuth client.
Usually a consumer must be registered with the service provider by the
developer of the consumer software. As part of that process, the service
provider gives the consumer a *key* and a *secret* with which the consumer
software can identify itself to the service. The consumer will include its
key in each request to identify itself, but will use its secret only when
signing requests, to prove that the request is from that particular
registered consumer.
Once registered, the consumer can then use its consumer credentials to ask
the service provider for a request token, kicking off the OAuth
authorization process.
"""
key = None
secret = None
def __init__(self, key, secret):
self.key = key
self.secret = secret
if self.key is None or self.secret is None:
raise ValueError("Key and secret must be set.")
def __str__(self):
data = {'oauth_consumer_key': self.key,
'oauth_consumer_secret': self.secret}
return urllib.urlencode(data)
class Token(object):
"""An OAuth credential used to request authorization or a protected
resource.
Tokens in OAuth comprise a *key* and a *secret*. The key is included in
requests to identify the token being used, but the secret is used only in
the signature, to prove that the requester is who the server gave the
token to.
When first negotiating the authorization, the consumer asks for a *request
token* that the live user authorizes with the service provider. The
consumer then exchanges the request token for an *access token* that can
be used to access protected resources.
"""
key = None
secret = None
callback = None
callback_confirmed = None
verifier = None
def __init__(self, key, secret):
self.key = key
self.secret = secret
if self.key is None or self.secret is None:
raise ValueError("Key and secret must be set.")
def set_callback(self, callback):
self.callback = callback
self.callback_confirmed = 'true'
def set_verifier(self, verifier=None):
if verifier is not None:
self.verifier = verifier
else:
self.verifier = generate_verifier()
def get_callback_url(self):
if self.callback and self.verifier:
# Append the oauth_verifier.
parts = urlparse.urlparse(self.callback)
scheme, netloc, path, params, query, fragment = parts[:6]
if query:
query = '%s&oauth_verifier=%s' % (query, self.verifier)
else:
query = 'oauth_verifier=%s' % self.verifier
return urlparse.urlunparse((scheme, netloc, path, params,
query, fragment))
return self.callback
def to_string(self):
"""Returns this token as a plain string, suitable for storage.
The resulting string includes the token's secret, so you should never
send or store this string where a third party can read it.
"""
data = {
'oauth_token': self.key,
'oauth_token_secret': self.secret,
}
if self.callback_confirmed is not None:
data['oauth_callback_confirmed'] = self.callback_confirmed
return urllib.urlencode(data)
@staticmethod
def from_string(s):
"""Deserializes a token from a string like one returned by
`to_string()`."""
if not len(s):
raise ValueError("Invalid parameter string.")
params = parse_qs(s, keep_blank_values=False)
if not len(params):
raise ValueError("Invalid parameter string.")
try:
key = params['oauth_token'][0]
except Exception:
raise ValueError("'oauth_token' not found in OAuth request.")
try:
secret = params['oauth_token_secret'][0]
except Exception:
raise ValueError("'oauth_token_secret' not found in "
"OAuth request.")
token = Token(key, secret)
try:
token.callback_confirmed = params['oauth_callback_confirmed'][0]
except KeyError:
pass # 1.0, no callback confirmed.
return token
def __str__(self):
return self.to_string()
def setter(attr):
name = attr.__name__
def getter(self):
try:
return self.__dict__[name]
except KeyError:
raise AttributeError(name)
def deleter(self):
del self.__dict__[name]
return property(getter, attr, deleter)
class Request(dict):
"""The parameters and information for an HTTP request, suitable for
authorizing with OAuth credentials.
When a consumer wants to access a service's protected resources, it does
so using a signed HTTP request identifying itself (the consumer) with its
key, and providing an access token authorized by the end user to access
those resources.
"""
version = OAUTH_VERSION
def __init__(self, method=HTTP_METHOD, url=None, parameters=None,
body='', is_form_encoded=False):
if url is not None:
self.url = to_unicode(url)
self.method = method
if parameters is not None:
for k, v in parameters.iteritems():
k = to_unicode(k)
v = to_unicode_optional_iterator(v)
self[k] = v
self.body = body
self.is_form_encoded = is_form_encoded
@setter
def url(self, value):
self.__dict__['url'] = value
if value is not None:
scheme, netloc, path, params, query, fragment = urlparse.urlparse(value)
# Exclude default port numbers.
if scheme == 'http' and netloc[-3:] == ':80':
netloc = netloc[:-3]
elif scheme == 'https' and netloc[-4:] == ':443':
netloc = netloc[:-4]
if scheme not in ('http', 'https'):
raise ValueError("Unsupported URL %s (%s)." % (value, scheme))
# Normalized URL excludes params, query, and fragment.
self.normalized_url = urlparse.urlunparse((scheme, netloc, path, None, None, None))
else:
self.normalized_url = None
self.__dict__['url'] = None
@setter
def method(self, value):
self.__dict__['method'] = value.upper()
def _get_timestamp_nonce(self):
return self['oauth_timestamp'], self['oauth_nonce']
def get_nonoauth_parameters(self):
"""Get any non-OAuth parameters."""
return dict([(k, v) for k, v in self.iteritems()
if not k.startswith('oauth_')])
def to_header(self, realm=''):
"""Serialize as a header for an HTTPAuth request."""
oauth_params = ((k, v) for k, v in self.items()
if k.startswith('oauth_'))
stringy_params = ((k, escape(str(v))) for k, v in oauth_params)
header_params = ('%s="%s"' % (k, v) for k, v in stringy_params)
params_header = ', '.join(header_params)
auth_header = 'OAuth realm="%s"' % realm
if params_header:
auth_header = "%s, %s" % (auth_header, params_header)
return {'Authorization': auth_header}
def to_postdata(self):
"""Serialize as post data for a POST request."""
d = {}
for k, v in self.iteritems():
d[k.encode('utf-8')] = to_utf8_optional_iterator(v)
# tell urlencode to deal with sequence values and map them correctly
# to resulting querystring. for example self["k"] = ["v1", "v2"] will
# result in 'k=v1&k=v2' and not k=%5B%27v1%27%2C+%27v2%27%5D
return urllib.urlencode(d, True).replace('+', '%20')
def to_url(self):
"""Serialize as a URL for a GET request."""
base_url = urlparse.urlparse(self.url)
try:
query = base_url.query
except AttributeError:
# must be python <2.5
query = base_url[4]
query = parse_qs(query)
for k, v in self.items():
query.setdefault(k, []).append(v)
try:
scheme = base_url.scheme
netloc = base_url.netloc
path = base_url.path
params = base_url.params
fragment = base_url.fragment
except AttributeError:
# must be python <2.5
scheme = base_url[0]
netloc = base_url[1]
path = base_url[2]
params = base_url[3]
fragment = base_url[5]
url = (scheme, netloc, path, params,
urllib.urlencode(query, True), fragment)
return urlparse.urlunparse(url)
def get_parameter(self, parameter):
ret = self.get(parameter)
if ret is None:
raise Error('Parameter not found: %s' % parameter)
return ret
def get_normalized_parameters(self):
"""Return a string that contains the parameters that must be signed."""
items = []
for key, value in self.iteritems():
if key == 'oauth_signature':
continue
# 1.0a/9.1.1 states that kvp must be sorted by key, then by value,
# so we unpack sequence values into multiple items for sorting.
if isinstance(value, basestring):
items.append((to_utf8_if_string(key), to_utf8(value)))
else:
try:
value = list(value)
except TypeError, e:
assert 'is not iterable' in str(e)
items.append((to_utf8_if_string(key), to_utf8_if_string(value)))
else:
items.extend((to_utf8_if_string(key), to_utf8_if_string(item)) for item in value)
# Include any query string parameters from the provided URL
query = urlparse.urlparse(self.url)[4]
url_items = self._split_url_string(query).items()
url_items = [(to_utf8(k), to_utf8(v)) for k, v in url_items if k != 'oauth_signature' ]
items.extend(url_items)
items.sort()
encoded_str = urllib.urlencode(items)
# Encode signature parameters per Oauth Core 1.0 protocol
# spec draft 7, section 3.6
# (http://tools.ietf.org/html/draft-hammer-oauth-07#section-3.6)
# Spaces must be encoded with "%20" instead of "+"
return encoded_str.replace('+', '%20').replace('%7E', '~')
def sign_request(self, signature_method, consumer, token):
"""Set the signature parameter to the result of sign."""
if not self.is_form_encoded:
# according to
# http://oauth.googlecode.com/svn/spec/ext/body_hash/1.0/oauth-bodyhash.html
# section 4.1.1 "OAuth Consumers MUST NOT include an
# oauth_body_hash parameter on requests with form-encoded
# request bodies."
self['oauth_body_hash'] = base64.b64encode(sha(self.body).digest())
if 'oauth_consumer_key' not in self:
self['oauth_consumer_key'] = consumer.key
if token and 'oauth_token' not in self:
self['oauth_token'] = token.key
self['oauth_signature_method'] = signature_method.name
self['oauth_signature'] = signature_method.sign(self, consumer, token)
@classmethod
def make_timestamp(cls):
"""Get seconds since epoch (UTC)."""
return str(int(time.time()))
@classmethod
def make_nonce(cls):
"""Generate pseudorandom number."""
return str(random.randint(0, 100000000))
@classmethod
def from_request(cls, http_method, http_url, headers=None, parameters=None,
query_string=None):
"""Combines multiple parameter sources."""
if parameters is None:
parameters = {}
# Headers
if headers and 'Authorization' in headers:
auth_header = headers['Authorization']
# Check that the authorization header is OAuth.
if auth_header[:6] == 'OAuth ':
auth_header = auth_header[6:]
try:
# Get the parameters from the header.
header_params = cls._split_header(auth_header)
parameters.update(header_params)
except:
raise Error('Unable to parse OAuth parameters from '
'Authorization header.')
# GET or POST query string.
if query_string:
query_params = cls._split_url_string(query_string)
parameters.update(query_params)
# URL parameters.
param_str = urlparse.urlparse(http_url)[4] # query
url_params = cls._split_url_string(param_str)
parameters.update(url_params)
if parameters:
return cls(http_method, http_url, parameters)
return None
@classmethod
def from_consumer_and_token(cls, consumer, token=None,
http_method=HTTP_METHOD, http_url=None, parameters=None,
body='', is_form_encoded=False):
if not parameters:
parameters = {}
defaults = {
'oauth_consumer_key': consumer.key,
'oauth_timestamp': cls.make_timestamp(),
'oauth_nonce': cls.make_nonce(),
'oauth_version': cls.version,
}
defaults.update(parameters)
parameters = defaults
if token:
parameters['oauth_token'] = token.key
if token.verifier:
parameters['oauth_verifier'] = token.verifier
return Request(http_method, http_url, parameters, body=body,
is_form_encoded=is_form_encoded)
@classmethod
def from_token_and_callback(cls, token, callback=None,
http_method=HTTP_METHOD, http_url=None, parameters=None):
if not parameters:
parameters = {}
parameters['oauth_token'] = token.key
if callback:
parameters['oauth_callback'] = callback
return cls(http_method, http_url, parameters)
@staticmethod
def _split_header(header):
"""Turn Authorization: header into parameters."""
params = {}
parts = header.split(',')
for param in parts:
# Ignore realm parameter.
if param.find('realm') > -1:
continue
# Remove whitespace.
param = param.strip()
# Split key-value.
param_parts = param.split('=', 1)
# Remove quotes and unescape the value.
params[param_parts[0]] = urllib.unquote(param_parts[1].strip('\"'))
return params
@staticmethod
def _split_url_string(param_str):
"""Turn URL string into parameters."""
parameters = parse_qs(param_str.encode('utf-8'), keep_blank_values=True)
for k, v in parameters.iteritems():
parameters[k] = urllib.unquote(v[0])
return parameters
class Client(httplib2.Http):
"""OAuthClient is a worker to attempt to execute a request."""
def __init__(self, consumer, token=None, cache=None, timeout=None,
proxy_info=None):
if consumer is not None and not isinstance(consumer, Consumer):
raise ValueError("Invalid consumer.")
if token is not None and not isinstance(token, Token):
raise ValueError("Invalid token.")
self.consumer = consumer
self.token = token
self.method = SignatureMethod_HMAC_SHA1()
httplib2.Http.__init__(self, cache=cache, timeout=timeout, proxy_info=proxy_info)
def set_signature_method(self, method):
if not isinstance(method, SignatureMethod):
raise ValueError("Invalid signature method.")
self.method = method
def request(self, uri, method="GET", body='', headers=None,
redirections=httplib2.DEFAULT_MAX_REDIRECTS, connection_type=None,
realm=None):
DEFAULT_POST_CONTENT_TYPE = 'application/x-www-form-urlencoded'
if not isinstance(headers, dict):
headers = {}
if method == "POST":
headers['Content-Type'] = headers.get('Content-Type',
DEFAULT_POST_CONTENT_TYPE)
is_form_encoded = \
headers.get('Content-Type') == 'application/x-www-form-urlencoded'
if is_form_encoded and body:
parameters = parse_qs(body)
else:
parameters = None
req = Request.from_consumer_and_token(self.consumer,
token=self.token, http_method=method, http_url=uri,
parameters=parameters, body=body, is_form_encoded=is_form_encoded)
req.sign_request(self.method, self.consumer, self.token)
schema, rest = urllib.splittype(uri)
if rest.startswith('//'):
hierpart = '//'
else:
hierpart = ''
host, rest = urllib.splithost(rest)
realm = realm if realm != None else schema + ':' + hierpart + host
if is_form_encoded:
body = req.to_postdata()
elif method == "GET":
uri = req.to_url()
else:
headers.update(req.to_header(realm=realm))
return httplib2.Http.request(self, uri, method=method, body=body,
headers=headers, redirections=redirections,
connection_type=connection_type)
class Server(object):
"""A skeletal implementation of a service provider, providing protected
resources to requests from authorized consumers.
This class implements the logic to check requests for authorization. You
can use it with your web server or web framework to protect certain
resources with OAuth.
"""
timestamp_threshold = 300 # In seconds, five minutes.
version = OAUTH_VERSION
signature_methods = None
def __init__(self, signature_methods=None):
self.signature_methods = signature_methods or {}
def add_signature_method(self, signature_method):
self.signature_methods[signature_method.name] = signature_method
return self.signature_methods
def verify_request(self, request, consumer, token):
"""Verifies an api call and checks all the parameters."""
self._check_version(request)
self._check_signature(request, consumer, token)
parameters = request.get_nonoauth_parameters()
return parameters
def build_authenticate_header(self, realm=''):
"""Optional support for the authenticate header."""
return {'WWW-Authenticate': 'OAuth realm="%s"' % realm}
def _check_version(self, request):
"""Verify the correct version of the request for this server."""
version = self._get_version(request)
if version and version != self.version:
raise Error('OAuth version %s not supported.' % str(version))
def _get_version(self, request):
"""Return the version of the request for this server."""
try:
version = request.get_parameter('oauth_version')
except:
version = OAUTH_VERSION
return version
def _get_signature_method(self, request):
"""Figure out the signature with some defaults."""
try:
signature_method = request.get_parameter('oauth_signature_method')
except:
signature_method = SIGNATURE_METHOD
try:
# Get the signature method object.
signature_method = self.signature_methods[signature_method]
except:
signature_method_names = ', '.join(self.signature_methods.keys())
raise Error('Signature method %s not supported try one of the following: %s' % (signature_method, signature_method_names))
return signature_method
def _get_verifier(self, request):
return request.get_parameter('oauth_verifier')
def _check_signature(self, request, consumer, token):
timestamp, nonce = request._get_timestamp_nonce()
self._check_timestamp(timestamp)
signature_method = self._get_signature_method(request)
try:
signature = request.get_parameter('oauth_signature')
except:
raise MissingSignature('Missing oauth_signature.')
# Validate the signature.
valid = signature_method.check(request, consumer, token, signature)
if not valid:
key, base = signature_method.signing_base(request, consumer, token)
raise Error('Invalid signature. Expected signature base '
'string: %s' % base)
def _check_timestamp(self, timestamp):
"""Verify that timestamp is recentish."""
timestamp = int(timestamp)
now = int(time.time())
lapsed = now - timestamp
if lapsed > self.timestamp_threshold:
raise Error('Expired timestamp: given %d and now %s has a '
'greater difference than threshold %d' % (timestamp, now,
self.timestamp_threshold))
class SignatureMethod(object):
"""A way of signing requests.
The OAuth protocol lets consumers and service providers pick a way to sign
requests. This interface shows the methods expected by the other `oauth`
modules for signing requests. Subclass it and implement its methods to
provide a new way to sign requests.
"""
def signing_base(self, request, consumer, token):
"""Calculates the string that needs to be signed.
This method returns a 2-tuple containing the starting key for the
signing and the message to be signed. The latter may be used in error
messages to help clients debug their software.
"""
raise NotImplementedError
def sign(self, request, consumer, token):
"""Returns the signature for the given request, based on the consumer
and token also provided.
You should use your implementation of `signing_base()` to build the
message to sign. Otherwise it may be less useful for debugging.
"""
raise NotImplementedError
def check(self, request, consumer, token, signature):
"""Returns whether the given signature is the correct signature for
the given consumer and token signing the given request."""
built = self.sign(request, consumer, token)
return built == signature
class SignatureMethod_HMAC_SHA1(SignatureMethod):
name = 'HMAC-SHA1'
def signing_base(self, request, consumer, token):
if not hasattr(request, 'normalized_url') or request.normalized_url is None:
raise ValueError("Base URL for request is not set.")
sig = (
escape(request.method),
escape(request.normalized_url),
escape(request.get_normalized_parameters()),
)
key = '%s&' % escape(consumer.secret)
if token:
key += escape(token.secret)
raw = '&'.join(sig)
return key, raw
def sign(self, request, consumer, token):
"""Builds the base signature string."""
key, raw = self.signing_base(request, consumer, token)
hashed = hmac.new(key, raw, sha)
# Calculate the digest base 64.
return binascii.b2a_base64(hashed.digest())[:-1]
class SignatureMethod_PLAINTEXT(SignatureMethod):
name = 'PLAINTEXT'
def signing_base(self, request, consumer, token):
"""Concatenates the consumer key and secret with the token's
secret."""
sig = '%s&' % escape(consumer.secret)
if token:
sig = sig + escape(token.secret)
return sig, sig
def sign(self, request, consumer, token):
key, raw = self.signing_base(request, consumer, token)
return raw
|
jgaul/python-oauth2
|
oauth2/__init__.py
|
Python
|
mit
| 29,092
|
# This file implements routines for extracting links from response objects.
import re
import lxml
import urlparse
import feedparser
# We have sought to disperse power, to set men and women free.
# That really means: to help them to discover that they are free.
# Everybody's free. The slave is free.
# The ultimate weapon isn't this plague out in Vegas, or any new super H-bomb.
# The ultimate weapon has always existed. Every man, every woman, and every child owns it.
# It's the ability to say No and take the consequences.
# 'Fear is failure.' 'The fear of death is the beginning of slavery.'
# "Thou hast no right but to do thy will.'
# The goose can break the bottle at any second.
# Socrates took the hemlock to prove it.
# Jesus went to the cross to prove it.
# It's in all history, all myth, all poetry.
# It's right out in the open all the time."
from goose import Goose
def extract_links(response):
urls = []
if ('content-type' in response.headers.keys()) and ('xml' in response.headers['content-type']):
f = feedparser.parse(response.text)
for entry in f.entries:
urls.append({entry.link: entry.title})
del f
else: # The following is a highly experimental feature.
url = urlparse.urlparse(response.url)
url = url.scheme + "://" + url.netloc
p = Parser(response.text, url=url)
urls = p.parse()
del url, p
return urls
class Parser(object):
"""
Build a list of relevant links from an HTML string and the root URL.
p = Parser(html_text, root_url)
urls = p.parse()
"""
def __init__(self,html=None,doc=None,url=None):
self.html=html
self.doc=doc
try: self.url = urlparse.urlparse(url).netloc
except: self.url = url
self.links=[]
def root_to_urls(self, doc, titles):
"""
Return a list of urls from an lxml root.
"""
if doc is None:
return []
a_tags = doc.xpath('//a')
# tries to find titles of link elements via tag text
if titles:
return [ (a.get('href'), a.text) for a in a_tags if a.get('href') ]
return [ a.get('href') for a in a_tags if a.get('href') ]
def get_urls(self,_input=None,titles=False,regex=False):
if (not _input) and (not self.html): return []
if not _input: _input = self.html
if regex:
text = re.sub('<[^<]+?>', ' ', _input)
text = re.findall('http[s]?://(?:[a-zA-Z]|[0-9]|[$-_@.&+]|[!*\(\),]|(?:%[0-9a-fA-F][0-9a-fA-F]))+', _input)
text = [i.strip() for i in _input]
return _input or []
if isinstance(_input, str) or isinstance(_input, unicode):
doc = self.fromstring(_input)
else:
doc = text
return self.root_to_urls(doc, titles)
def fromstring(self, html):
try:
self.doc = lxml.html.fromstring(html)
except Exception, e:
return None
return self.doc
def parse(self,html=None,url=None):
"""
Whittle a list of urls into things we're interested in.
"""
if self.links: self.links=[]
urls = self.get_urls(html)
if not urls: return urls
else: urls = set(urls)
if url: url = "http://%s/" % urlparse.urlparse(url).netloc
for u in urls:
if url:
if u == url: continue
if self.url:
if u == self.url: continue
if u.startswith('#'): continue
if not u.startswith('http'):
if url:
if (url[-1] == '/') and (u[0] == '/'): u = url + u[1:]
else: u = url+u
elif self.url:
if (self.url[-1] == '/') and (u[0] == '/'): u = self.url + u[1:]
else: u = self.url+u
else: continue
self.links.append(u)
return self.links
def extract_body(html):
"""
Extract the body text of a web page
"""
g = Goose({'enable_image_fetching':False})
article = g.extract(raw_html=html)
del g
return article.cleaned_text
def extract_title(html):
"""
Extract the body title of a web page
"""
g = Goose({'enable_image_fetching':False})
article = g.extract(raw_html=html)
del g
return article.title
def summarise(article):
stopnum = c = 0
for i,v in enumerate(article.split()):
if v.endswith('.'):
if c >= 2:
stopnum = i+1
break
else:
c += 1
return ' '.join(article.split()[:stopnum])
|
LukeB42/Emissary
|
emissary/controllers/parser.py
|
Python
|
mit
| 4,656
|
# coding: utf-8
"""
Talon.One API
The Talon.One API is used to manage applications and campaigns, as well as to integrate with your application. The operations in the _Integration API_ section are used to integrate with our platform, while the other operations are used to manage applications and campaigns. ### Where is the API? The API is available at the same hostname as these docs. For example, if you are reading this page at `https://mycompany.talon.one/docs/api/`, the URL for the [updateCustomerProfile][] operation is `https://mycompany.talon.one/v1/customer_profiles/id` [updateCustomerProfile]: #operation--v1-customer_profiles--integrationId--put # noqa: E501
The version of the OpenAPI document: 1.0.0
Generated by: https://openapi-generator.tech
"""
import pprint
import re # noqa: F401
import six
from talon_one.configuration import Configuration
class AccountAdditionalCost(object):
"""NOTE: This class is auto generated by OpenAPI Generator.
Ref: https://openapi-generator.tech
Do not edit the class manually.
"""
"""
Attributes:
openapi_types (dict): The key is attribute name
and the value is attribute type.
attribute_map (dict): The key is attribute name
and the value is json key in definition.
"""
openapi_types = {
'id': 'int',
'created': 'datetime',
'account_id': 'int',
'name': 'str',
'title': 'str',
'description': 'str',
'subscribed_applications_ids': 'list[int]'
}
attribute_map = {
'id': 'id',
'created': 'created',
'account_id': 'accountId',
'name': 'name',
'title': 'title',
'description': 'description',
'subscribed_applications_ids': 'subscribedApplicationsIds'
}
def __init__(self, id=None, created=None, account_id=None, name=None, title=None, description=None, subscribed_applications_ids=None, local_vars_configuration=None): # noqa: E501
"""AccountAdditionalCost - a model defined in OpenAPI""" # noqa: E501
if local_vars_configuration is None:
local_vars_configuration = Configuration()
self.local_vars_configuration = local_vars_configuration
self._id = None
self._created = None
self._account_id = None
self._name = None
self._title = None
self._description = None
self._subscribed_applications_ids = None
self.discriminator = None
self.id = id
self.created = created
self.account_id = account_id
self.name = name
self.title = title
self.description = description
if subscribed_applications_ids is not None:
self.subscribed_applications_ids = subscribed_applications_ids
@property
def id(self):
"""Gets the id of this AccountAdditionalCost. # noqa: E501
Unique ID for this entity. # noqa: E501
:return: The id of this AccountAdditionalCost. # noqa: E501
:rtype: int
"""
return self._id
@id.setter
def id(self, id):
"""Sets the id of this AccountAdditionalCost.
Unique ID for this entity. # noqa: E501
:param id: The id of this AccountAdditionalCost. # noqa: E501
:type: int
"""
if self.local_vars_configuration.client_side_validation and id is None: # noqa: E501
raise ValueError("Invalid value for `id`, must not be `None`") # noqa: E501
self._id = id
@property
def created(self):
"""Gets the created of this AccountAdditionalCost. # noqa: E501
The exact moment this entity was created. # noqa: E501
:return: The created of this AccountAdditionalCost. # noqa: E501
:rtype: datetime
"""
return self._created
@created.setter
def created(self, created):
"""Sets the created of this AccountAdditionalCost.
The exact moment this entity was created. # noqa: E501
:param created: The created of this AccountAdditionalCost. # noqa: E501
:type: datetime
"""
if self.local_vars_configuration.client_side_validation and created is None: # noqa: E501
raise ValueError("Invalid value for `created`, must not be `None`") # noqa: E501
self._created = created
@property
def account_id(self):
"""Gets the account_id of this AccountAdditionalCost. # noqa: E501
The ID of the account that owns this entity. # noqa: E501
:return: The account_id of this AccountAdditionalCost. # noqa: E501
:rtype: int
"""
return self._account_id
@account_id.setter
def account_id(self, account_id):
"""Sets the account_id of this AccountAdditionalCost.
The ID of the account that owns this entity. # noqa: E501
:param account_id: The account_id of this AccountAdditionalCost. # noqa: E501
:type: int
"""
if self.local_vars_configuration.client_side_validation and account_id is None: # noqa: E501
raise ValueError("Invalid value for `account_id`, must not be `None`") # noqa: E501
self._account_id = account_id
@property
def name(self):
"""Gets the name of this AccountAdditionalCost. # noqa: E501
The additional cost name that will be used in API requests and Talang. E.g. if `name == \"shipping\"` then you would set the shipping additional cost by including an `additionalCosts.shipping` property in your request payload. # noqa: E501
:return: The name of this AccountAdditionalCost. # noqa: E501
:rtype: str
"""
return self._name
@name.setter
def name(self, name):
"""Sets the name of this AccountAdditionalCost.
The additional cost name that will be used in API requests and Talang. E.g. if `name == \"shipping\"` then you would set the shipping additional cost by including an `additionalCosts.shipping` property in your request payload. # noqa: E501
:param name: The name of this AccountAdditionalCost. # noqa: E501
:type: str
"""
if self.local_vars_configuration.client_side_validation and name is None: # noqa: E501
raise ValueError("Invalid value for `name`, must not be `None`") # noqa: E501
self._name = name
@property
def title(self):
"""Gets the title of this AccountAdditionalCost. # noqa: E501
The human-readable name for the additional cost that will be shown in the Campaign Manager. Like `name`, the combination of entity and title must also be unique. # noqa: E501
:return: The title of this AccountAdditionalCost. # noqa: E501
:rtype: str
"""
return self._title
@title.setter
def title(self, title):
"""Sets the title of this AccountAdditionalCost.
The human-readable name for the additional cost that will be shown in the Campaign Manager. Like `name`, the combination of entity and title must also be unique. # noqa: E501
:param title: The title of this AccountAdditionalCost. # noqa: E501
:type: str
"""
if self.local_vars_configuration.client_side_validation and title is None: # noqa: E501
raise ValueError("Invalid value for `title`, must not be `None`") # noqa: E501
self._title = title
@property
def description(self):
"""Gets the description of this AccountAdditionalCost. # noqa: E501
A description of this additional cost. # noqa: E501
:return: The description of this AccountAdditionalCost. # noqa: E501
:rtype: str
"""
return self._description
@description.setter
def description(self, description):
"""Sets the description of this AccountAdditionalCost.
A description of this additional cost. # noqa: E501
:param description: The description of this AccountAdditionalCost. # noqa: E501
:type: str
"""
if self.local_vars_configuration.client_side_validation and description is None: # noqa: E501
raise ValueError("Invalid value for `description`, must not be `None`") # noqa: E501
self._description = description
@property
def subscribed_applications_ids(self):
"""Gets the subscribed_applications_ids of this AccountAdditionalCost. # noqa: E501
A list of the IDs of the applications that are subscribed to this additional cost # noqa: E501
:return: The subscribed_applications_ids of this AccountAdditionalCost. # noqa: E501
:rtype: list[int]
"""
return self._subscribed_applications_ids
@subscribed_applications_ids.setter
def subscribed_applications_ids(self, subscribed_applications_ids):
"""Sets the subscribed_applications_ids of this AccountAdditionalCost.
A list of the IDs of the applications that are subscribed to this additional cost # noqa: E501
:param subscribed_applications_ids: The subscribed_applications_ids of this AccountAdditionalCost. # noqa: E501
:type: list[int]
"""
self._subscribed_applications_ids = subscribed_applications_ids
def to_dict(self):
"""Returns the model properties as a dict"""
result = {}
for attr, _ in six.iteritems(self.openapi_types):
value = getattr(self, attr)
if isinstance(value, list):
result[attr] = list(map(
lambda x: x.to_dict() if hasattr(x, "to_dict") else x,
value
))
elif hasattr(value, "to_dict"):
result[attr] = value.to_dict()
elif isinstance(value, dict):
result[attr] = dict(map(
lambda item: (item[0], item[1].to_dict())
if hasattr(item[1], "to_dict") else item,
value.items()
))
else:
result[attr] = value
return result
def to_str(self):
"""Returns the string representation of the model"""
return pprint.pformat(self.to_dict())
def __repr__(self):
"""For `print` and `pprint`"""
return self.to_str()
def __eq__(self, other):
"""Returns true if both objects are equal"""
if not isinstance(other, AccountAdditionalCost):
return False
return self.to_dict() == other.to_dict()
def __ne__(self, other):
"""Returns true if both objects are not equal"""
if not isinstance(other, AccountAdditionalCost):
return True
return self.to_dict() != other.to_dict()
|
talon-one/talon_one.py
|
talon_one/models/account_additional_cost.py
|
Python
|
mit
| 10,814
|
# coding=utf-8
import inspect
def multiton(cls):
"""
Class decorator to make a class a multiton.
That is, there will be only (at most) one object existing for a given set
of initialization parameters.
"""
instances = {}
def getinstance(*args, **kwargs):
key = _gen_key(cls, *args, **kwargs)
if key not in instances:
instances[key] = cls(*args, **kwargs)
return instances[key]
return getinstance
kwd_mark = (object(), ) # seperate args and kwargs with a unique object
def _gen_key(cls, *args, **kwargs):
new_args, new_kwargs = _normalize_args(cls.__init__, *args, **kwargs)
key = new_args
if new_kwargs:
key += kwd_mark
sorted_items = sorted(new_kwargs.items())
for item in sorted_items:
key += item
return tuple(key)
def _normalize_args(func, *args, **kwargs):
try:
arg_names, _, _, arg_defaults = inspect.getargspec(func)
except AttributeError: # cls has no __init__
arg_names = ['self']
arg_defaults = ()
arg_names = arg_names[1:] # skip first arg (self)
if arg_defaults is None:
arg_defaults = ()
new_args = []
new_kwargs = {}
# match named args to names
for name, arg in zip(arg_names, args):
new_kwargs[name] = arg
# handle extra args from *
if len(args) > len(arg_names):
for arg in args[len(arg_names):]:
new_args.append(arg)
# or fill in default values
else:
for name, default in zip(arg_names[len(args):], arg_defaults):
new_kwargs[name] = default
# merge remaining **kwargs
new_kwargs.update(kwargs)
return new_args, new_kwargs
|
sait-berkeley-infosec/pynessus-api
|
nessusapi/utils.py
|
Python
|
mit
| 1,726
|
#!/usr/bin/python2.7
import os
import fileinput
# the version of the release
with open("version.txt") as f: version = f.read()
def getVersionTuple(v):
return tuple(map(int, (v.split("."))))
version_major = str(getVersionTuple(version)[0])
version_minor = str(getVersionTuple(version)[1])
version_patch = str(getVersionTuple(version)[2])
# update version in the header file
print("updating the version in the header file")
doctest_contents = ""
for line in fileinput.input(["../doctest/parts/doctest_fwd.h"]):
if line.startswith("#define DOCTEST_VERSION_MAJOR "):
doctest_contents += "#define DOCTEST_VERSION_MAJOR " + version_major + "\n"
elif line.startswith("#define DOCTEST_VERSION_MINOR "):
doctest_contents += "#define DOCTEST_VERSION_MINOR " + version_minor + "\n"
elif line.startswith("#define DOCTEST_VERSION_PATCH "):
doctest_contents += "#define DOCTEST_VERSION_PATCH " + version_patch + "\n"
else:
doctest_contents += line
readme = open("../doctest/parts/doctest_fwd.h", "w")
readme.write(doctest_contents)
readme.close()
# update meson file with version
meson_contents = ""
for line in fileinput.input(["../meson.build"]):
if line.startswith("project('doctest'"):
meson_contents += "project('doctest', ['cpp'], version: '" + version + "', meson_version:'>=0.50')\n"
else:
meson_contents += line
meson = open("../meson.build", "w")
meson.write(meson_contents)
meson.close()
# run generate_html.py
print("generating html documentation from markdown")
os.system("python generate_html.py")
|
onqtam/doctest
|
scripts/update_stuff.py
|
Python
|
mit
| 1,583
|
#! /usr/bin/python
# -*- coding: utf-8 -*-
#
from application import module
from application import agent as _agent
from libraries import irclib
import threading
import time
import sys
import Queue
class Module(module.AbstractModule):
thread = None
channels = None
targets = None
__password = None
def __init__(self, application):
super(Module, self).__init__(application)
self.channels = {}
self.targets = {}
def initialize(self):
self.__password = self.configuration.get("interaction", "password", "")
###
self.thread = _IRCThread(self)
self.thread.start()
# Registers global 'ircgate' function to RPC agents
for agent in self.application.agents.values():
if isinstance(agent, _agent.AbstractRPCAgent):
callback = lambda source, target, message, password = None: self.application.api.message(source, Message(target, message, password))
agent.register_function(callback, "ircgate")
# Loads targets configuration
for section in self.configuration.sections():
target_channels = []
if section.startswith("target."):
target = section[len("target."):]
items = self.configuration.items(section)
for key, item in items:
if item[0] == "#":
self.channels[item] = True
target_channels.append(item)
self.targets[target] = target_channels
def uninitialize(self):
self.thread.disconnect_event.set()
self.thread.done_event.wait(10)
def handle(self, message):
if isinstance(message.content, Message) and self.is_password_done(message.content.password):
self.thread.queue.put(message.content, True, 10)
def is_password_done(self, password):
return (self.__password and (password == self.__password)) or not self.__password
class Message:
target = None
content = None
password = None
def __init__(self, target, content, password = None):
self.target = target
self.content = content
self.password = password
class _IRCThread(threading.Thread):
disconnect_event = threading.Event()
done_event = threading.Event()
queue = None
parent = None
connection = None
irc = None
def __init__(self, parent):
threading.Thread.__init__(self)
self.setDaemon(True)
self.parent = parent
queue_size = self.parent.application.configuration.getint("queue", "size", 100000)
self.queue = Queue.Queue(queue_size)
def run(self):
pooling_interval = self.parent.application.pooling_interval
###
self.irc = irclib.IRC()
self.connect()
self.connection.add_global_handler("welcome", self.on_connect)
self.connection.add_global_handler("disconnect", self.on_disconnect)
self.connection.add_global_handler("privmsg", self.on_privmsg)
self.connection.add_global_handler("pubmsg", self.on_pubmsg)
while not self.disconnect_event.is_set():
self.irc.process_once()
while not self.queue.empty():
try:
message = self.queue.get(False)
self.send_targets(message.target, message.content)
except Queue.Empty:
break
time.sleep(pooling_interval)
###
if self.parent.application.reloading:
quit_message = self.parent.configuration.get("interaction", "reloading_message", "Reloading.")
else:
quit_message = None
self.disconnect(quit_message)
def send_targets(self, targets, message):
if isinstance(targets, str):
targets = [targets]
for target in targets:
if target in self.parent.targets:
items = self.parent.targets[target]
for item in items:
if ((item in self.parent.channels) and self.parent.channels[item]) or not (item in self.parent.channels):
self.connection.privmsg(item, message)
def join(self, channels):
if isinstance(channels, str):
channels = [channels]
for channel in channels:
self.connection.join(channel)
def disconnect(self, message = None):
if message is None:
quit_message = self.parent.configuration.get("interaction", "quit_message", "Bye!")
else:
quit_message = message
try:
self.connection.quit(quit_message)
self.connection.disconnect()
finally:
self.done_event.set()
def connect(self):
configuration = self.parent.configuration
server = configuration.get("connection", "server", "irc.freenode.net")
port = configuration.getint("connection", "port", 6667)
username = configuration.get("connection", "username", "ircbot")
password = configuration.get("connection", "password", "")
ipv6 = configuration.getboolean("connection", "ipv6", False)
ssl = configuration.getboolean("connection", "ssl", False)
###
while True:
try:
self.connection = self.irc.server().connect(server, port, username, ipv6 = ipv6, ssl = ssl)
break
except irclib.ServerConnectionError:
sys.stderr.write("Connection failed. Sleeping 20 seconds and then reconnect.\n")
time.sleep(20)
###
def on_connect(self, connection, event):
self.join(self.parent.channels)
def on_disconnect(self, connection, event):
if not self.disconnect_event.is_set():
self.connect()
def on_pubmsg(self, connection, event):
message = event.arguments()[0].strip().split(":", 1)
target_user = message[0].strip()
if target_user == connection.get_nickname():
message = message[1].strip()
else:
return
message = message.split(" ", 1)
source = event.source().split("!", 1)[0].strip()
target = event.target()
###
try:
if message[0] == "off":
# password = None
#
# if len(message) >= 2:
# password = message[1].strip()
if True: #self.parent.is_password_done(password):
self.parent.channels[target] = False
self.connection.privmsg(target, "%s: OK, off" % (source))
elif message[0] == "on":
# password = None
#
# if len(message) >= 2:
# password = message[1].strip()
if True: #self.parent.is_password_done(password):
self.parent.channels[target] = True
self.connection.privmsg(target, "%s: OK, on" % (source))
elif message[0] == "status":
if self.parent.channels[target]:
status = "on"
else:
status = "off"
self.connection.privmsg(target, "%s: %s" % (source, status))
except:
sys.stderr.write(source, "FAILED, exception: %s\n" % str(sys.exc_value))
def on_privmsg(self, connection, event):
if event.target() == connection.get_nickname():
message = event.arguments()[0].strip().split(" ", 1)
source = event.source().split("!", 1)[0].strip()
###
try:
if message[0] == "shutdown":
password = None
if len(message) >= 2:
password = message[1].strip()
module = self.parent.application.loads.modules["system"]
bus_message = module.Message(module.Message.SHUTDOWN, password)
self.parent.application.api.message(self.parent, bus_message)
self.connection.privmsg(source, "OK, i'm going to shutdown")
elif message[0] == "reload":
password = None
if len(message) >= 2:
password = message[1].strip()
module = self.parent.application.loads.modules["system"]
bus_message = module.Message(module.Message.RELOAD, password)
self.parent.application.api.message(self.parent, bus_message)
self.connection.privmsg(source, "OK, i'm going to reload")
except:
self.connection.privmsg(source, "FAILED, exception: " + str(sys.exc_value))
###
def factory(application):
return Module(application)
|
martinkozak/pyircgate-daemon
|
modules/ircgate/ircgate.py
|
Python
|
mit
| 7,493
|
with open("input.txt") as f:
data = f.read()
bots = 1000
inputs = [[False] for i in range(bots)]
outputs = [[] for i in range(bots)]
maps = [[] for i in range(bots)]
for line in data.splitlines():
if line.startswith("value"):
p = line.split()
val = int(p[1])
bot = int(p[-1])
inputs[bot].append(val)
elif line.startswith("bot"):
p = line.split()
fbot = int(p[1])
lowtype = p[5]
tolow = int(p[6])
hightype = p[-2]
tohigh = int(p[-1])
maps[fbot].append((lowtype, tolow))
maps[fbot].append((hightype, tohigh))
while True:
found = False
for bot_id, input_set in enumerate(inputs):
if not input_set[0] and len(input_set) == 3:
found = True
input_set[0] = True
low,high = sorted(input_set[1:])
if low == 17 and high == 61:
print("PART 1: {}".format(bot_id))
low_target,high_target = maps[bot_id]
if low_target[0] == "bot":
inputs[low_target[1]].append(low)
else:
outputs[low_target[1]].append(low)
if high_target[0] == "bot":
inputs[high_target[1]].append(high)
else:
outputs[high_target[1]].append(high)
#print("Bot {} gives {} to {} and {} to {}".format(bot_id, low, low_target, high, high_target))
if not found:
break
res = outputs[0][0] * outputs[1][0] * outputs[2][0]
print("PART 2: {}".format(res))
|
lamperi/aoc
|
2016/10/solve.py
|
Python
|
mit
| 1,547
|
import sys
import warnings
def saveit(func):
"""A decorator that caches the return value of a function"""
name = '_' + func.__name__
def _wrapper(self, *args, **kwds):
if not hasattr(self, name):
setattr(self, name, func(self, *args, **kwds))
return getattr(self, name)
return _wrapper
cacheit = saveit
def prevent_recursion(default):
"""A decorator that returns the return value of `default` in recursions"""
def decorator(func):
name = '_calling_%s_' % func.__name__
def newfunc(self, *args, **kwds):
if getattr(self, name, False):
return default()
setattr(self, name, True)
try:
return func(self, *args, **kwds)
finally:
setattr(self, name, False)
return newfunc
return decorator
def ignore_exception(exception_class):
"""A decorator that ignores `exception_class` exceptions"""
def _decorator(func):
def newfunc(*args, **kwds):
try:
return func(*args, **kwds)
except exception_class:
pass
return newfunc
return _decorator
def deprecated(message=None):
"""A decorator for deprecated functions"""
def _decorator(func, message=message):
if message is None:
message = '%s is deprecated' % func.__name__
def newfunc(*args, **kwds):
warnings.warn(message, DeprecationWarning, stacklevel=2)
return func(*args, **kwds)
return newfunc
return _decorator
def cached(size):
"""A caching decorator based on parameter objects"""
def decorator(func):
cached_func = _Cached(func, size)
return lambda *a, **kw: cached_func(*a, **kw)
return decorator
class _Cached(object):
def __init__(self, func, count):
self.func = func
self.cache = []
self.count = count
def __call__(self, *args, **kwds):
key = (args, kwds)
for cached_key, cached_result in self.cache:
if cached_key == key:
return cached_result
result = self.func(*args, **kwds)
self.cache.append((key, result))
if len(self.cache) > self.count:
del self.cache[0]
return result
def resolve(str_or_obj):
"""Returns object from string"""
from rope.base.utils.pycompat import string_types
if not isinstance(str_or_obj, string_types):
return str_or_obj
if '.' not in str_or_obj:
str_or_obj += '.'
mod_name, obj_name = str_or_obj.rsplit('.', 1)
__import__(mod_name)
mod = sys.modules[mod_name]
return getattr(mod, obj_name) if obj_name else mod
|
zrzka/blackmamba
|
blackmamba/lib/rope/base/utils/__init__.py
|
Python
|
mit
| 2,730
|
import sys, logging
logging.basicConfig(level=15, stream=sys.stderr, format="%(levelname)1s:%(filename)10s:%(lineno)3d:%(message)s")
# make log level names shorter so that we can show them
logging.addLevelName(50, 'C')
logging.addLevelName(40, 'E')
logging.addLevelName(30, 'W')
logging.addLevelName(20, 'I')
logging.addLevelName(10, 'D')
logging.info("test1")
import jpype
jpype.startJVM(convertStrings=False)
IPluginAtom = jpype.JClass("at.ac.tuwien.kr.hexlite.api.IPluginAtom")
ISolverContext = jpype.JClass("at.ac.tuwien.kr.hexlite.api.ISolverContext")
IAtom = jpype.JClass("at.ac.tuwien.kr.hexlite.api.IAtom")
ISymbol = jpype.JClass("at.ac.tuwien.kr.hexlite.api.ISymbol")
@jpype.JImplements(ISymbol)
class JavaSymbolImpl:
def __init__(self, what):
self.what = what
@jpype.JOverride
def getType(self):
return ISymbol.Type.CONSTANT
@jpype.JOverride
def getName(self):
return self.what
@jpype.JOverride
def getInteger(self):
return 4711
@jpype.JOverride
def getArguments(self):
return []
@jpype.JOverride
def getTuple(self):
return [self.getName()]
@jpype.JOverride
def hashCode(self):
return jpype.JInt(hash(self.what) & 0x7FFFFFFF)
@jpype.JOverride
def equals(self, other):
if self == other:
return True
elif self.what == other.what:
return True
else:
return False
@jpype.JImplements(IPluginAtom.IQuery)
class JavaQueryImpl:
def __init__(self):
pass
@jpype.JOverride
def getInterpretation(self):
logging.warning("TBD")
return None
@jpype.JOverride
def getInput(self):
ret = jpype.JClass("java.util.ArrayList")()
ret.add(JavaSymbolImpl('foo'))
ret.add(JavaSymbolImpl('bar'))
return ret
@jpype.JImplements(ISolverContext)
class JavaSolverContextImpl:
def __init__(self):
pass
@jpype.JOverride
def storeOutputAtom(self, atom):
logging.warning("TBD")
return jpype.JObject(None, IAtom)
@jpype.JOverride
def storeAtom(self, atom):
logging.warning("TBD")
return None
@jpype.JOverride
def storeConstant(self, s):
logging.warning("TBD store %s", s)
return JavaSymbolImpl('stored({})'.format(s))
@jpype.JOverride
def learn(self, nogood):
logging.warning("TBD")
def jmain():
logging.info("test2")
jls = jpype.JClass("java.lang.System")
jls.out.println("i am printing java.class.path")
print(jls.getProperty("java.class.path"))
JStringPlugin = jpype.JClass("at.ac.tuwien.kr.hexlite.stringplugin.StringPlugin")
logging.info("got JStringPlugin %s", JStringPlugin)
splugin = JStringPlugin()
logging.info("got splugin %s", splugin)
jatoms = splugin.createAtoms()
logging.info("got atoms %s", jatoms)
jconcat = jatoms[0]
jcontext = JavaSolverContextImpl()
jquery = JavaQueryImpl()
janswer = jconcat.retrieve(jcontext, jquery)
logging.info("answer is %s", janswer)
jpype.shutdownJVM()
logging.info("done")
def main():
try:
jmain()
except jpype.JClass("java.lang.Exception") as ex:
logging.error("exception: %s", ex.toString())
st = ex.getStackTrace()
for ste in st:
logging.error("\t at %s", ste.toString())
#sb.append(ex.getClass().getName() + ": " + ex.getMessage() + "\n");
main()
|
hexhex/hexlite
|
java-api/src/test/python/test-jpype.py
|
Python
|
mit
| 3,537
|
#!/usr/bin/env python
import os
import sys
import time
from subprocess import CalledProcessError, check_output, STDOUT
certs_dir = '{{ letsencrypt_certs_dir }}'
failed = False
sites = {{ sites }}
sites = (k for k, v in sites.items() if 'ssl' in v and v['ssl'].get('enabled', False) and v['ssl'].get('provider', 'manual') == 'letsencrypt')
for site in sites:
cert_path = os.path.join(certs_dir, site + '.cert')
bundled_cert_path = os.path.join(certs_dir, site + '-bundled.cert')
if os.access(cert_path, os.F_OK):
stat = os.stat(cert_path)
print 'Certificate file ' + cert_path + ' already exists'
if time.time() - stat.st_mtime < {{ letsencrypt_min_renewal_age }} * 86400:
print ' The certificate is younger than {{ letsencrypt_min_renewal_age }} days. Not creating a new certificate.\n'
continue
print 'Generating certificate for ' + site
cmd = ('/usr/bin/env python {{ acme_tiny_software_directory }}/acme_tiny.py '
'--ca {{ letsencrypt_ca }} '
'--account-key {{ letsencrypt_account_key }} '
'--csr {{ acme_tiny_data_directory }}/csrs/{0}.csr '
'--acme-dir {{ acme_tiny_challenges_directory }}'
).format(site)
try:
cert = check_output(cmd, stderr=STDOUT, shell=True)
except CalledProcessError as e:
failed = True
print 'Error while generating certificate for ' + site
print e.output
else:
with open(cert_path, 'w') as cert_file:
cert_file.write(cert)
with open('{{ letsencrypt_intermediate_cert_path }}') as intermediate_cert_file:
intermediate_cert = intermediate_cert_file.read()
with open(bundled_cert_path, 'w') as bundled_file:
bundled_file.write(''.join([cert, intermediate_cert]))
print 'Created certificate for ' + site
if failed:
sys.exit(1)
|
afonsoduarte/ansible-stacey
|
roles/letsencrypt/templates/renew-certs.py
|
Python
|
mit
| 1,905
|
# coding: utf-8
from django.conf.urls import patterns, url
urlpatterns = patterns(
'todo.core.views',
url(r'^user/list/$', 'user_list', name="user_list"),
url(r'^user/form/$', 'user_create', name="user_form"),
url(r'^user/form/(?P<pk>\d+)/$', 'user_update', name="user_form"),
url(r'^user/delete/(?P<pk>\d+)/$', 'user_delete', name="user_delete"),
)
|
zokis/TODO-agenda
|
todo/todo/core/urls.py
|
Python
|
mit
| 372
|
from . import pos_longpolling_controller
|
it-projects-llc/pos-addons
|
pos_longpolling/controllers/__init__.py
|
Python
|
mit
| 41
|
__author__ = 'vialette'
|
vialette/ultrastorage
|
ultrastorage/tools/__init__.py
|
Python
|
mit
| 24
|
from graph_tool.all import *
from sets import Set
import random
from geopy import geocoders, distance
from decimal import *
def randomize(iterable, bufsize=1000):
''' generator that randomizes an iterable. space: O(bufsize). time: O(n+bufsize). '''
buf = [None] * bufsize
for x in iterable:
i = random.randrange(bufsize)
if buf[i] is not None: yield buf[i]
buf[i] = x
for x in buf:
if x is not None: yield x
g = Graph()
g = load_graph("finalNodes.dot", "dot")
vprop_name = g.vertex_properties["nameP"] #new_vertex_property("string")
vprop_gps = g.vertex_properties["gpsP"] #g.new_vertex_property("vector<float>")
g.edge_properties["busP"] = g.new_edge_property("string")
g.edge_properties["departureP"] = g.new_edge_property("int")
g.edge_properties["travelTimeP"] = g.new_edge_property("int")
eprop_bus = g.edge_properties["busP"] #g.new_edge_property("string")
eprop_departure = g.edge_properties["departureP"] #g.new_edge_property("vector<int>")
eprop_travelTime = g.edge_properties["travelTimeP"] #g.new_edge_property("int") #"Weight"
## Todo: User interface
startAddress = raw_input('What is your starting destination? ')
endAddress = raw_input('What is your ending destination? ')
leaveTime = raw_input('What is your departure time (in 24-h format? ')
lt = leaveTime.split(':')
finalTime = int(lt[0]*60) + int(lt[1])
geo = geocoders.Google('YOUR_API_KEY')
sPlace, sGPS = geo.geocode(startAddress)
ePlace, eGPS = geo.geocode(endAddress)
closestS = -1
closestE = -1
shortestS = 1000.0
shortestE = 1000.0
for v in g.vertices():
coord = vprop_gps[v].split(', ')
sDist = distance.distance(sGPS, [float.fromhex(coord[0]), float.fromhex(coord[1])]).miles
eDist = distance.distance(eGPS, [float.fromhex(coord[0]), float.fromhex(coord[1])]).miles
if sDist < shortestS:
shortestS = sDist
closestS = v
if eDist < shortestE:
shortestE = eDist
closestE = v
print vprop_name[closestS]
print vprop_name[closestE]
START_TIME = finalTime
END_TIME = START_TIME+120
START_NODE = closestS
END_NODE = closestE
#TODO: Better way to add edges
# Specifically, read the file myself
f = open("busEdgesData.txt", "r")
d = f.readlines()
for line in d:
line = line.split(" ")
vData = line[0].split("->")
newSplit = line[1].split("\"")
bus = newSplit[1] # Bus number
times = newSplit[3]
travelTime = newSplit[5]
times = eval('['+times+']')
for t in times:
if t >= START_TIME and t <= END_TIME:
newE = g.add_edge(g.vertex(vData[0]), g.vertex(vData[1]))
eprop_bus[newE] = bus
eprop_departure[newE] = t
eprop_travelTime[newE] = int(travelTime)
successRoutes = []
for rider in range(150):
currentNode = g.vertex(START_NODE)
currentBus = "-1"
currentRoute = []
currentRoute.append([vprop_name[currentNode], "", START_TIME, 0])
print "Trial " + str(rider)
for choice in range(100):
if currentNode == g.vertex(END_NODE):
currentRoute.append(vprop_name[currentNode])
successRoutes.append(currentRoute)
break
else:
coin = random.randint(0,1)
if coin == 0 and currentBus != "-1":
# continue on bus
for e in currentNode.out_edges():
lastTry = currentRoute.pop()
currentRoute.append(lastTry)
#print eprop_bus[e] + " vs " + currentBus
#print str(eprop_departure[e]) + " vs " + str(lastTry[2] + lastTry[3])
if eprop_bus[e] == currentBus and eprop_departure[e] == (lastTry[2] + lastTry[3]):
currentNode = e.target()
currentRoute.append([vprop_name[currentNode], eprop_bus[e], eprop_departure[e], eprop_travelTime[e]])
break
else:
# try a different bus
found = -1
for e in randomize(currentNode.out_edges(), 1000):
lastTry = currentRoute.pop()
currentRoute.append(lastTry)
if eprop_departure[e] >= (lastTry[2] + lastTry[3]):
newE = e
found = 1
if found != -1:
currentBus = eprop_bus[newE]
currentNode = newE.target()
currentRoute.append([vprop_name[currentNode], eprop_bus[newE], eprop_departure[newE], eprop_travelTime[newE]])
#print currentRoute
threeShortest = []
thirdShortest = 10000
for r in successRoutes:
end = r.pop()
next = r.pop()
length = int(next[2])
r.append(next)
r.append(end)
if length < thirdShortest:
thirdShortest = length
if len(threeShortest) == 3:
threeShortest.pop()
threeShortest.append(r)
else:
threeShortest.append(r)
for r in threeShortest:
print "Attempt:"
print r
print str(len(successRoutes)) + "/" + str(100)
|
philipbjorge/WTA-Bus-Routing
|
WTA App.py
|
Python
|
mit
| 4,459
|
# -*- coding: utf-8 -*-
from __future__ import unicode_literals, absolute_import
from django.contrib.auth.models import AbstractUser
from django.core.urlresolvers import reverse
from django.db import models
from django.utils.encoding import python_2_unicode_compatible
from django.utils.translation import ugettext_lazy as _
import uuid
@python_2_unicode_compatible
class User(AbstractUser):
id = models.UUIDField(primary_key=True, default=uuid.uuid4, editable=False)
org_id = models.IntegerField(_('Organization ID'), default=-1, blank=False)
name = models.CharField(_('Name of User'), blank=False, max_length=255)
phone = models.CharField(_('Phone Number'), max_length=10, blank=True)
def __str__(self):
return self.username
def get_absolute_url(self):
return reverse('users:detail', kwargs={'username': self.username})
|
luftdanmark/fifo.li
|
fifo/users/models.py
|
Python
|
mit
| 883
|
#!/usr/bin/env python3
import os
import argparse
import sys
from time import sleep
import subprocess
import imp
import numpy as np
import pandas as pd
import matplotlib.pyplot as plt
import matplotlib
import datetime
import pickle
matplotlib.style.use('fivethirtyeight')
# print(plt.style.available)
# mypath = os.environ["PATH"]
# os.environ["PATH"] = "/home/wl45/python/bin:/home/wl45/opt:" + mypath
# my_env = os.environ.copy()
parser = argparse.ArgumentParser(
description="Plot my graphs quickly")
parser.add_argument("-s", "--save", action="store_true", default=False)
parser.add_argument("-r", "--reproduce", default=None)
parser.add_argument("-t", "--test", action="store_true", default=False)
parser.add_argument("-m", "--mode", default="pulling")
parser.add_argument("-d", "--debug", action="store_true", default=False)
args = parser.parse_args()
if(args.debug):
do = print
cd = print
else:
do = os.system
cd = os.chdir
if args.reproduce is not None:
print("Reproducing!")
with open(args.reproduce, "rb") as f:
args = pickle.load(f)
print(args)
args.save = False
if(args.test):
print("Hello Test World")
force_list = np.arange(1,2.5,0.1)
array = []
cwd = os.getcwd()
print(cwd)
for force in force_list:
folder = "2d_2_force_" + str(force)
cd(folder)
do("pulling_plotcontour.py")
do("cp test.png ../results/{}.png".format(force))
cd(cwd)
if(args.save):
# print(os.getcwd())
# print(args)
print("Saving")
# print(datetime.datetime.now().strftime("%Y%m%d-%H%M%S"))
with open("args"+datetime.datetime.now().strftime("%Y%m%d-%H%M"), "wb") as f:
pickle.dump(args, f)
os.system("cp ~/opt/plot.py plot_{}.py".format(datetime.datetime.now().strftime("%Y%m%d-%H%M")))
if(args.mode == 1):
do("pulling_prepare.py --data -m 6")
do("pulling_glue.py --move -m 4")
folder = "wham"
do("mkdir " + folder)
cd(folder)
do("make_metadata.py -k 200 -a 2 -t")
do("pulling_analysis.py -f -m 8")
do("sbatch freeEnergy.slurm")
|
luwei0917/awsemmd_script
|
pulling.py
|
Python
|
mit
| 2,102
|
import numpy as np
import cv2
import pickle
import os
import matplotlib.pyplot as plt
class cameraCalib():
def __init__(self, calib_image_path = 'camera_cal/'):
self.mtx = None
self.dist = None
self.calib_image_path = calib_image_path
self.calib_file = self.calib_image_path + 'camera_param.pickle'
if not os.path.isfile(self.calib_file):
calib_param = self.calc_calibration_parameters(self.calib_image_path, 6, 9, display_corners=False)
self.mtx = calib_param['mtx']
self.dist = calib_param['dist']
with open(self.calib_file, 'wb') as f:
pickle.dump(calib_param, file=f)
else:
with open(self.calib_file, "rb") as f:
calib_param = pickle.load(f)
self.mtx = calib_param['mtx']
self.dist = calib_param['dist']
print('Loaded calibration parameters from disk....')
def calc_calibration_parameters(self, calib_im_path, rows, cols, display_corners=False):
"""Compute calibration parameters from a set of calibration images.
Params:
calib_im_path: Directory of calibration images.
row: checkerboard row number
col: checkerboard col number
Return:
calib_param = {'objpoints': objpoints,
'imgpoints': imgpoints,
'mtx': mtx,
'dist': dist,
'rvecs': rvecs,
'tvecs': tvecs}
"""
# Object / image points collections.
objpoints = []
imgpoints = []
# Calibration points from images.
filenames = os.listdir(calib_im_path)
for fname in filenames:
img = cv2.imread(calib_im_path + fname)
gray = cv2.cvtColor(img, cv2.COLOR_BGR2GRAY)
# Theoretical Grid.
objp = np.zeros((rows * cols, 3), np.float32)
objp[:, :2] = np.mgrid[0:cols, 0:rows].T.reshape(-1, 2)
# Corners in the image.
ret, corners = cv2.findChessboardCorners(gray, (cols, rows), None)
if ret:
objpoints.append(objp)
imgpoints.append(corners)
if display_corners:
img = cv2.drawChessboardCorners(img, (cols, rows), corners, ret)
plt.imshow(img)
plt.show()
else:
print('Warning! Not chessboard found in image', fname)
# Calibration from image points.
ret, mtx, dist, rvecs, tvecs = cv2.calibrateCamera(objpoints,
imgpoints,
img.shape[0:2],
None, None)
calib_param = {'mtx': mtx,
'dist': dist}
return calib_param
def get_camera_parameters(self):
return self.mtx, self.dist
def undistort_image(self, img):
"""Undistort an image.
"""
return cv2.undistort(img, self.mtx, self.dist, None, self.mtx)
def show_undistorted_images(self, org, undist):
"""
Show orginal and undistorted images
:param org: The original image
:param undist: The undistorted image
:return:
"""
orgr = cv2.cvtColor(org, cv2.COLOR_BGR2RGB)
undistr = cv2.cvtColor(undist, cv2.COLOR_BGR2RGB)
f, (ax1, ax2) = plt.subplots(1, 2, figsize=(20, 10))
ax1.imshow(orgr)
ax1.set_title('Original Image', fontsize=30)
ax2.imshow(undistr)
ax2.set_title('Undistorted Image', fontsize=30)
plt.show()
def test_calibration(self, fname):
"""Test calibration on an example chessboard, and display the result.
"""
# Load image, draw chessboard and undistort.
img = cv2.imread(fname)
img_undist = self.undistort_image(img)
self.show_undistorted_images(img, img_undist)
|
sridhar912/Self-Driving-Car-NanoDegree
|
CarND-Advanced-Lane-Lines/CameraCalibration.py
|
Python
|
mit
| 4,101
|
# ./_sac.py
# -*- coding: utf-8 -*-
# PyXB bindings for NM:bd794131cb7c2b1e52ff4e6220a49c5d8509c55c
# Generated 2015-02-11 21:35:49.975586 by PyXB version 1.2.4 using Python 2.6.9.final.0
# Namespace urn:oasis:names:specification:ubl:schema:xsd:SignatureAggregateComponents-2 [xmlns:sac]
from __future__ import unicode_literals
import pyxb
import pyxb.binding
import pyxb.binding.saxer
import io
import pyxb.utils.utility
import pyxb.utils.domutils
import sys
import pyxb.utils.six as _six
# Unique identifier for bindings created at the same time
_GenerationUID = pyxb.utils.utility.UniqueIdentifier('urn:uuid:2b2e2fd1-b225-11e4-b26c-14109fe53921')
# Version of PyXB used to generate the bindings
_PyXBVersion = '1.2.4'
# Generated bindings are not compatible across PyXB versions
if pyxb.__version__ != _PyXBVersion:
raise pyxb.PyXBVersionError(_PyXBVersion)
# Import bindings for namespaces imported into schema
import pyxb.binding.datatypes
import _ds as _ImportedBinding__ds
import _sbc as _ImportedBinding__sbc
import _cbc as _ImportedBinding__cbc
# NOTE: All namespace declarations are reserved within the binding
Namespace = pyxb.namespace.NamespaceForURI('urn:oasis:names:specification:ubl:schema:xsd:SignatureAggregateComponents-2', create_if_missing=True)
Namespace.configureCategories(['typeBinding', 'elementBinding'])
_Namespace_ds = _ImportedBinding__ds.Namespace
_Namespace_ds.configureCategories(['typeBinding', 'elementBinding'])
_Namespace_cbc = _ImportedBinding__cbc.Namespace
_Namespace_cbc.configureCategories(['typeBinding', 'elementBinding'])
_Namespace_sbc = _ImportedBinding__sbc.Namespace
_Namespace_sbc.configureCategories(['typeBinding', 'elementBinding'])
def CreateFromDocument (xml_text, default_namespace=None, location_base=None):
"""Parse the given XML and use the document element to create a
Python instance.
@param xml_text An XML document. This should be data (Python 2
str or Python 3 bytes), or a text (Python 2 unicode or Python 3
str) in the L{pyxb._InputEncoding} encoding.
@keyword default_namespace The L{pyxb.Namespace} instance to use as the
default namespace where there is no default namespace in scope.
If unspecified or C{None}, the namespace of the module containing
this function will be used.
@keyword location_base: An object to be recorded as the base of all
L{pyxb.utils.utility.Location} instances associated with events and
objects handled by the parser. You might pass the URI from which
the document was obtained.
"""
if pyxb.XMLStyle_saxer != pyxb._XMLStyle:
dom = pyxb.utils.domutils.StringToDOM(xml_text)
return CreateFromDOM(dom.documentElement, default_namespace=default_namespace)
if default_namespace is None:
default_namespace = Namespace.fallbackNamespace()
saxer = pyxb.binding.saxer.make_parser(fallback_namespace=default_namespace, location_base=location_base)
handler = saxer.getContentHandler()
xmld = xml_text
if isinstance(xmld, _six.text_type):
xmld = xmld.encode(pyxb._InputEncoding)
saxer.parse(io.BytesIO(xmld))
instance = handler.rootObject()
return instance
def CreateFromDOM (node, default_namespace=None):
"""Create a Python instance from the given DOM node.
The node tag must correspond to an element declaration in this module.
@deprecated: Forcing use of DOM interface is unnecessary; use L{CreateFromDocument}."""
if default_namespace is None:
default_namespace = Namespace.fallbackNamespace()
return pyxb.binding.basis.element.AnyCreateFromDOM(node, default_namespace)
# Complex type {urn:oasis:names:specification:ubl:schema:xsd:SignatureAggregateComponents-2}SignatureInformationType with content type ELEMENT_ONLY
class SignatureInformationType (pyxb.binding.basis.complexTypeDefinition):
"""
"""
_TypeDefinition = None
_ContentTypeTag = pyxb.binding.basis.complexTypeDefinition._CT_ELEMENT_ONLY
_Abstract = False
_ExpandedName = pyxb.namespace.ExpandedName(Namespace, 'SignatureInformationType')
_XSDLocation = pyxb.utils.utility.Location('/Users/radu/Projects/esens/edocument/profiles/e_confirmation/xsd/common/UBL-SignatureAggregateComponents-2.1.xsd', 48, 3)
_ElementMap = {}
_AttributeMap = {}
# Base type is pyxb.binding.datatypes.anyType
# Element {http://www.w3.org/2000/09/xmldsig#}Signature uses Python identifier Signature
__Signature = pyxb.binding.content.ElementDeclaration(pyxb.namespace.ExpandedName(_Namespace_ds, 'Signature'), 'Signature', '__urnoasisnamesspecificationublschemaxsdSignatureAggregateComponents_2_SignatureInformationType_httpwww_w3_org200009xmldsigSignature', False, pyxb.utils.utility.Location('/Users/radu/Projects/esens/edocument/profiles/e_confirmation/xsd/common/UBL-xmldsig-core-schema-2.1.xsd', 54, 0), )
Signature = property(__Signature.value, __Signature.set, None, None)
# Element {urn:oasis:names:specification:ubl:schema:xsd:CommonBasicComponents-2}ID uses Python identifier ID
__ID = pyxb.binding.content.ElementDeclaration(pyxb.namespace.ExpandedName(_Namespace_cbc, 'ID'), 'ID', '__urnoasisnamesspecificationublschemaxsdSignatureAggregateComponents_2_SignatureInformationType_urnoasisnamesspecificationublschemaxsdCommonBasicComponents_2ID', False, pyxb.utils.utility.Location('/Users/radu/Projects/esens/edocument/profiles/e_confirmation/xsd/common/UBL-CommonBasicComponents-2.1.xsd', 341, 3), )
ID = property(__ID.value, __ID.set, None, None)
# Element {urn:oasis:names:specification:ubl:schema:xsd:SignatureBasicComponents-2}ReferencedSignatureID uses Python identifier ReferencedSignatureID
__ReferencedSignatureID = pyxb.binding.content.ElementDeclaration(pyxb.namespace.ExpandedName(_Namespace_sbc, 'ReferencedSignatureID'), 'ReferencedSignatureID', '__urnoasisnamesspecificationublschemaxsdSignatureAggregateComponents_2_SignatureInformationType_urnoasisnamesspecificationublschemaxsdSignatureBasicComponents_2ReferencedSignatureID', False, pyxb.utils.utility.Location('/Users/radu/Projects/esens/edocument/profiles/e_confirmation/xsd/common/UBL-SignatureBasicComponents-2.1.xsd', 28, 3), )
ReferencedSignatureID = property(__ReferencedSignatureID.value, __ReferencedSignatureID.set, None, None)
_ElementMap.update({
__Signature.name() : __Signature,
__ID.name() : __ID,
__ReferencedSignatureID.name() : __ReferencedSignatureID
})
_AttributeMap.update({
})
Namespace.addCategoryObject('typeBinding', 'SignatureInformationType', SignatureInformationType)
SignatureInformation = pyxb.binding.basis.element(pyxb.namespace.ExpandedName(Namespace, 'SignatureInformation'), SignatureInformationType, location=pyxb.utils.utility.Location('/Users/radu/Projects/esens/edocument/profiles/e_confirmation/xsd/common/UBL-SignatureAggregateComponents-2.1.xsd', 44, 3))
Namespace.addCategoryObject('elementBinding', SignatureInformation.name().localName(), SignatureInformation)
SignatureInformationType._AddElement(pyxb.binding.basis.element(pyxb.namespace.ExpandedName(_Namespace_ds, 'Signature'), _ImportedBinding__ds.SignatureType, scope=SignatureInformationType, location=pyxb.utils.utility.Location('/Users/radu/Projects/esens/edocument/profiles/e_confirmation/xsd/common/UBL-xmldsig-core-schema-2.1.xsd', 54, 0)))
SignatureInformationType._AddElement(pyxb.binding.basis.element(pyxb.namespace.ExpandedName(_Namespace_cbc, 'ID'), _ImportedBinding__cbc.IDType, scope=SignatureInformationType, location=pyxb.utils.utility.Location('/Users/radu/Projects/esens/edocument/profiles/e_confirmation/xsd/common/UBL-CommonBasicComponents-2.1.xsd', 341, 3)))
SignatureInformationType._AddElement(pyxb.binding.basis.element(pyxb.namespace.ExpandedName(_Namespace_sbc, 'ReferencedSignatureID'), _ImportedBinding__sbc.ReferencedSignatureIDType, scope=SignatureInformationType, location=pyxb.utils.utility.Location('/Users/radu/Projects/esens/edocument/profiles/e_confirmation/xsd/common/UBL-SignatureBasicComponents-2.1.xsd', 28, 3)))
def _BuildAutomaton ():
# Remove this helper function from the namespace after it is invoked
global _BuildAutomaton
del _BuildAutomaton
import pyxb.utils.fac as fac
counters = set()
cc_0 = fac.CounterCondition(min=0, max=1, metadata=pyxb.utils.utility.Location('/Users/radu/Projects/esens/edocument/profiles/e_confirmation/xsd/common/UBL-SignatureAggregateComponents-2.1.xsd', 55, 7))
counters.add(cc_0)
cc_1 = fac.CounterCondition(min=0, max=1, metadata=pyxb.utils.utility.Location('/Users/radu/Projects/esens/edocument/profiles/e_confirmation/xsd/common/UBL-SignatureAggregateComponents-2.1.xsd', 62, 7))
counters.add(cc_1)
cc_2 = fac.CounterCondition(min=0, max=1, metadata=pyxb.utils.utility.Location('/Users/radu/Projects/esens/edocument/profiles/e_confirmation/xsd/common/UBL-SignatureAggregateComponents-2.1.xsd', 70, 7))
counters.add(cc_2)
states = []
final_update = set()
final_update.add(fac.UpdateInstruction(cc_0, False))
symbol = pyxb.binding.content.ElementUse(SignatureInformationType._UseForTag(pyxb.namespace.ExpandedName(_Namespace_cbc, 'ID')), pyxb.utils.utility.Location('/Users/radu/Projects/esens/edocument/profiles/e_confirmation/xsd/common/UBL-SignatureAggregateComponents-2.1.xsd', 55, 7))
st_0 = fac.State(symbol, is_initial=True, final_update=final_update, is_unordered_catenation=False)
states.append(st_0)
final_update = set()
final_update.add(fac.UpdateInstruction(cc_1, False))
symbol = pyxb.binding.content.ElementUse(SignatureInformationType._UseForTag(pyxb.namespace.ExpandedName(_Namespace_sbc, 'ReferencedSignatureID')), pyxb.utils.utility.Location('/Users/radu/Projects/esens/edocument/profiles/e_confirmation/xsd/common/UBL-SignatureAggregateComponents-2.1.xsd', 62, 7))
st_1 = fac.State(symbol, is_initial=True, final_update=final_update, is_unordered_catenation=False)
states.append(st_1)
final_update = set()
final_update.add(fac.UpdateInstruction(cc_2, False))
symbol = pyxb.binding.content.ElementUse(SignatureInformationType._UseForTag(pyxb.namespace.ExpandedName(_Namespace_ds, 'Signature')), pyxb.utils.utility.Location('/Users/radu/Projects/esens/edocument/profiles/e_confirmation/xsd/common/UBL-SignatureAggregateComponents-2.1.xsd', 70, 7))
st_2 = fac.State(symbol, is_initial=True, final_update=final_update, is_unordered_catenation=False)
states.append(st_2)
transitions = []
transitions.append(fac.Transition(st_0, [
fac.UpdateInstruction(cc_0, True) ]))
transitions.append(fac.Transition(st_1, [
fac.UpdateInstruction(cc_0, False) ]))
transitions.append(fac.Transition(st_2, [
fac.UpdateInstruction(cc_0, False) ]))
st_0._set_transitionSet(transitions)
transitions = []
transitions.append(fac.Transition(st_1, [
fac.UpdateInstruction(cc_1, True) ]))
transitions.append(fac.Transition(st_2, [
fac.UpdateInstruction(cc_1, False) ]))
st_1._set_transitionSet(transitions)
transitions = []
transitions.append(fac.Transition(st_2, [
fac.UpdateInstruction(cc_2, True) ]))
st_2._set_transitionSet(transitions)
return fac.Automaton(states, counters, True, containing_state=None)
SignatureInformationType._Automaton = _BuildAutomaton()
|
getodacu/eSENS-eDocument
|
profiles/e_confirmation/xb_request/_sac.py
|
Python
|
mit
| 11,405
|
"""Endpoints for listing Projects, Versions, Builds, etc."""
import json
import logging
from allauth.socialaccount.models import SocialAccount
from django.conf import settings
from django.db.models import BooleanField, Case, Value, When
from django.shortcuts import get_object_or_404
from django.template.loader import render_to_string
from rest_framework import decorators, permissions, status, viewsets
from rest_framework.parsers import JSONParser, MultiPartParser
from rest_framework.renderers import BaseRenderer, JSONRenderer
from rest_framework.response import Response
from readthedocs.builds.constants import INTERNAL
from readthedocs.builds.models import Build, BuildCommandResult, Version
from readthedocs.oauth.models import RemoteOrganization, RemoteRepository
from readthedocs.oauth.services import GitHubService, registry
from readthedocs.projects.models import Domain, Project
from readthedocs.storage import build_commands_storage
from ..permissions import APIPermission, APIRestrictedPermission, IsOwner
from ..serializers import (
BuildAdminSerializer,
BuildCommandSerializer,
BuildSerializer,
DomainSerializer,
ProjectAdminSerializer,
ProjectSerializer,
RemoteOrganizationSerializer,
RemoteRepositorySerializer,
SocialAccountSerializer,
VersionAdminSerializer,
VersionSerializer,
)
from ..utils import (
ProjectPagination,
RemoteOrganizationPagination,
RemoteProjectPagination,
)
log = logging.getLogger(__name__)
class PlainTextBuildRenderer(BaseRenderer):
"""
Custom renderer for text/plain format.
charset is 'utf-8' by default.
"""
media_type = 'text/plain'
format = 'txt'
def render(self, data, accepted_media_type=None, renderer_context=None):
renderer_context = renderer_context or {}
response = renderer_context.get('response')
if not response or response.exception:
return data.get('detail', '').encode(self.charset)
data = render_to_string(
'restapi/log.txt',
{'build': data},
)
return data.encode(self.charset)
class DisableListEndpoint:
"""
Helper to disable APIv2 listing endpoint.
We are disablng the listing endpoint because it could cause DOS without
using any type of filtering.
This class disables these endpoints except:
- version resource when passing ``?project__slug=``
- build resource when using ``?commit=``
All the other type of listings are disabled and return 409 CONFLICT with an
error message pointing the user to APIv3.
"""
def list(self, *args, **kwargs):
# Using private repos will list resources the user has access to.
if settings.ALLOW_PRIVATE_REPOS:
return super().list(*args, **kwargs)
disabled = True
# NOTE: keep list endpoint that specifies a resource
if any([
self.basename == 'version' and 'project__slug' in self.request.GET,
self.basename == 'build'
and ('commit' in self.request.GET or 'project__slug' in self.request.GET),
self.basename == 'project' and 'slug' in self.request.GET,
]):
disabled = False
if not disabled:
return super().list(*args, **kwargs)
return Response(
{
'error': 'disabled',
'msg': (
'List endpoint have been disabled due to heavy resource usage. '
'Take into account than APIv2 is planned to be deprecated soon. '
'Please use APIv3: https://docs.readthedocs.io/page/api/v3.html'
)
},
status=status.HTTP_410_GONE,
)
class UserSelectViewSet(viewsets.ModelViewSet):
"""
View set that varies serializer class based on request user credentials.
Viewsets using this class should have an attribute `admin_serializer_class`,
which is a serializer that might have more fields that only admin/staff
users require. If the user is staff, this class will be returned instead.
"""
def get_serializer_class(self):
try:
if (
self.request.user.is_staff and
self.admin_serializer_class is not None
):
return self.admin_serializer_class
except AttributeError:
pass
return self.serializer_class
def get_queryset(self):
"""Use our API manager method to determine authorization on queryset."""
return self.model.objects.api(self.request.user)
class ProjectViewSet(DisableListEndpoint, UserSelectViewSet):
"""List, filter, etc, Projects."""
permission_classes = [APIPermission]
renderer_classes = (JSONRenderer,)
serializer_class = ProjectSerializer
admin_serializer_class = ProjectAdminSerializer
model = Project
pagination_class = ProjectPagination
filterset_fields = ('slug',)
@decorators.action(detail=True)
def translations(self, *_, **__):
translations = self.get_object().translations.all()
return Response({
'translations': ProjectSerializer(translations, many=True).data,
})
@decorators.action(detail=True)
def subprojects(self, request, **kwargs):
project = get_object_or_404(
Project.objects.api(request.user),
pk=kwargs['pk'],
)
rels = project.subprojects.all()
children = [rel.child for rel in rels]
return Response({
'subprojects': ProjectSerializer(children, many=True).data,
})
@decorators.action(detail=True)
def active_versions(self, request, **kwargs):
project = get_object_or_404(
Project.objects.api(request.user),
pk=kwargs['pk'],
)
versions = project.versions(manager=INTERNAL).filter(active=True)
return Response({
'versions': VersionSerializer(versions, many=True).data,
})
@decorators.action(
detail=True,
permission_classes=[permissions.IsAdminUser],
)
def token(self, request, **kwargs):
project = get_object_or_404(
Project.objects.api(request.user),
pk=kwargs['pk'],
)
token = GitHubService.get_token_for_project(project, force_local=True)
return Response({
'token': token,
})
@decorators.action(detail=True)
def canonical_url(self, request, **kwargs):
project = get_object_or_404(
Project.objects.api(request.user),
pk=kwargs['pk'],
)
return Response({
'url': project.get_docs_url(),
})
class VersionViewSet(DisableListEndpoint, UserSelectViewSet):
permission_classes = [APIRestrictedPermission]
renderer_classes = (JSONRenderer,)
serializer_class = VersionSerializer
admin_serializer_class = VersionAdminSerializer
model = Version
filterset_fields = (
'active',
'project__slug',
)
class BuildViewSet(DisableListEndpoint, UserSelectViewSet):
permission_classes = [APIRestrictedPermission]
renderer_classes = (JSONRenderer, PlainTextBuildRenderer)
serializer_class = BuildSerializer
admin_serializer_class = BuildAdminSerializer
model = Build
filterset_fields = ('project__slug', 'commit')
@decorators.action(
detail=False,
permission_classes=[permissions.IsAdminUser],
methods=['get'],
)
def concurrent(self, request, **kwargs):
project_slug = request.GET.get('project__slug')
project = get_object_or_404(Project, slug=project_slug)
limit_reached, concurrent, max_concurrent = Build.objects.concurrent(project)
data = {
'limit_reached': limit_reached,
'concurrent': concurrent,
'max_concurrent': max_concurrent,
}
return Response(data)
def retrieve(self, *args, **kwargs):
"""
Retrieves command data from storage.
This uses files from storage to get the JSON,
and replaces the ``commands`` part of the response data.
"""
if not settings.RTD_SAVE_BUILD_COMMANDS_TO_STORAGE:
return super().retrieve(*args, **kwargs)
instance = self.get_object()
serializer = self.get_serializer(instance)
data = serializer.data
if instance.cold_storage:
storage_path = '{date}/{id}.json'.format(
date=str(instance.date.date()),
id=instance.id,
)
if build_commands_storage.exists(storage_path):
try:
json_resp = build_commands_storage.open(storage_path).read()
data['commands'] = json.loads(json_resp)
except Exception:
log.exception(
'Failed to read build data from storage. path=%s.',
storage_path,
)
return Response(data)
@decorators.action(
detail=True,
permission_classes=[permissions.IsAdminUser],
methods=['post'],
)
def reset(self, request, **kwargs):
"""Reset the build so it can be re-used when re-trying."""
instance = self.get_object()
instance.reset()
return Response(status=status.HTTP_204_NO_CONTENT)
class BuildCommandViewSet(DisableListEndpoint, UserSelectViewSet):
parser_classes = [JSONParser, MultiPartParser]
permission_classes = [APIRestrictedPermission]
renderer_classes = (JSONRenderer,)
serializer_class = BuildCommandSerializer
model = BuildCommandResult
class DomainViewSet(DisableListEndpoint, UserSelectViewSet):
permission_classes = [APIRestrictedPermission]
renderer_classes = (JSONRenderer,)
serializer_class = DomainSerializer
model = Domain
class RemoteOrganizationViewSet(viewsets.ReadOnlyModelViewSet):
permission_classes = [IsOwner]
renderer_classes = (JSONRenderer,)
serializer_class = RemoteOrganizationSerializer
model = RemoteOrganization
pagination_class = RemoteOrganizationPagination
def get_queryset(self):
return (
self.model.objects.api(self.request.user).filter(
remote_organization_relations__account__provider__in=[
service.adapter.provider_id for service in registry
]
).distinct()
)
class RemoteRepositoryViewSet(viewsets.ReadOnlyModelViewSet):
permission_classes = [IsOwner]
renderer_classes = (JSONRenderer,)
serializer_class = RemoteRepositorySerializer
model = RemoteRepository
pagination_class = RemoteProjectPagination
def get_queryset(self):
if not self.request.user.is_authenticated:
return self.model.objects.none()
# TODO: Optimize this query after deployment
query = self.model.objects.api(self.request.user).annotate(
admin=Case(
When(
remote_repository_relations__user=self.request.user,
remote_repository_relations__admin=True,
then=Value(True)
),
default=Value(False),
output_field=BooleanField()
)
)
full_name = self.request.query_params.get('full_name')
if full_name is not None:
query = query.filter(full_name__icontains=full_name)
org = self.request.query_params.get('org', None)
if org is not None:
query = query.filter(organization__pk=org)
own = self.request.query_params.get('own', None)
if own is not None:
query = query.filter(
remote_repository_relations__account__provider=own,
organization=None,
)
query = query.filter(
remote_repository_relations__account__provider__in=[
service.adapter.provider_id for service in registry
],
).distinct()
# optimizes for the RemoteOrganizationSerializer
query = query.select_related('organization').order_by(
'organization__name', 'full_name'
)
return query
class SocialAccountViewSet(viewsets.ReadOnlyModelViewSet):
permission_classes = [IsOwner]
renderer_classes = (JSONRenderer,)
serializer_class = SocialAccountSerializer
model = SocialAccount
def get_queryset(self):
return self.model.objects.filter(user=self.request.user.pk)
|
rtfd/readthedocs.org
|
readthedocs/api/v2/views/model_views.py
|
Python
|
mit
| 12,622
|
# -- encoding: UTF-8 --
import sys
from click import echo
from colorama import Fore, Style
can_use_real_emoji = (
sys.stdout.isatty() and
sys.platform != "win32"
)
def success(msg):
if can_use_real_emoji:
sign = "\U0001F44C"
else:
sign = "[+] "
echo(Fore.GREEN + Style.BRIGHT + sign + msg)
def get_site_from_context(ctx):
try:
return ctx.meta["wurst.project"].site or None
except KeyError:
pass
|
wurstfabrik/wurst-cli
|
wurstc/cli/utils.py
|
Python
|
mit
| 461
|
import gammu.smsd
import thread
class smsd(object):
"""Starts gammu in another thread so the bot can interpret incoming sms"""
def __init__(self, configpath):
self.sms = gammu.smsd.SMSD(configpath)
self.thread = None
def start(self):
self.thread = thread.start_new_thread(self.sms.MainLoop, ())
def stop(self):
self.sms.Shutdown()
def inject_sms(self, phonenumber, message):
smsinfo = {
'Class': -1,
'Unicode': True,
'Entries': [
{
'ID': 'ConcatenatedTextLong',
'Buffer': message
}
]}
encoded = gammu.encode(smsinfo)
for message in encoded:
message['SMSC'] = {'Location': 1}
message['Number'] = phonenumber
self.sms.InjectSMS(encoded)
|
WaltonSimons/PhoneBot
|
smsd.py
|
Python
|
mit
| 904
|
#!/usr/bin/env python3
# Copyright (c) 2015-2017 The Doriancoin Core developers
# Distributed under the MIT software license, see the accompanying
# file COPYING or http://www.opensource.org/licenses/mit-license.php.
"""Test the ZMQ notification interface."""
import configparser
import os
import struct
from test_framework.test_framework import DoriancoinTestFramework, SkipTest
from test_framework.mininode import CTransaction
from test_framework.util import (assert_equal,
bytes_to_hex_str,
hash256,
)
from io import BytesIO
class ZMQSubscriber:
def __init__(self, socket, topic):
self.sequence = 0
self.socket = socket
self.topic = topic
import zmq
self.socket.setsockopt(zmq.SUBSCRIBE, self.topic)
def receive(self):
topic, body, seq = self.socket.recv_multipart()
# Topic should match the subscriber topic.
assert_equal(topic, self.topic)
# Sequence should be incremental.
assert_equal(struct.unpack('<I', seq)[-1], self.sequence)
self.sequence += 1
return body
class ZMQTest (DoriancoinTestFramework):
def set_test_params(self):
self.num_nodes = 2
def setup_nodes(self):
# Try to import python3-zmq. Skip this test if the import fails.
try:
import zmq
except ImportError:
raise SkipTest("python3-zmq module not available.")
# Check that doriancoin has been built with ZMQ enabled.
config = configparser.ConfigParser()
if not self.options.configfile:
self.options.configfile = os.path.abspath(os.path.join(os.path.dirname(__file__), "../config.ini"))
config.read_file(open(self.options.configfile))
if not config["components"].getboolean("ENABLE_ZMQ"):
raise SkipTest("doriancoind has not been built with zmq enabled.")
# Initialize ZMQ context and socket.
# All messages are received in the same socket which means
# that this test fails if the publishing order changes.
# Note that the publishing order is not defined in the documentation and
# is subject to change.
address = "tcp://127.0.0.1:28332"
self.zmq_context = zmq.Context()
socket = self.zmq_context.socket(zmq.SUB)
socket.set(zmq.RCVTIMEO, 60000)
socket.connect(address)
# Subscribe to all available topics.
self.hashblock = ZMQSubscriber(socket, b"hashblock")
self.hashtx = ZMQSubscriber(socket, b"hashtx")
self.rawblock = ZMQSubscriber(socket, b"rawblock")
self.rawtx = ZMQSubscriber(socket, b"rawtx")
self.extra_args = [["-zmqpub%s=%s" % (sub.topic.decode(), address) for sub in [self.hashblock, self.hashtx, self.rawblock, self.rawtx]], []]
self.add_nodes(self.num_nodes, self.extra_args)
self.start_nodes()
def run_test(self):
try:
self._zmq_test()
finally:
# Destroy the ZMQ context.
self.log.debug("Destroying ZMQ context")
self.zmq_context.destroy(linger=None)
def _zmq_test(self):
num_blocks = 5
self.log.info("Generate %(n)d blocks (and %(n)d coinbase txes)" % {"n": num_blocks})
genhashes = self.nodes[0].generate(num_blocks)
self.sync_all()
for x in range(num_blocks):
# Should receive the coinbase txid.
txid = self.hashtx.receive()
# Should receive the coinbase raw transaction.
hex = self.rawtx.receive()
tx = CTransaction()
tx.deserialize(BytesIO(hex))
tx.calc_sha256()
assert_equal(tx.hash, bytes_to_hex_str(txid))
# Should receive the generated block hash.
hash = bytes_to_hex_str(self.hashblock.receive())
assert_equal(genhashes[x], hash)
# The block should only have the coinbase txid.
assert_equal([bytes_to_hex_str(txid)], self.nodes[1].getblock(hash)["tx"])
# Should receive the generated raw block.
block = self.rawblock.receive()
assert_equal(genhashes[x], bytes_to_hex_str(hash256(block[:80])))
self.log.info("Wait for tx from second node")
payment_txid = self.nodes[1].sendtoaddress(self.nodes[0].getnewaddress(), 1.0)
self.sync_all()
# Should receive the broadcasted txid.
txid = self.hashtx.receive()
assert_equal(payment_txid, bytes_to_hex_str(txid))
# Should receive the broadcasted raw transaction.
hex = self.rawtx.receive()
assert_equal(payment_txid, bytes_to_hex_str(hash256(hex)))
if __name__ == '__main__':
ZMQTest().main()
|
doriancoins/doriancoin
|
test/functional/interface_zmq.py
|
Python
|
mit
| 4,806
|
import cv2
import skimage.io
import skimage.transform
import tensorflow as tf
import numpy as np
import matplotlib.pyplot as plt
from CatFinder import CatFinder
from CatPlayer import CatPlayer
from DrawGui import DrawArea
def load_image(path, size=224):
img = skimage.io.imread(path)
short_edge = min(img.shape[:2])
yy = int((img.shape[0] - short_edge) / 2)
xx = int((img.shape[1] - short_edge) / 2)
crop_img = img[yy : yy + short_edge, xx : xx + short_edge]
resized_img = skimage.transform.resize(crop_img, (size, size))
return resized_img
#img = load_image('/usr/local/data/ludde/DSC_0065.jpg')
#img = np.tile(img, (32, 1, 1, 1))
alpha = 0.6
draw = DrawArea()
drawing = draw.doDraw()
def nothing(x):
global alpha
alpha = x/100.
catPlayer = CatPlayer()
cv2.namedWindow('cam')
cv2.createTrackbar('alpha','cam',50,100,nothing)
cap = cv2.VideoCapture(0)
rval, frame = cap.read()
catFinder = CatFinder('/tmp/models/catnet12')
xrange = np.arange(224)
xv, yv = np.meshgrid(xrange, xrange)
up_scale_y = 640/224.
up_scale_x = 480/224.
cat_list = range(100)
cnt = 0
def saveImages(img_list, name):
for i in range(len(img_list)): cv2.imwrite("cat_" + str(i) + ".jpg", img_list[i])
while rval:
print 'alpha:', alpha
img = cv2.resize(frame, (224, 224))
img = img.astype(np.float32)/255
img = img[np.newaxis, :, :, ::-1]
frame[drawing] = (0, 0, 255)
segmentedCat = np.squeeze(catFinder.getProbability(img))# > alpha
segmentedCatBig = cv2.resize(segmentedCat, (640, 480)) > alpha
frame[segmentedCatBig] = 0.5*frame[segmentedCatBig] + frame[segmentedCatBig]*np.array([0.5, 0, 0])
if segmentedCat.max() > 0.90:
print 'IS CAT'
prod_dist = segmentedCat/segmentedCat.sum()
centerx = np.sum(prod_dist*xv)*up_scale_x
centery = np.sum(prod_dist*yv)*up_scale_y
cv2.circle(frame, (int(centery), int(centerx)), 4, (0, 255, 0))
if drawing[centerx, centery]:
catPlayer.play()
else:
catPlayer.pause()
#FOR VIDEO
cat_list[cnt%100] = frame
cnt += 1
cv2.imshow('cam', frame)
rval, frame = cap.read()
key = cv2.waitKey(1)
if key == 27:
break
segmentedCat = catFinder(img)
seg = np.zeros((224, 224, 3))
seg[:, :, 2] = segmentedCat
plt.imshow(0.5*seg + 0.5*img[0])
print segmentedCat
|
sigmunjr/VirtualPetFence
|
runSegmentation.py
|
Python
|
mit
| 2,367
|
#!/usr/bin/python
class Solution(object):
def reconstructQueue(self, people):
print people
people = sorted(people, key=lambda x: x[1])
print people
people = sorted(people, key=lambda x: -x[0])
print people
res = []
for p in people:
res.insert(p[1], p)
return res
class SolutionDumb(object):
def reconstructQueue(self, people):
"""
:type people: List[List[int]]
:rtype: List[List[int]]
"""
res = []
people = sorted(people, key = lambda x: x[0])
for i in xrange(len(people)):
found = False
j = i
while not found and j >= 0:
for p in people:
if p[1] == j:
taller = 0
for r in res:
if r[0] >= p[0]:
taller += 1
if taller == p[1]:
people.remove(p)
res.append(p)
found = True
break
j -= 1
return res
s = Solution()
x = [[7,0], [4,4], [7,1], [5,0], [6,1], [5,2]]
print s.reconstructQueue(x)
|
pisskidney/leetcode
|
medium/406.py
|
Python
|
mit
| 1,289
|
#!BPY
"""
Name: 'TikZ (.tex)...'
Blender: 245
Group: 'Export'
Tooltip: 'Export selected curves as TikZ paths for use with (La)TeX'
"""
__author__ = 'Kjell Magne Fauske'
__version__ = "1.0"
__url__ = ("Documentation, http://www.fauskes.net/code/blend2tikz/documentation/",
"Author's homepage, http://www.fauskes.net/",
"TikZ examples, http://www.fauskes.net/pgftikzexamples/")
__bpydoc__ = """\
This script exports selected curves and empties to TikZ format for use with TeX.
PGF and TikZ is a powerful macro package for creating high quality illustrations
and graphics for use with (La|Con)TeX.
Important: TikZ is primarily for creating 2D illustrations. This script will
therefore only export the X and Y coordinates. However, the Z coordinate is used
to determine draw order.
Usage:
Select the objects you want to export and invoke the script from the
"File->Export" menu[1]. Alternatively you can load and run the script from
inside Blender.
A dialog box will pop up with various options:<br>
- Draw: Insert a draw operation in the generated path.<br>
- Fill: Insert a fill operation in the generated path.<br>
- Transform: Apply translation and scale transformations.<br>
- Materials: Export materials assigned to curves.<br>
- Empties: Export empties as named coordinates.<br>
- Only properties: Use on the style property of materials if set.<br>
- Standalone: Create a standalone document.<br>
- Only code: Generate only code for drawing paths.<br>
- Clipboard: Copy generated code to the clipboard. <br>
Properties:
If an object is assigned a ID property or game property named 'style' of type
string, its value will be added to the path as an option. You can use the
Help->Property ID browser to set this value, or use the Logic panel to
add a game property.
Materials:
The exporter has basic support for materials. By default the material's RGB
value is used as fill or draw color. You can also set the alpha value for
transparency effects.
An alternative is to specify style options
directly by putting the values in a 'style' property assigned to the material.
You can use the Help->Property ID browser to set this value.
Issues:<br>
- Only bezier and polyline curves are supported.<br>
- A full Python install is required for clipboard support on Windows. Other platforms
need the standard subprocess module (requires Python 2.4 or later). Additionally:<br>
* Windows users need to install the PyWin32 module.<br>
* Unix-like users need the xclip command line tool or the PyGTK_ module installed.<br>
* OS X users need the pbcopy command line tool installed.<br>
[1] Requires you to put the script in Blender's scripts folder. Blender will
then automatically detect the script.
"""
import Blender
from Blender import sys as bsys
from itertools import izip
import itertools, math
from Blender import Mesh, Mathutils, Registry, Scene, Material, Group
from textwrap import wrap
from string import Template
# Curve types
TYPE_POLY = 0
TYPE_BEZIER = 1
TYPE_NURBS = 4
R2D = 180.0 / math.pi
# Start of configuration section -------
# Templates
standalone_template = r"""
\documentclass{article}
\usepackage{tikz}
%(preamble)s
%(materials)s
\begin{document}
\begin{tikzpicture}
%(pathcode)s
\end{tikzpicture}
\end{document}
"""
fig_template = r"""
%(materials)s
\begin{tikzpicture}
%(pathcode)s
\end{tikzpicture}
"""
REG_KEY = 'tikz_export'
# config options:
STANDALONE = True
CODE_ONLY = False
DRAW_CURVE = True
FILL_CLOSED_CURVE = True
TRANSFORM_CURVE = True
CLIPBOARD_OUTPUT = False
EMPTIES = True
EXPORT_MATERIALS = False
ONLY_PROPERTIES = False
USE_PLOTPATH = False
WRAP_LINES = True
tooltips = {
'STANDALONE': 'Output a standalone document',
'DRAW_CURVE':
'Draw curves',
'FILL_CLOSED_CURVE':
'Fill closed curves',
'TRANSFORM_CURVE':
'Apply transformations',
'CLIPBOARD_OUTPUT':
'Put generated code on clipboard',
'CODE_ONLY':
'Output pathcode only',
'EMPTIES': 'Export empties',
'EXPORT_MATERIALS': 'Apply materials to curves',
'ONLY_PROPERTIES':
'Use only properties for materials with the style property set',
'USE_PLOTPATH':
'Use the plot path operations for polylines',
'WRAP_LINES':
'Wrap long lines',
}
def update_registry():
d = {
'STANDALONE': STANDALONE,
'DRAW_CURVE': DRAW_CURVE,
'FILL_CLOSED_CURVE': FILL_CLOSED_CURVE,
'TRANSFORM_CURVE': TRANSFORM_CURVE,
'CLIPBOARD_OUTPUT': CLIPBOARD_OUTPUT,
'CODE_ONLY': CODE_ONLY,
'EMPTIES': EMPTIES,
'EXPORT_MATERIALS': EXPORT_MATERIALS,
'ONLY_PROPERTIES': ONLY_PROPERTIES,
'USE_PLOTPATH': USE_PLOTPATH,
'WRAP_LINES': WRAP_LINES,
}
Registry.SetKey(REG_KEY, d, True)
# Looking for a saved key in Blender.Registry dict:
rd = Registry.GetKey(REG_KEY, True)
if rd:
try:
STANDALONE = rd['STANDALONE']
DRAW_CURVE = rd['DRAW_CURVE']
FILL_CLOSED_CURVE = rd['FILL_CLOSED_CURVE']
TRANSFORM_CURVE = rd['TRANSFORM_CURVE']
CLIPBOARD_OUTPUT = rd['CLIPBOARD_OUTPUT']
CODE_ONLY = rd['CODE_ONLY']
EMPTIES = rd['EMPTIES']
EXPORT_MATERIALS = rd['EXPORT_MATERIALS']
ONLY_PROPERTIES = rd['ONLY_PROPERTIES']
USE_PLOTPATH = rd['USE_PLOTPATH']
WRAP_LINES = rd['WRAP_LINES']
except KeyError:
print "Keyerror"
update_registry()
else:
print "update registry"
update_registry()
# Start of GUI section ------------------------------------------------
from Blender import Draw
def draw_GUI():
global STANDALONE, DRAW_CURVE, FILL_CLOSED_CURVE, TRANSFORM_CURVE
global CLIPBOARD_OUTPUT, CODE_ONLY, EMPTIES, EXPORT_MATERIALS
global ONLY_PROPERTIES
global USE_PLOTPATH
global WRAP_LINES
standalonetog = Draw.Create(STANDALONE)
codeonlytog = Draw.Create(CODE_ONLY)
drawcurvetog = Draw.Create(DRAW_CURVE)
fillcurvetog = Draw.Create(FILL_CLOSED_CURVE)
transformcurvetog = Draw.Create(TRANSFORM_CURVE)
clipboardtog = Draw.Create(CLIPBOARD_OUTPUT)
emptiestog = Draw.Create(EMPTIES)
materialstog = Draw.Create(EXPORT_MATERIALS)
onlyproptog = Draw.Create(ONLY_PROPERTIES)
useplotpathtog = Draw.Create(USE_PLOTPATH)
wraplinestog = Draw.Create(WRAP_LINES)
block = []
#block.append("Export:")
block.append(("Draw", drawcurvetog, tooltips['DRAW_CURVE']))
block.append(("Fill", fillcurvetog, tooltips['FILL_CLOSED_CURVE']))
block.append(("Transform", transformcurvetog, tooltips['TRANSFORM_CURVE']))
block.append(("Use plot path", useplotpathtog, tooltips['USE_PLOTPATH']))
block.append("Export:")
block.append(("Materials", materialstog, tooltips['EXPORT_MATERIALS']))
block.append(("Empties", emptiestog, tooltips['EMPTIES']))
block.append("Material options:")
block.append(("Only properties", onlyproptog, tooltips['ONLY_PROPERTIES']))
block.append('Ouput options')
block.append(("Standalone", standalonetog, tooltips['STANDALONE']))
block.append(("Only code", codeonlytog, tooltips['CODE_ONLY']))
block.append(("Clipboard", clipboardtog, tooltips['CLIPBOARD_OUTPUT']))
block.append(("Wrap lines", wraplinestog, tooltips['WRAP_LINES']))
retval = Blender.Draw.PupBlock("Blend2TikZ options", block)
if retval:
# set options
STANDALONE = standalonetog.val
DRAW_CURVE = drawcurvetog.val
FILL_CLOSED_CURVE = fillcurvetog.val
TRANSFORM_CURVE = transformcurvetog.val
CLIPBOARD_OUTPUT = clipboardtog.val
CODE_ONLY = codeonlytog.val
EMPTIES = emptiestog.val
EXPORT_MATERIALS = materialstog.val
ONLY_PROPERTIES = onlyproptog.val
USE_PLOTPATH = useplotpathtog.val
WRAP_LINES = wraplinestog.val
update_registry()
return retval
# End of GUI section ----------------------
# End of configuration section ---------
X = 0
Y = 1
used_materials = {}
# Utility functions
def nsplit(seq, n=2):
"""Split a sequence into pieces of length n
If the lengt of the sequence isn't a multiple of n, the rest is discareded.
Note that nsplit will strings into individual characters.
Examples:
>>> nsplit('aabbcc')
[('a', 'a'), ('b', 'b'), ('c', 'c')]
>>> nsplit('aabbcc',n=3)
[('a', 'a', 'b'), ('b', 'c', 'c')]
# Note that cc is discarded
>>> nsplit('aabbcc',n=4)
[('a', 'a', 'b', 'b')]
"""
return [xy for xy in izip(*[iter(seq)] * n)]
def mreplace(s, chararray, newchararray):
for a, b in zip(chararray, newchararray):
s = s.replace(a, b)
return s
def tikzify(s):
if s.strip():
return mreplace(s, r'\,:.', '-+_*')
else:
return ""
def copy_to_clipboard(text):
"""Copy text to the clipboard
Returns True if successful. False otherwise.
Works on Windows, *nix and Mac. Tries the following:
1. Use the win32clipboard module from the win32 package.
2. Calls the xclip command line tool (*nix)
3. Calls the pbcopy command line tool (Mac)
4. Try pygtk
"""
# try windows first
try:
import win32clipboard
win32clipboard.OpenClipboard()
win32clipboard.EmptyClipboard()
win32clipboard.SetClipboardText(text)
win32clipboard.CloseClipboard()
return True
except:
pass
# try xclip
try:
import subprocess
p = subprocess.Popen(['xclip', '-selection', 'c'], stdin=subprocess.PIPE)
p.stdin.write(text)
p.stdin.close()
retcode = p.wait()
return True
except:
pass
# try pbcopy (Os X)
try:
import subprocess
p = subprocess.Popen(['pbcopy'], stdin=subprocess.PIPE)
p.stdin.write(text)
p.stdin.close()
retcode = p.wait()
return True
except:
pass
# try os /linux
try:
import subprocess
p = subprocess.Popen(['xsel'], stdin=subprocess.PIPE)
p.stdin.write(text)
p.stdin.close()
retcode = p.wait()
return True
except:
pass
# try pygtk
try:
# Code from
# http://www.vector-seven.com/2007/06/27/
# passing-data-between-gtk-applications-with-gtkclipboard/
import pygtk
pygtk.require('2.0')
import gtk
# get the clipboard
clipboard = gtk.clipboard_get()
# set the clipboard text data
clipboard.set_text(text)
# make our data available to other applications
clipboard.store()
except:
return False
def get_property(obj, name):
"""Get named object property
Looks first in custom properties, then game properties. Returns a list.
"""
prop_value = []
try:
prop_value.append(obj.properties[name])
except:
pass
try:
# look for game properties
prop = obj.getProperty(name)
if prop.type == "STRING" and prop.data.strip():
prop_value.append(prop.data)
except:
pass
return prop_value
def get_material(material):
"""Convert material to TikZ options"""
if not material:
return ""
opts = ""
mat_name = tikzify(material.name)
used_materials[mat_name] = material
return mat_name
def write_materials(used_materials):
"""Return code for the used materials"""
c = "% Materials section \n"
for material in used_materials.values():
mat_name = tikzify(material.name)
matopts = ''
proponly = ONLY_PROPERTIES
try:
proponly = material.properties['onlyproperties']
if proponly and type(proponly) == str:
proponly = proponly.lower() not in ('0', 'false')
except:
pass
try:
matopts = material.properties['style']
except:
pass
rgb = material.rgbCol
spec = material.specCol
alpha = material.alpha
flags = material.getMode()
options = []
if not (proponly and matopts):
c += "\\definecolor{%s_col}{rgb}{%s,%s,%s}\n" \
% tuple([mat_name] + rgb)
options.append('%s_col' % mat_name)
if alpha < 1.0:
options.append('opacity=%s' % alpha)
if matopts:
options += [matopts]
c += "\\tikzstyle{%s}= [%s]\n" % (mat_name, ",".join(options))
return c
def write_object(obj, empties):
"""Write Curves"""
s = ""
name = obj.name
prop = obj.properties
mtrx = obj.matrix.rotationPart()
x, y, z = obj.getLocation('worldspace')
rot = obj.getEuler('worldspace')
scale_x, scale_y, scale_z = obj.matrix.scalePart()
# Convert to degrees
rot_z = rot.z * R2D
if obj.type not in ["Curve", "Empty"]:
return s
ps = ""
if obj.type == 'Curve':
curvedata = obj.data
s += "%% %s\n" % name
for curnurb in curvedata:
if curnurb.type == TYPE_BEZIER:
knots = []
handles = []
# Build lists of knots and handles
for point in curnurb:
h1, knot, h2 = point.vec
handles.extend([h1, h2])
knots.append("(+%.4f,+%.4f)" % (knot[X], knot[Y]))
if curnurb.isCyclic():
# The curve is closed.
# Move the first handle to the end of the handles list.
handles = handles[1:] + [handles[0]]
# Repeat the first knot at the end of the knot list
knots.append(knots[0])
else:
# We don't need the first and last handles since the curve is
# not closed.
handles = handles[1:-1]
hh = []
for h1, h2 in nsplit(handles, 2):
hh.append("controls (+%.4f,+%.4f) and (+%.4f,+%.4f)" \
% (h1[X], h1[Y], h2[X], h2[Y]))
ps += "%s\n" % knots[0]
for h, k in zip(hh, knots[1:]):
ps += " .. %s .. %s\n" % (h, k)
if curnurb.isCyclic():
ps += " -- cycle\n"
elif curnurb.type == TYPE_POLY:
coords = ["(+%.4f,+%.4f)" % (point[X], point[Y]) for point in curnurb]
if USE_PLOTPATH:
plotopts = get_property(obj, 'plotstyle')
if plotopts:
poptstr = "[%s]" % ",".join(plotopts)
else:
poptstr = ''
ps += " plot%s coordinates {%s}" % (poptstr, " ".join(coords))
if curnurb.isCyclic():
ps += " -- cycle"
if WRAP_LINES:
ps = "\n".join(wrap(ps, 80, subsequent_indent=" ", break_long_words=False))
else:
if curnurb.isCyclic():
coords.extend([coords[0], 'cycle\n '])
# Join the coordinates. Could have used "--".join(coords), but
# have to add some logic for pretty printing.
if WRAP_LINES:
ps += "%s\n " % coords[0]
i = 0
for c in coords[1:]:
i += 1
if i % 3:
ps += "-- %s" % c
else:
ps += " -- %s\n " % c
else:
ps += "%s" % " -- ".join(coords)
else:
continue
if not ps:
return s
options = []
if DRAW_CURVE:
options += ['draw']
if FILL_CLOSED_CURVE:
if ps.find('cycle') > 0:
options += ['fill']
if TRANSFORM_CURVE:
if x <> 0: options.append('xshift=%.4fcm' % x)
if y <> 0: options.append('yshift=%.4fcm' % y)
if rot_z <> 0: options.append('rotate=%.4f' % rot_z)
if scale_x <> 1: options += ['xscale=%.4f' % scale_x]
if scale_y <> 1: options += ['yscale=%.4f' % scale_y]
if EXPORT_MATERIALS:
try:
materials = obj.data.getMaterials()
except:
materials = []
if materials:
# pick first material
for mat in materials:
if mat:
matopts = get_material(mat)
options.append(matopts)
break
extraopts = get_property(obj, 'style')
if extraopts:
options.extend(extraopts)
optstr = ",".join(options)
emptstr = ""
if EMPTIES:
if obj in empties:
for empty in empties[obj]:
# Get correct coordinate relative to the parent
if TRANSFORM_CURVE:
ex, ey, ez = (empty.mat * (obj.mat.copy()).invert()).translationPart()
else:
ex, ey, ez = (empty.matrix - obj.matrix).translationPart()
emptstr += " (+%.4f,+%.4f) coordinate (%s)\n" \
% (ex, ey, empty.name)
if not WRAP_LINES:
ps = ' '.join(ps.replace('\n', ' ').split())
if len(optstr) > 50 or emptstr:
s += "\\path[%s]\n%s %s;\n" % (optstr, emptstr, ps.rstrip())
else:
s += "\\path[%s] %s;\n" % (optstr, ps.rstrip())
elif obj.type == 'Empty' and EMPTIES and not obj.parent:
x, y, z = obj.loc
s += "\\coordinate (%s) at (%.4f,%.4f);\n" % (tikzify(obj.name), x, y)
return s
def write_objects(filepath):
"""Write all selected objects to filepath"""
def z_comp(a, b):
x, y, z1 = a.getLocation('worldspace')
x, y, z2 = b.getLocation('worldspace')
return cmp(z1, z2)
# get all selected objects
objects = Blender.Object.GetSelected()
# get current scene
scn = Blender.Scene.GetCurrent()
# iterate over each object
code = ""
# Find all empties with parents
empties_wp = [obj for obj in objects if obj.type == 'Empty' and obj.parent]
empties_dict = {}
for empty in empties_wp:
if empty.parent in empties_dict:
empties_dict[empty.parent] += [empty]
else:
empties_dict[empty.parent] = [empty]
for obj in sorted(objects, z_comp):
code += write_object(obj, empties_dict)
s = ""
if EXPORT_MATERIALS:
matcode = write_materials(used_materials)
else:
matcode = ""
try:
preamblecode = scn.properties['preamble']
except:
preamblecode = ''
templatevars = dict(pathcode=code, preamble=preamblecode, materials=matcode)
if STANDALONE:
extra = ""
try:
preambleopt = scn.properties['preamble']
templatevars['preamble'] = str(preambleopt)
except:
pass
template = standalone_template
elif CODE_ONLY:
template = "%(pathcode)s"
else:
template = fig_template
s = template % templatevars
if not CLIPBOARD_OUTPUT:
try:
f = file(filepath, 'w')
# write header to file
f.write('%% Generated by tikz_export.py v %s \n' % (__version__))
f.write(s)
print "Code written to %s" % filepath
finally:
f.close()
return
else:
success = copy_to_clipboard(s)
if not success:
print "Failed to copy code to the clipboard"
print "Pywin32, xclip, cbcopy or pygtk required for clipboard support"
Blender.Draw.PupMenu('ERROR: Failed to copy generated code to the clipboard')
# Start of script -----------------------------------------------------
# Ensure that at leas one object is selected
if len(Blender.Object.GetSelected()) == 0:
# no objects selected. Print error message and quit
Blender.Draw.PupMenu('ERROR: Please select at least one curve')
else:
fname = bsys.makename(ext=".tex")
retval = draw_GUI()
if retval and not CLIPBOARD_OUTPUT:
Blender.Window.FileSelector(write_objects, "Export TikZ", fname)
write_objects(fname)
print "tikz_export ended ..."
|
kjellmf/blend2tikz
|
tikz_export.py
|
Python
|
mit
| 20,546
|
import pymarkdownlint
from pymarkdownlint.filefinder import MarkdownFileFinder
from pymarkdownlint.lint import MarkdownLinter
from pymarkdownlint.config import LintConfig
import os
import click
DEFAULT_CONFIG_FILE = ".markdownlint"
def echo_files(files):
for f in files:
click.echo(f)
exit(0)
def get_lint_config(config_path=None):
""" Tries loading the config from the given path. If no path is specified, the default config path
is tried, and if that is not specified, we the default config is returned. """
# config path specified
if config_path:
config = LintConfig.load_from_file(config_path)
click.echo("Using config from {0}".format(config_path))
# default config path
elif os.path.exists(DEFAULT_CONFIG_FILE):
config = LintConfig.load_from_file(DEFAULT_CONFIG_FILE)
click.echo("Using config from {0}".format(DEFAULT_CONFIG_FILE))
# no config file
else:
config = LintConfig()
return config
@click.command()
@click.option('--config', type=click.Path(exists=True),
help="Config file location (default: {0}).".format(DEFAULT_CONFIG_FILE))
@click.option('--list-files', is_flag=True, help="List markdown files in given path and exit.")
@click.option('--ignore', default="", help="Ignore rules (comma-separated by id or name).")
@click.argument('path', type=click.Path(exists=True))
@click.version_option(version=pymarkdownlint.__version__)
def cli(list_files, config, ignore, path):
""" Markdown lint tool, checks your markdown for styling issues """
files = MarkdownFileFinder.find_files(path)
if list_files:
echo_files(files)
lint_config = get_lint_config(config)
lint_config.apply_on_csv_string(ignore, lint_config.disable_rule)
linter = MarkdownLinter(lint_config)
error_count = linter.lint_files(files)
exit(error_count)
if __name__ == "__main__":
cli()
|
jorisroovers/pymarkdownlint
|
pymarkdownlint/cli.py
|
Python
|
mit
| 1,924
|
"""
.. module:: counters
:synopsis: SFlow counter object interfaces
.. moduleauthor:: Colin Alston <colin@imcol.in>
"""
from construct import Struct, UBInt32, Array, Bytes
class InterfaceCounters(object):
"""Counters for network interfaces
"""
def __init__(self, u):
self.if_index = u.unpack_uint()
self.if_type = u.unpack_uint()
self.if_speed = u.unpack_uhyper()
self.if_mode = u.unpack_uint()
self.if_status = u.unpack_uint()
self.if_inOctets = u.unpack_uhyper()
self.if_inPackets = u.unpack_uint()
self.if_inMcast = u.unpack_uint()
self.if_inBcast = u.unpack_uint()
self.if_inDiscard = u.unpack_uint()
self.if_inError = u.unpack_uint()
self.if_unknown = u.unpack_uint()
self.if_outOctets = u.unpack_uhyper()
self.if_outPackets = u.unpack_uint()
self.if_outMcast = u.unpack_uint()
self.if_outBcast = u.unpack_uint()
self.if_outDiscard = u.unpack_uint()
self.if_outError = u.unpack_uint()
self.if_promisc = u.unpack_uint()
class EthernetCounters(object):
"""Counters for ethernet frames
"""
def __init__(self, u):
self.dot3StatsAlignmentErrors = u.unpack_uint()
self.dot3StatsFCSErrors = u.unpack_uint()
self.dot3StatsSingleCollisionFrames = u.unpack_uint()
self.dot3StatsMultipleCollisionFrames = u.unpack_uint()
self.dot3StatsSQETestErrors = u.unpack_uint()
self.dot3StatsDeferredTransmissions = u.unpack_uint()
self.dot3StatsLateCollisions = u.unpack_uint()
self.dot3StatsExcessiveCollisions = u.unpack_uint()
self.dot3StatsInternalMacTransmitErrors = u.unpack_uint()
self.dot3StatsCarrierSenseErrors = u.unpack_uint()
self.dot3StatsFrameTooLongs = u.unpack_uint()
self.dot3StatsInternalMacReceiveErrors = u.unpack_uint()
self.dot3StatsSymbolErrors = u.unpack_uint()
class VLANCounters(object):
"""Counters for VLANs
"""
def __init__(self, u):
self.vlan_id = u.unpack_uint()
self.octets = u.unpack_uhyper()
self.ucastPkts = u.unpack_uint()
self.multicastPkts = u.unpack_uint()
self.broadcastPkts = u.unpack_uint()
self.discards = u.unpack_uint()
class TokenringCounters(object):
"""Counters for Token ring networks
"""
def __init__(self, u):
self.dot5StatsLineErrors = u.unpack_uint()
self.dot5StatsBurstErrors = u.unpack_uint()
self.dot5StatsACErrors = u.unpack_uint()
self.dot5StatsAbortTransErrors = u.unpack_uint()
self.dot5StatsInternalErrors = u.unpack_uint()
self.dot5StatsLostFrameErrors = u.unpack_uint()
self.dot5StatsReceiveCongestions = u.unpack_uint()
self.dot5StatsFrameCopiedErrors = u.unpack_uint()
self.dot5StatsTokenErrors = u.unpack_uint()
self.dot5StatsSoftErrors = u.unpack_uint()
self.dot5StatsHardErrors = u.unpack_uint()
self.dot5StatsSignalLoss = u.unpack_uint()
self.dot5StatsTransmitBeacons = u.unpack_uint()
self.dot5StatsRecoverys = u.unpack_uint()
self.dot5StatsLobeWires = u.unpack_uint()
self.dot5StatsRemoves = u.unpack_uint()
self.dot5StatsSingles = u.unpack_uint()
self.dot5StatsFreqErrors = u.unpack_uint()
class VGCounters(object):
"""Counters for AnyLan frames
"""
def __init__(self, u):
self.dot5StatsLineErrors = u.unpack_uint()
self.dot5StatsBurstErrors = u.unpack_uint()
self.dot5StatsACErrors = u.unpack_uint()
self.dot5StatsAbortTransErrors = u.unpack_uint()
self.dot5StatsInternalErrors = u.unpack_uint()
self.dot5StatsLostFrameErrors = u.unpack_uint()
self.dot5StatsReceiveCongestions = u.unpack_uint()
self.dot5StatsFrameCopiedErrors = u.unpack_uint()
self.dot5StatsTokenErrors = u.unpack_uint()
self.dot5StatsSoftErrors = u.unpack_uint()
self.dot5StatsHardErrors = u.unpack_uint()
self.dot5StatsSignalLoss = u.unpack_uint()
self.dot5StatsTransmitBeacons = u.unpack_uint()
self.dot5StatsRecoverys = u.unpack_uint()
self.dot5StatsLobeWires = u.unpack_uint()
self.dot5StatsRemoves = u.unpack_uint()
self.dot5StatsSingles = u.unpack_uint()
self.dot5StatsFreqErrors = u.unpack_uint()
class HostCounters(object):
"""Counters for host endpoints
"""
format = 2000
def __init__(self, u):
self.hostname = u.unpack_string()
self.uuid = u.unpack_fopaque(16)
self.machine_type = u.unpack_uint()
self.os_name = u.unpack_uint()
self.os_release = u.unpack_string()
class HostAdapters(object):
"""Counters for HBAs
"""
format = 2001
def __init__(self, u):
self.adapters = Struct("adapters",
UBInt32("count"),
Array(lambda c: c.count,
Struct("adapter",
UBInt32("index"),
Bytes("MAC", 6)))
).parse(u.get_buffer())
class HostParent(object):
"""Counters for hosts
"""
format = 2002
def __init__(self, u):
self.container_type = u.unpack_uint()
self.container_index = u.unpack_uint()
class HostCPUCounters(object):
"""Counters for host CPU stats
"""
format = 2003
def __init__(self, u):
self.load_one = u.unpack_float()
self.load_five = u.unpack_float()
self.load_fifteen = u.unpack_float()
self.proc_run = u.unpack_uint()
self.proc_total = u.unpack_uint()
self.cpu_num = u.unpack_uint()
self.cpu_speed = u.unpack_uint()
self.uptime = u.unpack_uint()
self.cpu_user = u.unpack_uint()
self.cpu_nice = u.unpack_uint()
self.cpu_system = u.unpack_uint()
self.cpu_idle = u.unpack_uint()
self.cpu_wio = u.unpack_uint()
self.cpu_intr = u.unpack_uint()
self.cpu_sintr = u.unpack_uint()
self.interrupts = u.unpack_uint()
self.contexts = u.unpack_uint()
class HostMemoryCounters(object):
"""Counters for host memory
"""
format = 2004
def __init__(self, u):
self.mem_total = u.unpack_uhyper()
self.mem_free = u.unpack_uhyper()
self.mem_shared = u.unpack_uhyper()
self.mem_buffers = u.unpack_uhyper()
self.mem_cached = u.unpack_uhyper()
self.swap_total = u.unpack_uhyper()
self.swap_free = u.unpack_uhyper()
self.page_in = u.unpack_uint()
self.page_out = u.unpack_uint()
self.swap_in = u.unpack_uint()
self.swap_out = u.unpack_uint()
class DiskIOCounters(object):
"""Counters for disk IO
"""
format = 2005
def __init__(self, u):
self.disk_total = u.unpack_uhyper()
self.disk_free = u.unpack_uhyper()
self.part_max_used = u.unpack_uint()
self.reads = u.unpack_uint()
self.bytes_read = u.unpack_uhyper()
self.read_time = u.unpack_uint()
self.writes = u.unpack_uint()
self.bytes_written = u.unpack_uhyper()
self.write_time = u.unpack_uint()
class NetIOCounters(object):
"""Counters for network interface IO
"""
format = 2006
def __init__(self, u):
self.bytes_in = u.unpack_uhyper()
self.pkts_in = u.unpack_uint()
self.errs_in = u.unpack_uint()
self.drops_in = u.unpack_uint()
self.bytes_out = u.unpack_uhyper()
self.packets_out = u.unpack_uint()
self.errs_out = u.unpack_uint()
self.drops_out = u.unpack_uint()
class SocketIPv4Counters(object):
"""Counters for IPv4 sockets
"""
format = 2100
def __init__(self, u):
self.protocol = u.unpack_uint()
self.local_ip = u.unpack_fstring(4)
self.remote_ip = u.unpack_fstring(4)
self.local_port = u.unpack_uint()
self.remote_port = u.unpack_uint()
class SocketIPv6Counters(object):
"""Counters for IPv6 sockets
"""
format = 2101
def __init__(self, u):
self.protocol = u.unpack_uint()
self.local_ip = u.unpack_fstring(16)
self.remote_ip = u.unpack_fstring(16)
self.local_port = u.unpack_uint()
self.remote_port = u.unpack_uint()
class VirtMemoryCounters(object):
"""Counters for virtual memory
"""
format = 2102
def __init__(self, u):
self.memory = u.unpack_uhyper()
self.maxMemory = u.unpack_uhyper()
class VirtDiskIOCounters(object):
"""Counters for virtual disk IO
"""
format = 2103
def __init__(self, u):
self.capacity = u.unpack_uhyper()
self.allocation = u.unpack_uhyper()
self.available = u.unpack_uhyper()
self.rd_req = u.unpack_uint()
self.hyper = u.unpack_unsigend()
self.wr_req = u.unpack_uint()
self.wr_bytes = u.unpack_uhyper()
self.errs = u.unpack_uint()
class VirtNetIOCounters(object):
"""Counters for virtual network adapters
"""
format = 2104
def __init__(self, u):
self.rx_bytes = u.unpack_uhyper()
self.rx_packets = u.unpack_uint()
self.rx_errs = u.unpack_uint()
self.rx_drop = u.unpack_uint()
self.tx_bytes = u.unpack_uhyper()
self.tx_packets = u.unpack_uint()
self.tx_errs = u.unpack_uint()
self.tx_drop = u.unpack_uint()
def getDecoder(fmt):
"""Retrieve the decoder associated with a frame format ID
"""
decoders = {
1: InterfaceCounters,
2: EthernetCounters,
3: TokenringCounters,
4: VGCounters,
5: VLANCounters,
2000: HostCounters,
2001: HostAdapters,
2002: HostParent,
2003: HostCPUCounters,
2004: HostMemoryCounters,
2005: DiskIOCounters,
2006: NetIOCounters,
2101: SocketIPv6Counters,
2102: VirtMemoryCounters,
2103: VirtDiskIOCounters,
2104: VirtNetIOCounters
}
return decoders.get(fmt, None)
|
ducted/duct
|
duct/protocol/sflow/protocol/counters.py
|
Python
|
mit
| 10,161
|
import configparser
import importlib.util
import random
from pdb import set_trace
from warnings import warn
import numpy as np
import pygame
from PIL import Image
import os
from behaviours.Collide import Collide
from src.utils.CodeItWarning import CodeItWarning
from tiles.base.Tile import Tile
level_paths = {}
level_configs = {}
level_backgrounds = {}
entity_map = {}
image_file_formats = [".png", ".jpg", ".jpeg", ".bmp"]
def constructor_factory(constructor, name):
return lambda x, y: constructor(x, y, name)
def load_entity_map():
for path in [f.path for f in os.scandir("../entities") if f.is_dir()]:
path = [x for x in os.scandir(path) if os.path.splitext(x)[1] == ".py"][0]
class_name = os.path.splitext(os.path.basename(path))[0]
if ' ' in class_name:
raise ValueError("Entities cannot have spaces in their names, '" + class_name + "'")
spec = importlib.util.spec_from_file_location("dynamic_load.entities." + class_name, path)
foo = importlib.util.module_from_spec(spec)
spec.loader.exec_module(foo)
class_constructor = getattr(foo, class_name)
entity_map[class_name] = constructor_factory(class_constructor, class_name)
for path in [f.path for f in os.scandir("../tiles") if f.is_dir()]:
paths = [x for x in os.scandir(path) if os.path.splitext(x)[1] in image_file_formats]
if len(paths) <= 0:
continue
path = paths[0]
class_name = os.path.splitext(os.path.basename(path))[0]
if ' ' in class_name:
raise ValueError("Tiles cannot have spaces in their names, '" + class_name + "'")
entity_map[class_name] = constructor_factory(Tile, class_name)
load_entity_map()
def load_levels():
for path in [f.path for f in os.scandir("../levels") if f.is_dir()]:
config = configparser.ConfigParser()
config.read(path + "/config.ini")
level_name = config["General"]["Name"]
level_paths[level_name] = path
level_configs[level_name] = config
paths = [x for x in os.scandir(path) if "background" in os.path.splitext(x)[0]]
if len(paths) <= 0:
continue
path = paths[0]
level_backgrounds[level_name] = pygame.image.load(os.path.relpath(path, os.getcwd()))
def get_level_by_index(index):
level_list = []
for level_name in level_configs.keys():
level_index = int(level_configs[level_name]["General"]["Index"])
if level_index == index:
level_list.append(level_name)
if len(level_list) <= 0:
return
if len(level_list) == 1:
return Level(level_list[0])
return Level(random.choice(level_list))
load_levels()
def hex_to_rgb(h):
return tuple(int(h[i:i + 2], 16) for i in (0, 2, 4))
def rgb_to_hex(rgb: tuple) -> str:
return ('%02x%02x%02x' % rgb).upper()
def load_config(path):
config = configparser.ConfigParser()
config.read(path)
return config
default_color_map = load_config("../levels/default/color-map.ini")
# returns a dictionary that maps rgb tuples to entity constructors
# @param entity_map: a dictionary that maps entity names to entity constructors.
def get_color_map(level):
color_map = {}
config = combine_configs(load_config(level.path + "/color-map.ini"), default_color_map)
for key in config["Colors"]:
color_map[hex_to_rgb(key)] = (config["Colors"][key], entity_map[config["Colors"][key]])
return color_map
def combine_configs(c1, c2):
combined = {}
for category in list(set().union(c1.keys(), c2.keys())):
combined[category] = {**(c2[category] if category in c2 else {}), **(c1[category] if category in c1 else {})}
return combined
def load_map_image(level):
arr3d = np.array(Image.open(level.path + "/map.bmp"))
return arr3d, arr3d.shape[:2]
def load_entities(color_map, map_image, image_shape):
entities = []
entity_lookup = {}
tiles = []
for ix, iy in np.ndindex(image_shape):
rgb = tuple(map_image[ix, iy])
if rgb in color_map:
entity = color_map[rgb][1](iy, ix)
entity_name = color_map[rgb][0]
if isinstance(entity, Tile):
tiles.append(entity)
else:
entity.remain_on_reset = True
if entity_name in entity_lookup:
entity_lookup[entity_name].append(entity)
else:
entity_lookup[entity_name] = [entity]
entities.append(entity)
else:
warn("CodeIT:: No Tile or entity mapped to " + rgb_to_hex(rgb), category=CodeItWarning, stacklevel=0)
return entities, tiles, entity_lookup
class Level:
def __init__(self, name):
if name == "Default":
raise ValueError("Do not instantiate the default level")
self.path = level_paths[name]
self.config = combine_configs(level_configs[name], level_configs["Default"])
self.name = name
self.entities = list()
self.tiles = list()
self.color_map = {}
self.map_image = None
self.map_shape = (0, 0)
self.background = None
self.entity_lookup_map = {}
self.graveyard = []
def load(self):
try:
if self.config["General"]["type"] == "Pure GUI":
return
except KeyError:
pass
self.color_map = get_color_map(self)
self.map_image, self.map_shape = load_map_image(self)
self.entities, self.tiles, self.entity_lookup_map = load_entities(self.color_map, self.map_image,
self.map_shape)
self.level_time = int(self.config["General"]["timelimit"])
if self.name in level_backgrounds:
self.background = level_backgrounds[self.name]
else:
self.background = level_backgrounds["Default"]
return
def kill_entity(self, entity):
if entity in self.entities:
self.entities.remove(entity)
entity.get_behaviour(Collide).clear()
entity.is_dead = True
if entity.remain_on_reset:
self.graveyard.append(entity)
def revive_entities(self):
self.entities.extend(self.graveyard)
for entity in self.graveyard:
entity.is_dead = False
self.graveyard.clear()
def get_entities(self, entity_name):
if entity_name in self.entity_lookup_map:
return self.entity_lookup_map[entity_name]
raise Exception("Invalid entity!, " + entity_name)
def get_y(self, number, unit):
if unit == "percent":
return self.map_shape[1] * number / 100
return number
def get_x(self, number, unit):
if unit == "percent":
return self.map_shape[0] * number / 100
return number
def spawn_entity(self, entity_class, x, y):
entity_name = entity_class.__name__
entity = entity_class(x, y, entity_name)
self.entities.append(entity)
if entity_name in self.entity_lookup_map:
self.entity_lookup_map[entity_name].append(entity)
else:
self.entity_lookup_map[entity_name] = [entity]
return entity
def clear(self):
for tile in self.tiles:
tile.clear()
self.tiles.clear()
for entity in self.entities:
entity.clear()
self.entities.clear()
self.entity_lookup_map.clear()
def __del__(self):
self.clear()
|
cthit/CodeIT
|
src/level/Level.py
|
Python
|
mit
| 7,583
|
# -*- coding: utf-8 -*-
# Generated by Django 1.9.8 on 2016-07-29 18:16
from __future__ import unicode_literals
from django.db import migrations, models
class Migration(migrations.Migration):
dependencies = [
('django_sendgrid_parse', '0001_initial'),
]
operations = [
migrations.RenameField(
model_name='email',
old_name='to',
new_name='to_mailbox',
),
migrations.AddField(
model_name='email',
name='from_mailbox',
field=models.TextField(default=''),
preserve_default=False,
),
]
|
letops/django-sendgrid-parse
|
django_sendgrid_parse/migrations/0002_auto_20160729_1816.py
|
Python
|
mit
| 626
|
import datetime
import calendar
import dabo
dabo.ui.loadUI("wx")
from dabo.ui import dForm, dPanel, dSizer, dGridSizer, dButton, dEditBox, \
dTextBox, dControlMixin, callAfterInterval, dKeys, \
dLabel, dHyperLink
from dabo.lib.dates import goMonth, goDate
import biz
__all__ = ["FrmCalendar"]
class BaseMixin(dControlMixin):
"""Base behavior for all the relevant controls."""
def initProperties(self):
self.BorderStyle = None
def onGotFocus(self, evt):
callAfterInterval(10, self.Parent.gotFocus)
def onLostFocus(self, evt):
callAfterInterval(10, self.Parent.lostFocus)
class PnlNavigation(dPanel):
def afterInit(self):
hs = self.Sizer = dSizer("h")
but_props = {"FontBold": True, "ShowInBrowser": False,
"OnHit": self.onHit_but, "VisitedUnderline": False,
"LinkUnderline": False, "VisitedColor": "black",
"HoverUnderline": False, "LinkColor": "black"}
left_but = dHyperLink(self, Name="butLeft", Caption=" < ", **but_props)
right_but = dHyperLink(self, Name="butRight", Caption=" > ", **but_props)
lbl = dLabel(self, Name="lblMonthYear", FontBold=True)
hs.appendSpacer(20)
hs.append(left_but)
hs.appendSpacer(10)
hs.append(lbl, alignment="middle")
hs.appendSpacer(10)
hs.append(right_but)
hs.appendSpacer(20)
def setCaption(self, val):
self.lblMonthYear.Caption = val
self.layout()
callAfterInterval(100, self.Parent.Sizer.layout)
def onHit_but(self, evt):
pnlLayout = self.Form.pnlLayout
interval = {"butLeft": -1, "butRight": 1,}[evt.EventObject.Name]
new_date = goMonth(datetime.date(pnlLayout.Year, pnlLayout.Month, 1),
interval)
pnlLayout.Year = new_date.year
pnlLayout.Month = new_date.month
pnlLayout.setFocus()
class DummyTextBox(dTextBox, BaseMixin):
"""Invisible textbox simply to receive and respond to user actions."""
def initProperties(self):
## Dabo doesn't yet offer an interface to set wx.TE_PROCESS_ENTER.
import wx
self.Size = (0, 0)
self._addWindowStyleFlag(wx.TE_PROCESS_ENTER)
def onKeyDown(self, evt):
self.Parent.processDayKeyDown(evt)
class EditMixin(BaseMixin):
"""Base behavior for the edit boxes."""
def initProperties(self):
super(EditMixin, self).initProperties()
self.Height = 5
self.Width = 5
self.DynamicBackColor = self.getBackColor
def onLostFocus(self, evt):
super(EditMixin, self).onLostFocus(evt)
self.save()
def save(self):
"""Subclasses must override if they want entered data to be saved."""
pass
def getBackColor(self):
"""Color depends on whether this day is in the current month."""
try:
return (
"white" if self.Parent.Date.month == self.Parent.Parent.Month
else "lightgrey")
except StandardError:
return "darkgrey"
class Day(dLabel):
"""The day number in the upper-left of each day panel."""
def initProperties(self):
self.Width = 30
self.Height = 23
self.Name = "day"
self.ForeColor = "blue"
def onMouseLeftClick(self, evt):
self.Parent.setFocus()
def onMouseLeftDoubleClick(self, evt):
self.Parent.diary.setFocus()
class Static(dTextBox, EditMixin):
"""View/edit/save the diary entries that repeat on this day every year."""
def initProperties(self):
self.FontItalic = True
self.Name = "static"
super(Static, self).initProperties()
def save(self):
cal = self.Parent.Parent
day = self.Parent
bizStatic = cal.bizStatic
monthday = "%s%s" % (day.Date.month, day.Date.day)
if not bizStatic.locate(monthday, "monthday"):
bizStatic.new()
bizStatic.Record.monthday = monthday
bizStatic.Record.diary = self.Value
bizStatic.save()
class DiaryView(dEditBox, EditMixin):
"""The view of the day's diary (no unneeded scrollbars)."""
def initProperties(self):
## dabo doesn't yet allow overriding the display of scrollbars.
import wx
self.Name = "diary"
self._addWindowStyleFlag(wx.TE_NO_VSCROLL)
super(DiaryView, self).initProperties()
def onGotFocus(self, evt):
"""Switch to the edit control (with scrollbars)."""
edit = self.Parent.diaryedit
edit.Value = self.Value
edit.Visible = True
self.Visible = False
self.Parent.layout()
edit.setFocus()
def onLostFocus(self, evt):
## overridden to avoid unneccesary save()
pass
class DiaryEdit(dEditBox, EditMixin):
"""The editable view of the day's diary (with scrollbars)."""
def initProperties(self):
self.Name = "diaryedit"
self.Visible = False
super(DiaryEdit, self).initProperties()
def save(self):
cal = self.Parent.Parent
day = self.Parent
bizDaily = cal.bizDaily
date = day.Date
if not bizDaily.locate(date, "date"):
bizDaily.new()
bizDaily.Record.date = date
bizDaily.Record.diary = self.Value
bizDaily.save()
def onLostFocus(self, evt):
super(DiaryEdit, self).onLostFocus(evt)
view = self.Parent.diary
view.Value = self.Value
view.Visible = True
self.Visible = False
self.Value = ""
self.Parent.layout()
class PnlDay(dPanel):
"""Panel of a calendar day, including the day number and edit controls."""
def initProperties(self):
self.BorderStyle = "Raised"
self.DynamicBackColor = self.getBackColor
self._hadFocus = False
def afterInit(self):
self.dummy = DummyTextBox(self)
vs = self.Sizer = dSizer("v")
hs = dSizer("h")
hs.append(Day(self), "expand")
hs.append1x(Static(self))
vs.append(hs, "expand")
vs.append1x(DiaryView(self))
vs.append1x(DiaryEdit(self))
parent = self.Parent
diary = self.diary
self.key_actions = {
(dKeys.key_Enter, dKeys.key_Numpad_enter): (diary.setFocus, ()),
(ord("t"), ord("T")): (parent.setFocusToToday, ()),
(ord("+"), ord("=")): (self.setFocusToInterval, ("days", 1)),
(ord("-"),): (self.setFocusToInterval, ("days", -1)),
(ord("["),): (self.setFocusToInterval, ("months", -1)),
(ord("]"),): (self.setFocusToInterval, ("months", 1))}
def gotFocus(self):
day = self.day
dummy = self.dummy
if not self._hadFocus:
day.FontBold = True
day.ForeColor = "darkblue"
self._hadFocus = True
day.refresh()
def lostFocus(self):
if self.Form.ActiveControl not in self.Children:
self._hadFocus = False
day = self.day
day.FontBold = False
day.ForeColor = "blue"
self.refresh()
def setFocus(self):
self.dummy.setFocus()
def setFocusToInterval(self, mode, interval):
"""Go forward or backward by months or days."""
assert mode in ("days", "months")
func = {"days": goDate, "months": goMonth}[mode]
date = self.Date
self.Parent.setFocusToDate(func(date, interval))
def processDayKeyDown(self, evt):
"""User is navigating the calendar; respond appropriately."""
evtData = evt.EventData
kc = evtData["keyCode"]
ctrlDown = evtData["controlDown"]
layout = self.Form.CalendarLayout
for keys, func_args in self.key_actions.items():
if kc in keys:
func, args = func_args
func(*args)
evt.stop()
return
if kc not in [dKeys.key_Up, dKeys.key_Down,
dKeys.key_Left, dKeys.key_Right]:
return
evt.stop()
if not ctrlDown:
# move by day, wrapping around in the calendar
x,y = self.Pos
if kc == dKeys.key_Up and layout in ("month",):
y -= 1
if y < 0:
y = 5
elif kc == dKeys.key_Down and layout in ("month",):
y += 1
if y > 5:
y = 0
elif kc == dKeys.key_Left:
x -= 1
if x < 0:
x = 6
elif kc == dKeys.key_Right:
x += 1
if x > 6:
x = 0
new_ctrl = getattr(self.Parent, "day_%s_%s" % (x,y))
new_ctrl.setFocus()
else:
year, month = self.Parent.Year, self.Parent.Month
current_date = datetime.date(year, month, 1)
if kc == dKeys.key_Left:
new_date = goMonth(current_date, -1)
elif kc == dKeys.key_Right:
new_date = goMonth(current_date, 1)
elif kc == dKeys.key_Up:
new_date = goMonth(current_date, -12)
elif kc == dKeys.key_Down:
new_date = goMonth(current_date, 12)
self.Parent.Year = new_date.year
self.Parent.Month = new_date.month
def getBackColor(self):
try:
return ("white" if self.Date.month == self.Parent.Month
else "lightgrey")
except StandardError:
return "darkgrey"
def _getPos(self):
return self._pos
def _setPos(self, val):
self._pos = val
self.Name = "day_%s_%s" % val
def _getDate(self):
return self._date
def _setDate(self, val):
self._date = val
self.day.Caption = str(val.day)
Pos = property(_getPos, _setPos)
Date = property(_getDate, _setDate)
class PnlLayout(dPanel):
"""Superclass to handle common elements for views (day, month, week)."""
_week_range = None
def afterInit(self):
con = self.Form.Connection
gs = self.Sizer = dGridSizer(MaxCols=7)
self.bizStatic = biz.BizStatic(con)
self.bizDaily = biz.BizDaily(con)
header = calendar.weekheader(3).split()
for x in header:
gs.append(dLabel(self, Caption=x), alignment="center")
for y in range(self._week_range):
for x in range(7):
gs.append(PnlDay(self, Pos=(x,y)), "expand")
gs.setColExpand(True, x)
gs.setRowExpand(True, y+1)
self.setFocusToToday()
def afterDateChanged(self):
self.setFormCaption()
self.setDays()
def setFormCaption(self):
current_date = datetime.date(self.Year, self.Month, 1)
self.Form.setCaption("%s %s" % (
current_date.strftime(calendar.month_name.format), self.Year))
def setDays(self):
mv = biz.getMonthMatrix(self.Year, self.Month)
bizStatic = self.bizStatic
bizStatic.requery()
bizDaily = self.bizDaily
bizDaily.requery_for_dates(mv[0][0], mv[-1][-1])
self.date_obj_map = {}
for y in range(self._week_range):
for x in range(7):
o = getattr(self, "day_%s_%s" % (x,y))
o.Date = mv[y][x]
if bizStatic.locate("%s%s" % (
o.Date.month, o.Date.day), "monthday"):
o.static.Value = bizStatic.Record.diary
else:
o.static.Value = ""
if bizDaily.locate(o.Date, "date"):
o.diary.Value = bizDaily.Record.diary
else:
o.diary.Value = ""
self.date_obj_map[o.Date] = o
self.update()
def setFocusToToday(self):
"""Requerying the calendar if necessary, place cursor on today."""
self.setFocusToDate(datetime.date.today())
def setFocusToDate(self, date):
"""Requerying the calendar if necessary, place cursor on date."""
try:
self.date_obj_map[date].setFocus()
except (KeyError, AttributeError):
self.Year = date.year
self.Month = date.month
callAfterInterval(75, self.setFocusToDate, date)
def _getMonth(self):
return self._month
def _setMonth(self, val):
self._month = val
callAfterInterval(50, self.afterDateChanged)
def _getYear(self):
return self._year
def _setYear(self, val):
self._year = val
callAfterInterval(50, self.afterDateChanged)
Month = property(_getMonth, _setMonth)
Year = property(_getYear, _setYear)
class PnlMonth(PnlLayout):
"""Standard 7x6 month layout."""
_week_range = 6
class PnlWeek(PnlLayout):
"""7x1 week layout (INCOMPLETE)."""
_week_range = 1
class FrmCalendar(dForm):
"""The main containing form of the calendar."""
def afterInit(self):
self._appendCaption = ""
dcon = self.Connection
if dcon is None:
# use in-memory test sqlite database
dcon = self.Connection = dabo.db.connect(":memory:")
con = dcon._connection
con.executescript(open("./create_tables.sql").read())
self._appendCaption = "Temporary Database"
self._instantiatedLayouts = {}
vs = self.Sizer = dSizer("v")
vs.appendSpacer(5)
vs.append(PnlNavigation(self, Name="pnlNavigation"), alignment="center")
vs.appendSpacer(5)
self.updateLayout()
def updateLayout(self):
"""Draw the calendar on screen depending on self.Layout."""
pnls = self._instantiatedLayouts
PnlClass = {"month": PnlMonth, "week": PnlWeek}[self.CalendarLayout]
vs = self.Sizer
for pnl in pnls.values():
pnl.Visible = False
pnl = self.pnlLayout = pnls.setdefault(PnlClass, PnlClass(self))
if pnl not in vs.ChildWindows:
vs.append1x(pnl)
self.layout()
def setCaption(self, val):
appendCaption = self._appendCaption
if appendCaption:
appendCaption = "[%s]" % appendCaption
self.Caption = "%s %s" % (val, appendCaption)
self.pnlNavigation.setCaption(val)
def _getConnection(self):
return getattr(self, "_connection", None)
def _setConnection(self, val):
self._connection = val
def _getCalendarLayout(self):
return getattr(self, "_calendar_layout", "month")
def _setCalendarLayout(self, val):
self._calendar_layout = val
assert val in ("month", "week")
dabo.ui.callAfterInterval(10, self.updateLayout())
CalendarLayout = property(_getCalendarLayout, _setCalendarLayout, None,
"""Either "month" or "week".""")
Connection = property(_getConnection, _setConnection, None,
"Dabo dConnection instance.")
if __name__ == "__main__":
# If you want Sunday to be the first weekday, you need code like
# the following commented lines in your application prior to
# importing pmcalendar:
#import calendar
#calendar.setfirstweekday(6)
dabo.dApp(MainFormClass=FrmCalendar).start()
|
pmcnett/pmcalendar
|
pmcalendar/ui.py
|
Python
|
mit
| 15,413
|
# -*- coding: utf-8 -*-
# Define here the models for your scraped items
#
# See documentation in:
# http://doc.scrapy.org/en/latest/topics/items.html
import scrapy
class NewsItem(scrapy.Item):
# define the fields for your item here like:
# name = scrapy.Field()
# pass
news_thread=scrapy.Field()
news_title = scrapy.Field()
news_url = scrapy.Field()
news_time=scrapy.Field()
news_from=scrapy.Field()
from_url=scrapy.Field()
news_body=scrapy.Field()
|
DavidFnck/Python_Stock_Github
|
news/news/items.py
|
Python
|
mit
| 494
|
from django import forms
class SearchMovieForm(forms.Form):
search = forms.CharField(label='', max_length=25, widget=forms.TextInput(
attrs={'class': 'form-control'}))
choices = [('movie', 'MOVIES'),
('series', 'TV')]
choice = forms.ChoiceField(label='', choices=choices, widget=forms.Select(
attrs={'class': 'dropdown'}))
|
rish4bhn/movieclue
|
movieclue/forms.py
|
Python
|
mit
| 370
|
from django.contrib import admin
from cathedra.models import Lector, Room, Position, Science, Aspirant, Technician
admin.site.register(Lector)
admin.site.register(Aspirant)
admin.site.register(Technician)
admin.site.register(Room)
admin.site.register(Position)
admin.site.register(Science)
|
Axik/eapu
|
apps/cathedra/admin.py
|
Python
|
mit
| 293
|
import sublime_plugin
from php_coverage.data import CoverageDataFactory
from php_coverage.finder import CoverageFinder
from php_coverage.matcher import Matcher
class CoverageCommand(sublime_plugin.TextCommand):
"""
Base class for a text command which has a coverage file.
"""
def __init__(self, view, coverage_finder=None, matcher=None):
super(CoverageCommand, self).__init__(view)
self.coverage_finder = coverage_finder
self.matcher = matcher
def get_coverage_finder(self):
"""
Gets the coverage finder for the command. If none is set, it
instantiates an instance of the default CoverageFinder class.
"""
if not self.coverage_finder:
self.coverage_finder = CoverageFinder()
return self.coverage_finder
def coverage(self):
"""
Loads coverage data for the file open in the view which is
running this command.
"""
filename = self.view.file_name()
coverage_file = self.get_coverage_finder().find(filename)
if (coverage_file):
return CoverageDataFactory().factory(coverage_file)
return None
def get_matcher(self):
"""
Gets the matcher for the command. If none is set, it
instantiates an instance of the default Matcher class.
"""
if not self.matcher:
self.matcher = Matcher()
return self.matcher
def should_include(self, filename):
"""
Determines whether a file should be included or not.
"""
return self.get_matcher().should_include(filename)
|
bradfeehan/SublimePHPCoverage
|
php_coverage/command.py
|
Python
|
mit
| 1,636
|
from tests import TestCase
from werkzeug.urls import url_quote
from datamart.models import Role
from flask.ext.security import current_user
class TestRoles(TestCase):
def test_show_roles_anon(self):
"""Verify unathenticated users can't see the roles page."""
response = self.client.get('/roles/', follow_redirects=False)
new_location='/login?next=%s' % url_quote('/roles/', safe='')
self.assertRedirects(response, location=new_location)
response = self.client.get('/roles/', follow_redirects=True)
assert 'Please log in to access this page.' in response.data
self.assertTemplateUsed(name='login.html')
def test_show_roles_non_admin(self):
"""Make sure logged in non-admin users can't see the roles page."""
self.login('demo@example.com','123456')
assert current_user.is_authenticated
response = self._test_get_request('/roles/', 'index.html', follow_redirects=True)
assert 'Permission denied' in response.data
self.logout()
def test_show_roles_admin(self):
"""Make sure logged in admins can see the roles page."""
self.login('admin@example.com','123456')
response = self._test_get_request('/roles/', 'roles.html')
assert 'Please log in to access this page.' not in response.data
self.logout()
def test_role_add(self):
"""Adds a role using a post to /roles/add/"""
self.login('admin@example.com', '123456')
self._test_get_request('/roles/add/', 'role_edit.html')
data = {
'name': 'Ruler',
'description': "of all I survey.",
}
response = self.client.post('/roles/add/', data=data)
assert 'Please fix errors and resubmit.' not in response.data
new_role = Role.query.filter_by(name=data['name'])
assert new_role.count() == 1
self.logout()
def test_role_edit(self):
"""Edit a role using webforms."""
self.login('admin@example.com', '123456')
data = {
'name': 'Ruler',
'description': "of all I survey.",
}
new_role = Role.query.filter_by(name=data['name'])
if new_role.count() != 1:
response = self.client.post('/roles/add/', data=data)
assert 'Please fix errors and resubmit.' not in response.data
new_role = Role.query.filter_by(name=data['name'])
assert new_role.count() == 1
data['name'] = 'Peon'
response = self.client.post('/roles/%s/edit/' % new_role.first().id, data=data)
assert 'Please fix errors and resubmit.' not in response.data
new_role = Role.query.filter_by(name=data['name'])
assert new_role.count() == 1
response = self.client.get('/roles/')
assert data['name'] in response.data
self.logout()
|
msscully/datamart
|
tests/test_roles.py
|
Python
|
mit
| 2,860
|
#!/usr/bin/env python
import os
import sys
if __name__ == "__main__":
os.environ.setdefault("DJANGO_SETTINGS_MODULE", "seeker.settings.prod")
from django.core.management import execute_from_command_line
execute_from_command_line(sys.argv)
|
LomaxRx/service-seeker
|
manage.py
|
Python
|
mit
| 254
|
from voluptuous import All, Match, Coerce
from datatypes.core import SingleValueValidator
price_schema = All(Coerce(float), Coerce(unicode), Match('^[0-9]+(,[0-9]+)?(\.\d{1,2})?$'))
class Price(SingleValueValidator):
def define_schema(self):
return price_schema
def define_error_message(self):
return "Price must be a numeric value in pounds"
|
LandRegistry/datatypes-alpha
|
datatypes/validators/price_validator.py
|
Python
|
mit
| 373
|
#
# Walks every page that it can find in a website
#
# It's an excuse to play with BeautifulSoup
import re
from io import BytesIO
from bs4 import BeautifulSoup
from pycurl import Curl
from queue import Queue, Empty as QueueEmpty
from urllib.parse import urlsplit, urlunsplit, urljoin
from sys import stdout
class PageFetcher():
"""Fetches a page"""
def __init__(self):
self.curl = Curl()
self.url = None
self.headers = {}
self.status = ''
self.code = 0
self.charset_re = re.compile('charset=(\S+)')
def handle_headers(self, header):
"""Parses the headers from a HTTP response"""
header = header.decode('iso-8859-1') #headers always in iso-8859-1
if ':' in header:
#split out the headers name and value
n, v = header.split(': ', 1)
self.headers[n] = v.rstrip('\r\n')
elif 'HTTP' in header:
h, code, status = header.split(' ', 2)
self.code = int(code)
self.status = status.rstrip('\r\n')
def encoding(self):
"""Gets the encoding from the headers, otherwise assumes iso-8859-1"""
if 'Content-Type' in self.headers:
match = self.charset_re.search(self.headers['Content-Type'].lower())
if match:
return match.group(1)
return 'iso-8859-1'
def fetch(self, url, headers_only=False):
"""Gets the specified webpage"""
#reset the gathered data
self.headers = {}
self.code = 0
self.status = None
links = []
#get the page
buff = BytesIO()
self.curl.setopt(self.curl.URL, url)
if headers_only:
self.curl.setopt(self.curl.NOBODY, 1)
else:
self.curl.setopt(self.curl.NOBODY, 0)
self.curl.setopt(self.curl.WRITEDATA, buff)
self.curl.setopt(self.curl.HEADERFUNCTION, self.handle_headers)
self.curl.perform()
#decode the returned data to the correct type
body = buff.getvalue().decode(self.encoding())
return self.code, self.headers, body
class Spider:
"""Fetches every page within a website"""
def __init__(self):
self.htmlpage = re.compile('text/(html|xml)|application/(xhtml\+xml|xml)')
def walk(self, url, pagefilter = None):
"""Walks all pages in the given website"""
if not url.startswith('http'):
url = 'http://' + url
fetch = PageFetcher()
queued = Queue()
queued.put(url)
processed = set()
while not queued.empty():
url = queued.get()
if url not in processed:
have_body = False
processed.add(url)
code, headers, body = fetch.fetch(url, True) #only get headers
if code >= 300 and code <= 399 and 'Location' in headers:
locpart = urlsplit(headers['Location'])
urlpart = urlsplit(url)
if locpart.scheme == urlpart.scheme and locpart.netloc == urlpart.netloc:
queued.put(headers['Location'])
#parse out the links if it's a html page
if self.ishtml(headers):
code, headers, body = fetch.fetch(url) #get the whole page
have_body = True
links = self.sitelinks(body, url)
for l in links:
queued.put(l)
#pass the page onto the processor, if it passes the defined filter
if not pagefilter or (pagefilter and pagefilter(code, headers)):
if not have_body:
code, headers, body = fetch.fetch(url) #get the whole page
self.process_page(url, code, headers, body)
def ishtml(self, headers):
'''Determines if the retrieved page is a (x)html page'''
if 'Content-Type' in headers and self.htmlpage.search(headers['Content-Type']):
return True
return False
def sitelinks(self, html_page, url):
"""Finds all links in the provided html page"""
bs = BeautifulSoup(html_page)
links = set()
urlpart = urlsplit(url)
try:
for anchor in bs.find_all('a'):
linkpart = list(urlsplit(anchor['href']))
linkpart[4] = '' #remove the fragment
if linkpart[0] == '':
linkpart[0] = urlpart.scheme
if linkpart[1] == '':
linkpart[1] = urlpart.netloc
if linkpart[0] == urlpart.scheme and linkpart[1] == urlpart.netloc:
if linkpart[2].startswith('/'):
links.add(urlunsplit(linkpart))
elif linkpart[2] != '':
#relative URL.
links.add(urljoin(url, linkpart[2]))
except KeyError:
pass
return links
def process_page(self, url, code, headers, body):
'''Does things with the retrieved page'''
pass
|
nada-labs/sitemap-generator
|
spider.py
|
Python
|
mit
| 5,128
|
from decimal import Decimal
import pytest
from django.conf import settings
from django.contrib.auth import get_user_model
from django.urls import reverse
from rest_framework import status
from rest_framework.test import APIClient
from openslides.core.config import config
from openslides.motions.models import Motion, MotionOption, MotionPoll, MotionVote
from openslides.poll.models import BasePoll
from openslides.utils.auth import get_group_model
from openslides.utils.autoupdate import inform_changed_data
from tests.common_groups import GROUP_ADMIN_PK, GROUP_DEFAULT_PK, GROUP_DELEGATE_PK
from tests.count_queries import count_queries
from tests.test_case import TestCase
@pytest.mark.django_db(transaction=False)
def test_motion_poll_db_queries():
"""
Tests that only the following db queries are done:
* 1 request to get the polls,
* 1 request to get all options for all polls,
* 1 request to get all votes for all options,
* 1 request to get all users for all votes,
* 1 request to get all poll groups,
= 5 queries
"""
create_motion_polls()
assert count_queries(MotionPoll.get_elements)() == 5
@pytest.mark.django_db(transaction=False)
def test_motion_vote_db_queries():
"""
Tests that only 1 query is done when fetching MotionVotes
"""
create_motion_polls()
assert count_queries(MotionVote.get_elements)() == 1
@pytest.mark.django_db(transaction=False)
def test_motion_option_db_queries():
"""
Tests that only the following db queries are done:
* 1 request to get the options,
* 1 request to get all votes for all options,
= 2 queries
"""
create_motion_polls()
assert count_queries(MotionOption.get_elements)() == 2
def create_motion_polls():
"""
Creates 1 Motion with 5 polls with 5 options each which have 2 votes each
"""
motion = Motion.objects.create(title="test_motion_wfLrsjEHXBmPplbvQ65N")
group1 = get_group_model().objects.get(pk=1)
group2 = get_group_model().objects.get(pk=2)
for index in range(5):
poll = MotionPoll.objects.create(
motion=motion, title=f"test_title_{index}", pollmethod="YN", type="named"
)
poll.groups.add(group1)
poll.groups.add(group2)
for j in range(5):
option = MotionOption.objects.create(poll=poll)
for k in range(2):
user = get_user_model().objects.create_user(
username=f"test_username_{index}{j}{k}",
password="test_password_kbzj5L8ZtVxBllZzoW6D",
)
MotionVote.objects.create(
user=user,
option=option,
value=("Y" if k == 0 else "N"),
weight=Decimal(1),
)
poll.voted.add(user)
class CreateMotionPoll(TestCase):
"""
Tests creating polls of motions.
"""
def advancedSetUp(self):
self.motion = Motion(
title="test_title_Aiqueigh2dae9phabiqu",
text="test_text_Neekoh3zou6li5rue8iL",
)
self.motion.save()
def test_simple(self):
response = self.client.post(
reverse("motionpoll-list"),
{
"title": "test_title_ailai4toogh3eefaa2Vo",
"pollmethod": "YNA",
"type": "named",
"motion_id": self.motion.id,
"onehundred_percent_base": "YN",
"majority_method": "simple",
},
)
self.assertHttpStatusVerbose(response, status.HTTP_201_CREATED)
self.assertTrue(MotionPoll.objects.exists())
poll = MotionPoll.objects.get()
self.assertEqual(poll.title, "test_title_ailai4toogh3eefaa2Vo")
self.assertEqual(poll.pollmethod, "YNA")
self.assertEqual(poll.type, "named")
self.assertEqual(poll.motion.id, self.motion.id)
self.assertTrue(poll.options.exists())
def test_default_method(self):
response = self.client.post(
reverse("motionpoll-list"),
{
"title": "test_title_ailai4toogh3eefaa2Vo",
"type": "named",
"motion_id": self.motion.id,
"onehundred_percent_base": "YN",
"majority_method": "simple",
},
)
self.assertHttpStatusVerbose(response, status.HTTP_201_CREATED)
self.assertTrue(MotionPoll.objects.exists())
poll = MotionPoll.objects.get()
self.assertEqual(poll.pollmethod, "YNA")
def test_autoupdate(self):
response = self.client.post(
reverse("motionpoll-list"),
{
"title": "test_title_9Ce8OsdB8YWTVm5YOzqH",
"pollmethod": "YNA",
"type": "named",
"motion_id": self.motion.id,
"onehundred_percent_base": "YN",
"majority_method": "simple",
},
)
self.assertHttpStatusVerbose(response, status.HTTP_201_CREATED)
autoupdate = self.get_last_autoupdate(user=self.admin)
self.assertEqual(
autoupdate[0]["motions/motion-poll:1"],
{
"motion_id": 1,
"pollmethod": MotionPoll.POLLMETHOD_YNA,
"state": MotionPoll.STATE_CREATED,
"type": MotionPoll.TYPE_NAMED,
"title": "test_title_9Ce8OsdB8YWTVm5YOzqH",
"onehundred_percent_base": MotionPoll.PERCENT_BASE_YN,
"majority_method": MotionPoll.MAJORITY_SIMPLE,
"groups_id": [],
"votesvalid": "0.000000",
"votesinvalid": "0.000000",
"votescast": "0.000000",
"options_id": [1],
"id": 1,
"voted_id": [],
"user_has_voted": False,
"user_has_voted_for_delegations": [],
},
)
self.assertEqual(autoupdate[1], [])
def test_missing_keys(self):
complete_request_data = {
"title": "test_title_OoCh9aitaeyaeth8nom1",
"type": "named",
"motion_id": self.motion.id,
"onehundred_percent_base": "YN",
"majority_method": "simple",
}
for key in complete_request_data.keys():
request_data = {
_key: value
for _key, value in complete_request_data.items()
if _key != key
}
response = self.client.post(reverse("motionpoll-list"), request_data)
self.assertHttpStatusVerbose(response, status.HTTP_400_BAD_REQUEST)
self.assertFalse(MotionPoll.objects.exists())
def test_with_groups(self):
group1 = get_group_model().objects.get(pk=1)
group2 = get_group_model().objects.get(pk=2)
response = self.client.post(
reverse("motionpoll-list"),
{
"title": "test_title_Thoo2eiphohhi1eeXoow",
"pollmethod": "YNA",
"type": "named",
"motion_id": self.motion.id,
"onehundred_percent_base": "YN",
"majority_method": "simple",
"groups_id": [1, 2],
},
)
self.assertHttpStatusVerbose(response, status.HTTP_201_CREATED)
poll = MotionPoll.objects.get()
self.assertTrue(group1 in poll.groups.all())
self.assertTrue(group2 in poll.groups.all())
def test_with_empty_groups(self):
response = self.client.post(
reverse("motionpoll-list"),
{
"title": "test_title_Thoo2eiphohhi1eeXoow",
"pollmethod": MotionPoll.POLLMETHOD_YNA,
"type": MotionPoll.TYPE_NAMED,
"motion_id": self.motion.id,
"onehundred_percent_base": MotionPoll.PERCENT_BASE_YN,
"majority_method": MotionPoll.MAJORITY_SIMPLE,
"groups_id": [],
},
)
self.assertHttpStatusVerbose(response, status.HTTP_201_CREATED)
poll = MotionPoll.objects.get()
self.assertFalse(poll.groups.exists())
def test_not_supported_type(self):
response = self.client.post(
reverse("motionpoll-list"),
{
"title": "test_title_yaiyeighoh0Iraet3Ahc",
"pollmethod": MotionPoll.POLLMETHOD_YNA,
"type": "not_existing",
"motion_id": self.motion.id,
"onehundred_percent_base": MotionPoll.PERCENT_BASE_YN,
"majority_method": MotionPoll.MAJORITY_SIMPLE,
},
)
self.assertHttpStatusVerbose(response, status.HTTP_400_BAD_REQUEST)
self.assertFalse(MotionPoll.objects.exists())
def test_not_allowed_type(self):
setattr(settings, "ENABLE_ELECTRONIC_VOTING", False)
response = self.client.post(
reverse("motionpoll-list"),
{
"title": "test_title_3jdWIXbKBa7ZXutf3RYf",
"pollmethod": MotionPoll.POLLMETHOD_YN,
"type": MotionPoll.TYPE_NAMED,
"motion_id": self.motion.id,
"onehundred_percent_base": MotionPoll.PERCENT_BASE_YN,
"majority_method": MotionPoll.MAJORITY_SIMPLE,
},
)
self.assertHttpStatusVerbose(response, status.HTTP_400_BAD_REQUEST)
self.assertFalse(MotionPoll.objects.exists())
setattr(settings, "ENABLE_ELECTRONIC_VOTING", True)
def test_not_supported_pollmethod(self):
response = self.client.post(
reverse("motionpoll-list"),
{
"title": "test_title_SeVaiteYeiNgie5Xoov8",
"pollmethod": "not_existing",
"type": "named",
"motion_id": self.motion.id,
"onehundred_percent_base": MotionPoll.PERCENT_BASE_YN,
"majority_method": MotionPoll.MAJORITY_SIMPLE,
},
)
self.assertHttpStatusVerbose(response, status.HTTP_400_BAD_REQUEST)
self.assertFalse(MotionPoll.objects.exists())
def test_create_with_votes(self):
response = self.client.post(
reverse("motionpoll-list"),
{
"title": "test_title_0X5LifVkKiSh8OPGQM8e",
"pollmethod": MotionPoll.POLLMETHOD_YN,
"type": MotionPoll.TYPE_ANALOG,
"motion_id": self.motion.id,
"onehundred_percent_base": MotionPoll.PERCENT_BASE_YNA,
"majority_method": MotionPoll.MAJORITY_SIMPLE,
"votes": {
"Y": 1,
"N": 2,
"votesvalid": "-2",
"votesinvalid": "-2",
"votescast": "-2",
},
},
)
self.assertHttpStatusVerbose(response, status.HTTP_201_CREATED)
poll = MotionPoll.objects.get()
self.assertEqual(poll.state, MotionPoll.STATE_FINISHED)
self.assertTrue(MotionVote.objects.exists())
def test_create_with_votes_publish_immediately(self):
response = self.client.post(
reverse("motionpoll-list"),
{
"title": "test_title_iXhJX0jmNl3Nvadsi8JO",
"pollmethod": MotionPoll.POLLMETHOD_YN,
"type": MotionPoll.TYPE_ANALOG,
"motion_id": self.motion.id,
"onehundred_percent_base": MotionPoll.PERCENT_BASE_YNA,
"majority_method": MotionPoll.MAJORITY_SIMPLE,
"votes": {
"Y": 1,
"N": 2,
"votesvalid": "-2",
"votesinvalid": "-2",
"votescast": "-2",
},
"publish_immediately": "1",
},
)
self.assertHttpStatusVerbose(response, status.HTTP_201_CREATED)
poll = MotionPoll.objects.get()
self.assertEqual(poll.state, MotionPoll.STATE_PUBLISHED)
self.assertTrue(MotionVote.objects.exists())
def test_create_with_invalid_votes(self):
response = self.client.post(
reverse("motionpoll-list"),
{
"title": "test_title_phSl1IALPIoDyM9uI2Kq",
"pollmethod": MotionPoll.POLLMETHOD_YN,
"type": MotionPoll.TYPE_ANALOG,
"motion_id": self.motion.id,
"onehundred_percent_base": MotionPoll.PERCENT_BASE_YNA,
"majority_method": MotionPoll.MAJORITY_SIMPLE,
"votes": {"Y": 1, "N": 2, "votesvalid": "-2", "votesinvalid": "-2"},
"publish_immediately": "1",
},
)
self.assertHttpStatusVerbose(response, status.HTTP_400_BAD_REQUEST)
self.assertFalse(MotionPoll.objects.exists())
self.assertFalse(MotionVote.objects.exists())
def test_create_with_votes_wrong_type(self):
response = self.client.post(
reverse("motionpoll-list"),
{
"title": "test_title_PgvqRIvuKuVImEpQJAMZ",
"pollmethod": MotionPoll.POLLMETHOD_YN,
"type": MotionPoll.TYPE_NAMED,
"motion_id": self.motion.id,
"onehundred_percent_base": MotionPoll.PERCENT_BASE_YNA,
"majority_method": MotionPoll.MAJORITY_SIMPLE,
"votes": {"Y": 1, "N": 2, "votesvalid": "-2", "votesinvalid": "-2"},
"publish_immediately": "1",
},
)
self.assertHttpStatusVerbose(response, status.HTTP_400_BAD_REQUEST)
self.assertFalse(MotionPoll.objects.exists())
self.assertFalse(MotionVote.objects.exists())
class UpdateMotionPoll(TestCase):
"""
Tests updating polls of motions.
"""
def setUp(self):
self.client = APIClient()
self.client.login(username="admin", password="admin")
self.motion = Motion(
title="test_title_Aiqueigh2dae9phabiqu",
text="test_text_Neekoh3zou6li5rue8iL",
)
self.motion.save()
self.group = get_group_model().objects.get(pk=1)
self.poll = MotionPoll.objects.create(
motion=self.motion,
title="test_title_beeFaihuNae1vej2ai8m",
pollmethod="YNA",
type="named",
onehundred_percent_base="YN",
majority_method="simple",
)
self.poll.create_options()
self.poll.groups.add(self.group)
def test_patch_title(self):
response = self.client.patch(
reverse("motionpoll-detail", args=[self.poll.pk]),
{"title": "test_title_Aishohh1ohd0aiSut7gi"},
)
self.assertHttpStatusVerbose(response, status.HTTP_200_OK)
poll = MotionPoll.objects.get()
self.assertEqual(poll.title, "test_title_Aishohh1ohd0aiSut7gi")
def test_prevent_patching_motion(self):
motion = Motion(
title="test_title_phohdah8quukooHeetuz",
text="test_text_ue2yeisaech1ahBohhoo",
)
motion.save()
response = self.client.patch(
reverse("motionpoll-detail", args=[self.poll.pk]), {"motion_id": motion.id}
)
self.assertHttpStatusVerbose(response, status.HTTP_200_OK)
poll = MotionPoll.objects.get()
self.assertEqual(poll.motion.id, self.motion.id) # unchanged
def test_patch_pollmethod(self):
response = self.client.patch(
reverse("motionpoll-detail", args=[self.poll.pk]), {"pollmethod": "YN"}
)
self.assertHttpStatusVerbose(response, status.HTTP_200_OK)
poll = MotionPoll.objects.get()
self.assertEqual(poll.pollmethod, "YN")
self.assertEqual(poll.onehundred_percent_base, "YN")
def test_patch_invalid_pollmethod(self):
response = self.client.patch(
reverse("motionpoll-detail", args=[self.poll.pk]), {"pollmethod": "invalid"}
)
self.assertHttpStatusVerbose(response, status.HTTP_400_BAD_REQUEST)
poll = MotionPoll.objects.get()
self.assertEqual(poll.pollmethod, "YNA")
def test_patch_type(self):
response = self.client.patch(
reverse("motionpoll-detail", args=[self.poll.pk]), {"type": "analog"}
)
self.assertHttpStatusVerbose(response, status.HTTP_200_OK)
poll = MotionPoll.objects.get()
self.assertEqual(poll.type, "analog")
def test_patch_invalid_type(self):
response = self.client.patch(
reverse("motionpoll-detail", args=[self.poll.pk]), {"type": "invalid"}
)
self.assertHttpStatusVerbose(response, status.HTTP_400_BAD_REQUEST)
poll = MotionPoll.objects.get()
self.assertEqual(poll.type, "named")
def test_patch_not_allowed_type(self):
setattr(settings, "ENABLE_ELECTRONIC_VOTING", False)
response = self.client.patch(
reverse("motionpoll-detail", args=[self.poll.pk]),
{"type": BasePoll.TYPE_NAMED},
)
self.assertHttpStatusVerbose(response, status.HTTP_400_BAD_REQUEST)
poll = MotionPoll.objects.get()
self.assertEqual(poll.type, BasePoll.TYPE_NAMED)
setattr(settings, "ENABLE_ELECTRONIC_VOTING", True)
def test_patch_100_percent_base(self):
response = self.client.patch(
reverse("motionpoll-detail", args=[self.poll.pk]),
{"onehundred_percent_base": "cast"},
)
self.assertHttpStatusVerbose(response, status.HTTP_200_OK)
poll = MotionPoll.objects.get()
self.assertEqual(poll.onehundred_percent_base, "cast")
def test_patch_wrong_100_percent_base(self):
response = self.client.patch(
reverse("motionpoll-detail", args=[self.poll.pk]),
{"onehundred_percent_base": "invalid"},
)
self.assertHttpStatusVerbose(response, status.HTTP_400_BAD_REQUEST)
poll = MotionPoll.objects.get()
self.assertEqual(poll.onehundred_percent_base, "YN")
def test_patch_majority_method(self):
response = self.client.patch(
reverse("motionpoll-detail", args=[self.poll.pk]),
{"majority_method": "two_thirds"},
)
self.assertHttpStatusVerbose(response, status.HTTP_200_OK)
poll = MotionPoll.objects.get()
self.assertEqual(poll.majority_method, "two_thirds")
def test_patch_wrong_majority_method(self):
response = self.client.patch(
reverse("motionpoll-detail", args=[self.poll.pk]),
{"majority_method": "invalid majority method"},
)
self.assertHttpStatusVerbose(response, status.HTTP_400_BAD_REQUEST)
poll = MotionPoll.objects.get()
self.assertEqual(poll.majority_method, "simple")
def test_patch_groups_to_empty(self):
response = self.client.patch(
reverse("motionpoll-detail", args=[self.poll.pk]), {"groups_id": []}
)
self.assertHttpStatusVerbose(response, status.HTTP_200_OK)
poll = MotionPoll.objects.get()
self.assertFalse(poll.groups.exists())
def test_patch_groups(self):
group2 = get_group_model().objects.get(pk=2)
response = self.client.patch(
reverse("motionpoll-detail", args=[self.poll.pk]),
{"groups_id": [group2.id]},
)
self.assertHttpStatusVerbose(response, status.HTTP_200_OK)
poll = MotionPoll.objects.get()
self.assertEqual(poll.groups.count(), 1)
self.assertEqual(poll.groups.get(), group2)
def test_patch_title_started(self):
self.poll.state = 2
self.poll.save()
response = self.client.patch(
reverse("motionpoll-detail", args=[self.poll.pk]),
{"title": "test_title_1FjLGeQqsi9GgNzPp73S"},
)
self.assertHttpStatusVerbose(response, status.HTTP_200_OK)
poll = MotionPoll.objects.get()
self.assertEqual(poll.title, "test_title_1FjLGeQqsi9GgNzPp73S")
def test_patch_wrong_state(self):
self.poll.state = 2
self.poll.save()
response = self.client.patch(
reverse("motionpoll-detail", args=[self.poll.pk]),
{"type": BasePoll.TYPE_NAMED},
)
self.assertHttpStatusVerbose(response, status.HTTP_400_BAD_REQUEST)
poll = MotionPoll.objects.get()
self.assertEqual(poll.type, BasePoll.TYPE_NAMED)
def test_patch_majority_method_state_not_created(self):
self.poll.state = 2
self.poll.save()
response = self.client.patch(
reverse("motionpoll-detail", args=[self.poll.pk]),
{"majority_method": "two_thirds"},
)
self.assertHttpStatusVerbose(response, status.HTTP_200_OK)
poll = MotionPoll.objects.get()
self.assertEqual(poll.majority_method, "two_thirds")
def test_patch_100_percent_base_state_not_created(self):
self.poll.state = 2
self.poll.save()
response = self.client.patch(
reverse("motionpoll-detail", args=[self.poll.pk]),
{"onehundred_percent_base": "cast"},
)
self.assertHttpStatusVerbose(response, status.HTTP_200_OK)
poll = MotionPoll.objects.get()
self.assertEqual(poll.onehundred_percent_base, "cast")
def test_patch_wrong_100_percent_base_state_not_created(self):
self.poll.state = 2
self.poll.pollmethod = MotionPoll.POLLMETHOD_YN
self.poll.save()
response = self.client.patch(
reverse("motionpoll-detail", args=[self.poll.pk]),
{"onehundred_percent_base": MotionPoll.PERCENT_BASE_YNA},
)
self.assertHttpStatusVerbose(response, status.HTTP_200_OK)
poll = MotionPoll.objects.get()
self.assertEqual(poll.onehundred_percent_base, "YN")
class VoteMotionPollAnalog(TestCase):
def setUp(self):
self.client = APIClient()
self.client.login(username="admin", password="admin")
self.motion = Motion(
title="test_title_OoK9IeChe2Jeib9Deeji",
text="test_text_eichui1oobiSeit9aifo",
)
self.motion.save()
self.poll = MotionPoll.objects.create(
motion=self.motion,
title="test_title_tho8PhiePh8upaex6phi",
pollmethod="YNA",
type=BasePoll.TYPE_ANALOG,
)
self.poll.create_options()
def start_poll(self):
self.poll.state = MotionPoll.STATE_STARTED
self.poll.save()
def make_admin_delegate(self):
admin = get_user_model().objects.get(username="admin")
admin.groups.add(GROUP_DELEGATE_PK)
admin.groups.remove(GROUP_ADMIN_PK)
inform_changed_data(admin)
def test_start_poll(self):
response = self.client.post(reverse("motionpoll-start", args=[self.poll.pk]))
self.assertHttpStatusVerbose(response, status.HTTP_200_OK)
poll = MotionPoll.objects.get()
self.assertEqual(poll.state, MotionPoll.STATE_STARTED)
self.assertEqual(poll.votesvalid, None)
self.assertEqual(poll.votesinvalid, None)
self.assertEqual(poll.votescast, None)
self.assertFalse(poll.get_votes().exists())
def test_stop_poll(self):
self.start_poll()
response = self.client.post(reverse("motionpoll-stop", args=[self.poll.pk]))
self.assertHttpStatusVerbose(response, status.HTTP_400_BAD_REQUEST)
self.assertEqual(self.poll.state, MotionPoll.STATE_STARTED)
def test_vote(self):
self.start_poll()
response = self.client.post(
reverse("motionpoll-vote", args=[self.poll.pk]),
{
"data": {
"Y": "1",
"N": "2.35",
"A": "-1",
"votesvalid": "4.64",
"votesinvalid": "-2",
"votescast": "-2",
},
},
)
self.assertHttpStatusVerbose(response, status.HTTP_200_OK)
poll = MotionPoll.objects.get()
self.assertEqual(poll.votesvalid, Decimal("4.64"))
self.assertEqual(poll.votesinvalid, Decimal("-2"))
self.assertEqual(poll.votescast, Decimal("-2"))
self.assertEqual(poll.get_votes().count(), 3)
self.assertEqual(poll.state, MotionPoll.STATE_FINISHED)
option = poll.options.get()
self.assertEqual(option.yes, Decimal("1"))
self.assertEqual(option.no, Decimal("2.35"))
self.assertEqual(option.abstain, Decimal("-1"))
self.assertAutoupdate(poll)
def test_vote_no_permissions(self):
self.start_poll()
self.make_admin_delegate()
response = self.client.post(
reverse("motionpoll-vote", args=[self.poll.pk]), {"data": {}}
)
self.assertHttpStatusVerbose(response, status.HTTP_403_FORBIDDEN)
self.assertFalse(MotionPoll.objects.get().get_votes().exists())
def test_vote_no_data(self):
self.start_poll()
response = self.client.post(reverse("motionpoll-vote", args=[self.poll.pk]), {})
self.assertHttpStatusVerbose(response, status.HTTP_400_BAD_REQUEST)
self.assertFalse(MotionPoll.objects.get().get_votes().exists())
def test_vote_missing_data(self):
self.start_poll()
response = self.client.post(
reverse("motionpoll-vote", args=[self.poll.pk]),
{"data": {"Y": "4", "N": "22.6"}},
)
self.assertHttpStatusVerbose(response, status.HTTP_400_BAD_REQUEST)
self.assertFalse(MotionPoll.objects.get().get_votes().exists())
def test_vote_wrong_data_format(self):
self.start_poll()
response = self.client.post(
reverse("motionpoll-vote", args=[self.poll.pk]), {"data": [1, 2, 5]}
)
self.assertHttpStatusVerbose(response, status.HTTP_400_BAD_REQUEST)
self.assertFalse(MotionPoll.objects.get().get_votes().exists())
def test_vote_wrong_vote_data(self):
self.start_poll()
response = self.client.post(
reverse("motionpoll-vote", args=[self.poll.pk]),
{"data": {"Y": "some string", "N": "-2", "A": "3"}},
)
self.assertHttpStatusVerbose(response, status.HTTP_400_BAD_REQUEST)
self.assertFalse(MotionPoll.objects.get().get_votes().exists())
def test_vote_state_finished(self):
self.start_poll()
self.client.post(
reverse("motionpoll-vote", args=[self.poll.pk]),
{
"data": {
"Y": "3",
"N": "1",
"A": "5",
"votesvalid": "-2",
"votesinvalid": "1",
"votescast": "-1",
},
},
)
self.poll.state = 3
self.poll.save()
response = self.client.post(
reverse("motionpoll-vote", args=[self.poll.pk]),
{
"data": {
"Y": "1",
"N": "2.35",
"A": "-1",
"votesvalid": "4.64",
"votesinvalid": "-2",
"votescast": "3",
},
},
)
self.assertHttpStatusVerbose(response, status.HTTP_200_OK)
poll = MotionPoll.objects.get()
self.assertEqual(poll.votesvalid, Decimal("4.64"))
self.assertEqual(poll.votesinvalid, Decimal("-2"))
self.assertEqual(poll.votescast, Decimal("3"))
self.assertEqual(poll.get_votes().count(), 3)
option = poll.options.get()
self.assertEqual(option.yes, Decimal("1"))
self.assertEqual(option.no, Decimal("2.35"))
self.assertEqual(option.abstain, Decimal("-1"))
class VoteMotionPollNamed(TestCase):
def setUp(self):
self.client = APIClient()
self.client.login(username="admin", password="admin")
self.motion = Motion(
title="test_title_OoK9IeChe2Jeib9Deeji",
text="test_text_eichui1oobiSeit9aifo",
)
self.motion.save()
self.group = get_group_model().objects.get(pk=GROUP_DELEGATE_PK)
self.admin = get_user_model().objects.get(username="admin")
self.poll = MotionPoll.objects.create(
motion=self.motion,
title="test_title_tho8PhiePh8upaex6phi",
pollmethod="YNA",
type=BasePoll.TYPE_NAMED,
)
self.poll.create_options()
self.poll.groups.add(self.group)
def start_poll(self):
self.poll.state = MotionPoll.STATE_STARTED
self.poll.save()
def make_admin_delegate(self):
self.admin.groups.add(GROUP_DELEGATE_PK)
self.admin.groups.remove(GROUP_ADMIN_PK)
inform_changed_data(self.admin)
def make_admin_present(self):
self.admin.is_present = True
self.admin.save()
def test_start_poll(self):
response = self.client.post(reverse("motionpoll-start", args=[self.poll.pk]))
self.assertHttpStatusVerbose(response, status.HTTP_200_OK)
poll = MotionPoll.objects.get()
self.assertEqual(poll.state, MotionPoll.STATE_STARTED)
self.assertEqual(poll.votesvalid, Decimal("0"))
self.assertEqual(poll.votesinvalid, Decimal("0"))
self.assertEqual(poll.votescast, Decimal("0"))
self.assertFalse(poll.get_votes().exists())
def test_vote(self):
self.start_poll()
self.make_admin_delegate()
self.make_admin_present()
response = self.client.post(
reverse("motionpoll-vote", args=[self.poll.pk]), {"data": "N"}
)
self.assertHttpStatusVerbose(response, status.HTTP_200_OK)
poll = MotionPoll.objects.get()
self.assertEqual(poll.votesvalid, Decimal("1"))
self.assertEqual(poll.votesinvalid, Decimal("0"))
self.assertEqual(poll.votescast, Decimal("1"))
self.assertEqual(poll.get_votes().count(), 1)
option = poll.options.get()
self.assertEqual(option.yes, Decimal("0"))
self.assertEqual(option.no, Decimal("1"))
self.assertEqual(option.abstain, Decimal("0"))
vote = option.votes.get()
self.assertEqual(vote.user, self.admin)
self.assertEqual(vote.weight, Decimal("1"))
def test_vote_with_voteweight(self):
config["users_activate_vote_weight"] = True
self.start_poll()
self.make_admin_delegate()
self.make_admin_present()
self.admin.vote_weight = weight = Decimal("3.5")
self.admin.save()
response = self.client.post(
reverse("motionpoll-vote", args=[self.poll.pk]), {"data": "A"}
)
self.assertHttpStatusVerbose(response, status.HTTP_200_OK)
poll = MotionPoll.objects.get()
self.assertEqual(poll.votesvalid, weight)
self.assertEqual(poll.votesinvalid, Decimal("0"))
self.assertEqual(poll.votescast, Decimal("1"))
self.assertEqual(poll.get_votes().count(), 1)
self.assertEqual(poll.amount_users_voted_with_individual_weight(), weight)
option = poll.options.get()
self.assertEqual(option.yes, Decimal("0"))
self.assertEqual(option.no, Decimal("0"))
self.assertEqual(option.abstain, weight)
vote = option.votes.get()
self.assertEqual(vote.weight, weight)
def test_vote_without_voteweight(self):
self.admin.vote_weight = Decimal("3.5")
self.admin.save()
self.test_vote()
def test_change_vote(self):
self.start_poll()
self.make_admin_delegate()
self.make_admin_present()
response = self.client.post(
reverse("motionpoll-vote", args=[self.poll.pk]), {"data": "N"}
)
self.assertHttpStatusVerbose(response, status.HTTP_200_OK)
response = self.client.post(
reverse("motionpoll-vote", args=[self.poll.pk]), {"data": "A"}
)
self.assertHttpStatusVerbose(response, status.HTTP_400_BAD_REQUEST)
poll = MotionPoll.objects.get()
self.assertEqual(poll.votesvalid, Decimal("1"))
self.assertEqual(poll.votesinvalid, Decimal("0"))
self.assertEqual(poll.votescast, Decimal("1"))
self.assertEqual(poll.get_votes().count(), 1)
option = poll.options.get()
self.assertEqual(option.yes, Decimal("0"))
self.assertEqual(option.no, Decimal("1"))
self.assertEqual(option.abstain, Decimal("0"))
vote = option.votes.get()
self.assertEqual(vote.user, self.admin)
def test_vote_anonymous(self):
self.poll.groups.add(GROUP_DEFAULT_PK)
self.start_poll()
config["general_system_enable_anonymous"] = True
guest_client = APIClient()
response = guest_client.post(
reverse("motionpoll-vote", args=[self.poll.pk]), {"data": "Y"}
)
self.assertHttpStatusVerbose(response, status.HTTP_403_FORBIDDEN)
self.assertFalse(MotionPoll.objects.get().get_votes().exists())
def test_vote_wrong_state(self):
self.make_admin_present()
self.make_admin_delegate()
response = self.client.post(
reverse("motionpoll-vote", args=[self.poll.pk]), {"data": {}}
)
self.assertHttpStatusVerbose(response, status.HTTP_400_BAD_REQUEST)
self.assertFalse(MotionPoll.objects.get().get_votes().exists())
def test_vote_wrong_group(self):
self.start_poll()
self.make_admin_present()
response = self.client.post(
reverse("motionpoll-vote", args=[self.poll.pk]), {"data": {}}
)
self.assertHttpStatusVerbose(response, status.HTTP_403_FORBIDDEN)
self.assertFalse(MotionPoll.objects.get().get_votes().exists())
def test_vote_not_present(self):
self.start_poll()
self.make_admin_delegate()
response = self.client.post(
reverse("motionpoll-vote", args=[self.poll.pk]), {"data": {}}
)
self.assertHttpStatusVerbose(response, status.HTTP_403_FORBIDDEN)
self.assertFalse(MotionPoll.objects.get().get_votes().exists())
def test_vote_missing_data(self):
self.start_poll()
self.make_admin_delegate()
self.make_admin_present()
response = self.client.post(
reverse("motionpoll-vote", args=[self.poll.pk]), {"data": {}}
)
self.assertHttpStatusVerbose(response, status.HTTP_400_BAD_REQUEST)
self.assertFalse(MotionPoll.objects.get().get_votes().exists())
def test_vote_wrong_data_format(self):
self.start_poll()
self.make_admin_delegate()
self.make_admin_present()
response = self.client.post(
reverse("motionpoll-vote", args=[self.poll.pk]), {"data": [1, 2, 5]}
)
self.assertHttpStatusVerbose(response, status.HTTP_400_BAD_REQUEST)
self.assertFalse(MotionPoll.objects.get().get_votes().exists())
def setup_vote_delegation(self, with_delegation=True):
""" user -> admin """
self.start_poll()
self.make_admin_delegate()
self.make_admin_present()
self.user, self.user_password = self.create_user()
self.user.groups.add(GROUP_DELEGATE_PK)
if with_delegation:
self.user.vote_delegated_to = self.admin
self.user.save()
inform_changed_data(self.admin) # put the admin into the cache to update
# its vote_delegated_to_id field
def test_vote_delegation(self):
self.setup_vote_delegation()
response = self.client.post(
reverse("motionpoll-vote", args=[self.poll.pk]),
{"data": "N", "user_id": self.user.pk}, # user not present
)
self.assertHttpStatusVerbose(response, status.HTTP_200_OK)
poll = MotionPoll.objects.get()
self.assertEqual(poll.votesvalid, Decimal("1"))
self.assertEqual(poll.votesinvalid, Decimal("0"))
self.assertEqual(poll.votescast, Decimal("1"))
self.assertEqual(poll.get_votes().count(), 1)
option = poll.options.get()
self.assertEqual(option.yes, Decimal("0"))
self.assertEqual(option.no, Decimal("1"))
self.assertEqual(option.abstain, Decimal("0"))
vote = option.votes.get()
self.assertEqual(vote.user, self.user)
self.assertEqual(vote.delegated_user, self.admin)
autoupdate = self.get_last_autoupdate(user=self.admin)
self.assertIn("motions/motion-poll:1", autoupdate[0])
self.assertEqual(
autoupdate[0]["motions/motion-poll:1"]["user_has_voted_for_delegations"],
[self.user.pk],
)
def test_vote_delegation_and_self_vote(self):
self.test_vote_delegation()
response = self.client.post(
reverse("motionpoll-vote", args=[self.poll.pk]), {"data": "Y"}
)
self.assertHttpStatusVerbose(response, status.HTTP_200_OK)
poll = MotionPoll.objects.get()
self.assertEqual(poll.votesvalid, Decimal("2"))
self.assertEqual(poll.votesinvalid, Decimal("0"))
self.assertEqual(poll.votescast, Decimal("2"))
self.assertEqual(poll.get_votes().count(), 2)
option = poll.options.get()
self.assertEqual(option.yes, Decimal("1"))
self.assertEqual(option.no, Decimal("1"))
self.assertEqual(option.abstain, Decimal("0"))
vote = option.votes.get(user_id=self.admin.pk)
self.assertEqual(vote.user, self.admin)
self.assertEqual(vote.delegated_user, self.admin)
def test_vote_delegation_forbidden(self):
self.setup_vote_delegation(False)
response = self.client.post(
reverse("motionpoll-vote", args=[self.poll.pk]),
{"data": "N", "user_id": self.user.pk},
)
self.assertHttpStatusVerbose(response, status.HTTP_400_BAD_REQUEST)
self.assertFalse(MotionPoll.objects.get().get_votes().exists())
def test_vote_delegation_not_present(self):
self.setup_vote_delegation()
self.admin.is_present = False
self.admin.save()
response = self.client.post(
reverse("motionpoll-vote", args=[self.poll.pk]),
{"data": "N", "user_id": self.user.pk},
)
self.assertHttpStatusVerbose(response, status.HTTP_403_FORBIDDEN)
self.assertFalse(MotionPoll.objects.get().get_votes().exists())
def test_vote_delegation_delegate_not_in_group(self):
self.setup_vote_delegation()
self.admin.groups.remove(GROUP_DELEGATE_PK)
self.admin.save()
response = self.client.post(
reverse("motionpoll-vote", args=[self.poll.pk]),
{"data": "N", "user_id": self.user.pk},
)
self.assertHttpStatusVerbose(response, status.HTTP_200_OK)
poll = MotionPoll.objects.get()
self.assertEqual(poll.get_votes().count(), 1)
vote = poll.get_votes()[0]
self.assertEqual(vote.value, "N")
self.assertEqual(vote.user, self.user)
self.assertEqual(vote.delegated_user, self.admin)
def test_vote_delegation_delegator_not_in_group(self):
self.setup_vote_delegation()
self.user.groups.remove(GROUP_DELEGATE_PK)
self.user.save()
response = self.client.post(
reverse("motionpoll-vote", args=[self.poll.pk]),
{"data": "N", "user_id": self.user.pk},
)
self.assertHttpStatusVerbose(response, status.HTTP_403_FORBIDDEN)
self.assertFalse(MotionPoll.objects.get().get_votes().exists())
def test_vote_delegation_delegator_self_vote_not_allowed(self):
self.setup_vote_delegation()
# Make the user a delegate and present
self.admin.groups.add(GROUP_DELEGATE_PK)
self.admin.groups.remove(GROUP_ADMIN_PK)
self.user.is_present = True
self.user.save()
# Use the user to make the request to vote for himself
user_client = APIClient()
user_client.login(username=self.user.username, password=self.user_password)
response = user_client.post(
reverse("motionpoll-vote", args=[self.poll.pk]),
{"data": "N"},
)
self.assertHttpStatusVerbose(response, status.HTTP_400_BAD_REQUEST)
self.assertFalse(MotionPoll.objects.get().get_votes().exists())
class VoteMotionPollNamedAutoupdates(TestCase):
"""3 important users:
self.admin: manager, has can_see, can_manage, can_manage_polls (in admin group)
self.user1: votes, has can_see perms and in in delegate group
self.other_user: Just has can_see perms and is NOT in the delegate group.
"""
def advancedSetUp(self):
self.motion = Motion(
title="test_title_OoK9IeChe2Jeib9Deeji",
text="test_text_eichui1oobiSeit9aifo",
)
self.motion.save()
self.delegate_group = get_group_model().objects.get(pk=GROUP_DELEGATE_PK)
self.other_user, _ = self.create_user()
inform_changed_data(self.other_user)
self.user, user_password = self.create_user()
self.user.groups.add(self.delegate_group)
self.user.is_present = True
self.user.save()
self.user_client = APIClient()
self.user_client.login(username=self.user.username, password=user_password)
self.poll = MotionPoll.objects.create(
motion=self.motion,
title="test_title_tho8PhiePh8upaex6phi",
pollmethod="YNA",
type=BasePoll.TYPE_NAMED,
state=MotionPoll.STATE_STARTED,
onehundred_percent_base="YN",
majority_method="simple",
)
self.poll.create_options()
self.poll.groups.add(self.delegate_group)
def test_vote(self):
response = self.user_client.post(
reverse("motionpoll-vote", args=[self.poll.pk]), {"data": "A"}
)
self.assertHttpStatusVerbose(response, status.HTTP_200_OK)
poll = MotionPoll.objects.get()
vote = MotionVote.objects.get()
# Expect the admin to see the full data in the autoupdate
autoupdate = self.get_last_autoupdate(user=self.admin)
self.assertEqual(
autoupdate[0],
{
"motions/motion-poll:1": {
"motion_id": 1,
"pollmethod": "YNA",
"state": 2,
"type": "named",
"title": "test_title_tho8PhiePh8upaex6phi",
"onehundred_percent_base": "YN",
"majority_method": "simple",
"groups_id": [GROUP_DELEGATE_PK],
"votesvalid": "1.000000",
"votesinvalid": "0.000000",
"votescast": "1.000000",
"options_id": [1],
"id": 1,
"user_has_voted": False,
"user_has_voted_for_delegations": [],
"voted_id": [self.user.id],
},
"motions/motion-vote:1": {
"pollstate": 2,
"id": 1,
"weight": "1.000000",
"value": "A",
"user_id": self.user.id,
"delegated_user_id": self.user.id,
"option_id": 1,
},
"motions/motion-option:1": {
"abstain": "1.000000",
"id": 1,
"no": "0.000000",
"poll_id": 1,
"pollstate": 2,
"yes": "0.000000",
},
},
)
self.assertEqual(autoupdate[1], [])
# Expect user1 to receive his vote
autoupdate = self.get_last_autoupdate(user=self.user)
self.assertEqual(
autoupdate[0]["motions/motion-vote:1"],
{
"pollstate": 2,
"option_id": 1,
"id": 1,
"weight": "1.000000",
"value": "A",
"user_id": self.user.id,
"delegated_user_id": self.user.id,
},
)
self.assertEqual(
autoupdate[0]["motions/motion-option:1"],
{"id": 1, "poll_id": 1, "pollstate": 2},
)
self.assertEqual(autoupdate[1], [])
# Expect non-admins to get a restricted poll update
for user in (self.user, self.other_user):
self.assertAutoupdate(poll, user=user)
autoupdate = self.get_last_autoupdate(user=user)
self.assertEqual(
autoupdate[0]["motions/motion-poll:1"],
{
"motion_id": 1,
"pollmethod": "YNA",
"state": 2,
"type": "named",
"title": "test_title_tho8PhiePh8upaex6phi",
"onehundred_percent_base": "YN",
"majority_method": "simple",
"groups_id": [GROUP_DELEGATE_PK],
"options_id": [1],
"id": 1,
"user_has_voted": user == self.user,
"user_has_voted_for_delegations": [],
},
)
self.assertEqual(
autoupdate[0]["motions/motion-option:1"],
{
"id": 1,
"poll_id": 1,
"pollstate": 2,
}, # noqa black and flake are no friends :(
)
# Other users should not get a vote autoupdate
self.assertNoAutoupdate(vote, user=self.other_user)
self.assertNoDeletedAutoupdate(vote, user=self.other_user)
class VoteMotionPollPseudoanonymousAutoupdates(TestCase):
"""3 important users:
self.admin: manager, has can_see, can_manage, can_manage_polls (in admin group)
self.user: votes, has can_see perms and in in delegate group
self.other_user: Just has can_see perms and is NOT in the delegate group.
"""
def advancedSetUp(self):
self.motion = Motion(
title="test_title_OoK9IeChe2Jeib9Deeji",
text="test_text_eichui1oobiSeit9aifo",
)
self.motion.save()
self.delegate_group = get_group_model().objects.get(pk=GROUP_DELEGATE_PK)
self.other_user, _ = self.create_user()
inform_changed_data(self.other_user)
self.user, user_password = self.create_user()
self.user.groups.add(self.delegate_group)
self.user.is_present = True
self.user.save()
self.user_client = APIClient()
self.user_client.login(username=self.user.username, password=user_password)
self.poll = MotionPoll.objects.create(
motion=self.motion,
title="test_title_cahP1umooteehah2jeey",
pollmethod="YNA",
type=BasePoll.TYPE_PSEUDOANONYMOUS,
state=MotionPoll.STATE_STARTED,
onehundred_percent_base="YN",
majority_method="simple",
)
self.poll.create_options()
self.poll.groups.add(self.delegate_group)
def test_vote(self):
response = self.user_client.post(
reverse("motionpoll-vote", args=[self.poll.pk]), {"data": "A"}
)
self.assertHttpStatusVerbose(response, status.HTTP_200_OK)
poll = MotionPoll.objects.get()
vote = MotionVote.objects.get()
# Expect the admin to see the full data in the autoupdate
autoupdate = self.get_last_autoupdate(user=self.admin)
self.assertEqual(
autoupdate[0],
{
"motions/motion-poll:1": {
"motion_id": 1,
"pollmethod": "YNA",
"state": 2,
"type": "pseudoanonymous",
"title": "test_title_cahP1umooteehah2jeey",
"onehundred_percent_base": "YN",
"majority_method": "simple",
"groups_id": [GROUP_DELEGATE_PK],
"votesvalid": "1.000000",
"votesinvalid": "0.000000",
"votescast": "1.000000",
"options_id": [1],
"id": 1,
"user_has_voted": False,
"user_has_voted_for_delegations": [],
"voted_id": [self.user.id],
},
"motions/motion-vote:1": {
"pollstate": 2,
"option_id": 1,
"id": 1,
"weight": "1.000000",
"value": "A",
"user_id": None,
"delegated_user_id": None,
},
"motions/motion-option:1": {
"abstain": "1.000000",
"id": 1,
"no": "0.000000",
"poll_id": 1,
"pollstate": 2,
"yes": "0.000000",
},
},
)
self.assertEqual(autoupdate[1], [])
# Expect non-admins to get a restricted poll update and no autoupdate
# for a changed vote nor a deleted one
for user in (self.user, self.other_user):
self.assertAutoupdate(poll, user=user)
autoupdate = self.get_last_autoupdate(user=user)
self.assertEqual(
autoupdate[0]["motions/motion-poll:1"],
{
"motion_id": 1,
"pollmethod": "YNA",
"state": 2,
"type": "pseudoanonymous",
"title": "test_title_cahP1umooteehah2jeey",
"onehundred_percent_base": "YN",
"majority_method": "simple",
"groups_id": [GROUP_DELEGATE_PK],
"options_id": [1],
"id": 1,
"user_has_voted": user == self.user,
"user_has_voted_for_delegations": [],
},
)
self.assertNoAutoupdate(vote, user=user)
self.assertNoDeletedAutoupdate(vote, user=user)
class VoteMotionPollPseudoanonymous(TestCase):
def setUp(self):
self.client = APIClient()
self.client.login(username="admin", password="admin")
self.motion = Motion(
title="test_title_Chaebaenges1aebe8iev",
text="test_text_cah2aigh6ahc8OhNguQu",
)
self.motion.save()
self.group = get_group_model().objects.get(pk=GROUP_DELEGATE_PK)
self.admin = get_user_model().objects.get(username="admin")
self.poll = MotionPoll.objects.create(
motion=self.motion,
title="test_title_yohphei9Iegohqu9ki7m",
pollmethod="YNA",
type=BasePoll.TYPE_PSEUDOANONYMOUS,
)
self.poll.create_options()
self.poll.groups.add(self.group)
def start_poll(self):
self.poll.state = MotionPoll.STATE_STARTED
self.poll.save()
def make_admin_delegate(self):
self.admin.groups.add(GROUP_DELEGATE_PK)
self.admin.groups.remove(GROUP_ADMIN_PK)
inform_changed_data(self.admin)
def make_admin_present(self):
self.admin.is_present = True
self.admin.save()
def test_start_poll(self):
response = self.client.post(reverse("motionpoll-start", args=[self.poll.pk]))
self.assertHttpStatusVerbose(response, status.HTTP_200_OK)
poll = MotionPoll.objects.get()
self.assertEqual(poll.state, MotionPoll.STATE_STARTED)
self.assertEqual(poll.votesvalid, Decimal("0"))
self.assertEqual(poll.votesinvalid, Decimal("0"))
self.assertEqual(poll.votescast, Decimal("0"))
self.assertFalse(poll.get_votes().exists())
def test_vote(self):
self.start_poll()
self.make_admin_delegate()
self.make_admin_present()
response = self.client.post(
reverse("motionpoll-vote", args=[self.poll.pk]), {"data": "N"}
)
self.assertHttpStatusVerbose(response, status.HTTP_200_OK)
poll = MotionPoll.objects.get()
self.assertEqual(poll.votesvalid, Decimal("1"))
self.assertEqual(poll.votesinvalid, Decimal("0"))
self.assertEqual(poll.votescast, Decimal("1"))
self.assertEqual(poll.get_votes().count(), 1)
self.assertEqual(poll.amount_users_voted_with_individual_weight(), 1)
option = poll.options.get()
self.assertEqual(option.yes, Decimal("0"))
self.assertEqual(option.no, Decimal("1"))
self.assertEqual(option.abstain, Decimal("0"))
self.assertTrue(self.admin in poll.voted.all())
vote = option.votes.get()
self.assertEqual(vote.user, None)
def test_change_vote(self):
self.start_poll()
self.make_admin_delegate()
self.make_admin_present()
response = self.client.post(
reverse("motionpoll-vote", args=[self.poll.pk]), {"data": "N"}
)
self.assertHttpStatusVerbose(response, status.HTTP_200_OK)
response = self.client.post(
reverse("motionpoll-vote", args=[self.poll.pk]), {"data": "A"}
)
self.assertHttpStatusVerbose(response, status.HTTP_400_BAD_REQUEST)
option = MotionPoll.objects.get().options.get()
self.assertEqual(option.yes, Decimal("0"))
self.assertEqual(option.no, Decimal("1"))
self.assertEqual(option.abstain, Decimal("0"))
vote = option.votes.get()
self.assertEqual(vote.user, None)
def test_vote_anonymous(self):
self.poll.groups.add(GROUP_DEFAULT_PK)
self.start_poll()
config["general_system_enable_anonymous"] = True
guest_client = APIClient()
response = guest_client.post(
reverse("motionpoll-vote", args=[self.poll.pk]), {"data": "Y"}
)
self.assertHttpStatusVerbose(response, status.HTTP_403_FORBIDDEN)
self.assertFalse(MotionPoll.objects.get().get_votes().exists())
def test_vote_wrong_state(self):
self.make_admin_present()
self.make_admin_delegate()
response = self.client.post(
reverse("motionpoll-vote", args=[self.poll.pk]), {"data": {}}
)
self.assertHttpStatusVerbose(response, status.HTTP_400_BAD_REQUEST)
self.assertFalse(MotionPoll.objects.get().get_votes().exists())
def test_vote_wrong_group(self):
self.start_poll()
self.make_admin_present()
response = self.client.post(
reverse("motionpoll-vote", args=[self.poll.pk]), {"data": {}}
)
self.assertHttpStatusVerbose(response, status.HTTP_403_FORBIDDEN)
self.assertFalse(MotionPoll.objects.get().get_votes().exists())
def test_vote_not_present(self):
self.start_poll()
self.make_admin_delegate()
response = self.client.post(
reverse("motionpoll-vote", args=[self.poll.pk]), {"data": {}}
)
self.assertHttpStatusVerbose(response, status.HTTP_403_FORBIDDEN)
self.assertFalse(MotionPoll.objects.get().get_votes().exists())
def test_vote_missing_data(self):
self.start_poll()
self.make_admin_delegate()
self.make_admin_present()
response = self.client.post(
reverse("motionpoll-vote", args=[self.poll.pk]), {"data": {}}
)
self.assertHttpStatusVerbose(response, status.HTTP_400_BAD_REQUEST)
self.assertFalse(MotionPoll.objects.get().get_votes().exists())
def test_vote_wrong_data_format(self):
self.start_poll()
self.make_admin_delegate()
self.make_admin_present()
response = self.client.post(
reverse("motionpoll-vote", args=[self.poll.pk]), {"data": [1, 2, 5]}
)
self.assertHttpStatusVerbose(response, status.HTTP_400_BAD_REQUEST)
self.assertFalse(MotionPoll.objects.get().get_votes().exists())
class StopMotionPoll(TestCase):
def setUp(self):
self.client = APIClient()
self.client.login(username="admin", password="admin")
self.motion = Motion(
title="test_title_eiri4iipeemaeGhahkae",
text="test_text_eegh7quoochaiNgiyeix",
)
self.motion.save()
self.poll = MotionPoll.objects.create(
motion=self.motion,
title="test_title_Hu9Miebopaighee3EDie",
pollmethod="YNA",
type=BasePoll.TYPE_NAMED,
)
self.poll.create_options()
def test_stop_poll(self):
self.poll.state = MotionPoll.STATE_STARTED
self.poll.save()
response = self.client.post(reverse("motionpoll-stop", args=[self.poll.pk]))
self.assertHttpStatusVerbose(response, status.HTTP_200_OK)
self.assertEqual(MotionPoll.objects.get().state, MotionPoll.STATE_FINISHED)
def test_stop_wrong_state(self):
response = self.client.post(reverse("motionpoll-stop", args=[self.poll.pk]))
self.assertHttpStatusVerbose(response, status.HTTP_400_BAD_REQUEST)
self.assertEqual(MotionPoll.objects.get().state, MotionPoll.STATE_CREATED)
class PublishMotionPoll(TestCase):
def advancedSetUp(self):
self.motion = Motion(
title="test_title_lai8Ho5gai9aijahRasu",
text="test_text_KieGhosh8ahWiguHeu2D",
)
self.motion.save()
self.poll = MotionPoll.objects.create(
motion=self.motion,
title="test_title_Nufae0iew7Iorox2thoo",
pollmethod="YNA",
type=BasePoll.TYPE_PSEUDOANONYMOUS,
onehundred_percent_base="YN",
majority_method="simple",
)
self.poll.create_options()
option = self.poll.options.get()
self.user, _ = self.create_user()
self.vote = MotionVote.objects.create(
option=option, user=None, weight=Decimal(2), value="N"
)
def test_publish_poll(self):
self.poll.state = MotionPoll.STATE_FINISHED
self.poll.save()
response = self.client.post(reverse("motionpoll-publish", args=[self.poll.pk]))
self.assertHttpStatusVerbose(response, status.HTTP_200_OK)
self.assertEqual(MotionPoll.objects.get().state, MotionPoll.STATE_PUBLISHED)
# Test autoupdates: Every user should get the full data
for user in (self.admin, self.user):
autoupdate = self.get_last_autoupdate(user=user)
self.assertEqual(
autoupdate[0],
{
"motions/motion-poll:1": {
"motion_id": 1,
"pollmethod": "YNA",
"state": 4,
"type": "pseudoanonymous",
"title": "test_title_Nufae0iew7Iorox2thoo",
"onehundred_percent_base": "YN",
"majority_method": "simple",
"groups_id": [],
"votesvalid": "0.000000",
"votesinvalid": "0.000000",
"votescast": "0.000000",
"options_id": [1],
"id": 1,
"user_has_voted": False,
"user_has_voted_for_delegations": [],
"voted_id": [],
},
"motions/motion-vote:1": {
"pollstate": 4,
"option_id": 1,
"id": 1,
"weight": "2.000000",
"value": "N",
"user_id": None,
"delegated_user_id": None,
},
"motions/motion-option:1": {
"abstain": "0.000000",
"id": 1,
"no": "2.000000",
"poll_id": 1,
"pollstate": 4,
"yes": "0.000000",
},
},
)
self.assertEqual(autoupdate[1], [])
def test_publish_wrong_state(self):
response = self.client.post(reverse("motionpoll-publish", args=[self.poll.pk]))
self.assertHttpStatusVerbose(response, status.HTTP_400_BAD_REQUEST)
self.assertEqual(MotionPoll.objects.get().state, MotionPoll.STATE_CREATED)
class PseudoanonymizeMotionPoll(TestCase):
def setUp(self):
self.client = APIClient()
self.client.login(username="admin", password="admin")
self.motion = Motion(
title="test_title_lai8Ho5gai9aijahRasu",
text="test_text_KieGhosh8ahWiguHeu2D",
)
self.motion.save()
self.poll = MotionPoll.objects.create(
motion=self.motion,
title="test_title_Nufae0iew7Iorox2thoo",
pollmethod="YNA",
type=BasePoll.TYPE_NAMED,
state=MotionPoll.STATE_FINISHED,
)
self.poll.create_options()
self.option = self.poll.options.get()
self.user1, _ = self.create_user()
self.vote1 = MotionVote.objects.create(
user=self.user1, option=self.option, value="Y", weight=Decimal(1)
)
self.poll.voted.add(self.user1)
self.user2, _ = self.create_user()
self.vote2 = MotionVote.objects.create(
user=self.user2, option=self.option, value="N", weight=Decimal(1)
)
self.poll.voted.add(self.user2)
def test_pseudoanonymize_poll(self):
response = self.client.post(
reverse("motionpoll-pseudoanonymize", args=[self.poll.pk])
)
self.assertHttpStatusVerbose(response, status.HTTP_200_OK)
poll = MotionPoll.objects.get()
self.assertEqual(poll.get_votes().count(), 2)
self.assertEqual(poll.amount_users_voted_with_individual_weight(), 2)
self.assertEqual(poll.votesvalid, Decimal("2"))
self.assertEqual(poll.votesinvalid, Decimal("0"))
self.assertEqual(poll.votescast, Decimal("2"))
option = poll.options.get()
self.assertEqual(option.yes, Decimal("1"))
self.assertEqual(option.no, Decimal("1"))
self.assertEqual(option.abstain, Decimal("0"))
self.assertTrue(self.user1 in poll.voted.all())
self.assertTrue(self.user2 in poll.voted.all())
for vote in poll.get_votes().all():
self.assertTrue(vote.user is None)
def test_pseudoanonymize_wrong_state(self):
self.poll.state = MotionPoll.STATE_CREATED
self.poll.save()
response = self.client.post(
reverse("motionpoll-pseudoanonymize", args=[self.poll.pk])
)
self.assertHttpStatusVerbose(response, status.HTTP_400_BAD_REQUEST)
poll = MotionPoll.objects.get()
self.assertTrue(poll.get_votes().filter(user=self.user1).exists())
self.assertTrue(poll.get_votes().filter(user=self.user2).exists())
def test_pseudoanonymize_wrong_type(self):
self.poll.type = MotionPoll.TYPE_ANALOG
self.poll.save()
response = self.client.post(
reverse("motionpoll-pseudoanonymize", args=[self.poll.pk])
)
self.assertHttpStatusVerbose(response, status.HTTP_400_BAD_REQUEST)
poll = MotionPoll.objects.get()
self.assertTrue(poll.get_votes().filter(user=self.user1).exists())
self.assertTrue(poll.get_votes().filter(user=self.user2).exists())
class ResetMotionPoll(TestCase):
def advancedSetUp(self):
self.motion = Motion(
title="test_title_cheiJ1ieph5ohng9queu",
text="test_text_yahng6fiegaL7mooZ2of",
)
self.motion.save()
self.poll = MotionPoll.objects.create(
motion=self.motion,
title="test_title_oozie2Ui9xie0chaghie",
pollmethod="YNA",
type=BasePoll.TYPE_ANALOG,
state=MotionPoll.STATE_FINISHED,
)
self.poll.create_options()
self.option = self.poll.options.get()
self.user1, _ = self.create_user()
self.vote1 = MotionVote.objects.create(
user=self.user1, option=self.option, value="Y", weight=Decimal(1)
)
self.poll.voted.add(self.user1)
self.user2, _ = self.create_user()
self.vote2 = MotionVote.objects.create(
user=self.user2, option=self.option, value="N", weight=Decimal(1)
)
self.poll.voted.add(self.user2)
def test_reset_poll(self):
response = self.client.post(reverse("motionpoll-reset", args=[self.poll.pk]))
self.assertHttpStatusVerbose(response, status.HTTP_200_OK)
poll = MotionPoll.objects.get()
self.assertEqual(poll.get_votes().count(), 0)
self.assertEqual(poll.amount_users_voted_with_individual_weight(), 0)
self.assertEqual(poll.votesvalid, None)
self.assertEqual(poll.votesinvalid, None)
self.assertEqual(poll.votescast, None)
option = poll.options.get()
self.assertEqual(option.yes, Decimal("0"))
self.assertEqual(option.no, Decimal("0"))
self.assertEqual(option.abstain, Decimal("0"))
self.assertFalse(option.votes.exists())
def test_deleted_autoupdate(self):
response = self.client.post(reverse("motionpoll-reset", args=[self.poll.pk]))
self.assertHttpStatusVerbose(response, status.HTTP_200_OK)
poll = MotionPoll.objects.get()
option = poll.options.get()
self.assertAutoupdate(option, self.admin)
for user in (self.admin, self.user1, self.user2):
self.assertDeletedAutoupdate(self.vote1, user=user)
self.assertDeletedAutoupdate(self.vote2, user=user)
class TestMotionPollWithVoteDelegationAutoupdate(TestCase):
def advancedSetUp(self):
""" Set up user -> other_user delegation. """
self.motion = Motion(
title="test_title_dL91JqhMTiQuQLSDRItZ",
text="test_text_R7nURdXKVEfEnnJBXJYa",
)
self.motion.save()
self.delegate_group = get_group_model().objects.get(pk=GROUP_DELEGATE_PK)
self.other_user, _ = self.create_user()
self.user, user_password = self.create_user()
self.user.groups.add(self.delegate_group)
self.user.is_present = True
self.user.vote_delegated_to = self.other_user
self.user.save()
self.user_client = APIClient()
self.user_client.login(username=self.user.username, password=user_password)
self.poll = MotionPoll.objects.create(
motion=self.motion,
title="test_title_Q3EuRaALSCCPJuQ2tMqj",
pollmethod="YNA",
type=BasePoll.TYPE_NAMED,
onehundred_percent_base="YN",
majority_method="simple",
)
self.poll.create_options()
self.poll.groups.add(self.delegate_group)
self.poll.save()
def test_start_poll(self):
response = self.client.post(reverse("motionpoll-start", args=[self.poll.pk]))
self.assertHttpStatusVerbose(response, status.HTTP_200_OK)
# other_user has to receive an autoupdate because he was delegated
autoupdate = self.get_last_autoupdate(user=self.other_user)
assert "motions/motion-poll:1" in autoupdate[0]
|
jwinzer/OpenSlides
|
server/tests/integration/motions/test_polls.py
|
Python
|
mit
| 67,548
|
"""Auto-generated file, do not edit by hand. EE metadata"""
from ..phonemetadata import NumberFormat, PhoneNumberDesc, PhoneMetadata
PHONE_METADATA_EE = PhoneMetadata(id='EE', country_code=None, international_prefix=None,
general_desc=PhoneNumberDesc(national_number_pattern='1\\d{2,5}', possible_length=(3, 6)),
toll_free=PhoneNumberDesc(national_number_pattern='116(?:000|111)', example_number='116000', possible_length=(6,)),
emergency=PhoneNumberDesc(national_number_pattern='11[02]', example_number='112', possible_length=(3,)),
short_code=PhoneNumberDesc(national_number_pattern='1(?:\\d{2}|16(?:000|111))', example_number='116', possible_length=(3, 6)),
short_data=True)
|
samdowd/drumm-farm
|
drumm_env/lib/python2.7/site-packages/phonenumbers/shortdata/region_EE.py
|
Python
|
mit
| 699
|
#!/usr/bin/env python
# -*- coding: utf-8 -*-
#------------------------------------------------------------------------------
__author__ = 'James T. Dietrich'
__contact__ = 'james.t.dietrich@dartmouth.edu'
__copyright__ = '(c) James Dietrich 2016'
__license__ = 'MIT'
__date__ = 'Wed Nov 16 11:33:39 2016'
__version__ = '1.0'
__status__ = "initial release"
__url__ = "https://github.com/geojames/..."
"""
Name: Week6-1_Pandas.py
Compatibility: Python 3.5
Description: This program does stuff
URL: https://github.com/geojames/...
Requires: libraries
Dev ToDo:
AUTHOR: James T. Dietrich
ORGANIZATION: Dartmouth College
Contact: james.t.dietrich@dartmouth.edu
Copyright: (c) James Dietrich 2016
"""
#------------------------------------------------------------------------------
# Pandas provides a set of labeled array data structures
# Similar to what you're used to in Excel
# + 1-D "Series"
# + 2-D "Data Frames"
# Import Pandas
import pandas as pd
# You also need Numpy
import numpy as np
# and Matplotlib, if you want to plot data
import matplotlib.pyplot as plt
#%% Creating Pandas data
# chances are you'll be importing data into Pandas with read_csv, which we'll
# get to...but you'll also need to create Pandas data from scratch too
# Pandas uses two data structures:
# Series = 1-D data (single column)
# Data Frame = 2-D data (multiple columns)
# Terms:
# + Column = actual data
# + Index Column = first column, defaults to numbers, but also you can
# also name rows in the index column or use dates as the index column
# * Yes, it is confusing, since data frames still have indecies like
# arrays...
#
# The general structure of a data frame is:
#
# Index COL1 COL2 COL3 ...
# 1 data data data
# 2 data data data
# 3 data data data
# 4 data data data
# ...
# of for real data would look something like this:
# Index temp pres wind wind_dir ...
# 1 26.2 1012.3 0 nan
# 2 27.0 1011.1 5 "w"
# 3 27.8 1011.5 8 "nw"
# 4 24.3 1011.3 5 "w"
# ...
# 1-D series data are easy...
s = pd.Series([1,3,5,np.nan,6,8])
# 2-D data can be created in a couple different ways
# 1: typing in matrix vaules
df = pd.DataFrame([[1,2,3,4,5], [6,7,8,9,10]],columns=['A','B','C','D','E'])
# 2: from an exisitng Numpy Array
array = np.array([[1,2,3,4,5], [6,7,8,9,10]])
df2 = pd.DataFrame(array)
# 3: numpy array with column names
array = np.arange(1.,17.).reshape((4,4))
df3 = pd.DataFrame(array, columns=['A','B','C','D'])
# adding columns
# df['new column name'] = data to add...
colors = pd.Series(['blue','green','red','yellow'])
df3['color'] = colors
# directly...
df3['food'] = np.array(['fish','lettcue','apple','banana'])
#%% Accessing index and column name lists
df.index
df.columns
df.values
df.dtypes
df.describe()
#%% Accessng/adding/deleting column data
# access the column by name: string OR by dot notation
df3['A']
df3.A
# adding data
df3['color'] = pd.Series(['blue','green','red','yellow'])
df3['random'] = np.random.rand(4,1)
# new columns by math
df3['AxRand'] = df3.A * df3.random
# boolean tests as columns
df3['Blue_TF'] = df3.color == 'blue'
# adding one value to all rows (propogated value)
df3['ones'] = 'one'
# deleting columns
del df3['random']
# extracting values and deleting (pop out)
rand_vals = df3.pop('AxRand')
# Inserting new columns
# insert(position, col_name, values)
df3.insert(0,'AxR',rand_vals)
#%% Date ranges
# often your data will need dates. Pandas has some nifty short cuts for adding
# dates to your data frames
# Start and end dates with frequency
# pd.date_range(start, end, freq = 'D')
pd.date_range('1/1/2016','12/31/2018')
# Start, frequency and number or periods
# pd.date_range(start, periods=10, freq='D')
# Frequency Codes
# 'A' = year end - 'AS' = year start
# 'M' = month end - 'MS' = month start
# 'W' = week
# 'D' = days, default
# 'H' = hours
# 'min' = minutes
# 's' = seconds
#
# '5H' = 5 hours, '30s' = 30 seconds, '7D' = 7 days
pd.date_range('1/1/2016',periods=10, freq='D')
pd.date_range('1/1/2016',periods=10, freq='min')
# adding time (24-hour format)
pd.date_range('1/1/2016 07:30:00',periods=10, freq='30d')
#%% Data Frame Indexing
dates = pd.date_range('1/1/2016', periods=8)
df = pd.DataFrame(np.random.randn(8, 4), index=dates, columns=['A', 'B', 'C', 'D'])
# Getting Columns
df.A
df['B']
# Rows, all columns
# by classic index
df[0:4]
# by named index range
df['2016-01-02':'2016-01-04']
# using the location (loc) function
# dates index, row 0
df.loc[dates[0]]
# row and column
df.loc[2:4,['A','B']]
# index and columns
df.loc[dates[2],'A'] # or is you have a generic index: df.loc[index[2],'A']
# Pure index locations (iloc) function
# rows and columns, like arrays
df.iloc[3:5,0:2]
#%% Data Frame Math
# Math on an entire Data Frame
df2 = df * 3
# In place replacement (not recommended, but good for recalculating and overwriting)
df.B = df.B * 2
# better to create a new column
df['Btimes2'] = df.B * 2
# math between columns
df['convert'] = ((df.A * 9/5) + 32) - df.C
# math for specific indecies
df['indexMath'] = df.D[2:5] ** 2
# math with boolean masks
# only apply the equation to positive vaules in A
df['convertPos'] = ((df.A[df.A >=0] * 9/5) + 32) - df.C[df.A >=0]
# Apply a different equation to the negative values
df.convertPos[df.A < 0] = ((df.A[df.A < 0] * 9/5) + 32) + df.D[df.A < 0]
#%% Reduction
dates = pd.date_range('1/1/2016', periods=8)
df = pd.DataFrame(np.random.randn(8, 4), index=dates, columns=['A', 'B', 'C', 'D'])
# Basic stats (default is columns, and to skip NaN values)
# for row stats add (axis=0)
df.mean()
df.std()
df.median()
# Getting statistics for a specific column
df.A.mean()
# Sums (again by column)
df.sum()
df.B.sum()
#%% Combining / Adding Values
df1 = pd.DataFrame({'A': ['A0', 'A1', 'A2', 'A3'],
'B': ['B0', 'B1', 'B2', 'B3'],
'C': ['C0', 'C1', 'C2', 'C3'],
'D': ['D0', 'D1', 'D2', 'D3']},
index=[0, 1, 2, 3])
df2 = pd.DataFrame({'A': ['A4', 'A5', 'A6', 'A7'],
'B': ['B4', 'B5', 'B6', 'B7'],
'C': ['C4', 'C5', 'C6', 'C7'],
'D': ['D4', 'D5', 'D6', 'D7']},
index=[4, 5, 6, 7])
series = pd.Series(['X10', 'X11', 'X12', 'X13'], name='X')
result = pd.concat([df1, df2])
result2 = pd.concat([df1, df2], axis = 1)
result = df1.append(df2)
result = pd.concat([df1,series],axis=1)
# combining data with matching indicies
dates = pd.date_range('1/1/2016', periods=8)
df = pd.DataFrame(np.random.randn(8, 4), index=dates, columns=['A', 'B', 'C', 'D'])
dates = pd.date_range('1/4/2016', periods=8)
df2 = pd.DataFrame(np.random.randn(8, 2), index=dates, columns=['E', 'F'])
df3 = pd.concat([df, df2], axis=1)
#%% Converting Pandas to Numpy
# single column
np_from_pd = pd.np.array(df3.A)
# whole thing
np_from_pd = pd.np.array(df3)
# using numpy santax
array = np.array(df3.values)
# Monster statments with multiple instructions
np_with_reshape = pd.np.array(df3.A).reshape((2,2))
|
geojames/Dart_EnvGIS
|
Week6-1_Pandas.py
|
Python
|
mit
| 7,208
|
"""
Created on Sep 01, 2017
@author: StarlitGhost
"""
import datetime
import time
from twisted.plugin import IPlugin
from twisted.words.protocols.irc import assembleFormattedText, attributes as A
from zope.interface import implementer
from desertbot.message import IRCMessage
from desertbot.moduleinterface import IModule
from desertbot.modules.commandinterface import BotCommand
from desertbot.response import IRCResponse
from desertbot.utils import string
@implementer(IPlugin, IModule)
class Splatoon(BotCommand):
def triggers(self):
return ['splat']
def help(self, query):
return "splat [regular/ranked/league/fest]"
graySplitter = assembleFormattedText(A.normal[' ', A.fg.gray['|'], ' '])
def _fetch(self, j, short, mode, label):
r = j[mode]
data = []
t = A.normal[A.bold['{} {}: '.format(label, r[0]['rule']['name'])],
'/'.join([r[0]['stage_a']['name'], r[0]['stage_b']['name']])]
data.append(assembleFormattedText(t))
if not short:
# include next maps
now = int(time.time())
startTime = r[1]['startTime']
delta = startTime - now
d = datetime.timedelta(seconds=delta)
deltaStr = string.deltaTimeToString(d, resolution='m')
t = A.normal[A.bold['{} {} in {}: '.format(label, r[1]['rule']['name'], deltaStr)],
'/'.join([r[1]['stage_a']['name'], r[1]['stage_b']['name']])]
data.append(assembleFormattedText(t))
return ' | '.join(data)
def _regular(self, j, short):
return self._fetch(j, short, 'regular', 'Regular')
def _ranked(self, j, short):
return self._fetch(j, short, 'gachi', 'Ranked')
def _league(self, j, short):
return self._fetch(j, short, 'league', 'League')
def _fest(self, j, short):
if j['splatfests']:
pass
elif not short:
return 'No SplatFest is currently scheduled'
def execute(self, message: IRCMessage):
url = "https://splatoon2.ink/data/schedules.json"
response = self.bot.moduleHandler.runActionUntilValue('fetch-url', url)
j = response.json()
if len(message.parameterList) < 1:
# do everything
data = []
data += filter(None, [self._regular(j, short=True)])
data += filter(None, [self._ranked(j, short=True)])
data += filter(None, [self._league(j, short=True)])
data += filter(None, [self._fest(j, short=True)])
return IRCResponse(self.graySplitter.join(data), message.replyTo)
else:
subCommands = {
'regular': self._regular,
'ranked': self._ranked,
'league': self._league,
'fest': self._fest
}
subCommand = message.parameterList[0].lower()
if subCommand in subCommands:
return IRCResponse(subCommands[subCommand](j, short=False), message.replyTo)
else:
return IRCResponse(self.help(None), message.replyTo)
splatoon = Splatoon()
|
DesertBot/DesertBot
|
desertbot/modules/commands/Splatoon.py
|
Python
|
mit
| 3,155
|
# -*- coding: utf-8 -*-
"""
Created on Tue Apr 28 08:19:21 2015
@author: Roger
"""
from math import *
import numpy as np
import matplotlib.pyplot as plt
from matplotlib.patches import Polygon
wheelbase = 100 #inches
vel = 20 *12 # fps to inches per sec
steering_angle = radians(1)
t = 1 # second
orientation = 0. # radians
pos = np.array([0., 0.]
for i in range(100):
#if abs(steering_angle) > 1.e-8:
turn_radius = tan(steering_angle)
radius = wheelbase / tan(steering_angle)
dist = vel*t
arc_len = dist / (2*pi*radius)
turn_angle = 2*pi * arc_len
cx = pos[0] - radius * sin(orientation)
cy = pos[1] + radius * cos(orientation)
orientation = (orientation + turn_angle) % (2.0 * pi)
pos[0] = cx + (sin(orientation) * radius)
pos[1] = cy - (cos(orientation) * radius)
plt.scatter(pos[0], pos[1])
plt.axis('equal')
|
zaqwes8811/micro-apps
|
self_driving/deps/Kalman_and_Bayesian_Filters_in_Python_master/experiments/bicycle.py
|
Python
|
mit
| 875
|
# -*- coding: utf-8 -*-
import numpy as np
from scipy.spatial import distance_matrix
from scipy import sparse
import pandas as pd
from keras.models import Sequential
from keras.layers import Dense
from keras.layers import Dropout, Activation, Flatten
from keras.optimizers import SGD,RMSprop
#from keras.wrappers.scikit_learn import KerasRegressor
from keras import regularizers
from sklearn.model_selection import cross_val_score
from sklearn.model_selection import KFold
from sklearn.preprocessing import StandardScaler
from sklearn.preprocessing import MinMaxScaler
from sklearn.model_selection import GridSearchCV
from sklearn.model_selection import RandomizedSearchCV
from scipy.stats import randint as sp_randint
from sklearn import preprocessing
from sklearn.pipeline import Pipeline
from keras.utils import to_categorical
from sklearn.model_selection import train_test_split
import matplotlib.pyplot as plt
import matplotlib.mlab as mlab
from keras.layers import Convolution2D, MaxPooling2D
import time
#%%
# =============================================================================
# Gernerate data
# =============================================================================
sizeGrid = 5 # size of the grid
nObs = 6 # number of observations
K = 3 # number of features
# set coefficients
b1 = np.random.normal(5, 2, K).reshape(K,1) # coef for own characteristics
b2 = np.random.normal(5, 2, K).reshape(K,1) # coef for neighbor characteristics
errStd = 2 # error added to Y
cutW = 6 # cut distance in W
#%%
def generate_landscape(sizeGrid,nObs,K,b1,b2,errStd,cutW):
# Create location of observations
locOwn = np.array([[sizeGrid/2,sizeGrid/2]])
while True:
loc = np.random.randint(0,sizeGrid, size=(nObs,2))
locAll = np.concatenate((locOwn, loc), axis=0)
# Make sure that farm are not one same location
locAll = np.unique(locAll, axis=0)
if nObs+1 == locAll.shape[0]:
break
# Create own characteristics
X = np.random.randint(0,100, size=(nObs,K))
Xown = np.random.randint(0,100, size=(1,K))
# Create spatial weigthing matrix W
W = distance_matrix(locOwn, loc)<cutW
np.fill_diagonal(W,0) # fill diagonal with zeros
row_sum = W.sum(axis=1,keepdims=True) # calc row sum
row_sum[row_sum==0] = 1
W = W/row_sum # row standardize
# Create error
err = np.random.normal(0, errStd, 1)
# Calcualte Y
Y = np.matmul(Xown,b1)+ np.matmul(np.matmul(W,X),b2)+err
assert(Y.shape==(1,1))
maps = np.zeros((sizeGrid,sizeGrid,K))
#
for k in range(0,K):
I = np.concatenate((locOwn[:,0],loc[:,0]),axis=0)
J = np.concatenate((locOwn[:,1],loc[:,1]),axis=0)
V = np.concatenate((Xown[:,k],X[:,k]),axis=0)
A = sparse.coo_matrix((V,(I,J)),shape=(sizeGrid,sizeGrid))
maps[:,:,k] = A.todense()
#
return maps,Y
# %%
N = 100
X = np.zeros((N,sizeGrid*sizeGrid*K))
Y = np.zeros((N,1))
for i in range(0,N):
#
maps,y = generate_landscape(sizeGrid,nObs,K,b1,b2,errStd,cutW)
Y[i,:] = y
#X[i,:,:,:] = maps
X[i,:] = maps.T.reshape((1,maps.shape[0]*maps.shape[1]*maps.shape[2]))
# %%
# =============================================================================
# Train model
# =============================================================================
xDim = X.shape[1]
yDim = 1
def dropout2_model(lr=0.1,decay=1e-6,momentum=0.9,dropout_rate=0.2):
"""
Define a modelwith dropout.
Argument:
lr -- learning rate
decay --
momentum --
dropout_rate --
Returns:
model -- Keras model
"""
model = Sequential()
model.add(Dropout(dropout_rate, input_shape=(xDim,)))
model.add(Dropout(dropout_rate))
model.add(Dense(200, kernel_initializer='normal', activation='relu'))
model.add(Dropout(dropout_rate))
model.add(Dense(50, kernel_initializer='normal', activation='relu'))
model.add(Dropout(dropout_rate))
model.add(Dense(yDim, kernel_initializer='normal'))
# Compile model
#sgd = SGD(lr=lr, decay=decay, momentum=momentum, nesterov=True)
rmsprop = RMSprop(lr=0.001, rho=0.9, epsilon=1e-08, decay=0.0)
model.compile(loss='mean_squared_error', optimizer=rmsprop)
return model
# ## fix random seed for reproducibility
seed = 8
np.random.seed(seed)
# %% standardized features and targets
x_min_max = MinMaxScaler()
y_min_max = MinMaxScaler()
X_minmax = x_min_max.fit_transform(X)
Y_minmax = y_min_max.fit_transform(Y)
# %% Split sample in test and training_set
X_train, X_test, Y_train, Y_test = train_test_split(X_minmax, Y_minmax, test_size=0.10, random_state=42)
#%%
estimator= dropout2_model()
estimator.fit(X,Y,epochs=30, batch_size=124, validation_data=(X, Y))
# %%
Yhat_test = estimator.predict(X_test,batch_size=32)
oY_test = y_min_max.inverse_transform(Y_test)
oY_hat = y_min_max.inverse_transform(Yhat_test)
# %%
fig, ax = plt.subplots()
ax.scatter(oY_test, oY_hat, edgecolors=(0, 0, 0))
ax.plot([oY_test.min(), oY_test.max()], [oY_test.min(), oY_test.max()], 'k--', lw=4)
ax.set_xlabel('Measured')
ax.set_ylabel('Predicted')
plt.show()
# %%
def r2(y_true, y_pred):
"""Calcualte and return R2.
y_true -- the observed values
y_pred -- the prediced values
"""
SS_res = np.sum(np.square(y_true - y_pred))
SS_tot = np.sum(np.square(y_true - np.mean(y_true)))
return ( 1 - SS_res/SS_tot )
r2Model = r2(oY_test,oY_hat)
print("R2 Model: ",r2Model)
|
hstorm/nn_spatial
|
notebooks/test_spatial.py
|
Python
|
mit
| 5,525
|
import thread
from time import sleep, ctime
def loop0():
print 'start loop 0 at: ', ctime()
sleep(4)
print 'loop 0 done at: ', ctime()
def loop1():
print 'start loop 1 at: ', ctime()
sleep(2)
print 'loop 1 done at: ', ctime()
def main():
print 'starting at: ', ctime()
thread.start_new_thread(loop0, ())
thread.start_new_thread(loop1, ())
sleep(6)
print 'all DONE at: ', ctime()
if __name__ == '__main__':
main()
|
seerjk/reboot06
|
arch_pre_work/mtsleep1.py
|
Python
|
mit
| 467
|
"""
Provides a test case for issue 283 - "Inheritance breaks".
The issue is outlined here: https://github.com/neo4j-contrib/neomodel/issues/283
More information about the same issue at:
https://github.com/aanastasiou/neomodelInheritanceTest
The following example uses a recursive relationship for economy, but the
idea remains the same: "Instantiate the correct type of node at the end of
a relationship as specified by the model"
"""
import os
import neomodel
import datetime
import pytest
import random
try:
basestring
except NameError:
basestring = str
# Set up a very simple model for the tests
class PersonalRelationship(neomodel.StructuredRel):
"""
A very simple relationship between two basePersons that simply records
the date at which an acquaintance was established.
This relationship should be carried over to anything that inherits from
basePerson without any further effort.
"""
on_date = neomodel.DateProperty(default_now = True)
class BasePerson(neomodel.StructuredNode):
"""
Base class for defining some basic sort of an actor.
"""
name = neomodel.StringProperty(required = True, unique_index = True)
friends_with = neomodel.RelationshipTo("BasePerson", "FRIENDS_WITH",
model = PersonalRelationship)
class TechnicalPerson(BasePerson):
"""
A Technical person specialises BasePerson by adding their expertise.
"""
expertise = neomodel.StringProperty(required = True)
class PilotPerson(BasePerson):
"""
A pilot person specialises BasePerson by adding the type of airplane they
can operate.
"""
airplane = neomodel.StringProperty(required = True)
class BaseOtherPerson(neomodel.StructuredNode):
"""
An obviously "wrong" class of actor to befriend BasePersons with.
"""
car_color = neomodel.StringProperty(required = True)
class SomePerson(BaseOtherPerson):
"""
Concrete class that simply derives from BaseOtherPerson.
"""
pass
# Test cases
def test_automatic_object_resolution():
"""
Node objects at the end of relationships are instantiated to their
corresponding Python object.
"""
# Create a few entities
A = TechnicalPerson.get_or_create({"name":"Grumpy", "expertise":"Grumpiness"})[0]
B = TechnicalPerson.get_or_create({"name":"Happy", "expertise":"Unicorns"})[0]
C = TechnicalPerson.get_or_create({"name":"Sleepy", "expertise":"Pillows"})[0]
# Add connections
A.friends_with.connect(B)
B.friends_with.connect(C)
C.friends_with.connect(A)
# If A is friends with B, then A's friends_with objects should be
# TechnicalPerson (!NOT basePerson!)
assert type(A.friends_with[0]) is TechnicalPerson
A.delete()
B.delete()
C.delete()
def test_recursive_automatic_object_resolution():
"""
Node objects are instantiated to native Python objects, both at the top
level of returned results and in the case where they are returned within
lists.
"""
# Create a few entities
A = TechnicalPerson.get_or_create({"name":"Grumpier", "expertise":"Grumpiness"})[0]
B = TechnicalPerson.get_or_create({"name":"Happier", "expertise":"Grumpiness"})[0]
C = TechnicalPerson.get_or_create({"name":"Sleepier", "expertise":"Pillows"})[0]
D = TechnicalPerson.get_or_create({"name":"Sneezier", "expertise":"Pillows"})[0]
# Retrieve mixed results, both at the top level and nested
L, _ = neomodel.db.cypher_query("MATCH (a:TechnicalPerson) "
"WHERE a.expertise='Grumpiness' "
"WITH collect(a) as Alpha "
"MATCH (b:TechnicalPerson) "
"WHERE b.expertise='Pillows' "
"WITH Alpha, collect(b) as Beta "
"RETURN [Alpha, [Beta, [Beta, ['Banana', "
"Alpha]]]]", resolve_objects = True)
# Assert that a Node returned deep in a nested list structure is of the
# correct type
assert type(L[0][0][0][1][0][0][0][0]) is TechnicalPerson
# Assert that primitive data types remain primitive data types
assert issubclass(type(L[0][0][0][1][0][1][0][1][0][0]), basestring)
A.delete()
B.delete()
C.delete()
D.delete()
def test_validation_with_inheritance_from_db():
"""
Objects descending from the specified class of a relationship's end-node are
also perfectly valid to appear as end-node values too
"""
#Create a few entities
# Technical Persons
A = TechnicalPerson.get_or_create({"name":"Grumpy", "expertise":"Grumpiness"})[0]
B = TechnicalPerson.get_or_create({"name":"Happy", "expertise":"Unicorns"})[0]
C = TechnicalPerson.get_or_create({"name":"Sleepy", "expertise":"Pillows"})[0]
# Pilot Persons
D = PilotPerson.get_or_create({"name":"Porco Rosso", "airplane":"Savoia-Marchetti"})[0]
E = PilotPerson.get_or_create({"name":"Jack Dalton", "airplane":"Beechcraft Model 18"})[0]
# TechnicalPersons can befriend PilotPersons and vice-versa and that's fine
# TechnicalPersons befriend Technical Persons
A.friends_with.connect(B)
B.friends_with.connect(C)
C.friends_with.connect(A)
# Pilot Persons befriend Pilot Persons
D.friends_with.connect(E)
# Technical Persons befriend Pilot Persons
A.friends_with.connect(D)
E.friends_with.connect(C)
# This now means that friends_with of a TechnicalPerson can
# either be TechnicalPerson or Pilot Person (!NOT basePerson!)
assert (type(A.friends_with[0]) is TechnicalPerson) or (type(A.friends_with[0]) is PilotPerson)
assert (type(A.friends_with[1]) is TechnicalPerson) or (type(A.friends_with[1]) is PilotPerson)
assert type(D.friends_with[0]) is PilotPerson
A.delete()
B.delete()
C.delete()
D.delete()
E.delete()
def test_validation_enforcement_to_db():
"""
If a connection between wrong types is attempted, raise an exception
"""
#Create a few entities
# Technical Persons
A = TechnicalPerson.get_or_create({"name":"Grumpy", "expertise":"Grumpiness"})[0]
B = TechnicalPerson.get_or_create({"name":"Happy", "expertise":"Unicorns"})[0]
C = TechnicalPerson.get_or_create({"name":"Sleepy", "expertise":"Pillows"})[0]
# Pilot Persons
D = PilotPerson.get_or_create({"name":"Porco Rosso", "airplane":"Savoia-Marchetti"})[0]
E = PilotPerson.get_or_create({"name":"Jack Dalton", "airplane":"Beechcraft Model 18"})[0]
#Some Person
F = SomePerson(car_color = "Blue").save()
# TechnicalPersons can befriend PilotPersons and vice-versa and that's fine
A.friends_with.connect(B)
B.friends_with.connect(C)
C.friends_with.connect(A)
D.friends_with.connect(E)
A.friends_with.connect(D)
E.friends_with.connect(C)
# Trying to befriend a Technical Person with Some Person should raise an
# exception
with pytest.raises(ValueError):
A.friends_with.connect(F)
A.delete()
B.delete()
C.delete()
D.delete()
E.delete()
F.delete()
def test_failed_object_resolution():
"""
A Neo4j driver node FROM the database contains labels that are unaware to
neomodel's Database class. This condition raises ClassDefinitionNotFound
exception.
"""
class RandomPerson(BasePerson):
randomness = neomodel.FloatProperty(default = random.random)
# A Technical Person...
A = TechnicalPerson.get_or_create({"name":"Grumpy", "expertise":"Grumpiness"})[0]
# A Random Person...
B = RandomPerson.get_or_create({"name":"Mad Hatter"})[0]
A.friends_with.connect(B)
# Simulate the condition where the definition of class RandomPerson is not
# known yet.
del neomodel.db._NODE_CLASS_REGISTRY[frozenset(["RandomPerson","BasePerson"])]
# Now try to instantiate a RandomPerson
A = TechnicalPerson.get_or_create({"name":"Grumpy", "expertise":"Grumpiness"})[0]
with pytest.raises(neomodel.exceptions.ModelDefinitionMismatch):
for some_friend in A.friends_with:
print(some_friend.name)
A.delete()
B.delete()
def test_node_label_mismatch():
"""
A Neo4j driver node FROM the database contains a superset of the known
labels.
"""
class SuperTechnicalPerson(TechnicalPerson):
superness = neomodel.FloatProperty(default=1.0)
class UltraTechnicalPerson(SuperTechnicalPerson):
ultraness = neomodel.FloatProperty(default=3.1415928)
# Create a TechnicalPerson...
A = TechnicalPerson.get_or_create({"name":"Grumpy", "expertise":"Grumpiness"})[0]
# ...that is connected to an UltraTechnicalPerson
F = UltraTechnicalPerson(name = "Chewbaka", expertise="Aarrr wgh ggwaaah").save()
A.friends_with.connect(F)
# Forget about the UltraTechnicalPerson
del neomodel.db._NODE_CLASS_REGISTRY[frozenset(["UltraTechnicalPerson",
"SuperTechnicalPerson",
"TechnicalPerson",
"BasePerson"])]
# Recall a TechnicalPerson and enumerate its friends.
# One of them is UltraTechnicalPerson which would be returned as a valid
# node to a friends_with query but is currently unknown to the node class registry.
A = TechnicalPerson.get_or_create({"name":"Grumpy", "expertise":"Grumpiness"})[0]
with pytest.raises(neomodel.exceptions.ModelDefinitionMismatch):
for some_friend in A.friends_with:
print(some_friend.name)
def test_attempted_class_redefinition():
"""
A neomodel.StructuredNode class is attempted to be redefined.
"""
def redefine_class_locally():
# Since this test has already set up a class hierarchy in its global scope, we will try to redefine
# SomePerson here.
# The internal structure of the SomePerson entity does not matter at all here.
class SomePerson(BaseOtherPerson):
uid = neomodel.UniqueIdProperty()
with pytest.raises(neomodel.exceptions.ClassAlreadyDefined):
redefine_class_locally()
|
robinedwards/neomodel
|
test/test_issue283.py
|
Python
|
mit
| 10,529
|
"""distutils.command.build_py
Implements the Distutils 'build_py' command."""
# created 1999/03/08, Greg Ward
__revision__ = "$Id: build_py.py,v 1.34 2001/12/06 20:59:17 fdrake Exp $"
import sys, string, os
from types import *
from glob import glob
from distutils.core import Command
from distutils.errors import *
class build_py (Command):
description = "\"build\" pure Python modules (copy to build directory)"
user_options = [
('build-lib=', 'd', "directory to \"build\" (copy) to"),
('compile', 'c', "compile .py to .pyc"),
('no-compile', None, "don't compile .py files [default]"),
('optimize=', 'O',
"also compile with optimization: -O1 for \"python -O\", "
"-O2 for \"python -OO\", and -O0 to disable [default: -O0]"),
('force', 'f', "forcibly build everything (ignore file timestamps)"),
]
boolean_options = ['compile', 'force']
negative_opt = {'no-compile' : 'compile'}
def initialize_options (self):
self.build_lib = None
self.py_modules = None
self.package = None
self.package_dir = None
self.compile = 0
self.optimize = 0
self.force = None
def finalize_options (self):
self.set_undefined_options('build',
('build_lib', 'build_lib'),
('force', 'force'))
# Get the distribution options that are aliases for build_py
# options -- list of packages and list of modules.
self.packages = self.distribution.packages
self.py_modules = self.distribution.py_modules
self.package_dir = self.distribution.package_dir
# Ick, copied straight from install_lib.py (fancy_getopt needs a
# type system! Hell, *everything* needs a type system!!!)
if type(self.optimize) is not IntType:
try:
self.optimize = int(self.optimize)
assert 0 <= self.optimize <= 2
except (ValueError, AssertionError):
raise DistutilsOptionError, "optimize must be 0, 1, or 2"
def run (self):
# XXX copy_file by default preserves atime and mtime. IMHO this is
# the right thing to do, but perhaps it should be an option -- in
# particular, a site administrator might want installed files to
# reflect the time of installation rather than the last
# modification time before the installed release.
# XXX copy_file by default preserves mode, which appears to be the
# wrong thing to do: if a file is read-only in the working
# directory, we want it to be installed read/write so that the next
# installation of the same module distribution can overwrite it
# without problems. (This might be a Unix-specific issue.) Thus
# we turn off 'preserve_mode' when copying to the build directory,
# since the build directory is supposed to be exactly what the
# installation will look like (ie. we preserve mode when
# installing).
# Two options control which modules will be installed: 'packages'
# and 'py_modules'. The former lets us work with whole packages, not
# specifying individual modules at all; the latter is for
# specifying modules one-at-a-time. Currently they are mutually
# exclusive: you can define one or the other (or neither), but not
# both. It remains to be seen how limiting this is.
# Dispose of the two "unusual" cases first: no pure Python modules
# at all (no problem, just return silently), and over-specified
# 'packages' and 'py_modules' options.
if not self.py_modules and not self.packages:
return
if self.py_modules and self.packages:
raise DistutilsOptionError, \
"build_py: supplying both 'packages' and 'py_modules' " + \
"options is not allowed"
# Now we're down to two cases: 'py_modules' only and 'packages' only.
if self.py_modules:
self.build_modules()
else:
self.build_packages()
self.byte_compile(self.get_outputs(include_bytecode=0))
# run ()
def get_package_dir (self, package):
"""Return the directory, relative to the top of the source
distribution, where package 'package' should be found
(at least according to the 'package_dir' option, if any)."""
path = string.split(package, '.')
if not self.package_dir:
if path:
return apply(os.path.join, path)
else:
return ''
else:
tail = []
while path:
try:
pdir = self.package_dir[string.join(path, '.')]
except KeyError:
tail.insert(0, path[-1])
del path[-1]
else:
tail.insert(0, pdir)
return apply(os.path.join, tail)
else:
# Oops, got all the way through 'path' without finding a
# match in package_dir. If package_dir defines a directory
# for the root (nameless) package, then fallback on it;
# otherwise, we might as well have not consulted
# package_dir at all, as we just use the directory implied
# by 'tail' (which should be the same as the original value
# of 'path' at this point).
pdir = self.package_dir.get('')
if pdir is not None:
tail.insert(0, pdir)
if tail:
return apply(os.path.join, tail)
else:
return ''
# get_package_dir ()
def check_package (self, package, package_dir):
# Empty dir name means current directory, which we can probably
# assume exists. Also, os.path.exists and isdir don't know about
# my "empty string means current dir" convention, so we have to
# circumvent them.
if package_dir != "":
if not os.path.exists(package_dir):
raise DistutilsFileError, \
"package directory '%s' does not exist" % package_dir
if not os.path.isdir(package_dir):
raise DistutilsFileError, \
("supposed package directory '%s' exists, " +
"but is not a directory") % package_dir
# Require __init__.py for all but the "root package"
if package:
init_py = os.path.join(package_dir, "__init__.py")
if os.path.isfile(init_py):
return init_py
else:
self.warn(("package init file '%s' not found " +
"(or not a regular file)") % init_py)
# Either not in a package at all (__init__.py not expected), or
# __init__.py doesn't exist -- so don't return the filename.
return
# check_package ()
def check_module (self, module, module_file):
if not os.path.isfile(module_file):
self.warn("file %s (for module %s) not found" %
(module_file, module))
return 0
else:
return 1
# check_module ()
def find_package_modules (self, package, package_dir):
self.check_package(package, package_dir)
module_files = glob(os.path.join(package_dir, "*.py"))
modules = []
setup_script = os.path.abspath(self.distribution.script_name)
for f in module_files:
abs_f = os.path.abspath(f)
if abs_f != setup_script:
module = os.path.splitext(os.path.basename(f))[0]
modules.append((package, module, f))
else:
self.debug_print("excluding %s" % setup_script)
return modules
def find_modules (self):
"""Finds individually-specified Python modules, ie. those listed by
module name in 'self.py_modules'. Returns a list of tuples (package,
module_base, filename): 'package' is a tuple of the path through
package-space to the module; 'module_base' is the bare (no
packages, no dots) module name, and 'filename' is the path to the
".py" file (relative to the distribution root) that implements the
module.
"""
# Map package names to tuples of useful info about the package:
# (package_dir, checked)
# package_dir - the directory where we'll find source files for
# this package
# checked - true if we have checked that the package directory
# is valid (exists, contains __init__.py, ... ?)
packages = {}
# List of (package, module, filename) tuples to return
modules = []
# We treat modules-in-packages almost the same as toplevel modules,
# just the "package" for a toplevel is empty (either an empty
# string or empty list, depending on context). Differences:
# - don't check for __init__.py in directory for empty package
for module in self.py_modules:
path = string.split(module, '.')
package = string.join(path[0:-1], '.')
module_base = path[-1]
try:
(package_dir, checked) = packages[package]
except KeyError:
package_dir = self.get_package_dir(package)
checked = 0
if not checked:
init_py = self.check_package(package, package_dir)
packages[package] = (package_dir, 1)
if init_py:
modules.append((package, "__init__", init_py))
# XXX perhaps we should also check for just .pyc files
# (so greedy closed-source bastards can distribute Python
# modules too)
module_file = os.path.join(package_dir, module_base + ".py")
if not self.check_module(module, module_file):
continue
modules.append((package, module_base, module_file))
return modules
# find_modules ()
def find_all_modules (self):
"""Compute the list of all modules that will be built, whether
they are specified one-module-at-a-time ('self.py_modules') or
by whole packages ('self.packages'). Return a list of tuples
(package, module, module_file), just like 'find_modules()' and
'find_package_modules()' do."""
if self.py_modules:
modules = self.find_modules()
else:
modules = []
for package in self.packages:
package_dir = self.get_package_dir(package)
m = self.find_package_modules(package, package_dir)
modules.extend(m)
return modules
# find_all_modules ()
def get_source_files (self):
modules = self.find_all_modules()
filenames = []
for module in modules:
filenames.append(module[-1])
return filenames
def get_module_outfile (self, build_dir, package, module):
outfile_path = [build_dir] + list(package) + [module + ".py"]
return apply(os.path.join, outfile_path)
def get_outputs (self, include_bytecode=1):
modules = self.find_all_modules()
outputs = []
for (package, module, module_file) in modules:
package = string.split(package, '.')
filename = self.get_module_outfile(self.build_lib, package, module)
outputs.append(filename)
if include_bytecode:
if self.compile:
outputs.append(filename + "c")
if self.optimize > 0:
outputs.append(filename + "o")
return outputs
def build_module (self, module, module_file, package):
if type(package) is StringType:
package = string.split(package, '.')
elif type(package) not in (ListType, TupleType):
raise TypeError, \
"'package' must be a string (dot-separated), list, or tuple"
# Now put the module source file into the "build" area -- this is
# easy, we just copy it somewhere under self.build_lib (the build
# directory for Python source).
outfile = self.get_module_outfile(self.build_lib, package, module)
dir = os.path.dirname(outfile)
self.mkpath(dir)
return self.copy_file(module_file, outfile, preserve_mode=0)
def build_modules (self):
modules = self.find_modules()
for (package, module, module_file) in modules:
# Now "build" the module -- ie. copy the source file to
# self.build_lib (the build directory for Python source).
# (Actually, it gets copied to the directory for this package
# under self.build_lib.)
self.build_module(module, module_file, package)
# build_modules ()
def build_packages (self):
for package in self.packages:
# Get list of (package, module, module_file) tuples based on
# scanning the package directory. 'package' is only included
# in the tuple so that 'find_modules()' and
# 'find_package_tuples()' have a consistent interface; it's
# ignored here (apart from a sanity check). Also, 'module' is
# the *unqualified* module name (ie. no dots, no package -- we
# already know its package!), and 'module_file' is the path to
# the .py file, relative to the current directory
# (ie. including 'package_dir').
package_dir = self.get_package_dir(package)
modules = self.find_package_modules(package, package_dir)
# Now loop over the modules we found, "building" each one (just
# copy it to self.build_lib).
for (package_, module, module_file) in modules:
assert package == package_
self.build_module(module, module_file, package)
# build_packages ()
def byte_compile (self, files):
from distutils.util import byte_compile
prefix = self.build_lib
if prefix[-1] != os.sep:
prefix = prefix + os.sep
# XXX this code is essentially the same as the 'byte_compile()
# method of the "install_lib" command, except for the determination
# of the 'prefix' string. Hmmm.
if self.compile:
byte_compile(files, optimize=0,
force=self.force,
prefix=prefix,
verbose=self.verbose, dry_run=self.dry_run)
if self.optimize > 0:
byte_compile(files, optimize=self.optimize,
force=self.force,
prefix=prefix,
verbose=self.verbose, dry_run=self.dry_run)
# class build_py
|
MalloyPower/parsing-python
|
front-end/testsuite-python-lib/Python-2.2/Lib/distutils/command/build_py.py
|
Python
|
mit
| 15,090
|
# Python - 3.6.0
solution = lambda M1, M2, m1, m2, V, t: ((m1 / M1) + (m2 / M2)) * 0.082 * (t + 273.15) / V
|
RevansChen/online-judge
|
Codewars/8kyu/total-pressure-calculation/Python/solution1.py
|
Python
|
mit
| 109
|
from distutils.core import setup
setup(
name="tornado-stub-client",
version="0.2",
author="Danny Cosson",
author_email="support@venmo.com",
license="MIT",
description="Stubs out tornado AsyncHTTPClient.fetch with a nice interface, for testing code that relies on async code",
long_description=open("README.md").read(),
url="https://github.com/venmo/tornado-stub-client",
packages=["tornado_stub_client"],
requires=["tornado"],
)
|
dcosson/tornado-stub-client
|
setup.py
|
Python
|
mit
| 469
|
from django.shortcuts import render_to_response, get_object_or_404
from django.http import HttpResponseRedirect, Http404
from django.template import RequestContext
from django.core.urlresolvers import reverse
from django.utils.translation import ugettext_lazy as _
from django.contrib.auth.models import User
from django.contrib.auth.decorators import login_required
from django.views.generic import date_based
from django.conf import settings
from django.db.models import Q
from django.conf import settings
from swaps.models import Offer, Swap
from swaps.forms import OfferForm, ProposeSwapForm, ProposingOfferForm
if "notification" in settings.INSTALLED_APPS:
from notification import models as notification
else:
notification = None
try:
from threadedcomments.models import ThreadedComment
forums = True
except ImportError:
forums = False
@login_required
def offers(request, username=None):
offers = Offer.objects.filter(state=1).order_by("-offered_time")
return render_to_response("swaps/offers.html", {"offers": offers}, context_instance=RequestContext(request))
@login_required
def offer(request, offer_id):
offer = get_object_or_404(Offer, id=offer_id)
#deletable = offer.is_deletable()
return render_to_response("swaps/offer.html", {
"offer": offer,
#"deletable": deletable,
}, context_instance=RequestContext(request))
@login_required
def your_offers(request):
user = request.user
offers = Offer.objects.filter(offerer=user).order_by("-offered_time")
return render_to_response("swaps/your_offers.html", {"offers": offers}, context_instance=RequestContext(request))
@login_required
def swap(request, swap_id):
swap = get_object_or_404(Swap, id=swap_id)
return render_to_response("swaps/swap.html", {
"swap": swap,
}, context_instance=RequestContext(request))
@login_required
def proposed_by_you(request):
swaps = Swap.objects.filter(proposing_offer__offerer=request.user, state=1).order_by("-proposed_time")
return render_to_response("swaps/proposed_by_you.html", {"swaps": swaps}, context_instance=RequestContext(request))
@login_required
def proposed_to_you(request):
swaps = Swap.objects.filter(responding_offer__offerer=request.user, state=1).order_by("-proposed_time")
return render_to_response("swaps/proposed_to_you.html", {"swaps": swaps}, context_instance=RequestContext(request))
@login_required
def accepted_swaps(request):
swaps = Swap.objects.filter(
Q(state=2, proposing_offer__offerer=request.user) |
Q(state=2, responding_offer__offerer=request.user)).order_by("-accepted_time")
return render_to_response("swaps/accepted.html", {"swaps": swaps}, context_instance=RequestContext(request))
@login_required
def dead_swaps(request):
swaps = Swap.objects.filter(
Q(state__gt=3, proposing_offer__offerer=request.user) |
Q(state__gt=3, responding_offer__offerer=request.user)).order_by("-killed_time")
return render_to_response("swaps/dead.html", {"swaps": swaps}, context_instance=RequestContext(request))
@login_required
def new(request):
def is_valid(formsets):
valid = True
for name, formset in formsets:
valid = valid and formset.is_valid()
return valid
formsets = []
if request.method == "POST":
if request.POST["action"] == "create":
for name, formset_class in OfferForm.inlines:
formsets+=[(name, formset_class(request.POST, request.FILES)),]
offer_form = OfferForm(request.POST, request.FILES)
formsets_valid = is_valid(formsets)
offer_form.inlines = formsets
if offer_form.is_valid() and formsets_valid:
offer = offer_form.save(commit=False)
offer.offerer = request.user
offer.save()
for name, formset_class in OfferForm.inlines:
f = formset_class(request.POST, request.FILES, instance=offer)
if f.is_valid():
f.save()
request.user.message_set.create(message=_("Successfully saved offer '%s'") % offer.short_description)
#if notification:
# if friends: # @@@ might be worth having a shortcut for sending to all friends
# notification.send((x['friend'] for x in Friendship.objects.friends_for_user(offer.offerer)), "offer_friend_post", {"post": blog})
return HttpResponseRedirect(reverse("offer_list_yours"))
else:
offer_form = OfferForm()
else:
offer_form = OfferForm()
return render_to_response("swaps/new_offer.html", {
"offer_form": offer_form
}, context_instance=RequestContext(request))
@login_required
def edit_offer(request, offer_id):
def is_valid(formsets):
valid = True
for name, formset in formsets:
valid = valid and formset.is_valid()
return valid
offer = get_object_or_404(Offer, id=offer_id)
if offer.offerer != request.user:
request.user.message_set.create(message="You cannot edit offers that are not yours")
return HttpResponseRedirect(reverse("offer_list_yours"))
return_to = request.GET['returnto']
formsets = []
if request.method == "POST":
if request.POST["action"] == "update":
offer_form = OfferForm(request.POST, instance=offer)
for name, formset_class in OfferForm.inlines:
formsets+=[(name, formset_class(request.POST, request.FILES, instance=offer)),]
formsets_valid = is_valid(formsets)
offer_form.inlines = formsets
if offer_form.is_valid() and formsets_valid:
offer = offer_form.save(commit=False)
offer.save()
for name, formset_class in OfferForm.inlines:
f = formset_class(request.POST, request.FILES, instance=offer)
if f.is_valid():
f.save()
if notification:
for swap in offer.proposed_swaps.filter(state=1):
notification.send([swap.responding_offer.offerer,], "swaps_proposing_offer_changed",
{"creator": request.user,
"swap": swap,
"proposing_offer": swap.proposing_offer,
"responding_offer": swap.responding_offer})
for swap in offer.responding_swaps.filter(state=1):
notification.send([swap.proposing_offer.offerer,], "swaps_responding_offer_changed",
{"creator": request.user,
"swap": swap,
"proposing_offer": swap.proposing_offer,
"responding_offer": swap.responding_offer})
request.user.message_set.create(message=_("Successfully updated offer '%s'") % offer.short_description)
return HttpResponseRedirect(reverse(return_to))
else:
offer_form = OfferForm(instance=offer)
for name, formset_class in OfferForm.inlines:
formsets+=[(name, formset_class(None, None, instance=offer)),]
else:
offer_form = OfferForm(instance=offer)
for name, formset_class in OfferForm.inlines:
formsets+=[(name, formset_class(None, None, instance=offer)),]
offer_form.inlines = formsets
return render_to_response("swaps/edit_offer.html", {
"offer_form": offer_form,
"offer": offer,
}, context_instance=RequestContext(request))
@login_required
def delete_offer(request, offer_id):
offer = get_object_or_404(Offer, id=offer_id)
if offer.offerer != request.user:
request.user.message_set.create(message="You cannot delete offers that are not yours")
return HttpResponseRedirect(reverse("offer_list_yours"))
if request.method == "POST":
offer.delete()
return HttpResponseRedirect(reverse("offer_list_yours"))
@login_required
def cancel_offer(request, offer_id):
offer = get_object_or_404(Offer, id=offer_id)
if offer.offerer != request.user:
request.user.message_set.create(message="You cannot cancel offers that are not yours")
return HttpResponseRedirect(reverse("offer_list_yours"))
if request.method == "POST":
offer.cancel()
return HttpResponseRedirect(reverse("offer_list_yours"))
@login_required
def propose_swap(request, offer_id):
offer = get_object_or_404(Offer, id=offer_id)
if request.method == "POST":
swap_form = ProposeSwapForm(request.POST)
offer_form = ProposingOfferForm(request.POST)
swap = None
if swap_form.is_valid():
swap = swap_form.save(commit=False)
swap.responding_offer = offer
swap.save()
if offer_form.is_valid():
proposing_offer = offer_form.save(commit=False)
proposing_offer.offerer = request.user
proposing_offer.save()
swap = Swap(
proposing_offer=proposing_offer,
responding_offer=offer)
swap.save()
if swap:
if notification:
notification.send([offer.offerer,], "swaps_proposal",
{"creator": request.user,
"swap": swap,
"proposing_offer": swap.proposing_offer,
"responding_offer": swap.responding_offer})
return HttpResponseRedirect(reverse("proposed_by_you"))
else:
swap_form = ProposeSwapForm()
swap_form.fields['proposing_offer'].queryset = Offer.objects.filter(offerer=request.user, state=1)
offer_form = ProposingOfferForm()
return render_to_response("swaps/propose_swap.html", {
"offer": offer,
"swap_form": swap_form,
"offer_form": offer_form,
}, context_instance=RequestContext(request))
@login_required
def accept_swap(request, swap_id):
swap = get_object_or_404(Swap, id=swap_id)
swap.accept()
swap.save()
if notification:
notification.send([swap.proposing_offer.offerer,], "swaps_acceptance",
{"creator": request.user,
"swap": swap,
"proposing_offer": swap.proposing_offer,
"responding_offer": swap.responding_offer})
return HttpResponseRedirect(reverse("accepted_swaps"))
@login_required
def reject_swap(request, swap_id):
swap = get_object_or_404(Swap, id=swap_id)
swap.reject()
swap.save()
if notification:
notification.send([swap.proposing_offer.offerer,], "swaps_rejection",
{"creator": request.user,
"swap": swap,
"proposing_offer": swap.proposing_offer,
"responding_offer": swap.responding_offer})
return HttpResponseRedirect(reverse("dead_swaps"))
@login_required
def cancel_swap(request, swap_id):
swap = get_object_or_404(Swap, id=swap_id)
swap.cancel()
swap.save()
if notification:
notification.send([swap.responding_offer.offerer,], "swaps_cancellation",
{"creator": request.user,
"swap": swap,
"proposing_offer": swap.proposing_offer,
"responding_offer": swap.responding_offer})
return HttpResponseRedirect(reverse("dead_swaps"))
|
andriibekker/django-swaps
|
swaps/views.py
|
Python
|
mit
| 11,444
|
from os import remove
from flask import jsonify, request
from sqlalchemy import create_engine
from sqlalchemy.schema import MetaData
from sqlite3 import dbapi2 as sqlite
from cv2 import imread, COLOR_RGB2GRAY, cvtColor
from cv2.face import LBPHFaceRecognizer_create
from faces_recognizer import Recognizer
engine = create_engine('sqlite+pysqlite:///../sqlite/cafeDB.db', module=sqlite)
meta = MetaData()
meta.reflect(bind=engine)
meta.bind = engine
users_table = meta.tables['User']
recognizer = Recognizer(LBPHFaceRecognizer_create)
recognizer.set_recognizer_xml('faces.xml')
def add_route(app):
''' Add routes to a flask app Class. See API swagger doc'''
@app.route('/users/<id>', methods=['GET'])
def get_user(id):
try:
id = int(id)
except ValueError:
return "Invalid id", 400
print(id)
result = users_table.select().where(users_table.c.id == id).execute().first()
if result:
return jsonify(dict(result.items())), 200
else:
return "User not found", 404
@app.route('/users', methods=["POST"])
def new_user():
body = request.get_json()
try:
username, intensity, volume = [body[k] for k in ("username", "intensity", "volume")]
except (KeyError, TypeError):
return "Incorrect body", 400
# Because sqlite engine cannot return created tuple, it's not possible to get the id after creation.
# So we retrieve it.
id_max = engine.execute("SELECT seq FROM SQLITE_SEQUENCE WHERE name='User'").first()
new_id = id_max[0] + 1
stmt = users_table.insert().values(id=new_id, username=username, intensity=intensity, volume=volume)
stmt.execute()
# Get the create users.
result = users_table.select().where(users_table.c.id == new_id).execute().first()
return jsonify(dict(result.items()))
@app.route('/users/<id>', methods=["PUT"])
def modify_user(id):
try:
id = int(id)
except ValueError:
return "Invalid id", 400
find = users_table.select().where(users_table.c.id == id).execute().first()
if find:
body = request.get_json()
username, intensity, volume = [body.get(k) for k in ("username", "intensity", "volume")]
query = users_table.update().where(users_table.c.id == id)
if username:
query = query.values(username=username)
if intensity:
query = query.values(intensity=intensity)
if volume:
query = query.values(volume=volume)
query.execute()
result = users_table.select().where(users_table.c.id == id).execute().first()
return jsonify(dict(result.items()))
else:
return "User not found", 404
@app.route('/users/<id>', methods=["DELETE"])
def remove_user(id):
try:
id = int(id)
except ValueError:
return "Invalid id", 400
find = users_table.select().where(users_table.c.id == id).execute().first()
if find:
users_table.delete().where(users_table.c.id == id).execute()
return "User succesfuly destroyed"
else:
return "User not found", 404
@app.route('/recognition', methods=["POST"])
def recognize_faceimage():
if 'image' not in request.files:
return "no image", 401
file = request.files['image']
if file.filename == '':
return "Bad Request", 400
file.save(file.filename)
id, prob = recognizer.predict(cvtColor(imread(file.filename), COLOR_RGB2GRAY))
result = users_table.select().where(users_table.c.id == id).execute().first()
remove(file.filename)
if result:
return jsonify(dict(result.items()))
else:
return "User unsucessfully recognized", 400
|
Koisell/SmartCoffeeMachine
|
python/recognitionService_api/flask_app.py
|
Python
|
mit
| 3,964
|
# -*- coding: utf-8 -*-
import unittest
from .context import pso
class PSOTestSuite(unittest.TestCase):
def test_absolute_truth_and_meaning(self):
assert False
if __name__ == '__main__':
unittest.main()
|
JoaoGFarias/comparing_swarm_intelligence
|
tests/pso.py
|
Python
|
mit
| 223
|
"""
********
UnitTest
BayesNet
********
Method Checks that
assertEqual(a, b) a == b
assertNotEqual(a, b) a != b
assertTrue(x) bool(x) is True
assertFalse(x) bool(x) is False
assertIs(a, b) a is b
assertIsNot(a, b) a is not b
assertIsNone(x) x is None
assertIsNotNone(x) x is not None
assertIn(a, b) a in b
assertNotIn(a, b) a not in b
assertIsInstance(a, b) isinstance(a, b)
assertNotIsInstance(a, b) not isinstance(a, b)
assertAlmostEqual(a, b) round(a-b, 7) == 0
assertNotAlmostEqual(a, b) round(a-b, 7) != 0
assertGreater(a, b) a > b
assertGreaterEqual(a, b) a >= b
assertLess(a, b) a < b
assertLessEqual(a, b) a <= b
assertListEqual(a, b) lists
assertTupleEqual(a, b) tuples
assertSetEqual(a, b) sets or frozensets
assertDictEqual(a, b) dicts
"""
__author__ = """Nicholas Cullen <ncullen.th@dartmouth.edu>"""
import unittest
from pyBN.classes.bayesnet import BayesNet
from pyBN.readwrite.read import read_bn
import os
from os.path import dirname
class BayesNetTestCase(unittest.TestCase):
def setUp(self):
self.bn = BayesNet()
self.dpath = os.path.join(dirname(dirname(dirname(dirname(__file__)))),'data')
self.bn_bif = read_bn(os.path.join(self.dpath,'cancer.bif'))
self.bn_bn = read_bn(os.path.join(self.dpath,'cmu.bn'))
def tearDown(self):
pass
def test_isinstance(self):
self.assertIsInstance(self.bn,BayesNet)
def test_V_bif(self):
self.assertListEqual(self.bn_bif.V,
['Smoker', 'Pollution', 'Cancer', 'Xray', 'Dyspnoea'])
def test_E_bif(self):
self.assertDictEqual(self.bn_bif.E,
{'Cancer': ['Xray', 'Dyspnoea'],
'Dyspnoea': [],
'Pollution': ['Cancer'],
'Smoker': ['Cancer'],
'Xray': []})
def test_F_bif(self):
self.assertDictEqual(self.bn_bif.F,
{'Cancer': {'cpt': [0.03, 0.97, 0.05, 0.95, 0.001, 0.999, 0.02, 0.98],
'parents': ['Pollution', 'Smoker'],
'values': ['True', 'False']},
'Dyspnoea': {'cpt': [0.65, 0.35, 0.3, 0.7],
'parents': ['Cancer'],
'values': ['True', 'False']},
'Pollution': {'cpt': [0.9, 0.1], 'parents': [], 'values': ['low', 'high']},
'Smoker': {'cpt': [0.3, 0.7], 'parents': [], 'values': ['True', 'False']},
'Xray': {'cpt': [0.9, 0.1, 0.2, 0.8],
'parents': ['Cancer'],
'values': ['positive', 'negative']}})
def test_V_bn(self):
self.assertListEqual(self.bn_bn.V,
['Burglary', 'Earthquake', 'Alarm', 'JohnCalls', 'MaryCalls'])
def test_E_bn(self):
self.assertDictEqual(self.bn_bn.E,
{'Alarm': ['JohnCalls', 'MaryCalls'],
'Burglary': ['Alarm'],
'Earthquake': ['Alarm'],
'JohnCalls': [],
'MaryCalls': []})
def test_F_bn(self):
self.assertDictEqual(self.bn_bn.F,
{'Alarm': {'cpt': [0.999, 0.001, 0.71, 0.29, 0.06, 0.94, 0.05, 0.95],
'parents': ['Earthquake', 'Burglary'],
'values': ['No', 'Yes']},
'Burglary': {'cpt': [0.999, 0.001], 'parents': [], 'values': ['No', 'Yes']},
'Earthquake': {'cpt': [0.998, 0.002], 'parents': [], 'values': ['No', 'Yes']},
'JohnCalls': {'cpt': [0.95, 0.05, 0.1, 0.9],
'parents': ['Alarm'],
'values': ['No', 'Yes']},
'MaryCalls': {'cpt': [0.99, 0.01, 0.3, 0.7],
'parents': ['Alarm'],
'values': ['No', 'Yes']}})
def test_nodes(self):
n = list(self.bn_bn.nodes())
self.assertListEqual(n,
['Burglary', 'Earthquake', 'Alarm', 'JohnCalls', 'MaryCalls'])
def test_cpt(self):
cpt = list(self.bn_bn.cpt('Alarm'))
self.assertListEqual(cpt,
[0.999, 0.001, 0.71, 0.29, 0.06, 0.94, 0.05, 0.95])
def test_card(self):
self.assertEqual(self.bn_bn.card('Alarm'),2)
def test_scope(self):
self.assertListEqual(self.bn_bn.scope('Alarm'),
['Alarm', 'Earthquake', 'Burglary'])
def test_parents(self):
self.assertListEqual(self.bn_bn.parents('Alarm'),
['Earthquake','Burglary'])
def test_values(self):
self.assertListEqual(self.bn_bn.values('Alarm'),['No','Yes'])
def test_values_idx(self):
self.assertEqual(self.bn_bn.values('Alarm')[1],'Yes')
if __name__ == '__main__':
unittest.main(exit=False)
|
ncullen93/pyBN
|
pyBN/classes/_tests/test_bayesnet.py
|
Python
|
mit
| 4,002
|
from __future__ import unicode_literals
from django.utils.translation import ugettext_lazy as _
from rest_framework.compat import unicode_to_repr
from rest_framework.exceptions import ValidationError
from rest_framework.utils.representation import smart_repr
from csp import settings
class EqualityValidator(object):
message = _('The fields {field_names} must be equal.')
missing_message = _('This field is required.')
def __init__(self, fields, message=None):
self.fields = fields
self.serializer_field = None
self.message = message or self.message
def set_context(self, serializer):
"""
This hook is called by the serializer instance,
prior to the validation call being made.
"""
self.instance = getattr(serializer, 'instance', None)
self.initial_data = getattr(serializer,'initial_data', None)
self.validate_non_fields = getattr(serializer,'validate_non_fields', False)
def __call__(self,*args, **kwargs):
if self.validate_non_fields:
if self.fields[0] not in self.initial_data or self.fields[1] not in self.initial_data:
raise ValidationError("Both fields are required.")
if self.initial_data.get(self.fields[0],'Password1') != self.initial_data.get(self.fields[1],'Password2'):
field_names = ', '.join(self.fields)
raise ValidationError(self.message.format(field_names=field_names))
class LengthValidator(object):
message = _('Field {field_name} must be at least {length} characters long.')
missing_message = _('Field {field_name} is required.')
def __init__(self, field, length, message=None):
self.field = field
self.length = length
self.serializer_field = None
self.message = message or self.message
def set_context(self, serializer):
self.initial_data = getattr(serializer,'initial_data', None)
self.validate_non_fields = getattr(serializer,'validate_non_fields', False)
def __call__(self, *args, **kwargs):
if self.validate_non_fields:
if self.field not in self.initial_data:
raise ValidationError(self.missing_message.format(field_name=self.field))
if len(self.initial_data[self.field]) < self.length:
raise ValidationError(self.message.format(field_name=self.field, length=self.length))
class RegistrationAllowedValidator(object):
message = _('Currently registrations are not allowed.')
def __call__(self, *args, **kwargs):
if not args[0]['admin_override'] and not settings.REGISTRATION_ALLOWED:
raise ValidationError(self.message)
|
shriyanka/daemo-forum
|
crowdsourcing/validators/utils.py
|
Python
|
mit
| 2,693
|
#!/usr/bin/env python
# -*- coding: utf-8 -*-
"""An example CRUD CLI app."""
from __future__ import (
absolute_import,
print_function,
unicode_literals
)
import argparse
import json
import logging
import sys
TAGS = (
'red',
'orange',
'yellow',
'green',
'blue',
'violet',
'black',
'white',
'gray',
)
class CLI(object):
"""A really basic version of the famous Python Click library."""
log = logging.getLogger(__name__)
common_args = argparse.ArgumentParser(add_help=False)
log_group = common_args.add_mutually_exclusive_group()
log_group.add_argument(
'-v',
'--verbose',
dest='verbosity',
default=[logging.INFO],
action='append_const',
const=-10,
help='more verbose',
)
log_group.add_argument(
'-q',
'--quiet',
dest='verbosity',
action='append_const',
const=10,
help='less verbose',
)
def __init__(self):
self.parser = argparse.ArgumentParser()
self.subparsers = self.parser.add_subparsers(help='subcommands',
dest='command')
self.subparsers.required = True
def command(self, name, *args, **kwargs):
"""Register a function to the command-line interface."""
def wrapper(f):
f.parser = self.subparsers.add_parser(
name, *args, description=f.__doc__,
parents=[self.common_args], **kwargs)
if getattr(f, 'cli_args', None) is not None:
for fargs, fkwargs in f.cli_args:
f.parser.add_argument(*fargs, **fkwargs)
f.parser.set_defaults(action=f)
return f
return wrapper
def option(self, *args, **kwargs):
"""Register CLI arguments for function.
Accepts the same arguments as ArgumentParser().add_argument(...)
"""
def wrapper(f):
if getattr(f, 'cli_args', None) is None:
f.cli_args = []
f.cli_args.append((args, kwargs))
return f
return wrapper
def run(self):
"""Parse arguments and run the default action."""
args = self.parser.parse_args()
# init logging
log_level = max(logging.DEBUG, min(logging.CRITICAL, sum(args.verbosity)))
debug_on = log_level <= logging.DEBUG
logging.basicConfig(level=log_level)
kwargs = dict(vars(args))
# sanitize excess arguments, obiously there are better ways!
kwargs.pop('action', None)
kwargs.pop('command', None)
kwargs.pop('verbosity', None)
try:
# callback action
args.action(**kwargs)
except Exception as e:
self.log.error(e, exc_info=debug_on)
sys.exit(1)
sys.exit(0)
cli = CLI()
@cli.command('create')
@cli.option('name')
@cli.option('-s', '--size', type=int, choices=range(3))
@cli.option('-t', '--tag', choices=TAGS)
def create_command(name, size, tag):
"""Creates a resource."""
cli.log.info('creating resource')
cli.log.info(json.dumps({'name': name, 'size': size, 'tag': tag}))
@cli.command('update')
@cli.option('name')
@cli.option('-s', '--size', type=int, choices=range(3))
@cli.option('-t', '--tag', choices=TAGS)
def update_command(name, size=None, tag=None):
"""Updates a resource."""
cli.log.info('updating resource')
cli.log.info(json.dumps({'name': name, 'size': size, 'tag': tag}))
@cli.command('delete')
@cli.option('name')
def delete_command(name):
"""Delete a resource."""
cli.log.info('deleting resource %s' % name)
@cli.command('list')
def list_command():
"""Lists resources."""
cli.log.info('listing all resources')
@cli.command('find')
@cli.option('query')
def find_command(query):
"""Finds resources by query."""
cli.log.info('listing resources matching query %r' % query)
if __name__ == '__main__':
cli.run()
|
lukassup/python-cli
|
crud_cli/__init__.py
|
Python
|
mit
| 4,021
|
import uuid
import datetime
import time
import numpy as np
import threading
import vehicle
import azurehook
import json
class SpeedCamera(object):
TOPIC = "speedcamera"
EVENT_ACTIVATION = "ACTIVATION"
EVENT_DEACTIVATION = "DEACTIVATION"
EVENT_VEHICLE = "OBSERVATION"
def __init__(self, street, city, cloudhook = None, name = None):
self.id = str(uuid.uuid4())
self.street = street
self.city = city
self.isActive = False
self.speedLimit = None
self.rate = None
if cloudhook is None:
self.cloudhook = azurehook.AzureHook()
self.cloudhook.createTopic(self.TOPIC)
if name is not None:
self.name = name
def relocate(self, street, city = None):
self.street = street
if(city is not None):
self.city = city
# Most commonly executes on its own thread
def activate(self, speedLimit, rate):
if(self.isActive is True):
raise EnvironmentError('Speed camera is already active: deactivate first.')
self.speedLimit = speedLimit
self.rate = rate
self.datetime = datetime.datetime.now()
self.isActive = True
# Inform Azure of activated camera
self.__notifyCloudOfSelf()
# Event representing the passing of the next vehicle
self.nextVehicle = threading.Event()
# Loop until deactivate is called
# (preferably from a separate thread/process!!!!)
while self.isActive:
nextArrival = self.__genNextArrival()
self.nextVehicle.wait(timeout=nextArrival)
# Vehicle has passed - Create new vehicle
self.__onObservedVehicle()
# End of Loop
self.datetime = datetime.datetime.now()
self.__notifyCloudOfSelf()
# Preferably called from a separate thread
def deactivate(self):
if not self.isActive:
raise ValueError("Camera is not active")
self.isActive = False
self.nextVehicle.set()
def toDict(self):
return {"id" : self.id,
"street" : self.street,
"city" : self.city,
"rate" : self.rate,
"speedLimit" : self.speedLimit,
"isActive" : str(self.isActive),
"timestamp" : datetimeToTimestamp(self.datetime)}
def toJson(self):
return json.dumps(self.toDict(), indent = 4, sort_keys = True)
## Helping/Private methods
################################################
def __genNextArrival(self):
if(self.rate is None):
raise ValueError("Rate is undefined")
return np.random.exponential(1./self.rate)
def __notifyCloudOfSelf(self):
dic = {}
if self.isActive:
dic['event'] = self.EVENT_ACTIVATION
else:
dic['event'] = self.EVENT_DEACTIVATION
dic['camera'] = self.toDict()
json_string = json.dumps(dic, indent = 4, sort_keys = False)
self.cloudhook.publish(self.TOPIC, json_string, extra = {'event' : dic['event']})
def __notifyCloudOfVehicle(self, vehicle):
dic = {}
dic['event'] = self.EVENT_VEHICLE
dic['vehicle'] = vehicle.toDict()
dic['camera'] = self.toDict()
dic['timestamp'] = str(datetimeToTimestamp(datetime.datetime.now()))
json_string = json.dumps(dic, indent = 4, sort_keys = True)
self.cloudhook.publish(self.TOPIC, json_string,
extra = {'event' : dic['event'],
'isSpeeding' : dic['vehicle']['isSpeeding'],
'isPriority' : dic['vehicle']['isPriority']})
def __onObservedVehicle(self):
aVehicle = vehicle.NormalVehicle(self.speedLimit)
self.__notifyCloudOfVehicle(aVehicle)
##########################################################################################
##########################################################################################
##########################################################################################
#
# Helping Functions
#
##########################################################################################
##########################################################################################
##########################################################################################
def datetimeToTimestamp(dt):
return (dt - datetime.datetime(1970, 1, 1)).total_seconds()
## Global "factory" functions
def activateInNewThread(camera, speedLimit, rate, daemon = True):
thread = threading.Thread(target=camera.activate, args=(speedLimit, rate))
thread.daemon = daemon
thread.start()
return thread
|
PedrosWits/smart-cameras
|
smartcameras/speedcamera.py
|
Python
|
mit
| 4,814
|
""" License
MIT License
Copyright (c) 2017 OpenAdaptronik
Permission is hereby granted, free of charge, to any person obtaining a copy
of this software and associated documentation files (the "Software"), to deal
in the Software without restriction, including without limitation the rights
to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
copies of the Software, and to permit persons to whom the Software is
furnished to do so, subject to the following conditions:
The above copyright notice and this permission notice shall be included in all
copies or substantial portions of the Software.
THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
SOFTWARE.
"""
from django.contrib.auth import get_user_model
from django.contrib.auth.forms import UserCreationForm
from rattler.forms import BaseForm
from captcha.fields import ReCaptchaField
class RegisterForm(UserCreationForm, BaseForm):
""" User registration form.
Extends the django.contrib.auth.forms.UserCreationForm and rattler.forms.BaseForm.
"""
captcha = ReCaptchaField(
label="",
)
class Meta:
""" Meta informations.
"""
model = get_user_model()
fields = UserCreationForm.Meta.fields + ('email', 'password1', 'password2')
|
IT-PM-OpenAdaptronik/Webapp
|
apps/register/forms.py
|
Python
|
mit
| 1,676
|
import os
from pipeline import aminer
aminer_requirements = aminer.AminerNetworkData()
paths = [target.output().path for target in aminer_requirements.output()]
all_good = True
for path in paths:
if not os.path.exists(path):
all_good = False
print("Download was unsuccessful or config is invalid;"
" necessary required file does not exist: %s" % path)
if all_good:
print("Download and configuration successful; ready to run!")
|
macks22/dblp
|
verify_download.py
|
Python
|
mit
| 469
|
#!/usr/bin/env python3
import sys
from math import sqrt, hypot
from random import randint
from array import array
class Vector3d:
'''
//Define a vector class with constructor and operator: 'v'
struct v{
f x,y,z; // Vector has three float attributes.
v operator+(v r){return v(x+r.x,y+r.y,z+r.z);} //Vector add
v operator*(f r){return v(x*r,y*r,z*r);} //Vector scaling
f operator%(v r){return x*r.x+y*r.y+z*r.z;} //Vector dot product
v(){} //Empty constructor
v operator^(v r){return v(y*r.z-z*r.y,z*r.x-x*r.z,x*r.y-y*r.x);} //Cross-product
v(f a,f b,f c){x=a;y=b;z=c;} //Constructor
v operator!(){return *this*(1 /sqrt(*this%*this));} // Used later for normalizing the vector
};
'''
typecode = 'd'
__slots__ = ('x', 'y', 'z')
def __init__(self, x: float, y: float, z: float):
self.x = float(x)
self.y = float(y)
self.z = float(z)
@classmethod
def frombytes(cls, octets):
typecode = chr(octets[0])
memv = memoryview(octets[1:]).cast(typecode)
return cls(*memv)
def __iter__(self):
return (i for i in (self.x, self.y, self.z))
def __repr__(self):
class_name = type(self).__name__
return '{}({!r}, {!r}, {!r})'.format(class_name, *self)
def __str__(self):
return str(tuple(self))
def __bytes__(self):
return (bytes([ord(self.typecode)]) +
bytes(array(self.typecode, self)))
def __eq__(self, other: Vector3d):
return tuple(self) == tuple(other)
def __abs__(self):
return hypot(hypot(self.x, self.y), self.z)
def __bool__(self):
return bool(abs(self))
def __add__(self, other: Vector3d):
return Vector(self.x + other.x, self.y + other.y, self.z + other.z)
def __mul__(self, scale: Vector3d):
return Vector(self.x * scale, self.y * scale, self.z * scale)
def __matmul__(self, other: Vector3d):
return Vector(self.x * other.x, self.y * other.y, self.z * other.y)
def _cross_product(self, other: Vector3d):
return Vector(
self.y * other.z - self.z * other.y,
self.z * other.x - self.x * other.z,
self.x * other.y - self.y * other.x
)
def _normalisation(self):
return self * (1 / sqrt(self.__dot__(self)))
'''
//The set of sphere positions describing the world.
//Those integers are in fact bit vectors.
i G[]={247570,280596,280600,249748,18578,18577,231184,16,16};
/*
16 1
16 1
231184 111 111 1
18577 1 1 1 1 1
18578 1 1 1 1 1
249748 1111 11111 1 1
280600 1 1 1 11
280596 1 1 1 1 1
247570 1111 111 1 1
*/
'''
G = [247570, 280596, 280600, 249748, 18578, 18577, 231184, 16, 16]
def random() -> float:
'''
// Random generator, return a float within range [0-1]
f R(){return(f)rand()/RAND_MAX;}
Sadly Python's `random.random` is `[0.0, 1.0)` so we have to
use `random.randint` which is inclusive.
'''
return randint(0, 1000000) / 1000000.0
def intersection_test(o: Vector3d, d: Vector3d, t: float, n: Vector3d) -> int:
'''
//The intersection test for line [o,v].
// Return 2 if a hit was found (and also return distance t and bouncing ray n).
// Return 0 if no hit was found but ray goes upward
// Return 1 if no hit was found but ray goes downward
i T(v o,v d,f& t,v& n){
t=1e9;
i m=0;
f p=-o.z/d.z;
if(.01<p)
t=p,n=v(0,0,1),m=1;
//The world is encoded in G, with 9 lines and 19 columns
for(i k=19;k--;) //For each columns of objects
for(i j=9;j--;) //For each line on that columns
if(G[j]&1<<k){ //For this line j, is there a sphere at column i ?
// There is a sphere but does the ray hits it ?
v p=o+v(-k,0,-j-4);
f b=p%d,c=p%p-1,q=b*b-c;
//Does the ray hit the sphere ?
if(q>0){
//It does, compute the distance camera-sphere
f s=-b-sqrt(q);
if(s<t && s>.01)
// So far this is the minimum distance, save it. And also
// compute the bouncing ray vector into 'n'
t=s,
n=!(p+d*t),
m=2;
}
}
return m;
}
'''
t = 10^9
m = 0
p = -o.z / d.z
if .01 < p:
t = p
n = Vector3d(0., 0., 1.)
m = 1
for k in range(19, 0, -1):
for j in range(9, 0, -1):
if G[j] & 1<<k:
# There is a sphere but does the ray hit it ?
p = o + Vector3d(-k, 0.0, -j-4.0)
b = p @ d
c = p @ p - 1
q = b * b - c
# Does the ray hit the sphere
if q > 0:
# It does, compute the distance camera<->sphere
s = -b - sqrt(q)
if s < t and s > .01:
# This is the minimum distance, save it. And alos
# compute the bouncing ray vector into `n`.
t = s
n = !(p + d + t)
m = 2
return m
def sample(o: Vector3d, d: Vector3d) -> Vector3d:
"""
// (S)ample the world and return the pixel color for
// a ray passing by point o (Origin) and d (Direction)
v S(v o,v d){
f t;
v n;
//Search for an intersection ray Vs World.
i m=T(o,d,t,n);
if(!m) // m==0
//No sphere found and the ray goes upward: Generate a sky color
return v(.7,.6,1)*pow(1-d.z,4);
//A sphere was maybe hit.
v h=o+d*t, // h = intersection coordinate
l=!(v(9+R(),9+R(),16)+h*-1), // 'l' = direction to light (with random delta for soft-shadows).
r=d+n*(n%d*-2); // r = The half-vector
//Calculated the lambertian factor
f b=l%n;
//Calculate illumination factor (lambertian coefficient > 0 or in shadow)?
if(b<0||T(h,l,t,n))
b=0;
// Calculate the color 'p' with diffuse and specular component
f p=pow(l%r*(b>0),99);
if(m&1){ //m == 1
h=h*.2; //No sphere was hit and the ray was going downward: Generate a floor color
return((i)(ceil(h.x)+ceil(h.y))&1?v(3,1,1):v(3,3,3))*(b*.2+.1);
}
//m == 2 A sphere was hit. Cast an ray bouncing from the sphere surface.
return v(p,p,p)+S(h,r)*.5; //Attenuate color by 50% since it is bouncing (* .5)
}
"""
# Search for an intersection ray vs. world
m = intersection_test(o, d, t, n)
if not m:
# No sphere was found and the ray goes upward. Generate a sky colour
return Vector3d(.7, .6, 1.) * (1-d.z ** 4)
# A sphere was maybe hit
h = o + d * t # h = intersection co-ordinate
l = !(Vector3d(9 + random(), 9 + random(), 16) + h * -1) # l = direction to light (with random delta for soft-shadows)
r = d + n * (n @ d * -2) # r = The half-vector
b = l @ n # Calculate the lambertian factor
if b < 0 or intersection_test(h, l, t, n):
# Calculate the illumination factor (lambertian coefficient > 0 or in shadow) ?
b = 0
p = pow(l @ r * (b > 0), 99) # Calculate the colour `p` with diffuse and specular component
if m & 1:
h *= .2 # No sphere was hit and the ray was going downward. Generate a floor colour
return (ceil(h.x) + ceil(h.y)) & 1 ? Vector3d(3., 1., 1.) : Vector3d(3., 3., 3.) * (b * .2 + .1)
# m == 2 -> No sphere was hit. Cast a ray bounding from the sphere surface.
return Vector3d(p, p, p) + sample(h, r) * 0.5 # Attenuate colour by 50% since it is bouncing (* .5)
if __name__ == '__main__':
"""
i main(){
printf("P6 512 512 255 "); // The PPM Header is issued
// The '!' are for normalizing each vectors with ! operator.
v g=!v(-6,-16,0), // Camera direction
a=!(v(0,0,1)^g)*.002, // Camera up vector...Seem Z is pointing up :/ WTF !
b=!(g^a)*.002, // The right vector, obtained via traditional cross-product
c=(a+b)*-256+g; // WTF ? See https://news.ycombinator.com/item?id=6425965 for more.
for(i y=512;y--;) //For each column
for(i x=512;x--;){ //For each pixel in a line
//Reuse the vector class to store not XYZ but a RGB pixel color
v p(13,13,13); // Default pixel color is almost pitch black
//Cast 64 rays per pixel (For blur (stochastic sampling) and soft-shadows.
for(i r=64;r--;){
// The delta to apply to the origin of the view (For Depth of View blur).
v t=a*(R()-.5)*99+b*(R()-.5)*99; // A little bit of delta up/down and left/right
// Set the camera focal point v(17,16,8) and Cast the ray
// Accumulate the color returned in the p variable
p=S(v(17,16,8)+t, //Ray Origin
!(t*-1+(a*(R()+x)+b*(y+R())+c)*16) // Ray Direction with random deltas
// for stochastic sampling
)*3.5+p; // +p for color accumulation
}
printf("%c%c%c",(i)p.x,(i)p.y,(i)p.z);
}
}
"""
sys.stdout.write('P6 512 512 255 ')
# The `!` are for normalising each vectors with `!` operator
g = !Vector3d(-6., -16., 0.)
a = !(Vector3d(0., 0., 1.) ^ g) * .002
b = !(g ^ a) * .002
c = (a + b) * -256 + g
for y in range(512, 0, -1):
for x in range(512, 0, -1):
pixel = Vector3d(13., 13., 13.) # Reuse the vector class to store RGB values
# Cast 64 rays per pixel
for r in range(64, 0, -1):
t = a * (random() - .5) * 99 + b * (random() - .5) * 99
p = sample(
Vector3d(17., 16., 8.) + t, # Ray origin
# Ray direction with random deltas for stochastic sampling
!(t * -1 + (a * (random() + x) + b * (y + random()) + c ) * 16)
) * 3.5 + p # +p for colour accumulation
sys.stdout.write('{!c}{!c}{!c}')
|
WittscraftStudios/business-card-raytracer
|
python/card.py
|
Python
|
mit
| 10,620
|
from distutils.core import setup
setup(name="Complicated build",
version="1.4.2",
description="library which provides a simple interface to perform complex compilation tasks with multi-ligual sources.",
author="Joe Jordan",
author_email="tehwalrus@h2j9k.org",
url="https://github.com/joe-jordan/complicated_build",
packages=['cb']
)
|
joe-jordan/complicated_build
|
setup.py
|
Python
|
mit
| 346
|
#!/usr/bin/env python
#
# Use the raw transactions API to spend bitcoins received on particular addresses,
# and send any change back to that same address.
#
# Example usage:
# spendfrom.py # Lists available funds
# spendfrom.py --from=ADDRESS --to=ADDRESS --amount=11.00
#
# Assumes it will talk to a bitcoind or Bitcoin-Qt running
# on localhost.
#
# Depends on jsonrpc
#
from decimal import *
import getpass
import math
import os
import os.path
import platform
import sys
import time
from jsonrpc import ServiceProxy, json
BASE_FEE=Decimal("0.001")
def check_json_precision():
"""Make sure json library being used does not lose precision converting BTC values"""
n = Decimal("20000000.00000003")
satoshis = int(json.loads(json.dumps(float(n)))*1.0e8)
if satoshis != 2000000000000003:
raise RuntimeError("JSON encode/decode loses precision")
def determine_db_dir():
"""Return the default location of the bitcoin data directory"""
if platform.system() == "Darwin":
return os.path.expanduser("~/Library/Application Support/Bitcoin/")
elif platform.system() == "Windows":
return os.path.join(os.environ['APPDATA'], "Bitcoin")
return os.path.expanduser("~/.bitcoin")
def read_bitcoin_config(dbdir):
"""Read the bitcoin.conf file from dbdir, returns dictionary of settings"""
from ConfigParser import SafeConfigParser
class FakeSecHead(object):
def __init__(self, fp):
self.fp = fp
self.sechead = '[all]\n'
def readline(self):
if self.sechead:
try: return self.sechead
finally: self.sechead = None
else:
s = self.fp.readline()
if s.find('#') != -1:
s = s[0:s.find('#')].strip() +"\n"
return s
config_parser = SafeConfigParser()
config_parser.readfp(FakeSecHead(open(os.path.join(dbdir, "bitcoin.conf"))))
return dict(config_parser.items("all"))
def connect_JSON(config):
"""Connect to a bitcoin JSON-RPC server"""
testnet = config.get('testnet', '0')
testnet = (int(testnet) > 0) # 0/1 in config file, convert to True/False
if not 'rpcport' in config:
config['rpcport'] = 16452 if testnet else 6452
connect = "http://%s:%s@127.0.0.1:%s"%(config['rpcuser'], config['rpcpassword'], config['rpcport'])
try:
result = ServiceProxy(connect)
# ServiceProxy is lazy-connect, so send an RPC command mostly to catch connection errors,
# but also make sure the bitcoind we're talking to is/isn't testnet:
if result.getmininginfo()['testnet'] != testnet:
sys.stderr.write("RPC server at "+connect+" testnet setting mismatch\n")
sys.exit(1)
return result
except:
sys.stderr.write("Error connecting to RPC server at "+connect+"\n")
sys.exit(1)
def unlock_wallet(bitcoind):
info = bitcoind.getinfo()
if 'unlocked_until' not in info:
return True # wallet is not encrypted
t = int(info['unlocked_until'])
if t <= time.time():
try:
passphrase = getpass.getpass("Wallet is locked; enter passphrase: ")
bitcoind.walletpassphrase(passphrase, 5)
except:
sys.stderr.write("Wrong passphrase\n")
info = bitcoind.getinfo()
return int(info['unlocked_until']) > time.time()
def list_available(bitcoind):
address_summary = dict()
address_to_account = dict()
for info in bitcoind.listreceivedbyaddress(0):
address_to_account[info["address"]] = info["account"]
unspent = bitcoind.listunspent(0)
for output in unspent:
# listunspent doesn't give addresses, so:
rawtx = bitcoind.getrawtransaction(output['txid'], 1)
vout = rawtx["vout"][output['vout']]
pk = vout["scriptPubKey"]
# This code only deals with ordinary pay-to-bitcoin-address
# or pay-to-script-hash outputs right now; anything exotic is ignored.
if pk["type"] != "pubkeyhash" and pk["type"] != "scripthash":
continue
address = pk["addresses"][0]
if address in address_summary:
address_summary[address]["total"] += vout["value"]
address_summary[address]["outputs"].append(output)
else:
address_summary[address] = {
"total" : vout["value"],
"outputs" : [output],
"account" : address_to_account.get(address, "")
}
return address_summary
def select_coins(needed, inputs):
# Feel free to improve this, this is good enough for my simple needs:
outputs = []
have = Decimal("0.0")
n = 0
while have < needed and n < len(inputs):
outputs.append({ "txid":inputs[n]["txid"], "vout":inputs[n]["vout"]})
have += inputs[n]["amount"]
n += 1
return (outputs, have-needed)
def create_tx(bitcoind, fromaddresses, toaddress, amount, fee):
all_coins = list_available(bitcoind)
total_available = Decimal("0.0")
needed = amount+fee
potential_inputs = []
for addr in fromaddresses:
if addr not in all_coins:
continue
potential_inputs.extend(all_coins[addr]["outputs"])
total_available += all_coins[addr]["total"]
if total_available < needed:
sys.stderr.write("Error, only %f BTC available, need %f\n"%(total_available, needed));
sys.exit(1)
#
# Note:
# Python's json/jsonrpc modules have inconsistent support for Decimal numbers.
# Instead of wrestling with getting json.dumps() (used by jsonrpc) to encode
# Decimals, I'm casting amounts to float before sending them to bitcoind.
#
outputs = { toaddress : float(amount) }
(inputs, change_amount) = select_coins(needed, potential_inputs)
if change_amount > BASE_FEE: # don't bother with zero or tiny change
change_address = fromaddresses[-1]
if change_address in outputs:
outputs[change_address] += float(change_amount)
else:
outputs[change_address] = float(change_amount)
rawtx = bitcoind.createrawtransaction(inputs, outputs)
signed_rawtx = bitcoind.signrawtransaction(rawtx)
if not signed_rawtx["complete"]:
sys.stderr.write("signrawtransaction failed\n")
sys.exit(1)
txdata = signed_rawtx["hex"]
return txdata
def compute_amount_in(bitcoind, txinfo):
result = Decimal("0.0")
for vin in txinfo['vin']:
in_info = bitcoind.getrawtransaction(vin['txid'], 1)
vout = in_info['vout'][vin['vout']]
result = result + vout['value']
return result
def compute_amount_out(txinfo):
result = Decimal("0.0")
for vout in txinfo['vout']:
result = result + vout['value']
return result
def sanity_test_fee(bitcoind, txdata_hex, max_fee):
class FeeError(RuntimeError):
pass
try:
txinfo = bitcoind.decoderawtransaction(txdata_hex)
total_in = compute_amount_in(bitcoind, txinfo)
total_out = compute_amount_out(txinfo)
if total_in-total_out > max_fee:
raise FeeError("Rejecting transaction, unreasonable fee of "+str(total_in-total_out))
tx_size = len(txdata_hex)/2
kb = tx_size/1000 # integer division rounds down
if kb > 1 and fee < BASE_FEE:
raise FeeError("Rejecting no-fee transaction, larger than 1000 bytes")
if total_in < 0.01 and fee < BASE_FEE:
raise FeeError("Rejecting no-fee, tiny-amount transaction")
# Exercise for the reader: compute transaction priority, and
# warn if this is a very-low-priority transaction
except FeeError as err:
sys.stderr.write((str(err)+"\n"))
sys.exit(1)
def main():
import optparse
parser = optparse.OptionParser(usage="%prog [options]")
parser.add_option("--from", dest="fromaddresses", default=None,
help="addresses to get bitcoins from")
parser.add_option("--to", dest="to", default=None,
help="address to get send bitcoins to")
parser.add_option("--amount", dest="amount", default=None,
help="amount to send")
parser.add_option("--fee", dest="fee", default="0.0",
help="fee to include")
parser.add_option("--datadir", dest="datadir", default=determine_db_dir(),
help="location of bitcoin.conf file with RPC username/password (default: %default)")
parser.add_option("--testnet", dest="testnet", default=False, action="store_true",
help="Use the test network")
parser.add_option("--dry_run", dest="dry_run", default=False, action="store_true",
help="Don't broadcast the transaction, just create and print the transaction data")
(options, args) = parser.parse_args()
check_json_precision()
config = read_bitcoin_config(options.datadir)
if options.testnet: config['testnet'] = True
bitcoind = connect_JSON(config)
if options.amount is None:
address_summary = list_available(bitcoind)
for address,info in address_summary.iteritems():
n_transactions = len(info['outputs'])
if n_transactions > 1:
print("%s %.8f %s (%d transactions)"%(address, info['total'], info['account'], n_transactions))
else:
print("%s %.8f %s"%(address, info['total'], info['account']))
else:
fee = Decimal(options.fee)
amount = Decimal(options.amount)
while unlock_wallet(bitcoind) == False:
pass # Keep asking for passphrase until they get it right
txdata = create_tx(bitcoind, options.fromaddresses.split(","), options.to, amount, fee)
sanity_test_fee(bitcoind, txdata, amount*Decimal("0.01"))
if options.dry_run:
print(txdata)
else:
txid = bitcoind.sendrawtransaction(txdata)
print(txid)
if __name__ == '__main__':
main()
|
ikarutoko/Tradercoin
|
Contrib/Spendfrom/spendfrom.py
|
Python
|
mit
| 10,053
|
from operator import itemgetter
points = [{"coords":[14,48]},{"coords":[15,50]},
]
s = [p["coords"] for p in points]
s = [p["coords"] for p in points if p["coords"][0]>14]
print(s)
def twice(n):
return 2*n
numbers = ["12","14","23"]
nums = list(map(int,numbers))
doubled = list(map(twice,nums))
print(nums)
print(doubled)
s2 = list(map(itemgetter('coords'),points))
s3 = list(map(lambda p:p['coords'],points))
print(s2)
print(s3)
def bignum(n):
if n > 13:
return True
else:
return False
bignums = list(filter(bignum,nums))
bignums2 = list(filter(lambda n:n>13,nums))
print(bignums)
print(bignums2)
# TODO enumerate
|
xtompok/uvod-do-prg
|
cv12/cv12.py
|
Python
|
mit
| 660
|
import time
import urllib
import urllib2
from Queue import Queue, Empty
from functools import partial
from threading import Thread, Lock
import tornado.httpserver
import tornado.ioloop
from tornado.web import HTTPError, RequestHandler, Application, asynchronous
from msgbox import logger
from msgbox.sim import sim_manager, TxSmsReq
ioloop = tornado.ioloop.IOLoop.instance()
# application/x-www-form-urlencoded
# params:
# key: "asds7878" (optional)
# recipient: "+393482222222"
# sender: "+393481111111"
# imsi: "21312123232"
# text: "sms text"
class MTHandler(RequestHandler):
@asynchronous
def post(self):
sender = self.get_argument('sender', None)
recipient = self.get_argument('recipient')
text = self.get_argument('text')
imsi = self.get_argument('imsi', None)
key = self.get_argument('key', None)
if not ((sender is None) ^ (imsi is None)):
err_msg = 'Use either "sender" or "imsi" params'
raise HTTPError(400, err_msg)
sim_manager.send(TxSmsReq(sender, recipient, text, imsi, key,
callback=self.reply_callback))
def reply_callback(self, response_dict):
ioloop.add_callback(partial(self.handle_reply, response_dict))
def handle_reply(self, response_dict):
log_method = logger.warn if response_dict['status'] == 'ERROR' else \
logger.info
log_method(response_dict['desc'])
self.write(response_dict)
self.finish()
class HTTPServerManager(object):
def __init__(self, port=8080):
app = Application([
(r"/send_sms", MTHandler),
])
self.port = port
self.http_server = tornado.httpserver.HTTPServer(app)
def start(self):
logger.info('http listening on port %s', self.port)
self.http_server.listen(self.port)
def stop(self):
self.http_server.stop()
http_server_manager = HTTPServerManager()
# ~~~~~~~~ HTTP Client Manager ~~~~~~~~
class HTTPClientManagerStoppingError(Exception): pass
class HTTPClientManager(object):
N_WORKERS = 10
def __init__(self):
self.active = False
self.workers = []
self.queue = Queue()
self.mutex = Lock()
def start(self):
self.active = True
for i in xrange(self.N_WORKERS):
thread = Thread(name='HttpClient %d' % i, target=self._work)
self.workers.append(thread)
thread.start()
def stop(self):
with self.mutex:
self.active = False
def _work(self):
while self.active or not self.queue.empty:
try:
msg_dict = self.queue.get(timeout=2)
except Empty:
continue
url = msg_dict.pop('url')
for attempt in xrange(3):
try:
msg_dict['text'] = msg_dict['text'].encode('utf8')
data = urllib.urlencode(msg_dict)
urllib2.urlopen(url, data, timeout=20)
logger.info('forwarded sms - sender=%s recipient=%s',
msg_dict['sender'], msg_dict['recipient'])
break
except Exception, e:
logger.error('error while sending msg', exc_info=True)
time.sleep(10)
else:
logger.error('giving up sending message %s', msg_dict,
exc_info=1)
def enqueue(self, rx_sms):
with self.mutex:
if self.active:
self.queue.put(rx_sms)
else:
raise HTTPClientManagerStoppingError()
http_client_manager = HTTPClientManager()
|
paolo-losi/msgbox
|
msgbox/http.py
|
Python
|
mit
| 3,832
|
from django.contrib.contenttypes.models import ContentType
from django.db import models
from django_page_views.templatetags.django_page_views import views_count
from django_vote_system.templatetags.vote import downvote_count, upvote_count
class Common(models.Model):
class Meta:
abstract = True
@property
def content_type_obj(self):
return ContentType.objects.get_for_model(self)
@property
def app_label(self):
return self.content_type_obj.app_label
@property
def model_name(self):
return self.content_type_obj.model
class Vote(models.Model):
class Meta:
abstract = True
@property
def upvote_count(self):
return upvote_count(self.__class__, self.id)
@property
def downvote_count(self):
return downvote_count(self.__class__, self.id)
class View:
id = None
class Meta:
abstract = True
@property
def views(self):
return views_count(self.__class__, self.id)
|
hakancelik96/coogger
|
core/cooggerapp/models/common.py
|
Python
|
mit
| 1,003
|
# Validate TensorFlow installation
import tensorflow as tf
message = tf.constant('TensorFlow installation successful!')
sess = tf.Session()
print(sess.run(message))
|
guosibs/learning
|
src/validate.py
|
Python
|
mit
| 165
|
from splitwise import Splitwise
import unittest
try:
from unittest.mock import patch
except ImportError: # Python 2
from mock import patch
@patch('splitwise.Splitwise._Splitwise__makeRequest')
class GetCurrentUserTestCase(unittest.TestCase):
def setUp(self):
self.sObj = Splitwise('consumerkey', 'consumersecret')
def test_getCurrentUser_success(self, mockMakeRequest):
mockMakeRequest.return_value = '{"user":{"id":12345,"first_name":"Naman","last_name":"Aggarwal","picture":{"small":"https://splitwise.s3.amazonaws.com/uploads/user/avatar/12345/small_mypic.jpg","medium":"https://splitwise.s3.amazonaws.com/uploads/user/avatar/12345/medium_mypic.jpg","large":"https://splitwise.s3.amazonaws.com/uploads/user/avatar/12345/large_mypic.jpg"},"custom_picture":true,"email":"naman@naman.com","registration_status":"confirmed","force_refresh_at":"2017-03-18T11:41:36Z","locale":"en","country_code":"IN","date_format":"MM/DD/YYYY","default_currency":"SGD","default_group_id":null,"notifications_read":"2020-06-10T14:12:01Z","notifications_count":8,"notifications":{"added_as_friend":true,"added_to_group":true,"expense_added":false,"expense_updated":false,"bills":true,"payments":true,"monthly_summary":true,"announcements":true}}}' # noqa: E501
user = self.sObj.getCurrentUser()
mockMakeRequest.assert_called_with(
"https://secure.splitwise.com/api/v3.0/get_current_user")
self.assertEqual(user.getId(), 12345)
self.assertEqual(user.getFirstName(), "Naman")
self.assertEqual(user.getLastName(), "Aggarwal")
self.assertEqual(user.getEmail(), "naman@naman.com")
self.assertEqual(user.getPicture().getSmall(),
"https://splitwise.s3.amazonaws.com/uploads/user/avatar/12345/small_mypic.jpg")
self.assertEqual(user.getPicture().getMedium(),
"https://splitwise.s3.amazonaws.com/uploads/user/avatar/12345/medium_mypic.jpg")
self.assertEqual(user.getPicture().getLarge(),
"https://splitwise.s3.amazonaws.com/uploads/user/avatar/12345/large_mypic.jpg")
self.assertEqual(user.getRegistrationStatus(), "confirmed")
def test_getCurrentUser_exception(self, mockMakeRequest):
mockMakeRequest.side_effect = Exception(
"Invalid response %s. Please check your consumer key and secret." % 404)
with self.assertRaises(Exception):
self.sObj.getCurrentUser()
mockMakeRequest.assert_called_with(
"https://secure.splitwise.com/api/v3.0/get_current_user")
|
namaggarwal/splitwise
|
tests/test_getCurrentUser.py
|
Python
|
mit
| 2,597
|
from runipy.notebook_runner import NotebookRunner
from IPython.nbformat.current import read
from glob import glob
files = glob("./[01]*ipynb")
for file in files:
print("Doing file {}".format(file))
notebook = read(open(file), 'json')
r = NotebookRunner(notebook)
r.run_notebook(skip_exceptions=True)
from IPython.nbformat.current import write
write(r.nb, open("_runipy/"+file, 'w'), 'json')
print("Done file {}".format(file))
|
IanHawke/maths-with-python
|
run_notebooks.py
|
Python
|
mit
| 456
|
from datetime import timedelta
from celery import Celery
app = Celery('twitch',
backend='db+sqlite:///celeryresdb.sqlite',
broker='sqla+sqlite:///celerydb.sqlite',
include=['tasks'])
app.conf.CELERYBEAT_SCHEDULE = {
'clear-db': {
'task': 'twitch.clear_dbs',
'schedule': timedelta(minutes=15)
},
}
if __name__ == '__main__':
app.start()
|
Redpoint1/Twitch-SlinxBot
|
twitch.py
|
Python
|
mit
| 406
|
#coding: utf-8
"""
@Author: Well
@Date: 2014 - 04 - 14
"""
"""
求某一个英文文本中完整句子的数目,
文本中只包含大小写字母、空格、“,”和“.”,
完整的句子是指以“.”结束,且“.”号前必须出现至少一个字母。
"""
import os
# 文件名
name_ = os.path.basename(__file__).split('.')[0]
dir_ = os.path.dirname(__file__)
# 绝对文件夹路径
# file2 = os.path.dirname(__file__)
# print file2
#
# # 绝对路径
#file3 = os.path.abspath(__file__)
# print file3
# txt文件的绝对路径
file_ = dir_ + '\\' + name_ + '.txt'
txt_ = file(file_, 'r')
txt_2 = txt_.read().split('.')
print len(txt_2) - 1
txt_.close()
|
neiltest/neil_learn_python
|
src/learn_python/python_other/neil_06_txt_split.py
|
Python
|
mit
| 691
|
#!/usr/bin/env python
# encoding: utf-8
#
# The MIT License (MIT)
#
# Copyright (c) 2013-2015 CNRS
#
# Permission is hereby granted, free of charge, to any person obtaining a copy
# of this software and associated documentation files (the "Software"), to deal
# in the Software without restriction, including without limitation the rights
# to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
# copies of the Software, and to permit persons to whom the Software is
# furnished to do so, subject to the following conditions:
#
# The above copyright notice and this permission notice shall be included in
# all copies or substantial portions of the Software.
#
# THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
# IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
# FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
# AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
# LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
# OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
# SOFTWARE.
#
# AUTHORS
# Hervé BREDIN -- http://herve.niderb.fr/
from __future__ import unicode_literals
from .command import CommandWrapper
class AVConv(CommandWrapper):
"""Convert audio or video streams
Parameters
----------
avconv : str, optional.
Absolute path to `avconv` in case it is not reachable from PATH.
"""
def __init__(self, avconv=None):
if avconv is None:
avconv = 'avconv'
super(AVConv, self).__init__(avconv)
def audio_track(self, handbrake_to, stream, to):
"""
Extract audio track from HandBrake output
Parameters
----------
handbrake_to : str
Path to input file
stream : int
to : str
Path to output file
See savvyadmin.com/extract-audio-from-video-files-to-wav-using-ffmpeg
"""
options = [
'-y',
'-i', handbrake_to,
'-map', '0:{stream:d}'.format(stream=stream),
'-acodec', 'pcm_s16le',
'-ac', '1',
to
]
self.run_command(options=options, env=None)
# http://www.willus.com/author/streaming2.shtml
# -map option details can be foundhere :
# http://libav.org/avconv.html#Advanced-options
def mp4(self, handbrake_to, audio_stream, to):
options = [
'-y',
'-i', handbrake_to,
'-map', '0:0,0:0',
'-i_qfactor', '0.71',
'-qcomp', '0.6',
'-qmin', '10',
'-qmax', '63',
'-qdiff', '4',
'-trellis', '0',
'-vcodec', 'libx264',
'-b:v', '200k',
'-map', '0:{stream:d},0:0'.format(stream=audio_stream),
'-b:a', '56k',
'-ar', '22050',
'-ac', '2',
'-acodec', 'libvo_aacenc',
to
]
self.run_command(options=options, env=None)
def webm(self, handbrake_to, audio_stream, to):
options = [
'-y',
'-i', handbrake_to,
'-map', '0:0,0:0',
'-qmax', '63',
'-b:v', '200k',
'-vcodec', 'libvpx',
'-map', '0:{stream:d},0:0'.format(stream=audio_stream),
'-b:a', '56k',
'-ar', '22050',
'-ac', '2',
'-acodec', 'libvorbis',
to
]
self.run_command(options=options, env=None)
def ogv(self, handbrake_to, audio_stream, to):
options = [
'-y',
'-i', handbrake_to,
'-map', '0:0,0:0',
'-qmax', '63',
'-b:v', '200k',
'-vcodec', 'libtheora',
'-map', '0:{stream:d},0:0'.format(stream=audio_stream),
'-b:a', '56k',
'-ar', '22050',
'-ac', '2',
'-acodec', 'libvorbis',
to
]
self.run_command(options=options, env=None)
|
tvd-dataset/tvd
|
tvd/rip/avconv.py
|
Python
|
mit
| 4,097
|
#!/usr/bin/env python
from setuptools import setup, find_packages
# Get version
with open('compare_versions/version.py') as f:
exec(f.read())
# Get documentation
def readme():
with open('README.rst') as f:
return f.read()
setup(
name='compare_versions',
version=__version__,
author='Luke Yeager',
author_email='luke.yeager@gmail.com',
url='https://github.com/lukeyeager/compare-versions',
description='Compare versions using various versioning schemes',
long_description=readme(),
scripts=[
'bin/compare_versions',
],
packages=find_packages(exclude=['tests', 'tests.*']),
test_suite='tests',
classifiers=[
'License :: OSI Approved :: MIT License',
'Operating System :: OS Independent',
'Programming Language :: Python',
],
)
|
lukeyeager/compare-versions
|
setup.py
|
Python
|
mit
| 830
|
"""
WSGI config for codex project.
It exposes the WSGI callable as a module-level variable named ``application``.
For more information on this file, see
https://docs.djangoproject.com/en/1.7/howto/deployment/wsgi/
"""
import os
os.environ.setdefault("DJANGO_SETTINGS_MODULE", "codex.settings")
from django.core.wsgi import get_wsgi_application
application = get_wsgi_application()
|
mod2/codex
|
codex/wsgi.py
|
Python
|
mit
| 385
|
from .. import Provider as BaseProvider
class Provider(BaseProvider):
"""
A Faker provider for the Luxembourgish VAT IDs
"""
vat_id_formats = (
'LU########',
)
def vat_id(self):
"""
http://ec.europa.eu/taxation_customs/vies/faq.html#item_11
:return: a random Luxembourgish VAT ID
"""
return self.bothify(self.random_element(self.vat_id_formats))
|
danhuss/faker
|
faker/providers/ssn/lb_LU/__init__.py
|
Python
|
mit
| 423
|
"""
WSGI config for pyjobs project.
It exposes the WSGI callable as a module-level variable named ``application``.
For more information on this file, see
https://docs.djangoproject.com/en/1.8/howto/deployment/wsgi/
"""
import os
from django.core.wsgi import get_wsgi_application
os.environ.setdefault("DJANGO_SETTINGS_MODULE", "pyjobs.settings")
application = get_wsgi_application()
|
pyshop/pyjobs
|
src/pyjobs/wsgi.py
|
Python
|
mit
| 389
|
"""Value change dump stuff."""
class VCDObject:
"""Abstract VCD object class."""
class VCDScope(VCDObject):
"""VCD scope."""
def __init__(self, *scopes: str):
"""Initialize.
Parameters
----------
scopes
List of scope names
"""
self._scopes = []
for scope in scopes:
if not isinstance(scope, str):
raise TypeError("scope name must be string")
if len(scope) < 1:
# empty, ignore
continue
self._scopes.append(scope)
def __repr__(self):
"""Get representation."""
return "::".join(self._scopes)
def __len__(self):
"""Get scope length."""
return len(self._scopes)
def __getitem__(self, idx):
"""Get scope by index."""
return self._scopes[idx]
def __eq__(self, other):
"""Scope equality."""
# FIXME: raising TypeError is weird
if not isinstance(other, VCDScope):
raise TypeError(
f"other must be a VCDScope object, got {type(other)}"
)
return self._scopes == other._scopes
def __hash__(self):
"""Get hash."""
return hash(tuple(self._scopes))
def contains(self, other: "VCDScope") -> bool:
"""Get whether this scope contains other scope.
Arguments
----------
other
other scope to compare against
"""
if not isinstance(other, VCDScope):
raise TypeError("other must be a VCDScope object")
if len(self) >= len(other):
# cannot contain, length must be less
return False
for idx, this_subscope in enumerate(self._scopes):
if other[idx] != this_subscope:
return False
return True
@staticmethod
def from_str(scope_str: str) -> "VCDScope":
"""Build from string."""
if not isinstance(scope_str, str):
raise TypeError("must be a string")
scopes = scope_str.split("::")
inclusive = len(scopes[-1]) < 1
return (VCDScope(*scopes), inclusive)
def pack(self) -> str:
"""Pack."""
return str(self)
|
brunosmmm/hdltools
|
hdltools/vcd/__init__.py
|
Python
|
mit
| 2,238
|
# -*- coding: utf-8 -*-
# test/unit/stat/test_flushtosql.py
# Copyright (C) 2016 authors and contributors (see AUTHORS file)
#
# This module is released under the MIT License.
"""Test flushtosql()"""
# ============================================================================
# Imports
# ============================================================================
# Stdlib imports
import asyncio
from functools import partial
# Third-party imports
from pandas import DataFrame, read_sql_table, Series, Timestamp, to_timedelta
import pytest
from sqlalchemy import create_engine
# Local imports
import loadlimit.channel as channel
from loadlimit.core import BaseLoop
from loadlimit.result import SQLTotal
import loadlimit.stat as stat
from loadlimit.stat import CountStore, SendTimeData
from loadlimit.util import aiter
# ============================================================================
# Fixtures
# ============================================================================
@pytest.fixture
def fake_flushtosql(monkeypatch):
"""Setup fake flushtosql callable"""
fake_flushtosql = stat.FlushToSQL()
monkeypatch.setattr(stat, 'flushtosql', fake_flushtosql)
pytestmark = pytest.mark.usefixtures('fake_shutdown_channel',
'fake_timedata_channel')
# ============================================================================
# Tests
# ============================================================================
@pytest.mark.parametrize('num', [10, 12])
def test_flushtosql(testloop, num):
"""updateperiod updates statsdict with timeseries data points
num fixture allows testing the flushtosql_shutdown coro func for:
* all data has already been flushed to the sql db
* there's still some data remaining that needs to be flushed to sql db
"""
measure = CountStore()
# Setup sqlalchemy engine
engine = create_engine('sqlite://')
timetotal = SQLTotal(sqlengine=engine, countstore=measure)
# Create coro to time
@measure(name='churn')
async def churn(i):
"""Do nothing"""
await asyncio.sleep(0)
async def run():
"""run"""
async for i in aiter(range(num)):
await churn(i)
await channel.shutdown.send(0)
# Setup SendTimeData
send = SendTimeData(measure, flushwait=to_timedelta(0, unit='s'),
channel=stat.timedata)
# Add to shutdown channel
channel.shutdown(send.shutdown)
channel.shutdown(partial(stat.flushtosql_shutdown,
statsdict=timetotal.statsdict, sqlengine=engine))
channel.shutdown(stat.timedata.shutdown)
# Add flushtosql to timedata event
stat.timedata(stat.flushtosql)
# Run all the tasks
with BaseLoop() as main:
# Schedule SendTimeData coro
asyncio.ensure_future(send())
# Start every event, and ignore events that don't have any tasks
stat.timedata.open()
stat.timedata.start(asyncfunc=False, statsdict=timetotal.statsdict,
flushlimit=5, sqlengine=engine)
asyncio.ensure_future(run())
main.start()
assert timetotal.statsdict.numdata == 0
df = timetotal()
assert isinstance(df, DataFrame)
assert not df.empty
# ============================================================================
# Test FlushToSQL.flushdata()
# ============================================================================
@pytest.mark.parametrize('exctype', [Exception, RuntimeError, ValueError])
def test_flushdata_nodata(sqlengine, exctype):
"""Store error data in sqlite db"""
statsdict = stat.Period()
key = 'hello'
namekey = 'world'
sqltbl = 'period'
tblname = '{}_{}'.format(sqltbl, namekey)
end = Timestamp.now(tz='UTC')
delta = to_timedelta(5, unit='s').total_seconds()
err = exctype(42)
data = Series([end, 1/5, delta, repr(err), 1],
index=['end', 'rate', 'response', 'error', 'count'])
statsdict.adderror(key, data)
# Send data to sqlite db
with sqlengine.begin() as conn:
stat.flushtosql.flushdata(statsdict, key, sqltbl, namekey, sqlengine,
conn)
# Check sqlite db
with sqlengine.begin() as conn:
assert not sqlengine.dialect.has_table(conn, tblname)
# ============================================================================
# Test FlushToSQL.flusherror()
# ============================================================================
@pytest.mark.parametrize('exctype', [Exception, RuntimeError, ValueError])
def test_flusherror(sqlengine, exctype):
"""Store error data in sqlite db"""
statsdict = stat.Period()
key = 'hello'
namekey = 'world'
sqltbl = 'period'
tblname = '{}_error_{}'.format(sqltbl, namekey)
end = Timestamp.now(tz='UTC')
delta = to_timedelta(5, unit='s').total_seconds()
err = exctype(42)
data = Series([end, 1/5, delta, repr(err), 1],
index=['end', 'rate', 'response', 'error', 'count'])
statsdict.adderror(key, data)
# Send data to sqlite db
with sqlengine.begin() as conn:
stat.flushtosql.flusherror(statsdict, key, sqltbl, namekey, sqlengine,
conn)
# Check sqlite db
with sqlengine.begin() as conn:
assert sqlengine.dialect.has_table(conn, tblname)
df = read_sql_table(tblname, conn, index_col='index',
parse_dates={'end': dict(utc=True)})
assert len(df.index) == 1
assert df.iloc[0].error == repr(err)
# ============================================================================
# Test FlushToSQL.flushfailure()
# ============================================================================
def test_flushfailure(sqlengine):
"""Store error data in sqlite db"""
statsdict = stat.Period()
key = 'hello'
namekey = 'world'
sqltbl = 'period'
tblname = '{}_failure_{}'.format(sqltbl, namekey)
end = Timestamp.now(tz='UTC')
delta = to_timedelta(5, unit='s').total_seconds()
err = stat.Failure(42)
data = Series([end, 1/5, delta, str(err.args[0]), 1],
index=['end', 'rate', 'response', 'failure', 'count'])
statsdict.addfailure(key, data)
# Send data to sqlite db
with sqlengine.begin() as conn:
stat.flushtosql.flushfailure(statsdict, key, sqltbl, namekey,
sqlengine, conn)
# Check sqlite db
with sqlengine.begin() as conn:
assert sqlengine.dialect.has_table(conn, tblname)
df = read_sql_table(tblname, conn, index_col='index',
parse_dates={'end': dict(utc=True)})
assert len(df.index) == 1
assert df.iloc[0].failure == str(err.args[0])
# ============================================================================
#
# ============================================================================
|
arielmakestuff/loadlimit
|
test/unit/stat/test_flushtosql.py
|
Python
|
mit
| 7,022
|
#!/usr/bin/env python3
# Copyright (c) 2016-2019 The Bitcoin Core developers
# Distributed under the MIT software license, see the accompanying
# file COPYING or http://www.opensource.org/licenses/mit-license.php.
"""Test processing of feefilter messages."""
from decimal import Decimal
import time
from test_framework.messages import msg_feefilter
from test_framework.mininode import mininode_lock, P2PInterface
from test_framework.test_framework import BitcoinTestFramework
def hashToHex(hash):
return format(hash, '064x')
# Wait up to 60 secs to see if the testnode has received all the expected invs
def allInvsMatch(invsExpected, testnode):
for x in range(60):
with mininode_lock:
if (sorted(invsExpected) == sorted(testnode.txinvs)):
return True
time.sleep(1)
return False
class TestP2PConn(P2PInterface):
def __init__(self):
super().__init__()
self.txinvs = []
def on_inv(self, message):
for i in message.inv:
if (i.type == 1):
self.txinvs.append(hashToHex(i.hash))
def clear_invs(self):
with mininode_lock:
self.txinvs = []
class FeeFilterTest(BitcoinTestFramework):
def set_test_params(self):
self.num_nodes = 2
self.extra_args = [[
"-mintxfee=0.00001"
] for i in range(self.num_nodes)]
def skip_test_if_missing_module(self):
self.skip_if_no_wallet()
def run_test(self):
node1 = self.nodes[1]
node0 = self.nodes[0]
# Get out of IBD
node1.generate(1)
self.sync_blocks()
self.nodes[0].add_p2p_connection(TestP2PConn())
# Test that invs are received for all txs at feerate of 20 sat/byte
node1.settxfee(Decimal("0.00020000"))
txids = [node1.sendtoaddress(node1.getnewaddress(), 1) for x in range(3)]
assert allInvsMatch(txids, self.nodes[0].p2p)
self.nodes[0].p2p.clear_invs()
# Set a filter of 15 sat/byte
self.nodes[0].p2p.send_and_ping(msg_feefilter(15000))
# Test that txs are still being received (paying 20 sat/byte)
txids = [node1.sendtoaddress(node1.getnewaddress(), 1) for x in range(3)]
assert allInvsMatch(txids, self.nodes[0].p2p)
self.nodes[0].p2p.clear_invs()
# Change tx fee rate to 10 sat/byte and test they are no longer received
node1.settxfee(Decimal("0.00010000"))
[node1.sendtoaddress(node1.getnewaddress(), 1) for x in range(3)]
self.sync_mempools() # must be sure node 0 has received all txs
# Send one transaction from node0 that should be received, so that we
# we can sync the test on receipt (if node1's txs were relayed, they'd
# be received by the time this node0 tx is received). This is
# unfortunately reliant on the current relay behavior where we batch up
# to 35 entries in an inv, which means that when this next transaction
# is eligible for relay, the prior transactions from node1 are eligible
# as well.
node0.settxfee(Decimal("0.00020000"))
txids = [node0.sendtoaddress(node0.getnewaddress(), 1)]
assert allInvsMatch(txids, self.nodes[0].p2p)
self.nodes[0].p2p.clear_invs()
# Remove fee filter and check that txs are received again
self.nodes[0].p2p.send_and_ping(msg_feefilter(0))
txids = [node1.sendtoaddress(node1.getnewaddress(), 1) for x in range(3)]
assert allInvsMatch(txids, self.nodes[0].p2p)
self.nodes[0].p2p.clear_invs()
if __name__ == '__main__':
FeeFilterTest().main()
|
FeatherCoin/Feathercoin
|
test/functional/p2p_feefilter.py
|
Python
|
mit
| 3,641
|
import datetime
import os
from time import strftime
from reportlab.lib.colors import black
from reportlab.lib.pagesizes import letter
from reportlab.lib.enums import TA_JUSTIFY
from reportlab.pdfbase import pdfmetrics
from reportlab.platypus import Paragraph, Spacer, Table, PageTemplate, \
BaseDocTemplate, Frame, Image
from reportlab.lib.styles import getSampleStyleSheet, ParagraphStyle
from reportlab.lib.units import inch
from reportlab.platypus.flowables import HRFlowable
import requests
report_time = datetime.datetime.now().ctime()
class Report:
title = "Canary Cyber Security Report"
def __init__(self, hosts, router_status=True):
global report_time
self.hosts = hosts
self.router_status = router_status
self.title = str(report_time) + ".pdf"
self.template = BaseDocTemplate("reports/" + self.title,
pageTemplates=[PageTemplate(id="first_page", frames=[
Frame(inch, inch, 6.5 * inch, 9 * inch, topPadding=.6 * inch,
showBoundary=0)], onPage=FirstPageSetup, pagesize=letter)],
rightMargin=72, leftMargin=72,
topMargin=72, bottomMargin=18,
font='Courier', fontSize=12)
self.styles = getSampleStyleSheet()
@classmethod
def get_nvd_url(cls, host):
manufacturer_str = host.manufacturer
if " " in manufacturer_str:
manufacturer_str = manufacturer_str.split(" ")[0]
url = "https://web.nvd.nist.gov/view/vuln/search-results?query=%s&search_type=all&cves=on" % (manufacturer_str)
return url
def generate(self):
Story = []
self.styles.add(ParagraphStyle(name='Justify', alignment=TA_JUSTIFY))
# title = "<font size=24><b>%s</b></font>" % (Report.title)
# Story.append(Paragraph(title, self.styles['Normal']))
Story.append(Spacer(1, 12))
string = """<font size=12>Here is a status report about your home network.
Please review it and follow up with appropriate actions.
Remember that over two third of personal network intrusions occur
due to weak passwords on your routers and devices.</font>"""
Story.append(Paragraph(string, self.styles['Normal']))
Story.append(Spacer(2, 12))
sorted(self.hosts, key=type, reverse=True)
for i, host in enumerate(self.hosts):
if i == 0:
Story.append(Image('res/wifi.png'))
string = "<font size=16><b>Your router:</b></font>"
Story.append(Paragraph(string, self.styles['Normal']))
Story.append(Spacer(1, 16))
if not self.router_status:
string = """<font color=red size=12><b>Admin Page Status: </b>Not secured<br/><br/>Please change \
the password for your router admin page <a href='%s' color=blue>here.</a></font>""" % str(
"http://" + host.ip)
else:
string = "<font size=12><b>Admin Page Status: </b>" + "Secure</font>"
Story.append(Paragraph(string, self.styles['Normal']))
Story.append(Spacer(1, 16))
if i == 1:
Story.append(Image('res/map.png'))
string = "<font size=16><b>Your devices:</b></font>"
Story.append(Paragraph(string, self.styles['Normal']))
Story.append(Spacer(1, 16))
string = "<font size=12><b>MAC Address:</b> " + host.mac_address + "</font>"
Story.append(Paragraph(string, self.styles['Normal']))
Story.append(Spacer(1, 12))
string = "<font size=12><b>IP Address:</b> " + host.ip + "</font>"
Story.append(Paragraph(string, self.styles['Normal']))
Story.append(Spacer(1, 12))
string = "<font size=12><b>Manufacturer:</b> " + host.manufacturer + "</font>"
Story.append(Paragraph(string, self.styles['Normal']))
Story.append(Spacer(1, 12))
if host.manufacturer != "Unknown":
string = """<font size=12><b>Major issues associated with manufacturer:</b>
Please check out the <a href='%s' color='blue'>National Vulnerability
Database</a> for a list of current issues related to %s products.</font>
""" % (Report.get_nvd_url(host), host.manufacturer)
Story.append(Paragraph(string, self.styles['Normal']))
Story.append(Spacer(1, 12))
string = "<font size=12><b>Number of open ports:</b> " + str(len(host.open_ports)) + "</font>"
if len(host.open_ports) > 0:
if len(host.open_ports) > 2:
string += "<font color=red><br/>Please seek ways to close more ports.</font>"
Story.append(Paragraph(string, self.styles['Normal']))
Story.append(Spacer(1, 12))
port_data = [["NUMBER", "SERVICE", "NOTES"]]
L = 1
for port in host.open_ports:
row = [port.number, port.port_status]
port_site = "http://www.speedguide.net/port.php?port=%s" % port.number
google_port = "https://www.google.com/#q=port+%s" % port.number
check_site = requests.get(port_site)
if check_site.status_code == 200:
link_txt = 'More info here'
link = """<a href="%s">%s</a>""" % (port_site,link_txt)
else:
link_txt = 'Unknown port, search here'
link = """<a href="%s">%s</a>""" % (google_port,link_txt)
if pdfmetrics.stringWidth(link_txt,'Courier',12) > L:
L = pdfmetrics.stringWidth(link_txt,'Courier',12) + 1
inside_Table = Table([[Paragraph(link,self.styles['Normal'])]], colWidths=L)
row.append(inside_Table)
port_data.append(row)
port_table = Table(port_data)
Story.append(port_table)
Story.append(Spacer(3, 12))
if i < len(self.hosts) - 1:
Story.append(HRFlowable(color=black))
Story.append(Spacer(1, 12))
self.template.build(Story)
def FirstPageSetup(canvas, doc):
global report_time
canvas.saveState()
#Background image
canvas.drawImage(os.path.curdir + "/res/report_base.png", 0, 0, width=8.5 * inch, height=11.0 * inch)
#Footer
canvas.drawString(0.5 * inch, 0.5 * inch, 'Canary Security Report for ' + str(report_time))
canvas.drawRightString(8.0 * inch, 0.5 * inch, 'Page %d' % (doc.page))
canvas.restoreState()
class Host:
count = 0
def __init__(self, os="Unknown", ip="Unknown", manufacturer="Unknown", mac_address="Unknown", open_ports=[],
is_down=False):
self.os = os
self.ip = ip
self.manufacturer = manufacturer
self.open_ports = open_ports
self.mac_address = mac_address
self.is_down = False
Host.count += 1
def add_port(self, port):
self.open_ports.append(port)
@classmethod
def flag_router(cls, hosts):
for each in hosts:
return True
@classmethod
def return_num_hosts(cls):
return str(cls.count)
def display_summary(self):
port_string = ""
if len(self.open_ports) > 0:
for each in self.open_ports:
port_string += str(each.number) + " "
else:
port_string = "None detected yet."
print "OS: ", self.os, ",manufacturer: ", self.manufacturer, ", MAC Address: ", self.mac_address, ", Ports: ", port_string
class Router(Host):
def __init__(self, is_secured=True, os="Unknown", ip="Unknown", manufacturer="Unknown", mac_address="Unknown",
open_ports=[], is_down=False):
Host.__init__(self, os, ip, manufacturer, mac_address, open_ports, is_down)
self.is_secured = is_secured
class Port:
def __init__(self, number, port_service, port_status):
self.number = number
self.port_service = port_service
self.port_status = port_status
|
adriansoghoian/security-at-home
|
models.py
|
Python
|
mit
| 8,450
|
#!/usr/bin/env python
#
# import needed modules.
# pyzabbix is needed, see https://github.com/lukecyca/pyzabbix
#
import argparse
import ConfigParser
import os
import os.path
import sys
import distutils.util
from xml.etree import ElementTree as ET
from pyzabbix import ZabbixAPI
# define config helper function
def ConfigSectionMap(section):
dict1 = {}
options = Config.options(section)
for option in options:
try:
dict1[option] = Config.get(section, option)
if dict1[option] == -1:
DebugPrint("skip: %s" % option)
except:
print("exception on %s!" % option)
dict1[option] = None
return dict1
def PrintError(error):
if args.continue_on_error:
sys.stderr.write(error + '\n')
else:
sys.exit(error)
# set default vars
defconf = os.getenv("HOME") + "/.zbx.conf"
username = ""
password = ""
api = ""
noverify = ""
# Define commandline arguments
parser = argparse.ArgumentParser(formatter_class=argparse.RawDescriptionHelpFormatter,description='Imports Zabbix templates from XML files. Default behaviour is to add missing elements and update any existing elements. Optionally elements that are missing from the XML file can be removed from existing template(s) as well.', epilog="""
This program can use .ini style configuration files to retrieve the needed API connection information.
To use this type of storage, create a conf file (the default is $HOME/.zbx.conf) that contains at least the [Zabbix API] section and any of the other parameters:
[Zabbix API]
username=johndoe
password=verysecretpassword
api=https://zabbix.mycompany.com/path/to/zabbix/frontend/
no_verify=true
""")
parser.add_argument('-u', '--username', help='User for the Zabbix api')
parser.add_argument('-p', '--password', help='Password for the Zabbix api user')
parser.add_argument('-a', '--api', help='Zabbix API URL')
parser.add_argument('--no-verify', help='Disables certificate validation when using a secure connection',action='store_true')
parser.add_argument('-c','--config', help='Config file location (defaults to $HOME/.zbx.conf)')
parser.add_argument('-v', '--verbose', help='Enables verbose output.',action='store_true')
parser.add_argument('-T', '--templates', help='List of XML template files to import.',required=True, nargs="+")
parser.add_argument('-D', '--delete-missing', help='If a template already exists in Zabbix, any missing elements from the .XML will be removed from Zabbix as well.',action='store_true')
parser.add_argument('-U', '--update', help='If a template already exists in Zabbix, update any changes in template elements.',action='store_true')
parser.add_argument('-C', '--continue-on-error', help='Continue on error, use with caution.',action='store_true')
args = parser.parse_args()
# load config module
Config = ConfigParser.ConfigParser()
Config
# if configuration argument is set, test the config file
if args.config:
if os.path.isfile(args.config) and os.access(args.config, os.R_OK):
Config.read(args.config)
# if not set, try default config file
else:
if os.path.isfile(defconf) and os.access(defconf, os.R_OK):
Config.read(defconf)
# try to load available settings from config file
try:
username=ConfigSectionMap("Zabbix API")['username']
password=ConfigSectionMap("Zabbix API")['password']
api=ConfigSectionMap("Zabbix API")['api']
noverify=bool(distutils.util.strtobool(ConfigSectionMap("Zabbix API")["no_verify"]))
except:
pass
# override settings if they are provided as arguments
if args.username:
username = args.username
if args.password:
password = args.password
if args.api:
api = args.api
if args.no_verify:
noverify = args.no_verify
# test for needed params
if not username:
sys.exit("Error: API User not set")
if not password:
sys.exit("Error: API Password not set")
if not api:
sys.exit("Error: API URL is not set")
# Setup Zabbix API connection
zapi = ZabbixAPI(api)
if noverify is True:
zapi.session.verify = False
# Login to the Zabbix API
zapi.login(username, password)
##################################
# Start actual API logic
##################################
# We need the API version to know if valuemap importing is supported
zversion=zapi.apiinfo.version()
# set import modes
create=True
if args.update:
update=True
else:
update=False
if args.delete_missing:
delete=True
else:
delete=False
# set import rules, see https://www.zabbix.com/documentation/3.0/manual/api/reference/configuration/import
rules={}
rules['templates']={'createMissing': create, 'updateExisting':update}
rules['applications']={'createMissing': create, 'updateExisting': update, 'deleteMissing': delete}
rules['discoveryRules']={'createMissing': create, 'updateExisting': update, 'deleteMissing': delete}
rules['graphs']={'createMissing': create, 'updateExisting':update, 'deleteMissing': delete}
rules['groups']={'createMissing': create}
rules['items']={'createMissing': create, 'updateExisting':update, 'deleteMissing': delete}
rules['templateLinkage']={'createMissing': create}
rules['templateScreens']={'createMissing': create, 'updateExisting':update, 'deleteMissing': delete}
rules['triggers']={'createMissing': create, 'updateExisting':update, 'deleteMissing': delete}
# Valuemap imports are a Zabbix 3.x.x feature
if zversion.startswith('3.'):
rules['valueMaps']={'createMissing':create, 'updateExisting':update}
# Parse file list
for template in args.templates:
Continue=True
if Continue:
try:
# open file for reading
with file(template) as f:
xml = f.read()
f.close()
except:
# If the file can't be opened, exit with error
error="Error: Something went wrong when trying to read the file" + template
PrintError(error)
Continue=False
if Continue:
try:
# verify if the file is a valid XML
tree = ET.fromstring(xml)
except:
# If the file can't isn't a valid XML, exit with error
error="Error: XML is not valid in " + template
PrintError(error)
Continue=False
if Continue:
try:
# Everything looks good, let's try to import
result=zapi.configuration.Import(format="xml",rules=rules, source=xml)
if args.verbose:
print("Succesfully imported " + template)
except:
# Something went wrong with the API call or import
error="Error: Something went wrong while importing " + template + "\n" + str(sys.exc_info()[1][0])
PrintError(error)
Continue=False
# And we're done...
|
q1x/zabbix-gnomes
|
ztmplimport.py
|
Python
|
mit
| 6,615
|
__author__ = 'ilblackdragon@gmail.com'
from pymisc import log, decorators
class RegisterSystem(object):
interfaces = []
classes = []
@classmethod
@decorators.logprint(log)
def register(self, cls):
if cls.__name__[0] == 'I':
print("Regirstring interface `%s`" % cls.__name__)
RegisterSystem.interfaces.append(cls)
else:
print("Regirstring class `%s`" % cls.__name__)
RegisterSystem.classes.append(cls)
class InterfaceMeta(type):
def __new__(cls, name, bases, dict):
res = super(InterfaceMeta, cls).__new__(cls, name, bases, dict)
RegisterSystem.register(res)
if name[0] == 'I':
res.children = []
else:
for b in bases:
if b.children is not None:
b.children.append(res)
return res
class IBase(object):
__metaclass__ = InterfaceMeta
@classmethod
def send_signal(cls, method, *args, **kwargs):
res = []
for child in cls.children:
res.append((child, child.__dict__[method].__get__('')(*args, **kwargs)))
return res
|
ilblackdragon/pymisc
|
pymisc/abstract.py
|
Python
|
mit
| 1,152
|
import openmc
###############################################################################
# Simulation Input File Parameters
###############################################################################
# OpenMC simulation parameters
batches = 20
inactive = 10
particles = 10000
###############################################################################
# Exporting to OpenMC materials.xml File
###############################################################################
# Instantiate some Nuclides
h1 = openmc.Nuclide('H-1')
o16 = openmc.Nuclide('O-16')
u235 = openmc.Nuclide('U-235')
# Instantiate some Materials and register the appropriate Nuclides
fuel = openmc.Material(material_id=1, name='fuel')
fuel.set_density('g/cc', 4.5)
fuel.add_nuclide(u235, 1.)
moderator = openmc.Material(material_id=2, name='moderator')
moderator.set_density('g/cc', 1.0)
moderator.add_nuclide(h1, 2.)
moderator.add_nuclide(o16, 1.)
moderator.add_s_alpha_beta('HH2O', '71t')
# Instantiate a MaterialsFile, register all Materials, and export to XML
materials_file = openmc.MaterialsFile()
materials_file.default_xs = '71c'
materials_file.add_materials([moderator, fuel])
materials_file.export_to_xml()
###############################################################################
# Exporting to OpenMC geometry.xml File
###############################################################################
# Instantiate Surfaces
left = openmc.XPlane(surface_id=1, x0=-2, name='left')
right = openmc.XPlane(surface_id=2, x0=2, name='right')
bottom = openmc.YPlane(surface_id=3, y0=-2, name='bottom')
top = openmc.YPlane(surface_id=4, y0=2, name='top')
fuel1 = openmc.ZCylinder(surface_id=5, x0=0, y0=0, R=0.4)
fuel2 = openmc.ZCylinder(surface_id=6, x0=0, y0=0, R=0.3)
fuel3 = openmc.ZCylinder(surface_id=7, x0=0, y0=0, R=0.2)
left.boundary_type = 'vacuum'
right.boundary_type = 'vacuum'
top.boundary_type = 'vacuum'
bottom.boundary_type = 'vacuum'
# Instantiate Cells
cell1 = openmc.Cell(cell_id=1, name='Cell 1')
cell2 = openmc.Cell(cell_id=2, name='Cell 2')
cell3 = openmc.Cell(cell_id=101, name='cell 3')
cell4 = openmc.Cell(cell_id=102, name='cell 4')
cell5 = openmc.Cell(cell_id=201, name='cell 5')
cell6 = openmc.Cell(cell_id=202, name='cell 6')
cell7 = openmc.Cell(cell_id=301, name='cell 7')
cell8 = openmc.Cell(cell_id=302, name='cell 8')
# Use surface half-space to define regions
cell1.region = +left & -right & +bottom & -top
cell2.region = +left & -right & +bottom & -top
cell3.region = -fuel1
cell4.region = +fuel1
cell5.region = -fuel2
cell6.region = +fuel2
cell7.region = -fuel3
cell8.region = +fuel3
# Register Materials with Cells
cell3.fill = fuel
cell4.fill = moderator
cell5.fill = fuel
cell6.fill = moderator
cell7.fill = fuel
cell8.fill = moderator
# Instantiate Universe
univ1 = openmc.Universe(universe_id=1)
univ2 = openmc.Universe(universe_id=2)
univ3 = openmc.Universe(universe_id=3)
univ4 = openmc.Universe(universe_id=5)
root = openmc.Universe(universe_id=0, name='root universe')
# Register Cells with Universe
univ1.add_cells([cell3, cell4])
univ2.add_cells([cell5, cell6])
univ3.add_cells([cell7, cell8])
root.add_cell(cell1)
univ4.add_cell(cell2)
# Instantiate nested Lattices
lattice1 = openmc.RectLattice(lattice_id=4, name='4x4 assembly')
lattice1.dimension = [2, 2]
lattice1.lower_left = [-1., -1.]
lattice1.pitch = [1., 1.]
lattice1.universes = [[univ1, univ2],
[univ2, univ3]]
lattice2 = openmc.RectLattice(lattice_id=6, name='4x4 core')
lattice2.dimension = [2, 2]
lattice2.lower_left = [-2., -2.]
lattice2.pitch = [2., 2.]
lattice2.universes = [[univ4, univ4],
[univ4, univ4]]
# Fill Cell with the Lattice
cell1.fill = lattice2
cell2.fill = lattice1
# Instantiate a Geometry and register the root Universe
geometry = openmc.Geometry()
geometry.root_universe = root
# Instantiate a GeometryFile, register Geometry, and export to XML
geometry_file = openmc.GeometryFile()
geometry_file.geometry = geometry
geometry_file.export_to_xml()
###############################################################################
# Exporting to OpenMC settings.xml File
###############################################################################
# Instantiate a SettingsFile, set all runtime parameters, and export to XML
settings_file = openmc.SettingsFile()
settings_file.batches = batches
settings_file.inactive = inactive
settings_file.particles = particles
settings_file.set_source_space('box', [-1, -1, -1, 1, 1, 1])
settings_file.export_to_xml()
###############################################################################
# Exporting to OpenMC plots.xml File
###############################################################################
plot = openmc.Plot(plot_id=1)
plot.origin = [0, 0, 0]
plot.width = [4, 4]
plot.pixels = [400, 400]
plot.color = 'mat'
# Instantiate a PlotsFile, add Plot, and export to XML
plot_file = openmc.PlotsFile()
plot_file.add_plot(plot)
plot_file.export_to_xml()
###############################################################################
# Exporting to OpenMC tallies.xml File
###############################################################################
# Instantiate a tally mesh
mesh = openmc.Mesh(mesh_id=1)
mesh.type = 'regular'
mesh.dimension = [4, 4]
mesh.lower_left = [-2, -2]
mesh.width = [1, 1]
# Instantiate tally Filter
mesh_filter = openmc.Filter()
mesh_filter.mesh = mesh
# Instantiate the Tally
tally = openmc.Tally(tally_id=1)
tally.add_filter(mesh_filter)
tally.add_score('total')
# Instantiate a TalliesFile, register Tally/Mesh, and export to XML
tallies_file = openmc.TalliesFile()
tallies_file.add_mesh(mesh)
tallies_file.add_tally(tally)
tallies_file.export_to_xml()
|
mjlong/openmc
|
examples/python/lattice/nested/build-xml.py
|
Python
|
mit
| 5,849
|
# sublimelint.py
# SublimeLint is a code checking framework for Sublime Text
#
# Project: https://github.com/lunixbochs/sublimelint
# License: MIT
import sublime
import sublime_plugin
import os
from threading import Thread
import time
import json
from .lint.edit import apply_sublimelint_edit
from .lint.edit import Edit
from .lint.modules import Modules
from .lint.linter import Linter
from .lint.highlight import HighlightSet
from .lint.update import update
from .lint import persist
def plugin_loaded():
user_path = os.path.join(sublime.packages_path(), 'User', 'linters')
persist.modules = Modules(user_path).load_all()
persist.reinit()
Thread(target=update, args=(user_path,)).start()
class SublimeLint(sublime_plugin.EventListener):
def __init__(self, *args, **kwargs):
sublime_plugin.EventListener.__init__(self, *args, **kwargs)
self.loaded = set()
self.linted = set()
self.last_syntax = {}
persist.queue.start(self.lint)
# this gives us a chance to lint the active view on fresh install
window = sublime.active_window()
if window:
self.on_activated(window.active_view())
self.start = time.time()
@classmethod
def lint(cls, view_id, callback=None):
callback = callback or cls.finish
view = Linter.get_view(view_id)
sections = {}
for sel, _ in Linter.get_selectors(view_id):
sections[sel] = []
for result in view.find_by_selector(sel):
sections[sel].append(
(view.rowcol(result.a)[0], result.a, result.b)
)
if view is not None:
filename = view.file_name()
code = Linter.text(view)
args = (view_id, filename, code, sections, callback)
Linter.lint_view(*args)
@classmethod
def finish(cls, view, linters):
errors = {}
highlights = HighlightSet()
for linter in linters:
if linter.highlights:
highlights.add(linter.highlights)
if linter.errors:
errors.update(linter.errors)
highlights.clear(view)
highlights.draw(view)
persist.errors[view.id()] = errors
# helpers
def hit(self, view):
self.linted.add(view.id())
if view.size() == 0:
for l in Linter.get_linters(view.id()):
l.clear()
return
persist.queue.hit(view)
def check_syntax(self, view, lint=False):
vid = view.id()
syntax = view.settings().get('syntax')
# syntax either has never been set or just changed
if not vid in self.last_syntax or self.last_syntax[vid] != syntax:
self.last_syntax[vid] = syntax
# assign a linter, then maybe trigger a lint if we get one
if Linter.assign(view) and lint:
self.hit(view)
# callins
def on_modified(self, view):
self.check_syntax(view)
self.hit(view)
def on_modified_async(self, view):
self.on_selection_modified_async(view)
def on_load(self, view):
self.on_new(view)
def on_activated_async(self, view):
persist.reinit()
if not view:
return
self.check_syntax(view, True)
view_id = view.id()
if not view_id in self.linted:
if not view_id in self.loaded:
# it seems on_activated can be called before loaded on first start
if time.time() - self.start < 5: return
self.on_new(view)
self.hit(view)
self.on_selection_modified_async(view)
def on_open_settings(self, view):
# handle opening user preferences file
filename = view.file_name()
if filename:
dirname = os.path.basename(os.path.dirname(filename))
filename = os.path.basename(filename)
if filename != 'SublimeLint.sublime-settings':
return
if dirname.lower() == 'sublimelint':
return
persist.reinit()
settings = persist.settings
# fill in default plugin settings
plugins = settings.pop('plugins', {})
for name, language in persist.languages.items():
default = language.get_settings().copy()
default.update(plugins.pop(name, {}))
plugins[name] = default
settings['plugins'] = plugins
with Edit(view) as edit:
j = json.dumps({'user': settings}, indent=4, sort_keys=True)
j = j.replace(' \n', '\n')
edit.replace(sublime.Region(0, view.size()), j)
view.run_command('save')
def on_new(self, view):
self.on_open_settings(view)
vid = view.id()
self.loaded.add(vid)
self.last_syntax[vid] = view.settings().get('syntax')
Linter.assign(view)
def on_post_save(self, view):
# this will reload linters if they are saved with sublime text
for name, module in persist.modules.items():
if os.name == 'posix' and (
os.stat(module.__file__).st_ino == os.stat(view.file_name()).st_ino
) or module.__file__ == view.file_name():
persist.debug("reloading module '{}'".format(name))
persist.modules.reload(module)
Linter.reload(name)
break
# linting here doesn't matter, because we lint on load and on modify
# self.hit(view)
def on_selection_modified_async(self, view):
vid = view.id()
try:
lineno = view.rowcol(view.sel()[0].end())[0]
except IndexError:
lineno = -1
status = ''
if vid in persist.errors:
errors = persist.errors[vid]
if errors:
plural = 's' if len(errors) > 1 else ''
if lineno in errors:
if plural:
num = sorted(list(errors)).index(lineno) + 1
status += '%i/%i errors: ' % (num, len(errors))
# sublime statusbar can't hold unicode
status += '; '.join(set(errors[lineno]))
else:
status = '%i error%s' % (len(errors), plural)
view.set_status('sublimelint', status)
else:
view.erase_status('sublimelint')
persist.queue.delay()
|
lunixbochs/sublimelint
|
sublimelint.py
|
Python
|
mit
| 6,546
|
import time
from datetime import date
import numpy
from PIL import Image
import zbar
import os,sys
import wx # GUI
# Handle time lapse!
scanner = zbar.ImageScanner()
# configure the reader
scanner.parse_config('enable')
#scanner.set_config(0, zbar.Config.ENABLE, 0)
#scanner.set_config(zbar.Symbol.QRCODE, zbar.Config.ENABLE, 1)
label = ""
# TODO
# Read label better (crop enhance contrast etc...)
# copy files
# record previous file
def readQRCODE(ImageFile):
label = ""
pilImage = Image.open(ImageFile)
width, height = pilImage.size
pilImage = pilImage.crop((int(0.18*width), int(0.2*height),int(0.97*width), int(0.95*height)))
pilImage = pilImage.convert('L')
width, height = pilImage.size
raw = pilImage.tostring()
# wrap image data
image = zbar.Image(width, height, 'Y800', raw)
# scan the image for barcodes
scanner.scan(image)
# extract results
for symbol in image:
label = symbol.data
# clean up
del(image)
return label
class MainWindow(wx.Frame):
def __init__(self, parent, title):
wx.Frame.__init__(self, parent, title=title, size=(400,300))
self.control = wx.TextCtrl(self, style=wx.TE_MULTILINE | wx.TE_READONLY)
self.CreateStatusBar() # A Statusbar in the bottom of the window
# FOlders
self.dest_folder = os.path.dirname(sys.argv[0])
self.root_folder = os.path.dirname(sys.argv[0])
# Setting up the menu.
filemenu= wx.Menu()
# wx.ID_ABOUT and wx.ID_EXIT are standard IDs provided by wxWidgets.
filemenu.Append(1, "&Base Folders"," Set folders")
filemenu.Append(2, "&Run"," scan for files")
filemenu.Append(wx.ID_ABOUT, "&About"," Information about this program")
filemenu.AppendSeparator()
filemenu.Append(wx.ID_EXIT,"E&xit"," Terminate the program")
# xxxxxxx
self.Bind(wx.EVT_MENU, self.get_folder, id=1)
self.Bind(wx.EVT_MENU, self.scan_data, id=2)
# Creating the menubar.
menuBar = wx.MenuBar()
menuBar.Append(filemenu,"&File") # Adding the "filemenu" to the MenuBar
self.SetMenuBar(menuBar) # Adding the MenuBar to the Frame content.
self.Show(True)
def get_folder(self, id):
dlg = wx.DirDialog(self, "Choose Root Folder:")
if dlg.ShowModal() == wx.ID_OK:
self.root_folder = dlg.GetPath()
dlg.Destroy()
def scan_data(self, id):
#################################################################
# Find all suitable files in the current folder
#################################################################
dir = self.root_folder#os.path.dirname(sys.argv[0])
sys.path.append(dir)
for f in os.listdir(dir):
file, ext = os.path.splitext(f) # Handles no-extension files, etc.
if ext == '.JPG':
base_row = file.split("-")
base = base_row[0]
if len(base_row) == 1:
if os.path.isfile(dir+ "\\"+ file + "-QR-.jpg"):
genotype = readQRCODE(dir+ "\\"+ file + "-QR-.jpg")
# image properties
file_tmp1 = file.split('_')
file_id = file_tmp1[1]
#os.path.getmtime(dir+ "\\"+ file +ext)
# Image identifiers
identifyer = [None,None,None]
if len(genotype) > 5:
text = "Root directory: " + dir + "\n"
text += "File: " + file + "\n"
text += "Genotype: " + genotype
self.control.SetValue(text)
wx.Yield()
identifyer = genotype.split('_')
else:
pilImage = Image.open(dir+ "\\"+ file + "-QR-.jpg")
width, height = pilImage.size
pilImage = pilImage.crop((int(0.18*width), int(0.3*height),int(0.97*width), int(0.92*height)))
width, height = pilImage.size
sc = 0.6
pilImage = pilImage.resize((int(width*sc),int(height*sc)), Image.ANTIALIAS)
img = wx.EmptyImage( *pilImage.size )
pilImageCopy = pilImage.copy()
pilImageCopyRGB = pilImageCopy.convert( 'RGB' ) # Discard any alpha from the PIL image.
pilImageRgbData =pilImageCopyRGB.tostring()
img.SetData( pilImageRgbData )
identifyer_length = 0
while identifyer_length>-1:# !=3:
dlg = wx.TextEntryDialog(self, 'Type "Species Population Id" with space as separation', 'Could not read bar code', '')
dlg.SetValue("")
self.pnl = MyFrame(dlg, -1, "Label not read", size=(int(width*sc),int(height*sc)), pos = (800,100), style = wx.DEFAULT_FRAME_STYLE, pic = img)
self.pnl.Show(True)
if dlg.ShowModal() == wx.ID_OK:
txtvalue = dlg.GetValue() #genotype.split('_')
identifyer = txtvalue.split(' ')
identifyer_length = len(identifyer)
dlg.Destroy()
else:
text = "!!! Could not recover barcode for !!! :\n\n"
text += "Root directory: " + dir + "\n"
text += "File: " + file + "\n"
self.control.SetValue(text)
wx.Yield()
class MyFrame(wx.Frame):
def __init__(
self, parent, ID, title, pos=wx.DefaultPosition,
size=wx.DefaultSize, style=wx.DEFAULT_FRAME_STYLE, pic = None
):
wx.Frame.__init__(self, parent, ID, title, pos, size, style)
panel = wx.Panel(self, -1)
wx.StaticBitmap(panel, -1, pic.ConvertToBitmap(), (0, 0))
def OnCloseMe(self, event):
self.Close(True)
def OnCloseWindow(self, event):
self.Destroy()
app = wx.App(False)
frame = MainWindow(None, "Sample editor")
app.MainLoop()
|
LionelDupuy/ARCHI_PHEN
|
ImageJ/DatabaseInput_deprecated.py
|
Python
|
mit
| 5,558
|
import requests
import json
from decimal import Decimal
from i3pystatus import IntervalModule
from i3pystatus.core.util import internet, require
class Coin(IntervalModule):
"""
Fetches live data of all cryptocurrencies available at coinmarketcap <https://coinmarketcap.com/>.
Coin setting should be equal to the 'id' field of your coin in <https://api.coinmarketcap.com/v1/ticker/>.
Example coin settings: bitcoin, bitcoin-cash, ethereum, litecoin, dash, lisk.
Example currency settings: usd, eur, huf.
.. rubric:: Available formatters
* {symbol}
* {price}
* {rank}
* {24h_volume}
* {market_cap}
* {available_supply}
* {total_supply}
* {max_supply}
* {percent_change_1h}
* {percent_change_24h}
* {percent_change_7d}
* {last_updated} - time of last update on the API's part
* {status}
"""
settings = (
("format", "format string used for output."),
("color"),
("coin", "cryptocurrency to fetch"),
("decimal", "round coin price down to this decimal place"),
("currency", "fiat currency to show fiscal data"),
("symbol", "coin symbol"),
("interval", "update interval in seconds"),
("status_interval", "percent change status in the last: '1h' / '24h' / '7d'")
)
symbol = "¤"
color = None
format = "{symbol} {price}{status}"
coin = "ethereum"
currency = "USD"
interval = 600
status_interval = "24h"
decimal = 2
def fetch_data(self):
response = requests.get("https://api.coinmarketcap.com/v1/ticker/{}/?convert={}".format(self.coin, self.currency))
coin_data = response.json()[0]
coin_data["price"] = coin_data.pop("price_{}".format(self.currency.lower()))
coin_data["24h_volume"] = coin_data.pop("24h_volume_{}".format(self.currency.lower()))
coin_data["market_cap"] = coin_data.pop("market_cap_{}".format(self.currency.lower()))
coin_data["symbol"] = self.symbol
return coin_data
def set_status(self, change):
if change > 10:
return '⮅'
elif change > 0:
return '⭡'
elif change < -10:
return '⮇'
elif change < 0:
return '⭣'
else:
return ''
@require(internet)
def run(self):
fdict = self.fetch_data()
symbols = dict(bitcoin='฿', ethereum='Ξ', litecoin='Ł', dash='Đ')
if self.coin in symbols:
fdict["symbol"] = symbols[self.coin]
fdict["status"] = self.set_status(float(fdict["percent_change_{}".format(self.status_interval)]))
fdict["price"] = str(round(Decimal(fdict["price"]), self.decimal))
self.data = fdict
self.output = {"full_text": self.format.format(**fdict)}
if self.color is not None:
self.output['color'] = self.color
|
enkore/i3pystatus
|
i3pystatus/coin.py
|
Python
|
mit
| 2,896
|
#!/usr/bin/env python
# -*- coding: utf-8 -*-
"""from python cookbook 2rd edition"""
import sys
class Progressbar(object):
def __init__(self, finalcount, block_char = "."):
self.finalcount = finalcount
self.blockcount = 0
self.block = block_char
self.f = sys.stdout
if not self.finalcount: return
self.f.write("\n--------------------- % Progress ----------------1\n")
self.f.write(" 1 2 3 4 5 6 7 8 9 0\n")
self.f.write("----0----0----0----0----0----0----0----0----0----0\n")
def progress(self, count):
count = min(count, self.finalcount)
if self.finalcount:
percentcomplete = int(round(100.0 * count / self.finalcount))
if percentcomplete < 1: percentcomplete = 1
else:
percentcomplete = 100
blockcount = int(percentcomplete // 2)
if blockcount <= self.blockcount:
return
for i in range(self.blockcount, blockcount):
self.f.write(self.block)
self.f.flush()
self.blockcount = blockcount
if percentcomplete == 100:
self.f.write("\n")
def main():
from time import sleep
print "test:"
pb = Progressbar(8, "*")
for count in range(1, 9):
pb.progress(count)
sleep(0.2)
print "test 100:"
pb = Progressbar(100)
pb.progress(20)
sleep(0.3)
pb.progress(40)
sleep(0.3)
pb.progress(55)
sleep(0.3)
pb.progress(90)
sleep(0.3)
pb.progress(100)
sleep(0.3)
print "test 1:"
pb = Progressbar(1)
pb.progress(1)
if __name__=="__main__":
main()
|
ptrsxu/snippetpy
|
shdisplay/progressbar.py
|
Python
|
mit
| 1,671
|
import os
from onitu.api import Plug, ServiceError, DriverError
# A dummy library supposed to watch the file system
from fsmonitor import FSWatcher
plug = Plug()
@plug.handler()
def get_chunk(metadata, offset, size):
try:
with open(metadata.filename, 'rb') as f:
f.seek(offset)
return f.read(size)
except IOError as e:
raise ServiceError(
"Error reading '{}': {}".format(metadata.filename, e)
)
@plug.handler()
def upload_chunk(metadata, offset, chunk):
try:
with open(metadata.filename, 'r+b') as f:
f.seek(offset)
f.write(chunk)
except IOError as e:
raise ServiceError(
"Error writting '{}': {}".format(metadata.filename, e)
)
@plug.handler()
def end_upload(metadata):
metadata.revision = os.path.getmtime(metadata.filename)
metadata.write_revision()
class Watcher(FSWatcher):
def on_update(self, filename):
"""Called each time an update of a file is detected
"""
metadata = plug.get_metadata(filename)
metadata.revision = os.path.getmtime(metadata.filename)
metadata.size = os.path.getsize(metadata.filename)
plug.update_file(metadata)
def check_changes(self):
"""Check the changes on the file system since the last launch
"""
for filename in self.files:
revision = os.path.getmtime(filename)
metadata = plug.get_metadata(filename)
# If the file is more recent
if revision > metadata.revision:
metadata.revision = os.path.getmtime(metadata.filename)
metadata.size = os.path.getsize(metadata.filename)
plug.update_file(metadata)
def start():
try:
root = plug.options['root']
os.chdir(root)
except OSError as e:
raise DriverError("Can't access '{}': {}".format(root, e))
watcher = Watcher(root)
watcher.check_changes()
watcher.start()
plug.listen()
|
onitu/onitu
|
docs/examples/driver.py
|
Python
|
mit
| 2,035
|
from django.test import TestCase
from test_app.models import MonitorAllFields, MonitorSomeFields
class MonitoredModelGetChanges(TestCase):
def test_monitor_all_fields_no_changes___result_is_empty_dict(self):
m = MonitorAllFields(first_field=1)
self.assertEqual({}, m.get_changes())
def test_monitor_all_fields_all_fields_changed___result_contains_the_old_and_new_values_for_all_fields(self):
m = MonitorAllFields(first_field=1, second_field=3)
m.first_field = 2
m.second_field = 4
self.assertEqual({
"first_field": {"old": 1, "new": 2},
"second_field": {"old": 3, "new": 4}
}, m.get_changes())
def test_monitor_all_fields_single_field_changed___result_contains_the_changed_fields_old_and_new_values(self):
m = MonitorAllFields(first_field=1, second_field=3)
m.first_field = 2
self.assertEqual({
"first_field": {"old": 1, "new": 2},
}, m.get_changes())
def test_monitor_some_fields_change_monitored_fields___result_contains_the_changed_fields_old_and_new_values(self):
m = MonitorSomeFields(monitored_field=1)
m.monitored_field = 2
self.assertEqual({
"monitored_field": {"old": 1, "new": 2},
}, m.get_changes())
def test_monitor_some_fields_change_unmonitored_fields___result_is_empty_dict(self):
m = MonitorSomeFields(unmonitored_field=1)
m.unmonitored_field = 2
self.assertEqual({}, m.get_changes())
|
OmegaDroid/django-model-monitor
|
src/test_app/test/unit/test_monitored_model_get_changes.py
|
Python
|
mit
| 1,523
|
"""
__graph_MT_post__Model_T.py___________________________________________________________
Automatically generated graphical appearance ---> MODIFY DIRECTLY WITH CAUTION
_____________________________________________________________________________
"""
import tkFont
from graphEntity import *
from GraphicalForm import *
from ATOM3Constraint import *
class graph_MT_post__Model_T(graphEntity):
def __init__(self, x, y, semObject = None):
self.semanticObject = semObject
self.sizeX, self.sizeY = 173, 91
graphEntity.__init__(self, x, y)
self.ChangesAtRunTime = 0
self.constraintList = []
if self.semanticObject: atribs = self.semanticObject.attributesToDraw()
else: atribs = None
self.graphForms = []
self.imageDict = self.getImageDict()
def DrawObject(self, drawing, showGG = 0):
self.dc = drawing
if showGG and self.semanticObject: self.drawGGLabel(drawing)
h = drawing.create_oval(self.translate([209.0, 88.0, 209.0, 88.0]), tags = (self.tag, 'connector'), outline = '', fill = '' )
self.connectors.append( h )
h = drawing.create_rectangle(self.translate([38.0, 38.0, 209.0, 127.0]), tags = self.tag, stipple = '', width = 1, outline = 'black', fill = 'cyan')
self.gf5 = GraphicalForm(drawing, h, "gf5")
self.graphForms.append(self.gf5)
font = tkFont.Font( family='Arial', size=12, weight='normal', slant='roman', underline=0)
h = drawing.create_text(self.translate([116.0, 58.0, 116.0, 12.0])[:2], tags = self.tag, font=font, fill = 'black', anchor = 'center', text = 'MT_post__Model_T', width = '0', justify= 'left', stipple='' )
self.gf38 = GraphicalForm(drawing, h, 'gf38', fontObject=font)
self.graphForms.append(self.gf38)
helv12 = tkFont.Font ( family="Helvetica", size=12, weight="bold" )
h = drawing.create_text(self.translate([-3, -3]), font=helv12,
tags = (self.tag, self.semanticObject.getClass()),
fill = "black",
text=self.semanticObject.MT_label__.toString())
self.attr_display["MT_label__"] = h
self.gf_label = GraphicalForm(drawing, h, 'gf_label', fontObject=helv12)
self.graphForms.append(self.gf_label)
def postCondition( self, actionID, * params):
return None
def preCondition( self, actionID, * params):
return None
def getImageDict( self ):
imageDict = dict()
return imageDict
new_class = graph_MT_post__Model_T
|
levilucio/SyVOLT
|
UMLRT2Kiltera_MM/graph_MT_post__Model_T.py
|
Python
|
mit
| 2,614
|
PENDING = 0
RESOLVED = 1
STATUS = {
PENDING: 'pending',
RESOLVED: 'resolved',
}
|
kvchen/officehour-queue
|
oh_queue/entries/constants.py
|
Python
|
mit
| 84
|
# Generated by Django 2.1.4 on 2019-01-23 20:20
from django.db import migrations, models
class Migration(migrations.Migration):
dependencies = [
('guide', '0002_auto_20181116_1617'),
]
operations = [
migrations.AddField(
model_name='rule',
name='is_active',
field=models.BooleanField(default=True),
),
]
|
stefanw/froide
|
froide/guide/migrations/0003_rule_is_active.py
|
Python
|
mit
| 385
|
import win32com.client.makepy
import win32com.client
import os
# Generate type library so that we can access constants
win32com.client.makepy.GenerateFromTypeLibSpec('Acrobat')
# Use Unicode characters instead of their ascii psuedo-replacements
UNICODE_SNOB = 0
def convertHTML2PDF(htmlPath, pdfPath):
'Convert an HTML document to PDF format'
# Connect to Adobe Acrobat
avDoc = win32com.client.Dispatch('AcroExch.AVDoc')
# Open HTML document
avDoc.Open(os.path.abspath(htmlPath), 'html2pdf')
# Save in PDF format
pdDoc = avDoc.GetPDDoc()
pdDoc.Save(win32com.client.constants.PDSaveFull, os.path.abspath(pdfPath))
pdDoc.Close()
# Close HTML document without prompting to save
avDoc.Close(True)
def file_conversion():
directory = 'C:\\HTML'
for path, dirnames, files in os.walk(directory):
for file in files:
filepath, filename = path, file
if "clean" in filename and '.htm' in filename:
new_pdf = filename.replace('.html','.pdf')
pdf = os.path.join(filepath, new_pdf)
html = os.path.join(filepath, filename)
convertHTML2PDF(html, pdf)
else:
pass
|
ryan413/gotem
|
web_html_pdf_zip/html_to_pdf.py
|
Python
|
mit
| 1,227
|
#------------------------------------------------------------------------------
# Copyright (c) 2008 Richard W. Lincoln
#
# Permission is hereby granted, free of charge, to any person obtaining a copy
# of this software and associated documentation files (the "Software"), to
# deal in the Software without restriction, including without limitation the
# rights to use, copy, modify, merge, publish, distribute, sublicense, and/or
# sell copies of the Software, and to permit persons to whom the Software is
# furnished to do so, subject to the following conditions:
#
# The above copyright notice and this permission notice shall be included in
# all copies or substantial portions of the Software.
#
# THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
# IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
# FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
# AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
# LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
# FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS
# IN THE SOFTWARE.
#------------------------------------------------------------------------------
""" Defines traits used in more than on graph element.
"""
#------------------------------------------------------------------------------
# Imports:
#------------------------------------------------------------------------------
from enthought.traits.api import \
HasTraits, Color, Str, Enum, Float, Font, Any, Bool, Int, File, Trait, \
List, Tuple, ListStr, Property, Either, Regex
from enthought.enable.colors import ColorTrait
#------------------------------------------------------------------------------
# Syntactically concise alias trait:
#------------------------------------------------------------------------------
def Alias(name, **metadata):
""" Syntactically concise alias trait but creates a pair of lambda
functions for every alias you declare.
class MyClass(HasTraits):
line_width = Float(3.0)
thickness = Alias("line_width")
"""
return Property(lambda obj: getattr(obj, name),
lambda obj, val: setattr(obj, name, val),
**metadata)
#------------------------------------------------------------------------------
# Syntactically verbose alias trait:
#------------------------------------------------------------------------------
def _get_synced ( self, name ):
return getattr( self, self.trait( name ).sync_to )
def _set_synced ( self, name, value ):
setattr( self, self.trait( name ).sync_to, value )
Synced = Property( _get_synced, _set_synced )
#------------------------------------------------------------------------------
# Trait definitions
#------------------------------------------------------------------------------
# An ID is one of the following:
# * Any string of alphabetic ([a-zA-Z\200-\377]) characters, underscores
# ('_') or digits ([0-9]), not beginning with a digit;
# * a number [-]?(.[0-9]+ | [0-9]+(.[0-9]*)? );
# * any double-quoted string ("...") possibly containing escaped
# quotes (\")1;
# * an HTML string (<...>).
alphanum = "[a-zA-Z]"#\200-\377] "# | [0-9] "#| [_]"
number = "[-]?(.[0-9]+ | [0-9]+(.[0-9]*)? ) "
dquote = '\" '
html = "<...>"
id_trait = Regex(regex=alphanum+"|"+number+"|"+dquote+"|"+html)
pointf_trait = Tuple(Float, Float, desc="the point (x,y)", graphviz=True)
point_trait = pointf_trait #Either(
# pointf_trait, Tuple(Float, Float, Float, desc="the point (x,y,z)")
#)
color_schemes = ["X11", "Accent", "Blues", "BRBG", "BUGN", "BUPU", "Dark",
"GUBU", "Greens", "Greys", "Oranges", "OORD", "Paired", "Pastel", "PIYG",
"PRGN", "PUBU", "PUBUGN", "PUOR", "PURD", "Purples", "RDBU", "RDGY",
"RDPU", "RDYLBU", "RDYLGN", "Reds", "Set", "Spectral", "YLGN",
"YLGNBU", "YLORBR", "YLORRD"]
color_scheme_trait = Enum(color_schemes, desc="a color scheme namespace",
label="Color scheme", graphviz=True)
color_trait = Color("black", desc="drawing color for graphics, not text",
graphviz=True)
comment_trait = Str(desc="comments inserted into output", graphviz=True)
fontcolor_trait = Color("black", desc="color used for text",
label="Font color", graphviz=True)
fontname_trait = Font("Times-Roman", desc="font used for text",
label="Font name", graphviz=True)
fontsize_trait = Float(14.0, desc="size, in points, used for text",
label="Font size", graphviz=True)
label_trait = Str(desc="text label attached to objects", graphviz=True)
# FIXME: Implement layerRange
#
# layerId or layerIdslayerId,
# where layerId = "all", a decimal integer or a layer name. (An integer i
# corresponds to layer i.) The string s consists of 1 or more separator
# characters specified by the layersep attribute.
layer_trait = Str(desc="layers in which the node or edge is present",
graphviz=True)
#margin_trait = Either(
# Float, pointf_trait, desc="x and y margins of canvas or node label"
#)
margin_trait = Float(desc="x and y margins of canvas or node label",
graphviz=True)
nojustify_trait = Bool(False, label="No justify",
desc="multi-line labels will be justified in the context of itself",
graphviz=True)
peripheries_trait = Int(desc="number of peripheries used in polygonal shapes "
"and cluster boundaries", graphviz=True)
# FIXME: Implement splineType
#
# splineType
# spline ( ';' spline )*
# where spline = (endp)? (startp)? point (triple)+
# and triple = point point point
# and endp = "e,%d,%d"
# and startp = "s,%d,%d"
# If a spline has points p1 p2 p3 ... pn, (n = 1 (mod 3)), the points
# correspond to the control points of a B-spline from p1 to pn. If startp is
# given, it touches one node of the edge, and the arrowhead goes from p1 to
# startp. If startp is not given, p1 touches a node. Similarly for pn and endp.
pos_trait = Tuple(Float, Float,
desc="position of node, or spline control points", graphviz=True)
rectangle_trait = Tuple(Float, Float, Float, Float,
desc="The rect llx,lly,urx,ury gives the coordinates, in points, of the "
"lower-left corner (llx,lly) and the upper-right corner (urx,ury).",
graphviz=True)
root_trait = Str(desc="nodes to be used as the center of the layout and the "
"root of the generated spanning tree")
showboxes_trait = Trait("beginning", {"beginning": 1, "end": 2},
label="Show boxes", desc="guide boxes in PostScript output",
graphviz=True)
# Additional styles are available in device-dependent form. Style lists are
# passed to device drivers, which can use this to generate appropriate output.
edge_styles = ["dashed", "dotted", "solid", "invis", "bold"]
cluster_styles = ["filled", "rounded"]
node_styles = edge_styles + cluster_styles + ["diagonals"]
target_trait = Str(desc="if the object has a URL, this attribute determines "
"which window of the browser is used", graphviz=True)
tooltip_trait = Str(desc="tooltip annotation attached to the node or edge",
graphviz=True)
url_trait = Str(desc="hyperlinks incorporated into device-dependent output",
graphviz=True)
# EOF -------------------------------------------------------------------------
|
rwl/godot
|
godot/common.py
|
Python
|
mit
| 7,329
|
import warnings
from datapackage_pipelines.wrapper import ingest
from datapackage_pipelines.utilities.flow_utils import spew_flow
from datapackage_pipelines.lib.update_package import flow
if __name__ == '__main__':
warnings.warn(
'add_metadata will be removed in the future, use "update_package" instead',
DeprecationWarning
)
with ingest() as ctx:
spew_flow(flow(ctx.parameters), ctx)
|
frictionlessdata/datapackage-pipelines
|
datapackage_pipelines/lib/add_metadata.py
|
Python
|
mit
| 426
|
'''
Faça um Programa que peça dois números e imprima o maior deles.
'''
num1 = int (input('Primeiro numero: '))
num2 = int(input('Segundo numero: '))
if num1 > num2:
print(num1)
else:
print(num2)
|
GiordaneOliveira/Exercicios-resolvidos-em-python
|
01.py
|
Python
|
mit
| 222
|